| Job Position | Company | Posted | Location | Salary | Tags |
|---|---|---|---|---|---|
Ethena Labs | Remote | $98k - $103k | |||
Ledger | Paris, France | $91k - $100k | |||
Alpaca | Remote | $98k - $100k | |||
Anchorage Digital | United States | $84k - $100k | |||
| Learn job-ready web3 skills on your schedule with 1-on-1 support & get a job, or your money back. | | by Metana Bootcamp Info | |||
Bcbgroup | Remote | $62k - $64k | |||
Myshell | Remote | $98k - $110k | |||
Gsrmarkets | Remote | $80k - $95k | |||
Chainalysis | India | $79k - $99k | |||
Blockchain | Remote | $84k - $109k | |||
Bitgo | Remote | $95k - $111k | |||
Ledger | Paris, France | $50k | |||
Kraken | United States | $96k - $192k | |||
Certified | New York, NY, United States | $80k - $93k | |||
Binance | Taipei, Taiwan |
| |||
Zenith | Remote |
|
Senior Data Engineer
What You’ll Do
- Rapidly spin up the cloud environment. Deliver working historical backfill pipelines from Tardis.dev into a queryable database.
- Deliver a real-time Tardis WebSocket pipeline, ensuring data is normalized, cached for live consumption, accurate, replayable, and queryable by Day 60.
- Ensure all pipelines are idempotent, retryable, and use exactly-once semantics. Implement full CI/CD, Terraform, automated testing, and secrets management.
- Implement proper observability (structured logs, metrics, dashboards, alerting) from day one. Provide immediate self-service access to the MVP database for Trading and BI teams via tools like Tableau/Metabase, and through simple internal REST APIs.
- Develop specialized timeseries data, including USDe backing-asset and a full opportunity-surface timeseries for Delta-neutral/lending/borrow opportunities.
- Ingest data from additional sources (Kaiko, CoinAPI, on-chain via TheGraph/Dune). Plan for 10x+ data growth via schema evolution, partitioning, and performance tuning. Establish enterprise-grade governance, including a data quality framework, RBAC, audit logs, and a semantic layer.
- Create full architecture documentation, runbooks, and a data dictionary. Onboard and mentor future junior staff.
What We’re Looking For
- Proven track record of delivering working, production data in weeks, not months, with the ability to ruthlessly cut scope to hit a 60-day MVP while managing technical debt.
- Have built Tardis historical and real-time pipelines before (or equivalent high-quality crypto market data feeds), understanding specific quirks, rate limits, and WebSocket structures.
- Expert in large-scale, reliable ETL/ELT for financial or market data.
- Fluent in provisioning full environments with Terraform in days and expert in AWS/GCP serverless technologies.
- Expert Python and SQL skills and proficiency with time-series databases like TimescaleDB or ClickHouse, ensuring fast queries from day one.
- Advanced knowledge of WebSocket clients, message queues, and low-latency streaming, GitOps, automated testing/deploy and observability practices.
- Significant understanding of stablecoins, lending protocols, and opportunity surface concepts, or a proven ability to ramp up extremely quickly.
What does a data scientist in web3 do?
A data scientist in web3 is a type of data scientist who focuses on working with data related to the development of web-based technologies and applications that are part of the larger web3 ecosystem
This can include working with data from decentralized applications (DApps), blockchain networks, and other types of distributed and decentralized systems
In general, a data scientist in web3 is responsible for using data analysis and machine learning techniques to help organizations and individuals understand, interpret, and make decisions based on the data generated by these systems
Some specific tasks that a data scientist in web3 might be involved in include developing predictive models, conducting research, and creating data visualizations.