Data Science Jobs in Web3

1,662 jobs found

web3.career is now part of the Bondex Logo Bondex Ecosystem

Receive emails of Data Science Jobs in Web3
Job Position Company Posted Location Salary Tags

Integra

Remote

$21k - $64k

Integra

Remote

$31k

Keyrock

Brussels, Belgium

$36k - $75k

Polymarket

New York, NY, United States

$84k - $148k

Blockchain

Remote

$63k - $65k

Okx

Remote

$98k - $150k

Layerzerolabs

Vancouver, Canada

$75k - $77k

Launchpadtechnologiesinc

Latam

$86k - $109k

Integra

Remote

$88k - $101k

Integra

Remote

$72k - $84k

Bluecubeservices

Remote

$79k - $100k

Flipside

Boston, MA, United States

$98k - $171k

Crypto.com

Hong Kong, Hong Kong

$94k - $105k

Binance

Taipei, Taiwan

Who We Are

Stablecoins are beginning to reshape the global FX market, where more than $10 trillion trades every day. Hibachi is building the exchange designed for that shift.

We are building a modern central limit order book for global currencies with transparent prices, direct access to liquidity, and infrastructure designed for continuous global markets. Our goal is to open FX trading beyond the traditional interbank system and create a venue where global money can move freely.

We are a small team of engineers and traders who have built market infrastructure at Tower Research, Citadel, Coinbase, and Bloomberg. We care deeply about performance, correctness, and building systems that operate at global scale.

Hibachi is backed by Dragonfly Capital, Electric Capital, Coinbase Ventures, and Circle Ventures.




About The Technology

Hibachi runs a high performance off chain central limit order book built for fast, private trading and deep liquidity. Zero knowledge proofs allow anyone to verify the exchange’s solvency on chain without revealing user positions. The result is transparent infrastructure built for global markets.



The Role

We are seeking a Data Engineer with broad expertise in data modeling, advanced SQL, ETL/ELT development, and CDC (Change Data Capture). You will design and maintain end-to-end data solutions—covering batch and streaming ingestion, data warehousing with Iceberg, AWS DMS (or similar CDC tools). This role requires strong communication skills to ensure data initiatives align with and drive business objectives.




You’ll Be Responsible for:

  • Data Pipeline Development: Architect, build, and maintain batch and streaming data pipelines using PySpark, AWS Glue, and Airflow .Implement Change Data Capture (CDC) with AWS DMS (or comparable tools) to capture incremental updates from source systems.
  • Data Modeling & Architecture: Design modular, reusable, and scalable data models adhering to best practices. Work with iceberg backed Data Warehouse solution for performant storage, queries, and transformations. Ensure consistent data definitions and governance using frameworks like the Glue Catalog.
  • ETL/ELT: Manage ETL/ELT pipelines ensuring efficient data ingestion, cleansing, and aggregation. Monitor and debug performance bottlenecks, applying tuning techniques where necessary.
  • Data Visualization & Analytics: Develop QuickSight dashboards (or similar BI tools) to surface actionable insights for stakeholders.


You’ll Need to Have:

  • Bachelor’s or Master’s Degree in Computer Science, Engineering, or a related field (or equivalent experience).
  • 2+ years of hands-on experience with PySpark for batch and streaming pipelines. Familiarity with streaming ecosystems (Kafka, Kinesis, Spark Structured Streaming).
  • Strong proficiency in AWS Glue, Apache Airflow, and Iceberg
  • Experience with AWS DMS or other CDC tools to manage real-time or near-real-time data ingestion.
  • Advanced SQL knowledge, including performance tuning and complex transformations.
  • Proven background in data modeling and data architecture best practices (data warehouse/data lake).
  • Experience with BI platforms (QuickSight, Tableau, Power BI, etc.) for dashboard development.
  • Understanding of testing frameworks (e.g., Pytest) for data pipelines, unit testing, and QA processes
  • Excellent communication skills, with an ability to bridge technical and business requirements


We’d Love to See:

  • Background in trading, HFT, or capital markets infrastructure

What does a data scientist in web3 do?

A data scientist in web3 is a type of data scientist who focuses on working with data related to the development of web-based technologies and applications that are part of the larger web3 ecosystem

This can include working with data from decentralized applications (DApps), blockchain networks, and other types of distributed and decentralized systems

In general, a data scientist in web3 is responsible for using data analysis and machine learning techniques to help organizations and individuals understand, interpret, and make decisions based on the data generated by these systems

Some specific tasks that a data scientist in web3 might be involved in include developing predictive models, conducting research, and creating data visualizations.