| Job Position | Company | Posted | Location | Salary | Tags |
|---|---|---|---|---|---|
Binance | Brisbane, Australia |
| |||
Dune | London, United Kingdom | $140k - $150k | |||
Binance | Taipei, Taiwan |
| |||
Bitgo | Remote | $95k - $111k | |||
| Learn job-ready web3 skills on your schedule with 1-on-1 support & get a job, or your money back. | | by Metana Bootcamp Info | |||
Crypto Finance AG | Zurich, Switzerland | $91k - $115k | |||
Binance | Hong Kong, Hong Kong |
| |||
Bcbgroup | Remote | $62k - $64k | |||
Bitgo | Remote | $126k - $144k | |||
Bitgo | Remote | $126k - $144k | |||
Binance | Taipei, Taiwan |
| |||
Binance | Hong Kong, Hong Kong |
| |||
Gsrmarkets | Remote | $80k - $95k | |||
Binance | Brisbane, Australia |
| |||
Binance | Taipei, Taiwan |
| |||
Bitpanda | Remote | $105k - $150k |
Data Scientist, LLM & AI Agent Engineer (Applied AI)
Responsibilities
- Design “Sherlock”->
- Build reasoning-capable agents that analyze user behavior, device fingerprints, and trading history to assess whether withdrawals are fraudulent or safe
- Develop market manipulation detection agents that analyze unstructured social sentiment alongside order book data
- Identify pump-and-dump schemes and wash trading
- Build LLM-based triage systems to handle high volumes of risk alerts
- Use agents to pre-screen alerts by analyzing alert context and historical false positives
- Automatically decide whether to close alerts or escalate them to human analysts
- Reduce alert fatigue for Ops teams by filtering noise with high semantic understanding
- Use LLMs to extract features from messy, unstructured data such as chat logs, support tickets, KYC documents, and chat groups
- Convert vague risk signals into structured features (e.g., JSON) for downstream risk models or rule engines
- Build RAG pipelines that allow agents to query repositories of past fraud cases and identify recurring patterns
- Develop analyst copilot tools to draft investigation reports (SARs) and summarize complex cases
- Enable human-in-the-loop workflows where agents gather evidence, propose decisions, and humans review and approve outcomes
Requirements
- LLM application experience: 2+ years building applications with GPT, Claude, or other open-source LLM models
- Agent frameworks: Deep familiarity with LangGraph, LangChain, or CrewAI
- RAG for domain knowledge: Experience building retrieval systems that fetch relevant contextual information
- Evaluation (crucial): Proven ability to design evaluation sets to measure agent performance, such as correctly identifying fraudsters versus regular users
- Python mastery: Ability to write production-ready, maintainable code (not just scripts)
- Data engineering: Proficiency in SQL and experience working with data pipelines (e.g., Kafka, Spark) to provide agents with real-time data
- Domain interest (bonus): Understanding of Trust & Safety, anti-fraud, or financial risk domains
- Crypto familiarity (bonus): Knowledge of on-chain analysis, DeFi concepts, and wallet addresses
- Anomaly detection (bonus): Experience with anomaly detection concepts and techniques
- Educational background: Master’s degree with 2+ years of experience, or equivalent hands-on expertise in Data Science, AI Engineering, or Applied ML
- Scale & systems experience: Experience working with terabyte-scale datasets and real-time systems
- Financial systems knowledge (preferred): Understanding of financial markets, trading systems, or risk management
- Adaptability: Comfortable working in fast-changing, ambiguous 0→1 environments with the ability to prototype, iterate quickly, and drive execution
- Passion for AI: Strong interest in AI agents, autonomous systems, LLMs, or intelligent automation
- Communication skills: Strong English reading and writing skills for technical documentation and agent prompt engineering
What does a data scientist in web3 do?
A data scientist in web3 is a type of data scientist who focuses on working with data related to the development of web-based technologies and applications that are part of the larger web3 ecosystem
This can include working with data from decentralized applications (DApps), blockchain networks, and other types of distributed and decentralized systems
In general, a data scientist in web3 is responsible for using data analysis and machine learning techniques to help organizations and individuals understand, interpret, and make decisions based on the data generated by these systems
Some specific tasks that a data scientist in web3 might be involved in include developing predictive models, conducting research, and creating data visualizations.