About the company
We create foundational infrastructure to accelerate web3 adoption! Come join us! We are looking for researchers and engineers, product builders, entrepreneurs in fields spanning distributed systems, cryptography, compiler & VM, programming language, verification, and game theory.
Job Summary
Responsibilities:
šDesign and implement scalable ingestion pipelines with a reusable and modular framework šBuild robust, reusable frameworks to ingest data from internal sources (e.g., Prod DBs, cloud buckets, etc) and external APIs or files (e.g., CSVs, webhooks). šEnsure idempotency, backfill support, and error handling in pipeline design. šOptimize and own data warehousing, with clear table definitions, schemas, cost efficiency šArchitect a centralized data lake/warehouse with clear schemas and partitioning strategies. šSupport both batch and streaming workloads, and optimize for cost and performance. šEnable data discoverability, usability, and governance šImplement or integrate data cataloging and lineage tools šDefine naming conventions, documentation standards, and ownership metadata to make data self-serve and intuitive for data scientists, analysts, and product / GTM teams. šSet up a mechanism for scalable access controls (with RBAC or ABAC), PII tagging, and data obfuscation. šEnable approaches for data quality checks, validation pipelines, and alerting for broken or stale data šDevelop a strong understanding of how to use on-chain and off-chain data together
Required Qualifications:
š5+ years experience in data engineering šStrong SQL and Python šStrong and informed opinions on data orchestrators, catalogs, governance, and testing frameworks šExperience combining in-house and external data
If this role isnāt the perfect fit, there are plenty of exciting opportunities in blockchain technology, cryptocurrency startups, and remote crypto jobs to explore. Check them on our Jobs Board.