Binance is a leading global blockchain ecosystem behind the world’s largest cryptocurrency exchange by trading volume and registered users. We are trusted by over 280 million people in 100+ countries for our industry-leading security, user fund transparency, trading engine speed, deep liquidity, and an unmatched portfolio of digital-asset products. Binance offerings range from trading and finance to education, research, payments, institutional services, Web3 features, and more. We leverage the power of digital assets and blockchain to build an inclusive financial ecosystem to advance the freedom of money and improve financial access for people around the world.
About the Role
As a Senior Data Engineer, where you’ll architect and maintain a highly flexible, enterprise-scale data warehouse that accelerates insights and minimizes redundant work. Leveraging deep expertise in data modeling, governance and Big Data technologies (Hadoop, Spark, Hive, etc.), you’ll design end-to-end ETL pipelines, optimize performance, and build metadata and quality monitoring frameworks. You’ll collaborate closely across business and technical teams, mentor peers, and continuously evolve our data platform to meet the demands of petabyte-scale environments.
Responsibilities
Architect and implement a flexible, scalable data warehouse aligned with company specifications and business requirements, accelerating delivery and reducing redundant development.
Design, develop, test, deploy and monitor data models and ETL jobs; rapidly troubleshoot complex issues and optimize calculation logic and pipeline performance.
Lead data governance initiatives by building and maintaining metadata management and data quality monitoring systems.
Foster technical team growth through mentorship, knowledge sharing and continuous improvement of collective skills.
Requirements
5+ years of hands-on experience designing and developing data lakes and data warehouse solutions.
Deep expertise in data warehouse modeling and governance, including dimensional modeling, information factory (data vault) methodologies and “one data” principles.
Proficiency in at least one of Java, Scala or Python, plus strong Hive & Spark SQL programming skills.