Design, develop, and implement scalable ETL/ELT processes to transform and load data into Databricks
Develop and optimize data models for deployment in data lake architectures
Collaborate with cross-functional teams including data analysts, data scientists, and business stakeholders
Ensure high data quality, integrity, and compliance with governance standards
Document data flow, architecture decisions, and pipeline processes
Support performance tuning and troubleshooting across large-scale datasets
Demonstrated experience in at least two data engineering projects involving transforming and loading data into Databricks of similar size and complexity
Proven experience in developing data models for implementation in data lake environments
Resided in the United States for a minimum of 3 years
Able to obtain a Public Trust clearance
Must be local to the DMV area (Washington, D.C., Maryland, or Northern Virginia)
Experience with cloud platforms such as Azure, AWS, or GCP
Familiarity with Delta Lake, Apache Spark, and SQL
Experience working in government or public sector environments