Must-Have Skills & Experience:
· 3-5+ years of experience in data engineering or a similar role
· Strong hands-on expertise in Databricks (Spark, Delta Lake, Notebooks)
· Proficient in Python for data processing, scripting, and automation
· Solid experience with AWS infrastructure (S3, Lambda, Glue, EC2, IAM, etc.)
· Proven track record in building and maintaining data pipelines (ETL/ELT) for structured and unstructured data
· Familiarity with data quality, performance optimization, and orchestration tools (e.g., Airflow, DBT)
Nice to Have:
· Exposure to machine learning or AI workflows (e.g., feature engineering, model deployment pipelines)
· Experience working in MLOps or data science enablement teams
· Familiarity with ML frameworks like scikit-learn, TensorFlow, or PyTorch