Agilisium logo

Sr. Data Engineer

Agilisium
Full-time
On-site
Tamil Nadu, India

Introduction to the Role:

Are you passionate about unlocking the power of data to drive innovation and transform business outcomes? Join our cutting-edge Data Engineering team and be a key player in delivering scalable, secure, and high-performing data solutions across the enterprise. As a Data Engineer, you will play a central role in designing and developing modern data pipelines and platforms that support data-driven decision-making and AI-powered products. With a focus on Python, SQL, AWS, PySpark, and Databricks, you'll enable the transformation of raw data into valuable insights by applying engineering best practices in a cloud-first environment.

We are looking for a highly motivated professional who can work across teams to build and manage robust, efficient, and secure data ecosystems that support both analytical and operational workloads.


Accountabilities:
  • Design, build, and optimize scalable data pipelines using PySpark, Databricks, and SQL on AWS cloud platforms.

  • Collaborate with data analysts, data scientists, and business users to understand data requirements and ensure reliable, high-quality data delivery.

  • Implement batch and streaming data ingestion frameworks from a variety of sources (structured, semi-structured, and unstructured data).

  • Develop reusable, parameterized ETL/ELT components and data ingestion frameworks.

  • Perform data transformation, cleansing, validation, and enrichment using Python and PySpark.

  • Build and maintain data models, data marts, and logical/physical data structures that support BI, analytics, and AI initiatives.

  • Apply best practices in software engineering, version control (Git), code reviews, and agile development processes.

  • Ensure data pipelines are well-tested, monitored, and robust with proper logging and alerting mechanisms.

  • Optimize performance of distributed data processing workflows and large datasets.

  • Leverage AWS services (such as S3, Glue, Lambda, EMR, Redshift, Athena) for data orchestration and lakehouse architecture design.

  • Participate in data governance practices and ensure compliance with data privacy, security, and quality standards.

  • Contribute to documentation of processes, workflows, metadata, and lineage using tools such as Data Catalogs or Collibra (if applicable).

  • Drive continuous improvement in engineering practices, tools, and automation to increase productivity and delivery quality.

Essential Skills / Experience:

  • 4 to 6 years of professional experience in Data Engineering or a related field.

  • Strong programming experience with Python and experience using Python for data wrangling, pipeline automation, and scripting.

  • Deep expertise in writing complex and optimized SQL queries on large-scale datasets.

  • Solid hands-on experience with PySpark and distributed data processing frameworks.

  • Expertise working with Databricks for developing and orchestrating data pipelines.

  • Experience with AWS cloud services such as S3, Glue, EMR, Athena, Redshift, and Lambda.

  • Practical understanding of ETL/ELT development patterns and data modeling principles (Star/Snowflake schemas).

  • Experience with job orchestration tools like Airflow, Databricks Jobs, or AWS Step Functions.

  • Understanding of data lake, lakehouse, and data warehouse architectures.

  • Familiarity with DevOps and CI/CD tools for code deployment (e.g., Git, Jenkins, GitHub Actions).

  • Strong troubleshooting and performance optimization skills in large-scale data processing environments.

  • Excellent communication and collaboration skills, with the ability to work in cross-functional agile teams.

Desirable Skills / Experience:

  • AWS or Databricks certifications (e.g., AWS Certified Data Analytics, Databricks Data Engineer Associate/Professional).

  • Exposure to data observability, monitoring, and alerting frameworks (e.g., Monte Carlo, Datadog, CloudWatch).

  • Experience working in healthcare, life sciences, finance, or another regulated industry.

  • Familiarity with data governance and compliance standards (GDPR, HIPAA, etc.).

  • Knowledge of modern data architectures (Data Mesh, Data Fabric).

  • Exposure to streaming data tools like Kafka, Kinesis, or Spark Structured Streaming.

  • Experience with data visualization tools such as Power BI, Tableau, or QuickSight.

Work Environment & Collaboration:

We value a hybrid, collaborative environment that encourages shared learning and innovation. You will work closely with product owners, architects, analysts, and data scientists across geographies to solve real-world business problems using cutting-edge technologies and methodologies. We encourage flexibility while maintaining a strong in-office presence for better team synergy and innovation.

About Agilisium -  

  • Agilisium, is an AWS technology Advanced Consulting Partner that enables companies to accelerate their "Data-to-Insights-Leap. 
  • With $50+ million in annual revenue and over 30% year-over-year growth, Agilisium is one of the fastest-growing IT solution providers in Southern California. 
  • Our most important asset? People. 
  • Talent management plays a vital role in our business strategy. 
  • We’re looking for “drivers”; big thinkers with growth and strategic mindset — people who are committed to customer obsession, aren’t afraid to experiment with new ideas. 
  • And we are all about finding and nurturing individuals who are ready to do great work. 
  • At Agilisium, you’ll collaborate with great minds while being challenged to meet and exceed your potential