EXL logo

Senior Data Engineer

EXL
Full-time
On-site
Canada
Senior Jobs
Description

Overview /Objective:

 

We are seeking a seasoned Data Engineer to join our Sports Analytics & Engineering Practice. This role is pivotal in shaping and implementing our client’s vision for a cutting-edge, cloud-native data ecosystem in its privacy and consent domain. You will architect and build scalable data infrastructure that transforms raw data into high-value assets, powering analytics with consent data across digital products, fan engagement, and marketing domains. Your work will directly contribute to the development of a customer – privacy and consent centric data platform.

 

Responsibilities:

  • Design and build robust, scalable data transformation pipelines using SQL, DBT, and Jinja templating      
  • Develop and maintain data architecture and standards for Data Integration and Data Warehousing projects using DBT and Amazon Redshift
  • Monitor and ensure the smooth operation of data pipelines between OneTrust (OT) application and external data platforms (ESPs, CRMs, etc.).
  • Collaborate with cross-functional teams to gather requirements and deliver dimensional data models that serve as a single source of truth
  • Assist in the ongoing data integration tasks (e.g., connectors, APIs) between OT and various business systems.
  • Own the full stack of data modeling in DBT to empower analysts, data scientists, and BI engineers
  • Enhance and maintain the analytics codebase, including DBT models, SQL scripts, and ERD documentation.
  • Familiarize regulatory compliance guidelines across different geographical areas (CCPA, GDPR) and integrate them into current and new pipelines for consent collection.
  • Ensure data quality, governance alignment, and operational readiness of data pipelines
  • Apply software engineering best practices such as version control, CI/CD, and code reviews
  • Optimize SQL queries for performance, scalability, and maintainability across large datasets
  • Implement best practices for SQL performance tuning, including partitioning, clustering, and materialized views
  • Build and manage infrastructure as code using AWS CDK for scalable and repeatable deployments. Integrate and automate deployment workflows using AWS CodeCommit, CodePipeline, and related DevOps tools
  • Support Agile development processes and collaborate with offshore teams


Responsibilities

Responsibilities:

  • Design and build robust, scalable data transformation pipelines using SQL, DBT, and Jinja templating      
  • Develop and maintain data architecture and standards for Data Integration and Data Warehousing projects using DBT and Amazon Redshift
  • Monitor and ensure the smooth operation of data pipelines between OneTrust (OT) application and external data platforms (ESPs, CRMs, etc.).
  • Collaborate with cross-functional teams to gather requirements and deliver dimensional data models that serve as a single source of truth
  • Assist in the ongoing data integration tasks (e.g., connectors, APIs) between OT and various business systems.
  • Own the full stack of data modeling in DBT to empower analysts, data scientists, and BI engineers
  • Enhance and maintain the analytics codebase, including DBT models, SQL scripts, and ERD documentation.
  • Familiarize regulatory compliance guidelines across different geographical areas (CCPA, GDPR) and integrate them into current and new pipelines for consent collection.
  • Ensure data quality, governance alignment, and operational readiness of data pipelines
  • Apply software engineering best practices such as version control, CI/CD, and code reviews
  • Optimize SQL queries for performance, scalability, and maintainability across large datasets
  • Implement best practices for SQL performance tuning, including partitioning, clustering, and materialized views
  • Build and manage infrastructure as code using AWS CDK for scalable and repeatable deployments. Integrate and automate deployment workflows using AWS CodeCommit, CodePipeline, and related DevOps tools
  • Support Agile development processes and collaborate with offshore teams


Qualifications

Required Qualifications:

  • Bachelor’s or Master’s (preferred) degree in a quantitative or technical field such as Statistics, Mathematics, Computer Science, Information Technology, Computer Engineering or equivalent
  • 5+ years of experience in data engineering and analytics on modern data platforms
  • 3+ years’ extensive experience with DBT or similar data transformation tools, including building complex & maintainable DBT models and developing DBT packages/macros
  • Deep familiarity with dimensional modeling/data warehousing concepts and expertise in designing, implementing, operating, and extending enterprise dimensional models
  • Understand change data capture concepts 
  • Experience working with AWS Services (Lambda, Step Functions, MWAA, Glue, Redshift)
  • Hands-on experience with AWS CDK, CodeCommit, and CodePipeline for infrastructure automation and CI/CD
  • Python proficiency or general knowledge of Jinja templating in Python and/or PySpark
  • Agile experience and willingness to work with extended offshore teams and assist with design and code reviews with customer
  • Familiarity with OneTrust or similar consent management platforms is a plus.
  • A great teammate and self-starter, strong detail orientation is critical in this role.


Apply now
Share this job