At Alcon, we are driven by the meaningful work we do to help people see brilliantly. We innovate boldly, champion progress, and act with speed as the global leader in eye care. Here, you’ll be recognized for your commitment and contributions and see your career like never before. Together, we go above and beyond to make an impact in the lives of our patients and customers. We foster an inclusive culture and are looking for diverse, talented people to join Alcon.
About Your Future Team
You will join the Generative AI & Data Engineering team – a pioneering unit focused on delivering intelligent solutions that power AI-driven products and decision-making. We work at the intersection of structured enterprise data and unstructured content (text, images, documents), enabling next-gen applications in NLP, LLMs, and intelligent search.
In this role, a typical day will include:
Build and optimize robust data pipelines to ingest, process, and serve structured (SQL, tabular) and unstructured (PDFs, text, images) data to downstream GenAI systems.
Design scalable and modular ETL/ELT workflows using Azure Data Services (Azure Cognitive Search, Azure Document Intelligence, etc).
Develop reusable and modular Python components for preprocessing unstructured data for GenAI models.
Design and deploy data workflows using Docker containers for environment consistency and scalability.
Orchestrate data pipelines and jobs using Apache Airflow, ensuring reliable scheduling and monitoring.
Collaborate closely with LLM engineers, data scientists, and product teams to ensure data readiness for RAG, embeddings, and vector databases.
Handle large-scale data transformations, metadata tagging, and schema evolution across data formats (JSON, CSV, Parquet, images).
Integrate Azure OpenAI and other LLM APIs into the data workflow when required.
Contribute to the data layer that supports chatbots, document summarization, and intelligent assistants.
Work in Agile teams with global collaboration across data, AI, and product domains.
WHAT YOU’LL BRING TO ALCON:
3+ years of experience in Data Engineering, especially in cloud-first environments.
Strong hands-on skills in Python (data wrangling, file parsing, API integration) and SQL (complex queries, performance tuning).
Experience with unstructured data processing – PDFs, images, HTML, JSON, etc.
Solid understanding of Azure Data Stack: Data Lake, Azure Search, Azure Blob.
Comfortable working with large language models, vector stores.
Practical experience in preparing data for GenAI pipelines – at least 2 large-scale projects (e.g., chunking, vector embeddings).
Strong working knowledge of Docker and containerization best practices.
Hands-on experience in designing and maintaining DAGs in Apache Airflow.
Familiarity with data formats and standards relevant to AI (e.g., tokenization, embeddings, ML metadata).
Bonus: Experience with AWS or Hybrid Cloud Environments
Familiarity with CI/CD pipelines, version control (Git), and DevOps practices.
Proactive, curious, and adaptable with strong communication skills.
Fluent in English – written and spoken.
HOW YOU CAN THRIVE AT ALCON:
Alcon Careers See your impact at alcon.com/careers
ATTENTION: Current Alcon Employee/Contingent Worker
If you are currently an active employee/contingent worker at Alcon, please click the appropriate link below to apply on the Internal Career site.