Design and implement large-scale data solutions using Big Data and Cloud technologies
Design and develop conceptual data models for integrating, exploring and retrieving data from heterogeneous data sources and IT systems
Implement and manage ETL processes, data transformations, data flows and service APIs
Work with data virtualization, integration, and analytics technologies on our data platform
Implement optimized data representations (views) and aggregations from multiple data sources
Define and implement data ingestion, retention, lineage and access policies
Define and implement data service API management and usage policies
Define and implement data and data service catalog functionalities
Work in interdisciplinary, cross-functional teams according to agile methodology
Closely collaborate with business, digitalization and IT functions
Your Qualifications
Successfully completed studies in Computer Science or related fields
Proficient understanding of distributed and Cloud computing principles and technologies
Experience with integration of data from multiple data sources and IT systems, e.g. ERP/SAP, DWH/SAP-BW, Web Services, CRM/Salesforce, MES, etc.
Relevant knowledge in working with Cloud environments and services, preferably with Microsoft Azure, e.g. Azure Data Factory, Data Lake, IoT Hub, Databricks, etc.
Strong programming skills in Java and/or Python; Software development experience is a plus
Experience in data modeling (ER, UML) and proficiency in SQL
Experience with relational databases such as Microsoft SQL Server and Oracle Postgres as well as experience with NoSQL databases such as MongoDB, Cassandra, InfluxDB, Stardog
Knowledge of ETL techniques and frameworks, such as ADF, PowerCenter, NiFi, Sqoop
Experience with Hadoop Ecosystem (Spark, HDFS, Hive, HBase), Containers (Docker, Kubernetes) and Distributed Messaging Platforms (Kafka, NATS.io) is a plus
Fluent in English spoken and written, German is beneficial