Your Key Responsibilities Design and implement large-scale data solutions using Big Data and Cloud technologies Design and develop conceptual data models for integrating, exploring and retrieving data from heterogeneous data sources and IT systems Implement and manage ETL processes, data transformations, data flows and service APIs Work with data virtualization, integration, and analytics technologies on our data platform Implement optimized data representations (views) and aggregations from multiple data sources Define and implement data ingestion, retention, lineage and access policies Define and implement data service API management and usage policies Define and implement data and data service catalog functionalities Work in interdisciplinary, cross-functional teams according to agile methodology Closely collaborate with business, digitalization and IT functions Your Qualifications Successfully completed studies in Computer Science or related fields Proficient understanding of distributed and Cloud computing principles and technologies Experience with integration of data from multiple data sources and IT systems, e.g. ERP/SAP, DWH/SAP-BW, Web Services, CRM/Salesforce, MES, etc. Relevant knowledge in working with Cloud environments and services, preferably with Microsoft Azure, e.g. Azure Data Factory, Data Lake, IoT Hub, Databricks, etc. Strong programming skills in Java and/or Python; Software development experience is a plus Experience in data modeling (ER, UML) and proficiency in SQL Experience with relational databases such as Microsoft SQL Server and Oracle Postgres as well as experience with NoSQL databases such as MongoDB, Cassandra, InfluxDB, Stardog Knowledge of ETL techniques and frameworks, such as ADF, PowerCenter, NiFi, Sqoop Experience with Hadoop Ecosystem (Spark, HDFS, Hive, HBase), Containers (Docker, Kubernetes) and Distributed Messaging Platforms (Kafka, NATS.io) is a plus Fluent in English spoken and written, German is beneficial