Rhenus Logistics logo

Data Engineer

Rhenus Logistics
Full-time
On-site
Holzwickede, Germany

What You Can Expect:

About the Role:

We are looking for an experienced and solution-oriented Azure Data Platform Engineer to develop, operate, and optimize our modern Azure-based data platform. In this role, you will focus on Azure, Databricks, data infrastructure, and CI/CD, supporting multi-tenant environments and enabling reliable, scalable data solutions.

What You Can Expect:

• Develop a Modern Azure Data Platform: Design, build, and operate end-to-end data solutions using Azure Data Factory, Azure Data Lake Storage Gen2, Databricks, and Azure Synapse Analytics.

• Create Data Pipelines: Develop and maintain scalable ETL/ELT pipelines using PySpark and Spark, with a strong focus on data quality, reliability, and performance.

• Multi-Tenant & Environment Support: Support and operate multi-tenant data platforms across multiple environments (development, test, production) with clear separation and governance.

• Infrastructure & Platform Operations: Provision, configure, and maintain Azure data infrastructure, ensuring stability, security, and scalability.

• CI/CD for Data Platforms: Build and maintain CI/CD pipelines for data pipelines and Databricks workloads, enabling automated deployments across environments.

• Cost-Efficient & Best-Practice Azure Usage: Apply Azure best practices to optimize performance and cost, including resource sizing, lifecycle management, and cost monitoring.

• Collaboration with BI & Data Teams: Work closely with BI and data teams to support efficient data models and reporting solutions.

• Data Governance & Security Basics: Support data governance requirements such as access control, secure data handling, and basic metadata management.

What You Bring:

• Azure Data Platform Experience: Several years of hands-on experience with Azure Data Factory, ADLS Gen2, Databricks, and Azure Synapse Analytics.

• PySpark & Spark: Strong experience building distributed data processing pipelines using PySpark and Spark.

• ETL / ELT Knowledge: Solid understanding of ETL/ELT concepts and data modeling practices.

• CI/CD & Automation: Experience with CI/CD pipelines for data workloads and basic automation of deployments.

• SQL Skills: Strong SQL skills and experience optimizing analytical queries.

• Data Formats: Practical experience with Parquet and/or Avro.

• Infrastructure Awareness: Good understanding of Azure resource structure, environments, and operational best practices.

• Analytical & Team-Oriented Mindset: Solution-focused approach with the ability to work independently and collaboratively.

• Language Skills: Fluency in English is required; knowledge of German is an advantage.

This is a remote role in the European Union. Candidates need to be based in a country where Rhenus Overland Transport is already established.