Senior Data Engineer | Fully-Remote (Germany)
Experience: Senior – 5+ years
Package: €75,000 - €95,000 EUR
About the company:
We are partnered with an AI organisation based in Germany who have created a ground-breaking AI Platform. This product autonomously investigates underlying processes and helps companies deliver rapid ROI, improve efficiency and profitability. The platform even helps companies become more sustainable.
The AI team is the backbone of the business. As part of their continued growth, they’re looking for a Senior Data Engineer who has strong Data Modelling and Data Pipelines experience.
The role will be pivotal in bridging the gap between software engineering and data science, modernising legacy systems, and building scalable data infrastructure that powers advanced AI and ML products. This is an opportunity to join a collaborative and ambitious team that takes ownership of its entire data stack, from design and development to deployment and operations.
Key Responsibilities:
  - Design and implement complex data pipelines using Python, Spark, and SQL to process and transform large-scale data efficiently.
 
  - Build reusable, scalable components and ensure high-quality data transformation across batch and streaming workloads.
 
  - Implement data quality checks and validation frameworks to guarantee reliability and accuracy, essential for ML-driven applications.
 
  - Design and implement robust data models (dimensional, data vault, or similar) that support scalable analytical and operational workloads.
 
  - Collaborate closely with Data Scientists to deliver clean, well-structured datasets for model training and inference.
 
  - Participate in feature engineering discussions and implement scalable pipelines for ML workflows.
 
  - Build and optimise distributed computing workflows using Spark and related tools.
 
  - Contribute to infrastructure development on cloud platforms, leveraging DevOps and IaC practices.
 
  - Modernise existing systems and contribute to the long-term architecture for a new, greenfield data platform.
 
  - Mentor junior developers and promote best practices in data engineering and distributed systems.
 
Ideal Knowledge & Experience:
  - 5+ years of experience in data engineering, ideally within data science or AI-focused environments.
 
  - Strong hands-on proficiency with Python, Spark, and SQL.
 
  - Proven experience building and maintaining data pipelines and distributed computing systems.
 
  - Deep understanding of data modelling principles capable of designing and implementing new models from scratch (not just using Databricks).
 
  - Experience performing data quality checks and building validation mechanisms for ML applications.
 
  - Familiarity with Databricks and Lakehouse architectures is a strong plus.
 
  - Strong understanding of cloud data platforms (AWS, Azure, or GCP).
 
  - Exposure to DevOps, CI/CD, and infrastructure automation concepts.
 
Interview Process:
1st round: Introductory discussion with the hiring manager.
2nd round: Technical test + pair programming exercise.
3rd round: Cultural interview
What’s on offer:
Salary range: up to €95,000 
Permanent employment contract.
Fully Remote across Germany. You must have a permanent residence in DE.