Senior Data Engineer

Eindhoven, Netherlands
05-02-2025
Job Type
Permanent
Emp Type
Full Time
Industry
IT
Functional Expertise
Countries
Salary Type
Annual
Salary
Negotiable

Job Description

Senior Data Engineer – Scale-Up Opportunity

 

Are you passionate about building scalable, high-performance data architectures? We’re looking for a Senior Data Engineer to join a fast-paced, scale-up environment where problem-solving and adaptability are key. In this role, you’ll be responsible for designing, building, and optimizing data pipelines using open-source technologies, AWS, and ClickHouse (or equivalent OLAP databases) to support evolving business needs.

 

This is an opportunity to make a real impact by driving efficiency, reliability, and scalability in data-driven decision-making. If you thrive in a hands-on role and enjoy collaborating with engineering teams to create cutting-edge data solutions, we’d love to hear from you.

 

Key Responsibilities:

Data Pipeline Design: Architect scalable, high-performance data solutions using AWS and OLAP databases like ClickHouse.
ETL/ELT Development: Design, build, and maintain efficient data pipelines for large-scale data ingestion, transformation, and storage.
Cloud Infrastructure: Utilize a combination of open-source and AWS technologies to optimize data workflows.
Data Governance & Quality: Ensure data accuracy, consistency, and reliability through best practices in data governance, lineage, and monitoring.
Performance Optimization: Improve data storage, retrieval, and processing through partitioning, indexing, and query tuning.
Collaboration & Leadership: Work closely with data scientists, analysts, and engineers while mentoring junior team members.
Automation & CI/CD: Implement automated deployment and monitoring strategies for data pipelines.

 

What We’re Looking For:

  • 5+ years of experience in data engineering, with a focus on open-source and cloud-native solutions.
  • Expertise in ClickHouse (or similar OLAP databases).
  • Strong experience in designing and architecting large-scale data systems.
  • Proficiency in Python (or other scripting languages) for data processing and pipeline development.
  • Deep understanding of ETL/ELT processes, data modeling (Star Schema, Snowflake Schema), and data warehousing.
  • Hands-on experience with orchestration tools like Apache Airflow, Step Functions, or Dagster.
  • Familiarity with infrastructure-as-code tools (Terraform, CloudFormation, CDK) is a plus.
  • Strong problem-solving skills and ability to work in a fast-paced environment.
  • Knowledge of SQL performance tuning, materialized views, and sharding strategies for large datasets.
  • Experience with containerization (Docker, Kubernetes) and serverless computing.
  • Familiarity with monitoring tools like Prometheus, Grafana, AWS CloudWatch.

 

This is an exciting opportunity to join a market leading E-Learning provider based in Eindhoven. If you are interested, please apply through the Leap29 website with your updated CV.

Consultant

Ben Quinn