Senior Data Engineer
Job Description
Senior Data Engineer – Scale-Up Opportunity
Are you passionate about building scalable, high-performance data architectures? We’re looking for a Senior Data Engineer to join a fast-paced, scale-up environment where problem-solving and adaptability are key. In this role, you’ll be responsible for designing, building, and optimizing data pipelines using open-source technologies, AWS, and ClickHouse (or equivalent OLAP databases) to support evolving business needs.
This is an opportunity to make a real impact by driving efficiency, reliability, and scalability in data-driven decision-making. If you thrive in a hands-on role and enjoy collaborating with engineering teams to create cutting-edge data solutions, we’d love to hear from you.
Key Responsibilities:
Data Pipeline Design: Architect scalable, high-performance data solutions using AWS and OLAP databases like ClickHouse.
ETL/ELT Development: Design, build, and maintain efficient data pipelines for large-scale data ingestion, transformation, and storage.
Cloud Infrastructure: Utilize a combination of open-source and AWS technologies to optimize data workflows.
Data Governance & Quality: Ensure data accuracy, consistency, and reliability through best practices in data governance, lineage, and monitoring.
Performance Optimization: Improve data storage, retrieval, and processing through partitioning, indexing, and query tuning.
Collaboration & Leadership: Work closely with data scientists, analysts, and engineers while mentoring junior team members.
Automation & CI/CD: Implement automated deployment and monitoring strategies for data pipelines.
What We’re Looking For:
- 5+ years of experience in data engineering, with a focus on open-source and cloud-native solutions.
- Expertise in ClickHouse (or similar OLAP databases).
- Strong experience in designing and architecting large-scale data systems.
- Proficiency in Python (or other scripting languages) for data processing and pipeline development.
- Deep understanding of ETL/ELT processes, data modeling (Star Schema, Snowflake Schema), and data warehousing.
- Hands-on experience with orchestration tools like Apache Airflow, Step Functions, or Dagster.
- Familiarity with infrastructure-as-code tools (Terraform, CloudFormation, CDK) is a plus.
- Strong problem-solving skills and ability to work in a fast-paced environment.
- Knowledge of SQL performance tuning, materialized views, and sharding strategies for large datasets.
- Experience with containerization (Docker, Kubernetes) and serverless computing.
- Familiarity with monitoring tools like Prometheus, Grafana, AWS CloudWatch.
This is an exciting opportunity to join a market leading E-Learning provider based in Eindhoven. If you are interested, please apply through the Leap29 website with your updated CV.