ABOUT YOUR JOB

  • Develop big data ETL-Pipelines with Python and Spark (Pyspark) in an agile environment
  • Connect a variety of data sources (SAP-HANA, Google, (non)relational databases, Kafka Stream etc.) and integrate it into our cloud based DataLake (Microsoft Azure)
  • Work closely with the product owner and design the data model to match their needs for sourced tables and structured Datamart’s)
  • Make sure the Data Warehouse keeps its data integrity and availability

ABOUT YOUR PROFILE

  • University degree in economics with a technical focus, business informatics or any other related area of study
  • 3 years of working experience in the data and business intelligence field
  • Very good Python knowledge
  • Solid experience with Apache Spark and big data
  • Experience in agile Software-Development (Git, Jira, Jenkins CI)
  • Know-how of the latest toolsets on Data integration, storage and modelling in the context of Data Warehouse and Data Lake (preferably Databricks, Kafka Streaming, Airflow)
  • Sound understanding of distributed systems

We ware looking for a (Senior) Data Engineer/Architect to initiate, manage and automate our routines to gather (un)structured data from several different data sources and load it into our cloud-based Data Warehouse.

You will work within our Business Intelligence team, which consists of Data Engineers, Data Scientist and Data Analysts. Furthermore, you will work closely with other IT departments to ensure the operation of our data warehouse. Our Data Engineers work with the latest technologies to build our Data Warehouse as the single point of truth for the company.

ABOUT your benefits