Our toolchain democratizes access to data for everyone and makes it easy and painless to run experiments to establish cause and effect. The team focus is on the complete data life-cycle to ensure any data leaving Kiwi.com is of the highest quality.

If you are interested in putting our data loads and collectors on the next level including both the batch and real-time processing of our data routines, you are the ONE we looking for! If you love to experiment and build on top technologies like airflow, custom python apps or anything else from the open source world come to see us!

Few examples of our Data Engineers’ work:

Data workflow management: to manage our data loads for the Analytics world we’re using the Apache Airflow. Apache Airflow enables scheduling data-related workflows with a code-as-configuration model and web front end, we driving our data routines to feed up data provisioning customers.

Real-time streaming infrastructure: to enable our analytics teams to move quickly, getting accurate data with minimal delay is a core focus in data provisioning & engineering. Currently, we are building out real-time infrastructure to allow for easy development of streaming applications that includes anomaly detections and forecasts.

Interactive dimensional analysis: our data analysts have a strong need to query data and compute aggregates on various dimensional cuts in “yesterday was too late” frame. To address this, we are building a query tool stack to allow users to interactively slice-and-dice large datasets.

What will you do?

  • Develop, monitor and support our data workflow management environments and ELT/ETL routines tooling as a service, as well as decommission any no longer used service/tool in order to perform reasonable data associated with those decommissioned tools
  • Provide continuous support on data workflow management end ETL jobs for our data infrastructure services; maintain and provide all relevant information on current infrastructure and tools
  • Educate on the current toolings and data used within the data provisioning stack in order to make the access easier for anyone in the company
  • Regularly update and clearly communicate on the team achievements and main projects progress

What we expect?

  • 2+ years of full-time, industry experience
  • Strong coding skills in Python
  • Experienced & interested in technologies like Airflow, Postgres, Redis/Kafka, or Presto
  • Working knowledge of relational databases and query authoring (SQL)
  • Working with batch and real-time data processing routines
  • Rigor in high code quality, automated testing, and other engineering best practices
  • Operations of robust distributed systems in cloud (AWS, Google Cloud) is the best fit
  • BS/MS in Computer Science or a related field (ideal)

Why it rocks to be at Kiwi.com?

  • We deploy immediately after a job is completed, not after months of QA
  • Do, fail, learn – repeat! We understand that mistakes happen and we learn fast
  • We decide which cutting-edge technologies are appropriate for the task
  • We love contributing to and using Open Source
  • We visit and speak at conferences and tech events worldwide
  • We code at hackathons and other competitions
  • We support the local tech community
  • We use our work time wisely with a very friendly vacation policy and work schedule
  • We also like to party and hang out together whilst enjoying benefits such as 6 weeks of paid vacation, gym&wellness membership pass, private health insurance, flight vouchers…
  • Besides a fair salary, we can also look forward to quarterly bonuses dependent on our performance

Interested? Join us and hack the traditional ways of travel!

Recruiter

Gabriela Manová

Your manager to be

Michael Štencl

Brno, Czech Republic | Full-Time
Bratislava, Slovakia | Full-Time
Brno, Czech Republic | Full-Time
Brno, Czech Republic | Full-Time
Brno, Czech Republic | Full-Time