We are looking for an independent and motivated individual who is not afraid of taking ownership of complex tasks, creates working solutions quickly, can creatively come up with innovative ideas, but also knows how and when to use already existing solutions. As a data engineer, your mission is to build a strong data foundation on which our ML products can stand and thrive.
Challenges:
- Become an essential part of AI Services team focused on building and operating ML applications for customers across DT group.
- Proactively identify opportunities to improve our data infrastructure and processes, proposing and implementing innovative solutions.
- Develop and maintain our data processing and ETL pipelines.
- Design, maintain and update data models.
- Integrate data from various internal and external sources.
- Build and maintain scalable, efficient, and robust data infrastructure solutions for our ML products.
- Collaborate with data scientists and front-end developers in creating curated and production ready datasets for training our models and building visualization tools.
- Work on improving data quality and implementing automated data quality monitoring solutions.
- Contribute to development and adoption of best data engineering practices.
- Stay up to date with recent data engineering technologies and developments.
- Collaborate with engineers from other teams and make decisions about the future of our technical stack and infrastructure.
- Write clean, production ready, well-documented and tested code.
Skills:
- Bachelor or Master of Science in Computer Science or similar fields
- 5+ years of experience as a data engineer or a Python software developer with focus on data-heavy tasks
- Fluent in English (German / Czech / Slovak is a plus)
- Advanced SQL knowledge
- Advanced database systems knowledge (SQL, NoSQL, database optimization techniques)
- Advanced Python knowledge (OOP, design patterns, testing, good coding practices and popular data processing libraries)
- Working experience with data modeling
- Working experience with ETL/ELT processes and development of data pipelines
- Experience with using and/or building APIs and integrating data from various sources.
- Knowledge of orchestration tools such as Dagster or Airflow
- Knowledge of cloud data solutions (preferably GCP, but AWS or Azure are also fine)
- Working knowledge of git (we use Gitlab)
- Strong analytical skills, ability to understand complex data structures and produce efficient data solutions.
- Independence and resilience when facing obstacles.
- Good communications skills, positive attitude towards working in an international environment.
Bonuses:
- Experience with agile work (Scrum)
- Working knowledge of Gitlab CI/CD
- Working knowledge of Docker and/or Kubernetes
- Experience with modern data warehouses (BigQuery, Snowflake)
- Experience with data engineering tools such as Databricks, Spark or Kafka
- Experience developing machine learning applications