.
Machine Learning Data Engineer
  • Kraków
Machine Learning Data Engineer
Kraków, Kraków, Lesser Poland Voivodeship, Polska
ITDS Polska Sp. z o.o.
15. 3. 2025
Informacje o stanowisku

technologies-expected :


  • Python
  • Pyspark
  • AWS
  • Microsoft Azure
  • Google Cloud Platform
  • Spark
  • HDFS
  • HIVE
  • Databricks
  • SQL
  • NoSQL
  • Parquet
  • ORC
  • Avro
  • Git

technologies-optional :


  • Kafka
  • Apache Flink

about-project :


  • As a Data Engineer, you will be working for our client, a leading global financial institution, known for building innovative digital solutions and transforming the banking industry. You will play a key role in supporting their data and digital transformation initiatives by developing and optimizing data engineering processes. Working with cutting-edge technologies, you’ll contribute to the development of robust and scalable data solutions for critical financial services, handling everything from data pipelines to cloud integrations. You’ll be part of a dynamic team working on both greenfield projects and established banking applications.
  • Join us, and enhance data solutions with the latest technologies and tools!
  • Krakow-based opportunity with the possibility to work 80% remote (8 days per month from the office).

responsibilities :


  • Developing and optimizing data engineering processes
  • Building robust, fault-tolerant data solutions for both cloud and on-premise environments
  • Automating data pipelines to ensure seamless data flow from ingestion to serving
  • Creating well-tested, clean code in line with modern software engineering principles
  • Working with cloud technologies (AWS, Azure, GCP) to support large-scale data operations
  • Supporting data transformation and migration efforts from on-premise to cloud ecosystems
  • Designing and implementing scalable data models and schemas
  • Maintaining and enhancing big data technologies such as Hadoop, HDFS, Spark, and Cloudera
  • Collaborating with cross-functional teams to solve complex technical problems
  • Contributing to the development of CI/CD pipelines and version control practices

requirements-expected :


  • Strong experience in the Data Engineering Lifecycle, especially in building data pipelines
  • Proficiency in Python, Pyspark, and the Python ecosystem
  • Experience with cloud platforms such as AWS, Azure, or GCP (preferably GCP)
  • Expertise in Hadoop on-premise distributions, particularly Cloudera
  • Experience with big data tools such as Spark, HDFS, HIVE, and Databricks
  • Knowledge of data lake formation, data warehousing, and schema design
  • Strong understanding of SQL and NoSQL databases
  • Ability to work with data formats like Parquet, ORC, and Avro
  • Familiarity with CI/CD pipelines and version control tools like Git
  • Strong communication skills to collaborate with diverse teams

offered :


  • Stable and long-term cooperation with very good conditions
  • Enhance your skills and develop your expertise in the financial industry
  • Work on the most strategic projects available in the market
  • Define your career roadmap and develop yourself in the best and fastest possible way by delivering strategic projects for different clients of ITDS over several years
  • Participate in Social Events, training, and work in an international environment
  • Access to attractive Medical Package
  • Access to Multisport Program
  • Access to Pluralsight
  • Flexible hours

benefits :


  • sharing the costs of sports activities
  • private medical care
  • flexible working time
  • fruits
  • integration events
  • corporate gym
  • mobile phone available for private use
  • computer available for private use
  • saving & investment scheme
  • no dress code
  • coffee / tea
  • drinks
  • christmas gifts
  • birthday celebration
  • sharing the costs of a streaming platform subscription
  • access to +100 projects
  • access to Pluralsight

  • Praca Kraków
  • Kraków - Oferty pracy w okolicznych lokalizacjach


    89 125
    9 748