.
Data Engineer
  • Kraków
Data Engineer
Kraków, Kraków, Lesser Poland Voivodeship, Polska
HAVI Global Business Services
5. 10. 2025
Informacje o stanowisku

technologies-expected :


  • Azure Databricks
  • PySpark
  • Scala
  • Spark SQL
  • SQL
  • Delta Lake
  • Azure Data Factory
  • ADLS Gen2
  • Key Vault
  • Great Expectations
  • Deequ
  • pytest
  • dbx
  • Azure DevOps
  • GitHub Actions
  • Terraform
  • Bicep
  • Azure Monitor
  • Log Analytics

about-project :


  • Data Engineer, Integrated Planning Analytics and Insights
  • Architect, design, implement, enhance, and maintain highly scalable, available, secure, and elastic cloud-ready data solutions using cutting-edge technologies to support our predictive and prescriptive analytics needs. Be an expert in our data domains, act as a trusted partner and advisor to solutions architects and data scientists and become a crucial part of the analytics solution lifecycle – from prototype to production and operations of our data science and advanced analytics solutions in areas such as promotions, supply and demand planning, item/menu level analytics, supply chain simulations, and optimization, competitive benchmarking, and root cause analysis. Continuously improve and advance our data solutions.

responsibilities :


  • Responsible for working with the data management, data science, decision science, and technology teams to address supply chain data needs in demand and supply planning, replenishment, pricing, and optimization
  • Develops/refines the data requirements, designs/develops data deliverables, and optimizes data pipelines in non-production and production environments
  • Designs, builds, and manages/monitors data pipelines for data structures encompassing data transformation, data models, schemas, metadata, and workload management. The ability to work with both IT and business
  • Integrates analytics and data science output into business processes and workflows
  • Builds and optimizes data pipelines, pipeline architectures, and integrated datasets. These should include ETL/ELT, data replication/CI-CD, API design, and access
  • Works with and optimizes existing ETL processes and data integration and preparation flows and help move them to production
  • Works with popular data discovery, analytics, and BI and AI tools in semantic-layer data discovery
  • Adepts in agile methodologies and capable of applying DevOps and DataOps principles to data pipelines to improve communication, integration, reuse, and automation of data flows between data managers and data consumers across the organization
  • Implements Agentic AI capability to drive efficiency and opportunity

requirements-expected :


  • Bachelor’s degree in computer science, data management, information systems, information science or a related field; advanced degree in computer science, data management, information systems, information science or a related field preferred.
  • 3+ years in data engineering building production data pipelines (batch and/or streaming) with Spark on cloud.
  • 2+ years hands-on Azure Databricks (PySpark/Scala, Spark SQL, Delta Lake) including:
  • Delta Lake operations (MERGE/CDC, OPTIMIZE/Z-ORDER, VACUUM, partitioning, schema evolution).
  • Unity Catalog (RBAC, permissions, lineage, data masking/row-level access).
  • Databricks Jobs/Workflows or Delta Live Tables.
  • Azure Data Factory for orchestration (pipelines, triggers, parameterization, IRs) and integration with ADLS Gen2, Key Vault.
  • Strong SQL across large datasets; performance tuning (joins, partitions, file sizing).
  • Data quality at scale (e.g., Great Expectations/Deequ), monitoring and alerting; debug/backfill playbooks.
  • DevOps for data: Git branching, code reviews, unit/integration testing (pytest/dbx), CI/CD (Azure DevOps/GitHub Actions).
  • Infrastructure as Code (Terraform or Bicep) for Databricks workspaces, cluster policies, ADF, storage.
  • Observability & cost control: Azure Monitor/Log Analytics; cluster sizing, autoscaling, Photon; cost/perf trade-offs.
  • Proven experience collaborating with cross-functional stakeholders (analytics, data governance, product, security) to ship and support data products.

offered :


  • Possibility of turning your own ideas into success
  • Diverse development opportunities
  • Varied and interesting field of work
  • Responsible task with plenty of leeway
  • Collegial working atmosphere
  • Open corporate culture
  • Cooperation with a dynamic team
  • Attractive remuneration models with performance-related pay
  • Flat hierarchies and short decision-making processes
  • Successful and rapidly growing employer
  • Comprehensive, individual familiarization with the work
  • Offer for health promotion
  • Modern work equipment
  • Diverse development opportunities in an international environment
  • Training according to training schedule and training regulations in the relevant specialist field

benefits :


  • private medical care
  • sharing the costs of foreign language classes
  • life insurance
  • remote work opportunities
  • flexible working time
  • fruits
  • retirement pension plan
  • saving & investment scheme
  • no dress code
  • coffee / tea
  • parking space for employees
  • extra social benefits
  • christmas gifts
  • employee referral program
  • charity initiatives
  • modern office
  • annual bonuses
  • lunch subsidies
  • language courses reimburstments

  • Praca Kraków
  • Kraków - Oferty pracy w okolicznych lokalizacjach


    128 331
    24 626