Our Client is one of the leading telecommunications operators in Poland, providing mobile telephony, internet, and advanced digital solutions. As part of a large international group, the company focuses on innovation, digital transformation, and continuous technology development to stay competitive in a rapidly evolving market.
Long-term project focused on the migration of an Oracle Data Warehouse to Google Cloud Platform (GCP).
You will join a data-focused team responsible for building modern, scalable data pipelines and ensuring high data quality in a cloud-based environment.
responsibilities :
Build and maintain data ingestion processes from various source systems into a Data Lake
Design, develop, and optimize complex data pipelines ensuring reliable and efficient data flow
Create, enhance, and maintain frameworks supporting data pipeline development
Implement comprehensive testing frameworks for data pipelines
Collaborate closely with data analysts and data scientists to deliver high-quality, trusted data
Ensure strong data governance, security, and compliance practices
Research and implement new technologies to improve pipeline performance
Integrate data from diverse source systems such as Kafka, MQ, SFTP, databases, APIs, and file shares
requirements-expected :
Strong hands-on experience with Google Cloud Platform (GCP)
Advanced knowledge of SQL and PL/SQL
Proficiency in Python
Minimum 3 years of experience in a data-focused role (Data Analyst / Data Quality Analyst / Data Engineer)
Solid understanding of data quality management and data governance practices
Experience with ETL processes and real-time data processing
Hands-on experience with Linux and bash scripting
Practical knowledge of Apache Spark and Apache Kafka (Cloudera stack)
Bachelor’s or Master’s degree in Data Science, Computer Science, Statistics, Economics, or a related field