The project involves integrating and processing data from various sources using ETL tools both in GCP cloud and on-premise. The main goal is to optimize data loading, processing, and exposing via APIs, along with integration into modern database and queuing solutions.
responsibilities :
Design and optimize ETL processes for on-prem and GCP environments
Develop and expose REST APIs for seamless data access
Manage data pipelines and scheduling with tools like Informatica Power Center and Apache AirFlow
requirements-expected :
5+ years of experience in ETL development for data integration on on-premise systems
Expertise with relational databases
Ability to develop and expose RESTful APIs
3+ years of experience as a Data Engineer
Proven experience with Google Cloud Platform
Advanced proficiency in Python for data processing and analysis
offered :
B2B contract
Hybrid job - 1 time per week from office in Warsaw