The project involves integrating and processing data from various sources using ETL tools both in GCP cloud and on-premise. The main goal is to optimize data loading, processing, and exposing via APIs, along with integration into modern database and queuing solutions.
Data Engineer with GCP
Your responsibilities
- Design and optimize ETL processes for on-prem and GCP environments
- Develop and expose REST APIs for seamless data access
- Manage data pipelines and scheduling with tools like Informatica Power Center and Apache AirFlow
Our requirements
- 5+ years of experience in ETL development for data integration on on-premise systems
- Expertise with relational databases
- Ability to develop and expose RESTful APIs
- 3+ years of experience as a Data Engineer
- Proven experience with Google Cloud Platform
- Advanced proficiency in Python for data processing and analysis
What we offer
- B2B contract
- Hybrid job - 1 time per week from office in Warsaw