We are seeking a skilled Data Engineer to build and maintain our data warehouse and efficiently process data from various 3rd party platforms, ensuring it is transformed into well-structured data products for further business use. You will play a pivotal role in shaping the data infrastructure for a finance-related project, focusing on transforming raw data into reliable, accessible formats that meet the organization’s needs.
As a Data Engineer, you will design and maintain scalable data pipelines that ingest and transform data from external platforms. You’ll ensure data quality, reliability, and availability to support business requirements. You will collaborate directly with engineering, product, and business teams to understand their data needs and deliver tailored technical solutions that address those requirements.
We are seeking a proactive individual who can be responsible for a big part of our data platform, drive innovation, and continuously enhance the data architecture to support the growing needs of the business. Your ability to effectively communicate technical concepts, provide constructive feedback, and facilitate knowledge sharing among team members at all levels will be highly valued.
About the project:
The client is a renowned venture capital firm that manages its fund using Salesforce as the central data repository, alongside various other platforms for additional data sources. These sources range from Excel spreadsheets to external commercial tools and custom-built user-facing applications. Different departments rely on multiple siloed systems for data management, leading to a fragmented data landscape.
The major challenge is the suboptimal time to make the final decision as analysts cant rely on the existing data due to possible human errors and complicated data management. The solution is to build a unified data platform that will be integrated with Salesforce, fund accounting, and other existing systems to efficiently transform financial, legal, and ops data to data marts that are ready for interaction and comprehensive visualization with BI tools on the web and mobile platforms.
Tech stack: Google Cloud, BigQuery, GCS, DBT, SQL, Airbyte, Airflow, Python, FastAPI, Pandas, Terraform, GitHub Actions, Sigma Computing, Open Metadata, Monte Carlo Data, OpenAI, LLMs
Location: Remote Poland/Europe
Nice to have
We are seeking a skilled Data Engineer to build and maintain our data warehouse and efficiently process data from various 3rd party platforms, ensuring it is transformed into well-structured data products for further business use. You will play a pivotal role in shaping the data infrastructure for a finance-related project, focusing on transforming raw data into reliable, accessible formats that meet the organization’s needs.
As a Data Engineer, you will design and maintain scalable data pipelines that ingest and transform data from external platforms. You’ll ensure data quality, reliability, and availability to support business requirements. You will collaborate directly with engineering, product, and business teams to understand their data needs and deliver tailored technical solutions that address those requirements.
We are seeking a proactive individual who can be responsible for a big part of our data platform, drive innovation, and continuously enhance the data architecture to support the growing needs of the business. Your ability to effectively communicate technical concepts, provide constructive feedback, and facilitate knowledge sharing among team members at all levels will be highly valued.
About the project:
The client is a renowned venture capital firm that manages its fund using Salesforce as the central data repository, alongside various other platforms for additional data sources. These sources range from Excel spreadsheets to external commercial tools and custom-built user-facing applications. Different departments rely on multiple siloed systems for data management, leading to a fragmented data landscape.
The major challenge is the suboptimal time to make the final decision as analysts cant rely on the existing data due to possible human errors and complicated data management. The solution is to build a unified data platform that will be integrated with Salesforce, fund accounting, and other existing systems to efficiently transform financial, legal, and ops data to data marts that are ready for interaction and comprehensive visualization with BI tools on the web and mobile platforms.
Tech stack: Google Cloud, BigQuery, GCS, DBT, SQL, Airbyte, Airflow, Python, FastAPI, Pandas, Terraform, GitHub Actions, Sigma Computing, Open Metadata, Monte Carlo Data, OpenAI, LLMs
Location: Remote Poland/Europe
,[Collaborate with cross-functional teams to understand business requirements , Design, implement, and maintain scalable and reliable data pipelines, data warehouses, and data lakes. , Develop and enforce best practices for data governance, data quality, and data security. , Help maintain code quality, organization, and automation. , Collaborate with other teams as needed to ensure interoperability, stability, and code reusability. , Optimize data processing and querying for better performance and cost-efficiency. , Stay up-to-date with the latest trends and technologies in the data engineering field (Modern Data Stack) and propose improvements to the existing architecture. Requirements: Data warehouse, Data pipelines, Google cloud, BigQuery, dbt, SQL, Python, pandas, PostgreSQL, Redis, Docker, Communication skills, Data engineering, Data warehouses, Data Lake, BI, Terraform, FastAPI, Airflow Additionally: Knowledge sharing.