Informacje o stanowisku
For our business partner (trading company from USA) , we are looking for a Data Infrastructure Engineer!
Responsibilities on the project:
- Build and run data platform using technologies such as public cloud infrastructure (AWS and GCP), Kafka, databases, and containers.
- Develop data science platform based on open source software and Cloud services.
- Build and run ETL tools and frameworks to onboard data into the platform, define schema, build DAG processing pipelines, and monitor data quality.
- Help develop machine learning development framework and pipelines.
- Manage and run mission-critical production services.
Requirements:
- 3+ years of experience in a Data Platform/Infrastructure Engineering role.
- Strong experience with Python .
- Experience building ETL and stream processing tools and frameworks using Kafka, Spark, Flink, Airflow/Prefect, etc.
- Experience with SQL and databases/engines such as MySQL, PostgreSQL, MS SQL, Snowflake, Redshift, Presto, etc.
- Familiarity with data science stack: e.g. Jupyter, Pandas, Scikit-learn, Dask, PyTorch, MLFlow, Kubeflow, etc.
- Experience using AWS/GCP (S3/GCS, EC2/GCE, IAM, etc.) , Kubernetes , and Linux in production.
- Strong proclivity for automation and DevOps practices .
- Experience with managing increasing data volume, velocity, and variety.
- English at least B2+/C1 level.
Nice to have:
- University degree (preferred in IT or similar).
- Experience with: Java, C++, Rust, Go .
- Understanding of TCP/IP and distributed systems .
- Experience managing time series data .
- Familiarity with working with open source communities .
We offer:
Location: 100% remote.
Wages: 160 PLN/H - 210 PLN/H net + VAT / B2B.
Working hours: Overlap with US at least 5h (11:00/12:00 - 19:00/20:00 Polish time).
#J-18808-Ljbffr
Praca WarszawaWarszawa - Oferty pracy w okolicznych lokalizacjach