Wrocław, Wrocław, Lower Silesian Voivodeship, Polska
Comscore
18. 9. 2025
Informacje o stanowisku
technologies-expected :
Spark
AWS
SQL
Java
Python
Scala
Git
about-project :
We have multiple Java + Spark, Scala + Spark, Python + Spark teams and we may try to match you to multiple teams or just find you single best fit depends on your skills and experience.
responsibilities :
Design, implement, and maintain petabyte-scale big data pipelines using Spark (Java, Python, or Scala - depending on the team), Apache Airflow, Kubernetes, and other technologies
Optimize performance - working with big data is highly specific: sometimes IO-bound, sometimes CPU-bound. You’ll help figure out the most efficient approaches
Collaborate closely with other big data teams
Work with technologies such as AWS, Kubernetes, Airflow, EMR, Hadoop, Linux/Ubuntu, Kafka, and Spark
requirements-expected :
Solid understanding of Spark basics, building blocks, and mechanics
Strong knowledge of Python, Java, or Scala
1+ years of experience with Spark
Good SQL skills
Understanding of data quality issues in large datasets
1+ years of experience with Linux
Professional working proficiency in English
Understanding of HTTP API communication patterns and protocol