.
Streaming Data Platform Engineer @ Spyrosoft
  • Wrocław
Streaming Data Platform Engineer @ Spyrosoft
Wrocław, Wrocław, Lower Silesian Voivodeship, Polska
Spyrosoft
22. 1. 2026
Informacje o stanowisku

Project description

Our customer is a leading German producer of customized solutions for the self-supply of solar-powered electricity. This includes photovoltaic, energy storage systems as well as cloud technology systems helping individuals to become energetically independent.

Please note that this position involves occasional on-call duty to resolve potential customer-critical incidents.

Recruitment process:

  • Recruitment screening and experience survey (ca. 1 hour) 
  • Profile-specific offline assignment (approx. 2–3 evenings of concentrated work, depending on experience) 
  • Client technical interview (ca. 1 hour) 
  • Client team interview (ca. 1 hour) 
  • Offer meeting (ca. 30 mins) 

Requirements:

  • InfluxDB 
  • Python/Java 
  • Ansible/Terraform/Terragrunt 
  • Apache Kafka/Confluent Cloud 
  • Docker/Kubernetes/ArgoCD/AKS 
  • SCMS 
  • CI/CD/Gitlab-CI 
  • Cloud 

Required skills:

  • At least 3 years of experience working with InfluxDB. 
  • At least 3 years of experience with programming languages such as Python and Java. 
  • Familiarity with Ansible, Terraform or Terragrunt. 
  • Knowledge of Apache Kafka or Confluent Cloud. 
  • Experience with containers in scalable environments such as Docker, Kubernetes, ArgoCD and AKS. 
  • Experience with SCMS and CI/CD systems like Gitlab-CI. 
  • Knowledge of cloud environments and services such as Azure. 

Project description

Our customer is a leading German producer of customized solutions for the self-supply of solar-powered electricity. This includes photovoltaic, energy storage systems as well as cloud technology systems helping individuals to become energetically independent.

Please note that this position involves occasional on-call duty to resolve potential customer-critical incidents.

Recruitment process:

  • Recruitment screening and experience survey (ca. 1 hour) 
  • Profile-specific offline assignment (approx. 2–3 evenings of concentrated work, depending on experience) 
  • Client technical interview (ca. 1 hour) 
  • Client team interview (ca. 1 hour) 
  • Offer meeting (ca. 30 mins) 
,[Configure and operate our time-series database infrastructure to store and process telemetry data. , Continuously develop and optimize our stream-processing pipelines. , Continuously develop and optimize our Python and Java-based applications that are close to the TSDB. , Automate build and deployment processes in hybrid environments (e.g., public cloud and self-hosted). , Work closely with development teams and architects to enhance performance and features. , Python development to adapt to telemetry database components, API design for scalable infrastructure and API development for interfaces from software development projects. , Manage, scale, and handle access management for telemetry databases.  Requirements: InfluxDB, Python, Java, Ansible, Terraform, Kafka, Cloud, Docker, Kubernetes, ArgoCD, AKS, GitLab CI/CD, Storage, API, Apache Kafka, Azure, CA

  • Praca Wrocław
  • Wrocław - Oferty pracy w okolicznych lokalizacjach


    138 684
    20 326