4+ years of software development experience in writing performant, commercial-grade systems and applications
Experience with monitoring and troubleshooting production environments
Proficiency in programming languages used in high volume data processing and applications like Java or Scala and Python
Experience building data pipelines with distributed compute frameworks like Hadoop. Spark, orDask
Knowledge of Linux/Unix systems, Docker/Kubernetes and CI/CD including scripting in Python or other scripting languages to automate build and deployment processes
Knowledge of professional software engineering practices & software development life cycle, including coding standards, code reviews, source control management, build processes, testing, and operations
Leverages best practices and past experiences to mentor and improve the productivity of the team
We’d particularly love it if you have
Deep experience building and debugging distributed data pipelines
Experience with columnar databases and storage formats like Delta Lake and Parquet
Experience deploying and managing services on Kubernetes
Experience building with Rust
If you don’t meet 100% of the above qualifications, you should still seriously consider applying.