Emporix is an AI-driven Autonomous Commerce Execution Platform that gives enterprises full control over how their commerce works. It orchestrates operations end-to-end—so you can scale faster, respond smarter, and deliver better customer experiences. We are looking for a skilled Machine Learning Engineer to join our team in Poland, with hands-on experience in fine-tuning and deploying custom large and small language models (LLMs/SLMs) to production. In this role, you’ll build and optimize cutting-edge AI models that power intelligent commerce experiences. If you’re excited about production-grade AI, retrieval-augmented generation (RAG), and cloud-native ML engineering, we’d love to hear from you.
What You’ll Be Doing
LLM & SLM Development
- Fine-tune and optimize LLMs and SLMs on domain-specific datasets to enhance relevance and performance.
- Build and iterate on custom models tailored to business-specific needs and customer workflows.
- Evaluate models for accuracy, latency, cost, and robustness in real-world scenarios.
RAG & System Integration
- Design and implement retrieval-augmented generation (RAG) pipelines.
- Work closely with backend engineers to integrate models into microservice-based systems via APIs and orchestration layers.
Production-Grade ML Operations
- Deploy and monitor ML models in cloud environments such as GCP or equivalent platforms.
- Build scalable, reproducible training and inference pipelines using MLOps best practices.
- Track performance with monitoring, logging, and feedback loops for continuous improvement.
Collaboration and Optimization
- Collaborate with AI developers, product managers, and domain experts to understand business use cases and translate them into ML solutions.
Documentation
- Write and maintain clean, well-structured documentation for LLMs/SLMs.
- Contribute to knowledge bases and architecture diagrams for long-term maintainability and onboarding.
What We’re Looking For
- At least 3+ years of hands-on experience in machine learning, with a focus on NLP and language models.
- Proven experience with LLM/SLM fine-tuning, training, and optimization, including LoRA, QLoRA, or similar techniques.
- Strong understanding of transformer architectures and experience working with open-source models (e.g., Hugging Face, Mistral, LLaMA, Gemma).
- Experience designing and deploying RAG systems in production environments.
- Hands-on experience with Python, PyTorch or TensorFlow, and ML frameworks and libraries.
- Experience deploying and operating ML models in cloud environments, ideally GCP.
- Familiarity with vector databases, embedding models, and semantic search.
- Ability to bridge research and engineering: experiment fast, ship stable solutions.
- Strong English communication skills (B2/C1 or above).
Why You’ll Love It Here
- Real Impact: Your work will directly power next-gen autonomous commerce experiences used by top-tier businesses.
- Cutting-Edge Technology: Work with advanced AI frameworks like LangChain and LangGraph, and bring LLM-driven features to production in a cloud-native environment.
- Autonomy and Innovation: Explore agentic AI, RAG systems, and next-gen architectures with the freedom to propose and test new ideas.
- Collaborative Culture: Join a team of talented professionals who value knowledge-sharing, open communication, and a passion for building exceptional systems.
- Flexible Work Setup: Enjoy a hybrid work model that promotes flexibility and personal well-being while encouraging collaboration and meaningful team connections.
- Impactful Contribution: Be part of a company where your work directly impacts the success of a next-generation commerce platform used by top-tier businesses.