Senior Web Data Acquisition Engineer (Resilient Crawling) Miejsce pracy: Rzeszów Technologies we use Optional Python Java Kotlin Docker K8s About the project TL;DR Checklist: •[ Scrape JS & non-JS (headless, VMs) •[ Emulate human like behavior (proxies, rate, CAPTCHA, fingerprints) •[ Parsers resilient to layout changes •[ Monitor & alert on blocks/errors; retry/backoff •[ Python or Java/Kotlin •[ REST APIs creation and consumption •[ Remote-first work All checked? In that case, send us your CV! And if you want more details, read on. Help us turn public web data into clean, timely product intelligence that powers the systems that help our customers across many countries. You will design and scale resilient data acquisition pipelines — combining browser automation and high‑performance collection—to deliver reliable, cost‑efficient data to downstream systems. We operate with a startup mindset: small, senior, fast-moving. The team is energetic, collaborative, and pragmatic. We use AI at multiple layers (change detection, parser generation, anomaly detection, agentic ops)—you’ll both use and shape these production‑ready applications and learn a lot along the way. Department/Team: Technology - ECOM Team Employment: B2B Location: Remote-first (EU-friendly time zones); Warsaw/Rzeszów meetups Your responsibilities Design, build, and run high‑throughput web data pipelines across diverse, modern web stacks (rendered and non‑rendered). Make smart render vs. fetch decisions to balance accuracy, latency, and cost. Create robust extractors/parsers (HTML/DOM/XPath/JSON) with auto‑healing patterns and clear schemas. Engineer resilience against dynamic access controls (rate limits, traffic shaping, session/identity rotation) while following legal/ethical guidelines. Own observability: success rate, freshness, latency, cost per successful unit. Collaborate with AI/Platform to integrate LLM‑assisted maintenance, anomaly detection, and triage bots. Our requirements Expertise in Python or Java/Kotlin; comfort across browser runtimes, and modern web protocols. Deep understanding of headless execution, DOM/JS, and network behavior; experience with browser automation frameworks is a plus. Practical experience handling rate limits/challenges and building systems that degrade gracefully. Solid grounding in data quality: normalization, validation, deduplication, and schema/versioning. Experience with anti‑detection browser stacks/fingerprinting mitigation. Optional Systems mindset: queues/schedulers/workers; containerized deploys (Docker/K8s); CI/CD; secrets management. Auto‑healing selector strategies (heuristics/ML), layout change detection. Multi‑region traffic management and cost optimization. This is how we organize our work This is how we work in house you have influence on the choice of tools and technologies you have influence on the technological solutions applied you have influence on the product What we offer Impact: Your work becomes the backbone of pricing & availability intelligence used across products and markets. Learning: Hands‑on with production AI (LLMs, embeddings, agents) and modern data engineering. Team: Talented, supportive peers; high autonomy and ownership; clear problem statements. Pace & Pragmatism: We ship, measure, and improve. Benefits sharing the costs of sports activities private medical care remote work opportunities flexible working time Recruitment stages Recruitment Interview ZEN.COM sp. z o.o. ZEN.COM is a smart financial app designed for your everyday life – at home and on the go. We make payments, online shopping, and personal finance management fast, secure, and effortless. With ZEN, you can enjoy cashback on purchases, full control over your spending, and peace of mind thanks to purchase protection. A built-in multi-currency account lets you spend abroad or shop internationally with great exchange rates and no hidden fees.