Project for pharmacy sector - a key project in the field of creating Data Products and reporting solutions supporting decisions for organizations servicing diagnostic instruments.
responsibilities :
Analyze and organize raw data.
Build data systems and pipelines.
Evaluate business needs and objectives.
Interpret trends and patterns.
Conduct complex data analysis and report on results.
Prepare data for prescriptive and predictive modeling.
Build algorithms and prototypes.
requirements-expected :
4+ years of working with programming language focused on data pipelines,eg. Python or R
2+ years working with Snowflake
4+ years of experience working with SQL
3+ years of experience in data pipelines maintenance
3+ years of experience with different types of storage (filesystem, relation, MPP, NoSQL) and working with various kinds of data (structured, unstructured, metrics, logs, etc.)
3+ years of experience in working in data architecture concepts (in any of following areas data modeling, metadata mng., workflow management, ETL/ELT, real-time streaming, data quality, distributed systems)
3+ years of cloud technologies with emphasis on data pipelines (Airflow, Glue, Dataflow - but also other smart solutions of handling data in the cloud - elastic, redshift, bigquery, lambda, s3, EBS etc.)
Very good knowledge of data serialization languages such as JSON, XML, YAML
Excellent knowledge of Git, Gitflow and DevOps tools (e.g. Docker, Bamboo, Jenkins, Terraform
Capability to conduct performance analysis, troubleshooting and remediation (optional)