In Cyclad we work with top international IT companies in order to boost their potential in delivering outstanding, cutting edge technologies that shape the world of the future. For our customer, leader in consulting, technology services and digital transformation we are seeking an Solution Architect to join our international team.
Solution Architect
Numer referencyjny: #2299
Project information:
- Location: 100% remotely (with occasional visits to the office in Warsaw-once every two months)
- Type of employment: B2B contract
- Renumeration: up to 210 PLN NET/H
- Project language: English
Responsibilities:
- Designing new solutions and taking initiatives to improve existing solutions within data platforms - both as part of orders from the business (functional changes) and from technology (architectural changes)
- Coordination of architectural implementation with client-side Architects and other suppliers
- Coordinating the work of a team of Data Engineers (responsible for the development of data platforms and ETL/ELT processes) and Data Analysts (responsible for data modeling and report development) within multiple projects implemented in parallel for the same client; close cooperation with Project Managers
- Optimization of the data solution production process by introducing process and tool changes (code management, tests, documentation; selection of programming tools)
- Ensuring adherence to existing platform development standards, as well as setting them (where gaps are identified) and improving them
Requirements:
- At least 5 years of experience designing and delivering scalable BI, ETL/ELT, DWH, data lake, or big data architectures
- Hands-on experience working with data services in a AWS cloud environment
- AWS Redshift (Redshift Spectrum / external tables, stored procedures, performancedriven table design (sort keys, distribution styles, …), materialized views, temporary tables,)
- AWS S3 experience (storage classes/tiers, S3 buckets, prefixes, objects versioning)
- Apache Spark 9AWS Glue or other)
- AWS EMR, Databricks, Azure Synapse Spark Pools,PySpark)
- Git and Parquet knowledge
- Proficient in both SQL and Python for data processing and analysis
- Hive Metastore (HMS; AWS Glue Data Catalog, Databricks, Apache Nifi, Presto, Apache Atlas, Hortonworks DataPlane, Cloudera Navigator, …)
Additional qualifications that will be an advantage:
- Airflow
- AWS CloudFormation
- Ansible
- Azure Resource Manager
- Chef
- GCP Deployment Manager
- Terraform
- AWS CodePipeline
- Bitbucket Pipelines
- GitHub Actions
- GitLab Pipelines
- Jenkins
- TeamCity
- Travis CI
- AWS Glue
- AWS Lambda
- Azure Functions
- Google Cloud Functions
- Azure Synapse
- Databricks
- Google Cloud BigQuery
- Snowflake
- AWS Step Functions
- dbt
- Delta Lake,Apache Iceberg or Apache
- Hudi
- Hadoop File System (HDFS)
- Scala
- Data Lakehouse
- Data Governance
- Data Quality
- Data Lineage / Data Provenance
- Streaming Data / Real-time Data
- Star Schema / Dimensional Modelling / Kimball
- Data Vault
- Common Data Model / Corporate Data Model
- Master Data Management
We offer:
- B2B contract with rate up to 210 PLN NET/H
- 100% remote job
- Additional benefits
- Innovative working environment
- Opportunity to develop professional skills