- snowflake
- Python
- ETL
- SQL
- Git
- Gitflow
- Docker
- Bamboo
- Jenkins
- Terraform
technologies-optional :
- DBT
about-project :
- Project for pharmacy sector - a key project in the field of creating Data Products and reporting solutions supporting decisions for organizations servicing diagnostic instruments.
responsibilities :
- Analyze and organize raw data.
- Build data systems and pipelines.
- Evaluate business needs and objectives.
- Interpret trends and patterns.
- Conduct complex data analysis and report on results.
- Prepare data for prescriptive and predictive modeling.
- Build algorithms and prototypes.
requirements-expected :
- 4+ years of working with programming language focused on data pipelines,eg. Python or R
- 2+ years' working with Snowflake
- 4+ years of experience working with SQL
- 3+ years of experience in data pipelines maintenance
- 3+ years of experience with different types of storage (filesystem, relation, MPP, No SQL) and working with various kinds of data (structured, unstructured, metrics, logs, etc.)
- 3+ years of experience in working in data architecture concepts (in any of following areas data modeling, metadata mng., workflow management, ETL/ELT, real-time streaming, data quality, distributed systems)
- 3+ years of cloud technologies with emphasis on data pipelines (Airflow, Glue, Dataflow - but also other smart solutions of handling data in the cloud - elastic, redshift, bigquery, lambda, s3, EBS etc.)
- Very good knowledge of data serialization languages such as JSON, XML, YAML
- Excellent knowledge of Git, Gitflow and Dev Ops tools (e.g. Docker, Bamboo, Jenkins, Terraform
- Capability to conduct performance analysis, troubleshooting and remediation (optional)
- Excellent knowledge of Unix