Responsible for at-scale infrastructure design, build and deployment with focus on distributed systems
Build and maintain architecture patterns for data processing, workflow definitions, and system-to-system integrations using Big Data and Cloud technologies
Evaluate and translate technical design to workable technical solutions/code and technical specifications
Drive creation of reusable artifacts and establish scalable, efficient, automated processes for data analysis
Define data models, validation and implementation and work closely with analysts/data scientists to understand downstream impact
Write efficient and well-organized software to ship products in an iterative, continual release environment
Contribute and promote good software engineering practices across the team and communicate to technical and non-technical audiences
Define data retention policies, monitor performance and advise necessary infrastructure changes
Requirements
ready to start immediately
3+ years’ experience with Azure (Data Factory, Databricks, SQL, Data Lake, Power BI, DevOps, Delta Lake, CosmosDB)
5+ years’ experience with data engineering or backend/fullstack software development
strong SQL skills
Python scripting proficiency
experience with data transformation tools - Databricks and Spark
experience with data manipulation libraries (such as Pandas, NumPy, PySpark)
experience in structuring and modelling data in both relational and non-relational forms
ability to elaborate and propose relational/non-relational approach, normalization/denormalization and data warehousing concepts (star, snowflake schemas)
designing for transactional and analytical operations
experience with CI/CD tooling (GitHub, Azure DevOps, Harness etc)
good verbal and written communication skills in English
work from the European Union region and a work permit are required
candidates must have an active VAT status in the EU VIES registry
Applicant Tracking System Keywords
Tip: use these terms in your resume and cover letter to boost ATS matches.