Tech Stack
ApacheAzureCloudETLPythonSparkSQL
About the role
- Gather and analyze business requirements; advise business and external IT suppliers and participate in project meetings to ensure compliance with IT Governance standards and the Institution’s technical stack
- Create and maintain Enterprise Data Warehouses (EDW) and Business Intelligence solutions, including Data Lakes and Data Lakehouses
- Develop and maintain data pipelines and foundational datasets to support analytics, modeling, experimentation, and business/product needs
- Design and build database architectures for large and complex datasets, balancing usability, computational efficiency, and cost
- Design and develop reporting applications and scalable dashboards, primarily using Power BI, to support business objectives and enable data-driven decision-making
- Design cloud (primarily Azure), hybrid, and on-premises architectures, including ETL processes using corporate-approved tools such as Microsoft Synapse
- Perform physical database design and collaborate with DevSecOps teams for DB performance analysis
- Design data quality and performance monitoring processes
- Produce technical documentation, including architecture overviews and technical specifications
- Support the Service Manager of the Institution’s Data Platform in daily activities
- Collaborate closely with data science, engineering, and DevSecOps teams to enhance instrumentation and monitoring coverage, accuracy, and reliability
- Assist DevSecOps teams with system deployment and configuration
Requirements
- Master's degree with minimum of 9 years of relevant experience (or Bachelor's degree with minimum of 13 years of relevant experience)
- Minimum 2 years of experience in designing ETL/data pipelines using Microsoft Synapse or certification on Microsoft Synapse (e.g., Microsoft Azure Data Engineer Associate) or minimum 3 years hands-on with similar market products
- At least 2 years of experience with data warehouse, data lake, and data lakehouse design
- Minimum 2 years of experience with relational database systems applied to data warehouses
- Excellent ability to produce requirements documents for data analytics systems
- Strong knowledge of non-relational database technologies; specific knowledge of vector databases used to store embeddings in AI-based solutions (e.g., RAG) is considered an asset
- At least 2 years of experience with access rights management in data warehouses/data lakes, including row-level access segregation based on user roles
- Minimum 2 years of experience with SQL
- Very good knowledge of Python and Apache Spark; knowledge of R is considered an asset
- Minimum 2 years of experience in business intelligence reporting tools; specific knowledge of Microsoft Power BI is considered an asset
- At least 1 year of experience with cloud-based design; knowledge of Microsoft Azure is considered an asset
- Minimum 2 years of experience with online analytical processing (OLAP) and data mining tools
- Good knowledge of written and spoken English (B2 level)