Salary
💰 CA$100 - CA$120 per hour
Tech Stack
AirflowApacheBigQueryCloudERPETLGoogle Cloud PlatformInformaticaPythonSQLTerraformVault
About the role
- Define and implement enterprise-wide data strategy aligned with business goals, including governance, classification, retention, and privacy policies
- Design conceptual, logical, and physical data models to support analytics and operational workloads; implement star, snowflake, and data vault models
- Implement S4 CDS views in Google BigQuery
- Architect data solutions on GCP using BigQuery, Cloud Storage, Dataflow, and Dataproc and apply cost optimization strategies
- Design and orchestrate ETL/ELT pipelines using Apache Airflow (Cloud Composer) and Dataflow
- Integrate data from multiple systems including SAP BW, SAP HANA, Business Objects using tools like SAP SLT and Google Cortex Framework; leverage Boomi for interoperability
- Develop complex SQL queries for analytics, transformations, and performance tuning and build automation scripts/utilities in Python
- Lead on-premise to cloud migrations for enterprise data platforms, managing migration of SAP datasets to GCP ensuring data integrity and minimal downtime
- Implement CI/CD pipelines for data workflows and apply infrastructure-as-code principles using GitHub Actions, Cloud Build, and Terraform
Requirements
- Strong SAP data integration expertise
- Proven experience with GCP BigQuery, Cloud Storage, Pub/Sub, Dataflow, Dataproc
- Hands-on experience with SAP data extraction, modeling, and integration from ERP, BW, and/or HANA systems
- Experience with SAP SLT and Google Cortex Framework
- Experience with integration tools such as Boomi, Informatica, or MuleSoft
- Experience implementing S4 CDS views in Google BigQuery
- Strong SQL and Python programming skills
- Good understanding of CDS views and ABAP language
- Experience designing and orchestrating ETL/ELT pipelines using Apache Airflow (Cloud Composer) and Dataflow
- Experience with data modeling techniques: star, snowflake, data vault
- Knowledge of data governance frameworks, data classification, retention, and privacy policies (GDPR, HIPAA, PIPEDA)
- Experience leading on-premise to cloud migrations for enterprise data platforms (SAP BW/Bobj)
- Experience implementing CI/CD pipelines for data workflows using GitHub Actions, Cloud Build, and Terraform
- Ability to implement cost optimization strategies for GCP workloads
- Experience collaborating with business stakeholders and engineering teams