Architect and lead the development of large-scale systems for data pipelines, data lakes that handle billions of daily events.
Design and implement solutions that ensure data is available, secure, and scalable across the platform, enabling real-time and batch processing.
Make high-level architectural decisions about system design, technology choices, and platform evolution, ensuring scalability and long-term sustainability.
Collaborate with key stakeholders, such as product teams, data engineers, back-end developers, and ML engineers, to build tools & frameworks that power analytics, product features, and data-driven workflows.
Work with business stakeholders such as Analytics Eng, Data Science teams to build high-impact data products enabling business-critical features, research, and experimentation with a focus on efficiency and alignment.
Collaborate with Leadership in devising the charter and technical roadmaps.
Mentor engineers across back-end infrastructure and data engineering disciplines, fostering a culture of collaboration and technical excellence.
Requirements
Brings at least 10 years of experience in large-scale distributed computing systems, with expertise in both data engineering and data modeling.
Has 5+ years of experience managing live production environments, including high-load systems or data-intensive workflows with a focus on uptime and reliability.
Expert in SQL and proficient in technologies like Spark, Kafka, Terraform, and in at least one programming language commonly used for data engineering (e.g., Python, Scala, or Java).
Strong knowledge of ETL/ELT design patterns, orchestration tools (e.g., Airflow, dbt, Dagster), and data quality frameworks.
Ability to design scalable, secure, and maintainable data models and architectures, and understanding of data governance, privacy regulations (GDPR/CCPA).
Hands-on experience with modern data storage technologies (e.g., Delta Lake, Snowflake, BigQuery, Redshift).
Is a strategic thinker, able to evaluate whether to build internally or leverage third-party solutions, ensuring the right balance between speed, scalability, and long-term viability.
Is an excellent communicator and collaborator, able to work across diverse teams and stakeholders, translates business needs into robust data solutions, and ensures alignment on technical goals.
Has a demonstrated ability to work independently with minimal guidance, proactively manages tasks and priorities across multiple projects, analyzes and executes work efficiently, collaborates effectively with cross-functional teams, and thrives in fast-paced, results-driven environments.
Embodies our EAGER values—is ethical, adaptable, gritty, empathetic, and remarkable.
Is inspired by our MOVE principles: move fast and learn faster; obsess about creating customer value; value impact over activity; and embrace healthy disagreement rooted in trust.
Benefits
Excellent health care (including a wide range of medical, dental, vision, mental health, and fertility benefits)
Disability and life insurance options
401(k) and RRSP matching
Paid parental leave
20 days of paid time off per year, 12 days of paid holidays per year, two floating holidays per year, and flexible sick time
Generous stipends (including those for caregiving, pet care, wellness, your home office, and more)
Annual professional development budget and opportunities
Applicant Tracking System Keywords
Tip: use these terms in your resume and cover letter to boost ATS matches.
Hard skills
data engineeringdata modelingSQLSparkKafkaTerraformPythonScalaJavaETL/ELT design patterns