
Senior Data Engineer – Databricks
Fluent, Inc
full-time
Posted on:
Location Type: Hybrid
Location: Toronto • 🇨🇦 Canada
Visit company websiteSalary
💰 CA$110,000 - CA$150,000 per year
Job Level
Senior
Tech Stack
ApacheAWSKafkaPySparkSparkSQLUnity
About the role
- Design, build, and support scalable real-time and batch data pipelines using PySpark and Spark Structured Streaming.
- Develop pipelines following the Bronze → Silver → Gold architecture using Delta Lake and Enterprise Data Model best practices.
- Integrate with Kafka for event-driven ingestion and stream processing.
- Orchestrate workflows with Databricks Workflows/Jobs and DABs.
- Implement monitoring and observability—Databricks metrics, dashboards, and alerts to ensure pipeline reliability and performance.
- Collaborate cross-functionally in agile sprints with Product Managers, Data Scientists, and downstream data consumers.
- Partner closely with Data Architects to translate Enterprise Data Models into performant physical data models.
- Write clean, modular, and version-controlled code in Git-based CI/CD environments; perform rigorous peer reviews.
- Implement robust logging, error handling, and data quality validation throughout pipelines.
- Utilize AWS services (S3, IAM, Secrets Manager) for storage and infrastructure.
- Evangelize engineering best practices through brown bags, tech talks, and documentation.
- Stay current on emerging trends within the Databricks and data engineering ecosystem.
Requirements
- 5+ years of experience in Data Engineering, including strong Spark (PySpark) and SQL expertise.
- 3+ years of hands-on experience building pipelines on Databricks (Workflows, Notebooks, Delta Lake).
- Deep understanding of Apache Spark distributed processing model and internals.
- Strong experience with streaming data architectures and event-driven processing using Kafka.
- Familiarity with Databricks metrics, observability, and monitoring features.
- Understanding of Unity Catalog and Lakehouse architecture.
- Knowledge of idempotent processing patterns and robust data modeling.
- Proficiency in Git-based, CI/CD-driven development workflows.
- Strong debugging, optimization, and performance tuning skills.
- Proven experience building large-scale data pipelines handling massive volumes of data.
Benefits
- Competitive compensation
- Ample career and professional growth opportunities
- New Headquarters with an open floor plan to drive collaboration
- Health, dental, and vision insurance
- Pre-tax savings plans and transit/parking programs
- 401K with competitive employer match
- Volunteer and philanthropic activities throughout the year
- Educational and social events
- The amazing opportunity to work for a high-flying performance marketing company!
Applicant Tracking System Keywords
Tip: use these terms in your resume and cover letter to boost ATS matches.
Hard skills
PySparkSpark Structured StreamingSQLDelta LakeKafkaGitCI/CDdata quality validationdebuggingperformance tuning
Soft skills
collaborationcommunicationagile methodologypeer reviewsengineering best practices