Tech Stack
CloudKafkaPythonSQL
About the role
- Design and own real‑time and batch pipelines (Confluent Cloud/Kafka, Flink SQL, Snowflake, SQLmesh) with reliability and observability built‑in.
- Model domain‑rich schemas for sports/video/ticketing/subscriptions/ads that serve both analytics and ML feature needs.
- Implement data quality, contracts, testing, and lineage (dbt tests, Great Expectations, OpenLineage/Marquez or similar).
- Drive SLOs, alerting, and remediation playbooks to minimize data downtime and MTTR; lead blameless postmortems.
- Scale event ingestion/CDC from sources like GA4/Segment, app stores, payments (Stripe), 3rd‑party sports systems; manage schema evolution.
Requirements
- 8–12+ years in data engineering (or equivalent) with ownership of mission‑critical pipelines and data products at scale.
- Expert SQL and strong Python; deep knowledge of incremental processing, partitioning, compaction, and schema design.
- Hands‑on with real‑time systems (Kafka + Flink or similar) and modern ELT in a warehouse (Snowflake preferred).
- Proven reliability practice: contracts, tests, lineage, SLAs/SLOs, and effective incident response.
- Experience modeling complex domains and enabling both BI and ML/feature store use cases.
- Clear, concise design docs and stakeholder communication; ability to lead cross‑team initiatives.
- Multiple medical insurance plans to choose from
- Dental, vision life and disability insurance
- Employee Emergency Fund
- Company equity (stock options)
- Open PTO policy
- 401K plan with company match
- Hybrid/flexible work environment
Applicant Tracking System Keywords
Tip: use these terms in your resume and cover letter to boost ATS matches.
Hard skills
SQLPythonKafkaFlink SQLSnowflakedbtGreat ExpectationsOpenLineagedata qualityschema design
Soft skills
stakeholder communicationleadershipcross-team initiativesdesign documentationincident response