
Senior Data Engineer
Zeta Global
full-time
Posted on:
Location Type: Remote
Location: United States
Visit company websiteExplore more
Salary
💰 $165,000 - $175,000 per year
Job Level
Tech Stack
About the role
- Build data pipelines: Develop robust batch and streaming pipelines (Kafka/Kinesis) to ingest, transform, and enrich large-scale event data (impressions, clicks, conversions, costs, identity signals).
- Create data aggregates & marts: Design and maintain curated aggregates and dimensional models for multiple consumers—prediction models, agents, BI dashboards, and measurement workflows.
- Data modeling & contracts: Define schemas, data contracts, and versioning strategies to keep downstream systems stable as sources evolve.
- Data quality & reliability: Implement validation, anomaly detection, backfills, and reconciliation to ensure completeness, correctness, and timeliness (SLAs/SLOs).
- Performance & cost optimization: Optimize compute/storage for scale (partitioning, file sizing, incremental processing, indexing), balancing latency, throughput, and cost.
- Orchestration & automation: Build repeatable workflows with scheduling/orchestration (e.g., Airflow, Dagster, Step Functions) and CI/CD for data pipelines.
- Observability for data systems: Instrument pipelines with metrics, logs, lineage, and alerting to accelerate detection and root-cause analysis of data issues.
- Security & governance: Apply least-privilege access, PII-aware handling, and governance controls aligned with enterprise standards.
Requirements
- 5+ years building production data pipelines and data products (batch and/or streaming) in a high-scale environment.
- Strong experience with SQL and data modeling (dimensional modeling, star/snowflake schemas, event modeling).
- Hands-on experience with streaming systems (Kafka preferred) and/or AWS Kinesis, including event-driven designs.
- Proficiency in one or more languages used for data engineering (Python, Java, Scala, or Go).
- Experience with distributed data processing (Spark, Flink, or equivalent) and performance tuning at scale.
- Experience with AWS data services and cloud-native patterns (S3, Glue/EMR, Athena, Redshift, etc. as applicable).
- Familiarity with lakehouse/table formats and large-scale storage patterns (e.g., Parquet; Iceberg/Hudi/Delta are a plus).
- Experience with orchestration/workflow tooling (Airflow/Dagster/Step Functions) and CI/CD for data workloads.
- Strong data quality/observability practices (tests, monitoring, lineage; understanding of SLAs/SLOs).
- Experience with SQL + NoSQL data stores (e.g., Postgres/MySQL; DynamoDB/Cassandra/Redis) and choosing the right store per use case.
- Clear communicator and collaborator; able to work with mixed audiences and translate needs into reliable data interfaces.
Benefits
- Unlimited PTO
- Excellent medical, dental, and vision coverage
- Employee Equity
- Employee Discounts, Virtual Wellness Classes, and Pet Insurance And more!!
Applicant Tracking System Keywords
Tip: use these terms in your resume and cover letter to boost ATS matches.
Hard Skills & Tools
data pipelinesbatch processingstreaming processingSQLdata modelingPythonJavaScalaSparkAWS
Soft Skills
clear communicatorcollaborator