1upHealth, Inc.

Senior Data Engineer, Core Data Platform

1upHealth, Inc.

full-time

Posted on:

Location Type: Remote

Location: United States

Visit company website

Explore more

AI Apply
Apply

Job Level

About the role

  • Design, develop, and implement robust ETL/ELT data pipelines for large-scale data ingestion, transformation, and storage.
  • Ensure data quality, integrity, and governance by implementing validation techniques, data monitoring, and automated testing.
  • Collaborate with cross-functional teams, including data scientists, analysts, platform engineers, and business stakeholders, to develop scalable and reusable data solutions.
  • Automate deployments and testing using CI/CD pipelines with Git, Terraform, GitHub Actions, or Jenkins.
  • Design and build custom data tools and abstractions to support analytics, machine learning, and real-time data processing.
  • Work with DevOps and platform teams to establish efficient deployment and monitoring processes for internal and external data products.
  • Develop and implement alerting, monitoring, and observability frameworks for data pipelines to ensure reliability and proactive issue resolution.
  • Contribute to the data architecture and strategy, driving improvements in scalability, performance, and cost optimization.
  • Stay up to date with emerging technologies and industry best practices to continuously enhance data engineering capabilities.

Requirements

  • 5+ years of hands-on experience in data engineering, with expertise in distributed data processing and big data frameworks (e.g., Apache Spark, Apache Iceberg, Trino, Apache Airflow, dbt, Dagster).
  • Advanced programming skills in Scala or Python for data transformation and automation.
  • Experience with real-time data streaming technologies such as Apache Flink, Spark Streaming, or Kafka.
  • Strong experience in performance tuning for Spark and optimizing large-scale data workflows.
  • Proficiency in SQL and database management, with hands-on experience in Massively Parallel Processing (MPP) databases such as Amazon Redshift, Snowflake, Teradata.
  • Familiarity with cloud-based data services (AWS, RDS, DynamoDB) and containerized infrastructure (EKS, Docker, Kubernetes).
  • Hands-on experience integrating DevOps and CI/CD practices in data engineering using GitHub Actions, Jenkins, or Terraform.
  • Proven ability to build monitoring, alerting, and observability tools for data pipelines to ensure high availability and reliability.
  • Experience in data mapping, validation, and testing frameworks to ensure accuracy and consistency.
  • Exposure in machine learning/deep learning using PyTorch, Tensorflow, or Keras (PyTorch preferred)
  • Exposure to machine learning workflows and familiarity with MLOps tools for model deployment and lifecycle management.
  • Self-starter with strong problem-solving skills who thrives in a fast-paced, agile environment.
Benefits
  • 100% Paid BCBS Medical and Dental Insurance for Employees
  • Vision Insurance
  • Flexible PTO
  • Equity for All Employees
  • 401(k)
  • Home Office Stipend
  • Lifestyle Savings Account
  • Parental Leave (6 weeks bonding leave for all new parents + 10 weeks for birthing parents)
Applicant Tracking System Keywords

Tip: use these terms in your resume and cover letter to boost ATS matches.

Hard Skills & Tools
ETLELTdata pipelinesdata transformationApache SparkApache IcebergTrinoApache AirflowScalaPython
Soft Skills
problem-solvingcollaborationself-starteragilecommunication