Tech Stack
AirflowAmazon RedshiftApacheAWSAzureBigQueryCloudDockerETLGoogle Cloud PlatformJavaKafkaKubernetesPythonScalaSQLTerraform
About the role
- Design, build, and maintain data pipelines and workflows
- Transform raw sources into clean, reliable, and scalable data streams
- Build and optimize data pipelines that ingest, transform, and deliver data from diverse sources (EHRs, claims, APIs, CRM) into analytics-ready structures
- Guide customers through complex data challenges and partner with healthcare providers
- Deliver high-quality results and teach/mentor counterparts to support/expand deliverables
- Work in a fully remote team environment focused on healthcare IT strategy and analytics
Requirements
- 2–5+ years of professional experience in data engineering or related roles
- Strong SQL skills, including query optimization and debugging
- Proficiency in Python (or another programming/scripting language such as Scala or Java)
- Hands-on experience with at least one of: Snowflake; Databricks; Azure Data Factory; AWS Redshift; Google BigQuery; dbt or similar transformation tools; Apache Airflow or other orchestration frameworks
- Familiarity with ETL/ELT principles, data warehousing, and data modeling concepts
- Experience with cloud services (AWS, Azure, or GCP)
- Healthcare industry knowledge and experience (Epic, HL7, FHIR, claims) (desired)
- Experience with CI/CD pipelines, Git, and DevOps workflows (desired)
- Familiarity with Infrastructure-as-Code tools (Terraform, CloudFormation) (desired)
- Experience with real-time/streaming data tools (Kafka, Kinesis, Pub/Sub) (desired)
- Containerization experience (Docker, Kubernetes) (desired)
- Cloud or data tool certifications (desired)
- Full-time salaried role; no relocation required; candidates required to have a suitable home office to operate from