Own the full ingestion path from edge to cloud, ensuring robot telemetry, sensor data, and warehouse events are reliably captured, transported, and made available for downstream systems
Design, build, and operate scalable pipelines and foundational data layers (streaming and batch) that deliver low-latency, reliable data for analytics, AI/ML, and product features
Implement observability, monitoring, and CI/CD practices to ensure pipeline quality and keep data flows robust, maintainable, and trustworthy
Scale and optimize multi-tenant infrastructure, balancing performance, reliability, and cost-efficiency
Collaborate directly with robotics, AI/ML, and product teams to translate product requirements into resilient data systems that unlock features in Vista, Portal, and ScoutMap
Establish and enforce best practices for data engineering, reliability, and security while enabling analytics engineers to deliver marts, metrics, and dashboards
Shape how humans manage and interact with robotic fleets by powering Vista (AI insights) and ScoutMap (3D mapping and environment intelligence)
Requirements
5+ years of professional experience in data engineering or data infrastructure roles
Strong proficiency in Python and SQL, with the ability to write production-quality, scalable, and well-tested code
Proven experience designing and operating ingestion pipelines and staging layers (streaming and batch)
Experience deploying and managing cloud data infrastructure in AWS using infrastructure-as-code (e.g., Terraform, Kubernetes, Docker)
Hands-on experience with cloud-based data platforms, storage systems, and infrastructure
Familiarity with data quality practices, testing frameworks, and CI/CD for data pipelines
Highly motivated teammate with excellent oral and written communication skills
Enjoy working in a fast paced, collaborative and dynamic start-up environment
Willingness to travel occasionally for on-site support or testing, as needed
Must have and maintain US work authorization
Preferred: Proven experience as the technical lead or primary owner of a data pipeline or platform project
Preferred: Experience with Databricks (Delta Live Tables, SQL Warehouse) and familiarity with dbt or similar tools
Preferred: Strong understanding of multi-tenant architectures and cost/performance/reliability tradeoffs
Preferred: Background in streaming systems (Kafka, Flink, Kinesis, or Spark Structured Streaming)
Preferred: Familiarity with data quality and observability tools (e.g., Great Expectations, Monte Carlo)
Preferred: Exposure to IoT/robotics telemetry or 3D/spatial data processing (e.g., point clouds, LiDAR, time-series)
Preferred: Experience working in a product-facing data role, collaborating with product, engineering, and AI/ML teams
Benefits
equity
comprehensive benefits
Option to work remotely within the United States
Preferred office locations: Santa Clara, CA or Seattle, WA
Applicant Tracking System Keywords
Tip: use these terms in your resume and cover letter to boost ATS matches.
Hard skills
PythonSQLdata engineeringdata infrastructureingestion pipelinescloud data infrastructureinfrastructure-as-codeTerraformKubernetesDocker
Soft skills
communication skillscollaborationmotivationadaptabilityteamwork