Architect and implement robust, scalable data pipelines to process, synchronize, and package robotics data (e.g., LiDAR, camera, IMU, proprietary maps) for third-party consumption.
Build a data processing and egress platform, ensuring the timely and accurate delivery of datasets according to strict partner SLAs.
Create data lifecycle policies to control cloud data costs. Build and maintain a universal data catalogue of all raw robot data. Create cost monitoring, attribution and alerting systems.
Build data discoverability platform features, use ml models to generate new attributes and maintain efficient, highly scalable search indexes.
Setup data access audit trails and strong security controls managed through IaC. Create lineage maps and expose data traceability capabilities to internal consumers.
Requirements
5+ years of professional experience in software or data engineering.
Strong programming proficiency in Python, SQL
Hands-on experience building and maintaining large-scale data processing pipelines using cloud technologies
Proficiency with data warehousing and ETL/ELT concepts
Solid understanding of system design, along with data privacy and security best practices
Hands on experience setting up IaC to orchestrate cloud resources and security policies
Experience with GCP and solid understanding of fully managed cloud infrastructure
Familiarity with robotics data such as lidar, multi-modal camera, mapping, etc
Experience working in a fast paced startup environment
Experience building and optimizing terabyte scale data pipelines
Benefits
Offers Equity
📊 Resume Score
Upload your resume to see if it passes auto-rejection tools used by recruiters
Check Resume Score
Applicant Tracking System Keywords
Tip: use these terms in your resume and cover letter to boost ATS matches.