Tech Stack
AndroidCloudDockerGoogle Cloud PlatformiOSKubernetesPython
About the role
- Work alongside machine learning researchers, engineers, and product managers to bring AI Voices to customers for diverse use cases
- Deploy and operate the core ML inference workloads for the AI Voices serving pipeline
- Introduce new techniques, tools, and architecture to improve performance, latency, throughput, and efficiency of deployed models
- Build tools to identify bottlenecks and instability sources and design/implement solutions for high-priority issues
- Collaborate within a distributed team and contribute to product decisions and user experience design
Requirements
- Experience shipping Python-based services
- Experience being responsible for the successful operation of a critical production service
- Experience with public cloud environments, GCP preferred
- Experience with Infrastructure as Code, Docker, and containerized deployments
- Preferred: Experience deploying high-availability applications on Kubernetes
- Preferred: Experience deploying ML models to production
- Strong work ethic and solid communication skills
- Ability to think strategically, make product decisions, and work in fast-paced environments