Tech Stack
AndroidCloudDockerGoogle Cloud PlatformiOSKubernetesPython
About the role
- Work alongside machine learning researchers, engineers, and product managers to bring AI Voices to customers for diverse use cases
- Deploy and operate the core ML inference workloads for AI Voices serving pipeline
- Introduce techniques, tools, and architecture to improve the performance, latency, throughput, and efficiency of deployed models
- Build tools to give visibility into bottlenecks and sources of instability and implement solutions to address highest priority issues
- Collaborate across distributed, cross-functional teams in a 100% remote environment
Requirements
- Experience shipping Python-based services
- Experience being responsible for the successful operation of a critical production service
- Experience with public cloud environments, GCP preferred
- Experience with Infrastructure such as Code, Docker, and containerized deployments
- Preferred: Experience deploying high-availability applications on Kubernetes
- Preferred: Experience deploying ML models to production
- Ability to work in a fast-paced, strategic environment
- Strong work ethic and solid communication skills