Salary
💰 $168,000 - $322,000 per year
Tech Stack
AnsibleAWSAzureCloudDistributed SystemsDockerGoogle Cloud PlatformKubernetesLinuxPythonPyTorchServiceNowTensorflowTerraform
About the role
- Design and deploy custom AI solutions including distributed training, inference optimization, and MLOps pipelines across customer environments
Provide remote technical support to strategic customers, optimize AI workloads, diagnose and resolve performance issues, and guide technical implementations through virtual collaboration
Deploy and manage AI workloads across DGX Cloud, customer data centers, and CSP environments using Kubernetes, Docker, and scheduling systems for GPU
Profile and optimize large-scale model training and inference workloads, implement monitoring solutions, and resolve scaling challenges
Build custom integrations with customer systems, develop APIs and data pipelines, and implement enterprise software connections
Create implementation guides, documentation for resolution approaches and standard methodologies for complex AI deployments
Requirements
- 8+ years of experience in customer-facing technical roles (Solutions Engineering, DevOps, ML Infrastructure Engineering)
BS, MS, or Ph.D. in CS, CE, EE or equivalent experience
Strong proficiency with Linux systems, distributed computing, Kubernetes, and GPU scheduling
AI/ML experience supporting inference workloads and training at large-scale
Programming skills in Python, with experience in PyTorch, TensorFlow, or similar AI frameworks
Customer engagement ability to work effectively with technical teams under high-pressure situations