Tech Stack
AnsibleAWSAzureCloudGoGoogle Cloud PlatformKubernetesPythonRustTerraform
About the role
- Lead the design and deployment of AI agents and orchestration frameworks for dynamic task automation.
- Develop and optimize Micro LLM deployments for performance at the edge.
- Design infrastructure-as-code systems to support scalable AI/ML deployments.
- Collaborate with DevOps and MLOps teams to build reliable, automated pipelines.
- Drive architectural decisions that impact AI workflows, model deployment, and system monitoring.
- Troubleshoot and enhance AI deployments in distributed environments, focusing on performance and cost-efficiency.
- Stay current on emerging trends in AI agents, micro-models, and edge technologies.
Requirements
- BS/MS in Computer Science, AI, Machine Learning, or a related field.
- 5+ years of experience in AI engineering, infrastructure automation, or backend systems.
- Strong experience with LLMs and orchestration frameworks (LangChain, AutoGPT, AgentGPT, etc.).
- Proven knowledge in deploying and optimizing Micro LLMs in low-latency or edge environments.
- Expertise with edge computing technologies and architectures (e.g., NVIDIA Jetson, Coral, or similar).
- Deep experience with infrastructure automation tools such as Terraform, Ansible, or Kubernetes.
- Strong cloud experience (AWS, Azure, or GCP) with an emphasis on scalable AI infrastructure.
- Proficiency in Python (preferred), Go, or Rust.
- Familiarity with vector databases (e.g., Pinecone, Weaviate, FAISS).
- Excellent communication skills in English (C1 preferred, strong B2 may be considered).
- Must reside and have work authorization in Latin America.
- Must be available to work with significant overlap with Mountain Standard Time (MST).