Salary
💰 $200,000 - $220,000 per year
Tech Stack
AWSAzureCloudDockerGoogle Cloud PlatformJavaJavaScriptKubernetesNeo4jPythonPyTorchReactTerraformTypeScript
About the role
- Work alongside clients, embedding with their teams to tackle technical and operational problems and deploy AI systems in production
- Operationalizing AI PoCs by transforming demos and prototypes into robust, production-grade AI systems
- Designing and deploying data pipelines and retrieval frameworks that prepare enterprise data for LLM training, fine-tuning, and RAG-based applications
- Fine-tuning, evaluating, and deploying large language models and agentic workflows, ensuring outputs remain grounded, explainable, and compliant
- Integrating SeekrFlow AI agents with customer environments, containerizing and orchestrating them across cloud, hybrid, and on-prem deployments (AWS, Azure, GCP, private cloud)
- Embedding with client engineering teams to debug model behavior, optimize inference performance, and implement monitoring/guardrails for safe operation
- Codifying best practices into reusable playbooks, evaluation frameworks, and deployment accelerators that scale AI adoption across industries
Requirements
- Bachelor’s degree in Computer Science, Engineering, Mathematics, Physics, or Data Science (advanced degree a plus)
- 6+ years of professional experience in software engineering, data engineering, or applied ML, with a track record of building and deploying production systems
- Hands-on expertise with AI/ML and LLM frameworks (e.g., PyTorch, Hugging Face, LangChain, vLLM), including model fine-tuning and agent deployment
- Proven experience designing, deploying, and maintaining production-grade systems that meet enterprise security, compliance, and performance standards
- Strong background in on-premise, private cloud, and multi-cloud environments (AWS, Azure, GCP), including experience with containerization and orchestration (Docker, Kubernetes)
- Proficiency in at least one modern programming language (Python, Java, C++, TypeScript/JavaScript, or similar), with the ability to learn and adapt quickly
- Ability to work directly with client stakeholders — technical teams, business leaders, and end users — to translate needs into deployed solutions
- Willingness to travel 25–75%, depending on client and team needs
- Familiar with common graph databases and graph languages (e.g., Neo4j, AGE, SPARQL, Cypher)
- Familiar with front end frameworks (e.g., React) and capable of building insightful and engaging application front ends
- Experience with CI/CD pipelines, version control (GitHub/GitLab), and infrastructure as code (Terraform, CloudFormation, etc.)
- Familiarity with vector databases, retrieval-augmented generation (RAG) architectures, and enterprise data integration patterns
- Exposure to monitoring, logging, and observability frameworks for deployed AI/ML systems
- Background in responsible AI practices and familiarity with Trustworthy AI principles
- Experience building custom client-facing applications or agentic workflows from prototype to scalable deployment
- Strong communication skills — ability to present complex technical concepts clearly to diverse audiences