Tech Stack
CloudDistributed SystemsPython
About the role
- Design, test, and deploy AI systems that are embedded at scale.
- Leverage cutting-edge models to build robust, production-ready AI applications.
- Lead the design of data and inference pipelines.
- Orchestrate the integration of LLM-based components such as retrieval-augmented generation (RAG).
- Balance engineering constraints with business priorities.
- Own all stages of product iteration and development from system design to deployment.
- Collaborate with data scientists and may work in forward-deployed roles with clients or centralized roles to abstract solutions into scalable assets.
Requirements
- BSc, MSc or PhD in Computer Science, Engineering, Physics, or a related technical or STEM field from a leading university.
- Experience working with relevant models and/ or architectures within the following fields: Computer Vision, Natural Language Processing, or Graph Neural Networks.
- Experience writing production-ready code.
- Experience programming in Python.
- Familiarity with the frontier labs APIs (Gemini, OpenAI, Anthropic), and their advanced features like structured outputs and tool calling.
- Understanding of LLM inference considerations e.g. input vs. output tokens, prompt caching.
- Strong computer science fundamentals in areas such as data structures, cloud infrastructure, distributed systems, machine learning, and high-performance computing.
- Experience and willingness to work through the challenges associated with large, unstructured, real-world data sets.
- Ability to develop custom tooling for efficient experimentation and evaluation of LLM applications / agentic workflows