Salary
💰 $460,000 - $555,000 per year
About the role
- Lead research into embedding models and retrieval systems optimized for grounding, relevance, and adaptive reasoning
- Manage a team of researchers and engineers building end-to-end infrastructure for training, evaluating, and integrating embeddings into frontier models
- Design new embedding training objectives, scalable vector store architectures, and dynamic indexing methods
- Drive innovation in dense, sparse, and hybrid representation techniques, metric learning, and learning-to-retrieve systems
- Collaborate closely with Pretraining, Inference, and other Research teams to integrate retrieval throughout the model lifecycle
- Support retrieval across OpenAI products and internal research efforts with opportunities for scientific publication and deep technical impact
- Contribute to OpenAI’s long-term vision of AI systems with memory and knowledge access capabilities rooted in learned representations
Requirements
- Proven experience leading high-performance teams of researchers or engineers in ML infrastructure or foundational research
- Deep technical expertise in representation learning, embedding models, or vector retrieval systems
- Familiarity with transformer-based LLMs and how embedding spaces can interact with language model objectives
- Research experience in areas such as contrastive learning, supervised or unsupervised embedding learning, or metric learning
- A track record of building or scaling large machine learning systems, particularly embedding pipelines in production or research contexts
- A first-principles mindset for challenging assumptions about how retrieval and memory should work for large models
- Ability to work from OpenAI's US office three days per week (hybrid in San Francisco)