Lead teams of researchers to produce original research in LLM evaluation methodologies, interpretability, and human-AI knowledge alignment.
Develop novel frameworks and assessment techniques that reveal deep insights into model capabilities, limitations, and emergent behaviors.
Collaborate with engineers to translate research breakthroughs into scalable benchmarks, evaluation systems, and standards.
Pioneer new approaches to measuring reasoning, alignment, and trustworthiness in frontier AI systems.
Author high-quality code to enable large-scale experimentation, reproducible evaluation, and knowledge assessment workflows.
Publish in top-tier conferences and journals, establishing new directions in the science of AI evaluation.
Work cross-functionally with leadership, engineers, and external partners to set industry standards for responsible AI evaluation and alignment.
Requirements
PhD or equivalent research experience in machine learning, computer science, cognitive science, or related fields with focus on AI evaluation, interpretability, or model understanding.
6+ years of academic or industry experience post-doc in a research-first environment
Strong background in LLM research, evaluation methodologies, and/or foundational AI assessment techniques.
Proven ability to independently design, lead, and execute evaluation research programs with novel data types end-to-end.
Deep proficiency in Python and PyTorch for large-scale model analysis, benchmarking, and evaluation.
Experience building or leading novel benchmark development, systematic model assessment, or interpretability studies.
Strong publication record in post-training, evaluation, or interpretability that demonstrates field-defining contributions.
Ability to clearly communicate complex insights and influence both technical and non-technical stakeholders.