OpenAI is seeking a Research Scientist for their Chain-of-thought Interpretability Team, part of the Science of Deep Learning team. This role focuses on developing innovative machine learning techniques to monitor models' latent thinking, particularly in chain-of-thought (CoT) reasoning. The ideal candidate will work on technical approaches to determine when model CoTs are monitorable, faithful, and legible, and explore interventions to improve or degrade monitorability.
Key responsibilities include:
The ideal candidate should:
OpenAI offers a competitive compensation package, including a salary range of $295K – $440K plus equity. They are committed to diversity, equal opportunity, and providing reasonable accommodations to applicants with disabilities.
Join OpenAI in shaping the future of technology and ensuring that the benefits of AI are widely shared.