Anthropic is seeking a Software Engineer for their Trust & Safety team to help build safety and oversight mechanisms for AI systems. This role focuses on developing systems to monitor models, prevent misuse, and ensure user well-being. Key responsibilities include building abuse detection infrastructure, creating monitoring systems for API partners, and implementing real-time safety mechanisms at scale.
The ideal candidate will have 3-10+ years of software engineering experience, preferably in integrity, spam, fraud, or abuse detection. Proficiency in SQL, Python, and data analysis tools is required, along with strong communication skills. A background in AI/ML systems, experience with machine learning frameworks, and familiarity with prompt engineering and adversarial inputs are considered strong assets.
Anthropic offers a competitive compensation package, including a salary range of £265,000 - £370,000 GBP, equity options, and comprehensive benefits. The company follows a hybrid work model, expecting staff to be in the office at least 25% of the time.
This role presents an opportunity to work on cutting-edge AI safety and ethics challenges, contributing to Anthropic's mission of creating reliable, interpretable, and steerable AI systems that are safe and beneficial for users and society. Join a collaborative team of researchers, engineers, and policy experts working to shape the future of AI technology.