Taro Logo

Software Engineer, Safeguards (London)

Anthropic creates reliable, interpretable, and steerable AI systems, focusing on safe and beneficial AI development through research, engineering, and policy work.
$240,000 - $325,000
Security
Senior Software Engineer
Hybrid
501 - 1,000 Employees
3+ years of experience
AI · Cybersecurity
This job posting may no longer be active. You may be interested in these related jobs instead:

Description For Software Engineer, Safeguards (London)

Anthropic is seeking a Software Engineer for their Safeguards team in London, focusing on building critical safety and oversight mechanisms for AI systems. This role sits at the intersection of AI safety and security engineering, working to ensure AI systems remain reliable and beneficial.

The position involves developing sophisticated monitoring and abuse detection systems to maintain the integrity of Anthropic's AI platforms. Key responsibilities include building real-time defense mechanisms, analyzing user behavior patterns, and implementing automated enforcement actions when necessary. The role requires collaboration with research teams to improve model safety at the training stage.

The ideal candidate will bring 3-8+ years of software engineering experience, with particular emphasis on integrity, spam, fraud, or abuse detection systems. Strong technical skills in Python, SQL, and data analysis are essential, as is the ability to communicate complex technical concepts to diverse stakeholders.

Anthropic offers a competitive compensation package ranging from £240,000 to £325,000 GBP annually, along with benefits including equity options, parental leave, and visa sponsorship for international candidates. The company maintains a hybrid work model requiring at least 25% in-office presence.

As a public benefit corporation, Anthropic is committed to developing safe and beneficial AI systems. The company values diverse perspectives and encourages applications from candidates of all backgrounds, even if they don't meet every listed qualification. The role offers an opportunity to work on meaningful challenges in AI safety while contributing to the development of trustworthy AI systems.

The position combines technical expertise with the mission-critical goal of ensuring AI system safety and reliability. Successful candidates will join a collaborative team working on large-scale research efforts, contributing to the advancement of steerable, trustworthy AI technology. The role offers significant impact potential in shaping the future of AI safety mechanisms and oversight.

Last updated 13 days ago

Responsibilities For Software Engineer, Safeguards (London)

  • Develop monitoring systems to detect unwanted behaviors from API partners and take automated enforcement actions
  • Build abuse detection mechanisms and infrastructure
  • Surface abuse patterns to research teams to harden models at the training stage
  • Build robust and reliable multi-layered defenses for real-time improvement of safety mechanisms at scale
  • Analyze user reports of inappropriate content or accounts

Requirements For Software Engineer, Safeguards (London)

Python
  • Bachelor's degree in Computer Science, Software Engineering or comparable experience
  • 3-8+ years of experience in software engineering, preferably with focus on integrity, spam, fraud, or abuse detection
  • Proficiency in SQL, Python, and data analysis tools
  • Strong communication skills and ability to explain complex technical concepts to non-technical stakeholders

Benefits For Software Engineer, Safeguards (London)

Visa Sponsorship
Equity
Parental Leave
  • Visa Sponsorship
  • Competitive compensation
  • Equity
  • Parental Leave
  • Vacation

Interested in this job?