Impact Careers Full-time

ML/Research Engineer, Safeguards

Anthropic

Posted

Jan 15, 2026

Location

USA

Type

Full-time

Compensation

$350000 - $500000

Mission

What you will drive

  • Develop classifiers to detect misuse and anomalous behavior at scale, including synthetic data pipelines and methods to automatically source representative evaluations
  • Build systems to monitor for harms that span multiple exchanges, such as coordinated cyber attacks and influence operations, and develop new methods for aggregating and analyzing signals across contexts
  • Evaluate and improve the safety of agentic products by developing threat models, test environments for agentic risks, and mitigations for prompt injection attacks
  • Conduct research on automated red-teaming, adversarial robustness, and other research that helps test for or find misuse

Impact

The difference you'll make

This role helps detect and mitigate misuse of AI systems, building safeguards that identify harmful use and develop defenses to keep products safe as capabilities advance, directly feeding into Anthropic's Responsible Scaling Policy commitments.

Profile

What makes you a great fit

  • 4+ years of experience in ML engineering, research engineering, or applied research in academia or industry
  • Proficiency in Python and experience building ML systems
  • Comfort working across the research-to-deployment pipeline, from exploratory experiments to production systems
  • Strong communication skills and ability to explain complex technical concepts to non-technical stakeholders

Benefits

What's in it for you

  • Competitive compensation: $350,000–$500,000 USD annual salary
  • Optional equity donation matching
  • Generous vacation and parental leave
  • Flexible working hours
  • Lovely office space for collaboration

About

Inside Anthropic

Visit site →

Anthropic is a frontier AI research and product company, with teams working on alignment, policy, and security. We post specific opportunities at Anthropic that we think may be high impact. We do not necessarily recommend working at other positions at Anthropic. You can read concerns about doing harm by working at a frontier AI company in our career review on the topic.