Impact Careers Full-time

Research Engineer / Scientist, Alignment Science

Anthropic

Posted

Dec 13, 2025

Location

USA

Type

Full-time

Compensation

$350000 - $500000

Mission

What you will drive

Core responsibilities:

  • Build and run elegant and thorough machine learning experiments to understand and steer the behavior of powerful AI systems
  • Contribute to exploratory experimental research on AI safety with a focus on risks from powerful future systems
  • Work on projects like testing the robustness of safety techniques, running multi-agent reinforcement learning experiments, and building evaluation tooling
  • Contribute ideas, figures, and writing to research papers, blog posts, and talks

Impact

The difference you'll make

This role contributes to making AI systems safe, helpful, honest, and harmless by conducting research on AI alignment and safety techniques, directly addressing risks from powerful future AI systems.

Profile

What makes you a great fit

Required qualifications:

  • Significant software, ML, or research engineering experience
  • Some experience contributing to empirical AI research projects
  • Some familiarity with technical AI safety research
  • Preference for fast-moving collaborative projects over extensive solo efforts
  • At least a Bachelor's degree in a related field or equivalent experience

Benefits

What's in it for you

Competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible working hours, and a lovely office space for collaboration.

About

Inside Anthropic

Visit site →

Anthropic is a frontier AI research and product company, with teams working on alignment, policy, and security. We post specific opportunities at Anthropic that we think may be high impact. We do not necessarily recommend working at other positions at Anthropic. You can read concerns about doing harm by working at a frontier AI company in our career review on the topic.