Research Engineer / Scientist, Alignment Science
Anthropic
Posted
Dec 13, 2025
Location
USA
Type
Full-time
Compensation
$350000 - $500000
Mission
What you will drive
Core responsibilities:
- Build and run elegant and thorough machine learning experiments to understand and steer the behavior of powerful AI systems
- Contribute to exploratory experimental research on AI safety with a focus on risks from powerful future systems
- Work on projects like testing the robustness of safety techniques, running multi-agent reinforcement learning experiments, and building evaluation tooling
- Contribute ideas, figures, and writing to research papers, blog posts, and talks
Impact
The difference you'll make
This role contributes to making AI systems safe, helpful, honest, and harmless by conducting research on AI alignment and safety techniques, directly addressing risks from powerful future AI systems.
Profile
What makes you a great fit
Required qualifications:
- Significant software, ML, or research engineering experience
- Some experience contributing to empirical AI research projects
- Some familiarity with technical AI safety research
- Preference for fast-moving collaborative projects over extensive solo efforts
- At least a Bachelor's degree in a related field or equivalent experience
Benefits
What's in it for you
Competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible working hours, and a lovely office space for collaboration.
About
Inside Anthropic
Anthropic is a frontier AI research and product company, with teams working on alignment, policy, and security. We post specific opportunities at Anthropic that we think may be high impact. We do not necessarily recommend working at other positions at Anthropic. You can read concerns about doing harm by working at a frontier AI company in our career review on the topic.