AI Safety Fellow
Anthropic
Posted
Mar 11, 2026
Location
Remote (US)
Type
Full-time
Compensation
Up to $3850
Mission
What you will drive
- In this fellowship, you'll conduct empirical AI safety research for four months with the goal of producing a publishable paper.
- Work with mentors from Anthropic's research team on projects aligned with their research priorities.
- Use external infrastructure like open-source models and APIs to implement and test ideas.
- Collaborate within the broader AI safety research community in shared workspaces.
- Produce public research outputs that advance understanding in areas like scalable oversight or mechanistic interpretability.
Profile
What makes you a great fit
- In this fellowship, you'll conduct empirical AI safety research for four months with the goal of producing a publishable paper.
- Work with mentors from Anthropic's research team on projects aligned with their research priorities.
- Use external infrastructure like open-source models and APIs to implement and test ideas.
- Collaborate within the broader AI safety research community in shared workspaces.
- Produce public research outputs that advance understanding in areas like scalable oversight or mechanistic interpretability.
About
Inside Anthropic
Anthropic is a frontier AI research and product company, with teams working on alignment, policy, and security. We post specific opportunities at Anthropic that we think may be high impact. We do not necessarily recommend working at other positions at Anthropic. You can read concerns about doing harm by working at a frontier AI company in our career review on the topic.