Product Manager, Safeguards (Beneficial Deployments)
Anthropic
Posted
Feb 07, 2026
Location
USA
Type
Full-time
Compensation
$305000 - $385000
Mission
What you will drive
- Determine how to build in safety by design upstream and leverage downstream defenses for Anthropic's frontier models, AI products, and customers across various platforms
- Drive impact via ruthless prioritization by clearly defining problems, solution options, and requirements for MVP vs. ideal state
- Align and collaborate with policy, enforcement, research, engineering, and cross-functional stakeholders
- Lead the development of metrics to understand performance, blindspots, and inform future project planning
Impact
The difference you'll make
This role creates positive change by developing safeguards systems that protect users from the risks of powerful AI systems and ensuring AI deployment is safe and beneficial for humanity.
Profile
What makes you a great fit
- 5+ years in product management with focus on fast problem understanding, building roadmaps, and getting into details on data, detection, interventions, infrastructure, tools, and/or evals
- Ability to make technical tradeoff decisions with experience working across policy experts, AI/ML research engineers, and software engineering teams
- Demonstrated ability to build product and engineering strategy across multiple cross-functional teams for a rapidly changing space
- Very strong ability to navigate and prioritize amidst rapidly changing product specs, and flex into different domains to bring clarity and execute
Benefits
What's in it for you
Competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible working hours, and a lovely office space for collaboration.
About
Inside Anthropic
Anthropic is a frontier AI research and product company, with teams working on alignment, policy, and security. We post specific opportunities at Anthropic that we think may be high impact. We do not necessarily recommend working at other positions at Anthropic. You can read concerns about doing harm by working at a frontier AI company in our career review on the topic.