[Expression of Interest] Research Scientist/Engineer, Alignment Finetuning
Anthropic
Posted
Dec 13, 2025
Location
USA
Type
Full-time
Compensation
$350000 - $500000
Mission
What you will drive
Core responsibilities:
- Develop and implement novel finetuning techniques using synthetic data generation and advanced training pipelines
- Use these to train models to have better alignment properties including honesty, character, and harmlessness
- Create and maintain evaluation frameworks to measure alignment properties in models
- Collaborate across teams to integrate alignment improvements into production models
- Develop processes to help automate and scale the work of the team
Impact
The difference you'll make
This role creates positive change by developing techniques to train language models that are more aligned with human values, demonstrating better moral reasoning, improved honesty, and good character, ultimately working to build reliable, interpretable, and steerable AI systems that are safe and beneficial for society.
Profile
What makes you a great fit
Required skills and qualifications:
- MS/PhD in Computer Science, ML, or related field, or equivalent experience
- Strong programming skills, especially in Python
- Experience with ML model training and experimentation
- Track record of implementing ML research
- Strong analytical skills for interpreting experimental results
- Experience with ML metrics and evaluation frameworks
- Ability to turn research ideas into working code
- Ability to identify and resolve practical implementation challenges
Benefits
What's in it for you
Competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible working hours, and a lovely office space for collaboration.
About
Inside Anthropic
Anthropic is a frontier AI research and product company, with teams working on alignment, policy, and security. We post specific opportunities at Anthropic that we think may be high impact. We do not necessarily recommend working at other positions at Anthropic. You can read concerns about doing harm by working at a frontier AI company in our career review on the topic.