Impact Careers Full-time

[Expression of Interest] Research Scientist/Engineer, Alignment Finetuning

Anthropic

Posted

Dec 13, 2025

Location

USA

Type

Full-time

Compensation

$350000 - $500000

Mission

What you will drive

Core responsibilities:

  • Develop and implement novel finetuning techniques using synthetic data generation and advanced training pipelines
  • Use these to train models to have better alignment properties including honesty, character, and harmlessness
  • Create and maintain evaluation frameworks to measure alignment properties in models
  • Collaborate across teams to integrate alignment improvements into production models
  • Develop processes to help automate and scale the work of the team

Impact

The difference you'll make

This role creates positive change by developing techniques to train language models that are more aligned with human values, demonstrating better moral reasoning, improved honesty, and good character, ultimately working to build reliable, interpretable, and steerable AI systems that are safe and beneficial for society.

Profile

What makes you a great fit

Required skills and qualifications:

  • MS/PhD in Computer Science, ML, or related field, or equivalent experience
  • Strong programming skills, especially in Python
  • Experience with ML model training and experimentation
  • Track record of implementing ML research
  • Strong analytical skills for interpreting experimental results
  • Experience with ML metrics and evaluation frameworks
  • Ability to turn research ideas into working code
  • Ability to identify and resolve practical implementation challenges

Benefits

What's in it for you

Competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible working hours, and a lovely office space for collaboration.

About

Inside Anthropic

Visit site →

Anthropic is a frontier AI research and product company, with teams working on alignment, policy, and security. We post specific opportunities at Anthropic that we think may be high impact. We do not necessarily recommend working at other positions at Anthropic. You can read concerns about doing harm by working at a frontier AI company in our career review on the topic.