Technical Scaled Abuse Threat Investigator
Anthropic
Location
USA
Type
Full-time
Posted
Jan 15, 2026
Compensation
USD 200000 – 200000
Mission
What you will drive
Core responsibilities:
- Detect and investigate large-scale abuse patterns including model distillation, unauthorized API access, account farming, fraud schemes, and scam operations
- Develop abuse signals and tracking strategies to proactively identify scaled adversarial activity and coordinated abuse networks
- Conduct technical investigations using SQL, Python, and data science methodologies to analyze large datasets and uncover sophisticated abuse patterns
- Create actionable intelligence reports on new attack vectors, vulnerabilities, and threat actor TTPs targeting AI systems at scale
Impact
The difference you'll make
Your work will directly inform our defenses against threat actors who seek to exploit our products for financial gain, competitive advantage, or to cause widespread harm, helping to ensure AI systems are safe and beneficial for society.
Profile
What makes you a great fit
Required qualifications:
- Strong proficiency in SQL and Python with a data science background
- Experience with large language models and understanding of how AI technology could be exploited at scale
- Subject matter expertise in abusive user behavior detection, fraud patterns, account abuse, or platform integrity
- Experience tracking threat actors across surface, deep, and dark web environments
Benefits
What's in it for you
Competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible working hours, and a lovely office space in which to collaborate with colleagues. Annual salary range: $230,000—$290,000 USD.
About
Inside Anthropic
Anthropic's mission is to create reliable, interpretable, and steerable AI systems. We want AI to be safe and beneficial for our users and for society as a whole.