14 days ago

Red Team Engineer Safeguards

Anthropic

On Site
Full Time
$310,000
Washington, DC

Job Overview

Job TitleRed Team Engineer Safeguards
Job TypeFull Time
CategoryCommerce
Experience5 Years
DegreeMaster
Offered Salary$310,000
LocationWashington, DC

Who's the hiring manager?

Sign up to PitchMeAI to discover the hiring manager's details for this job. We will also write them an intro email for you.

Uncover Hiring Manager

Job Description

About Anthropic

Anthropic’s mission is to create reliable, interpretable, and steerable AI systems that are safe and beneficial for users and society. The company is a quickly growing team of researchers, engineers, policy experts, and business leaders.

About the Role

As a Red Team Engineer Safeguards at Anthropic, you will take an adversarial approach to uncover vulnerabilities in deployed AI systems and products before they can be exploited. You will work across technical infrastructure and handle emergent risks from advanced AI capabilities.

Responsibilities

  • Conduct comprehensive adversarial testing across Anthropic’s product surfaces.
  • Research and implement novel testing approaches for emerging capabilities.
  • Design and execute 'full kill chain' attacks emulating real-world threat actors.
  • Develop and maintain systematic testing methodologies and automated frameworks.
  • Collaborate with Product, Engineering, and Policy teams to translate findings into concrete improvements.
  • Help establish metrics for measuring detection effectiveness of novel abuse.

Qualifications

You may be a good fit if you have:

  • Experience in penetration testing, red teaming, or application security.
  • Strong technical skills in web application security with hands-on expertise using tools like Burp Suite, Metasploit.
  • Demonstrated ability to discover novel attack vectors and chain vulnerabilities creatively.
  • Public work such as CVEs, blog posts, or bug bounty reports.
  • Ability to build custom automation and adaptability to emerging threats.
  • Excellent communication skills to explain technical concepts to varied audiences.

Strong candidates may also have experience with AI/ML security, API security, anti-fraud systems, distributed systems, and a background in abuse prevention.

Compensation & Logistics

  • Annual Salary: 300,000 - 320,000 USD
  • Education: Bachelor’s degree or equivalent experience required.
  • Work Arrangement: Hybrid (at least 25% in-office presence).
  • Visa Sponsorship: Available for select candidates.

How We're Different

At Anthropic, work is collaborative and focused on large-scale research efforts that advance trustworthy and steerable AI systems. Communication and impact are valued in a cohesive team environment.

Key skills/competency

  • Penetration Testing
  • Red Teaming
  • Application Security
  • Adversarial Testing
  • Automation
  • Vulnerability Assessment
  • Web Security
  • Custom Scripting
  • Collaboration
  • AI Safety

Tags:

Red Team Engineer Safeguards
penetration testing
application security
adversarial testing
automation
vulnerability assessment
Burp Suite
Metasploit
AI safety
custom scripting
security testing tools
advanced AI

Share Job:

How to Get Hired at Anthropic

  • Customize your resume: Align skills with red teaming and security testing.
  • Highlight relevant experience: Showcase penetration testing and automation work.
  • Research Anthropic: Understand its mission and AI safety focus.
  • Prepare technical examples: Illustrate past full kill chain simulations clearly.

Frequently Asked Questions

Find answers to common questions about this job opportunity

Explore similar opportunities that match your background