Policy Manager Harmful Persuasion
Anthropic
Job Overview
Who's the hiring manager?
Sign up to PitchMeAI to discover the hiring manager's details for this job. We will also write them an intro email for you.

Job Description
About Anthropic
Anthropic’s mission is to create reliable, interpretable, and steerable AI systems. We work to ensure AI is safe and beneficial for users and society.
About The Role
As a Safeguards Product Policy Manager for Harmful Persuasion, you will develop, refine, and maintain policies to prevent AI misuse in influence operations, harmful manipulation, and fraud. You will own the policy framework in areas including election integrity, information integrity, and fraud, translating risk frameworks into clear, enforceable guidelines and supporting regulatory compliance.
Responsibilities
- Develop and maintain policy frameworks for harmful persuasion risks.
- Design enforceable policy language for consistent application.
- Oversee evaluations assessing AI misuse and deceptive tactics.
- Write and refine external-facing usage policy language.
- Develop training guidelines, assessment rubrics, and evaluation protocols.
- Validate enforcement decisions and provide qualitative policy guidance.
- Coordinate with external experts and gather feedback.
- Advise on UX design to align with policy intent and minimize friction.
- Contribute to model safety improvements and regulatory compliance.
- Function as an escalation point for complex harmful persuasion cases.
You May Be a Good Fit If You Have
- 5+ years in policy development, trust & safety, or platform policy.
- Knowledge of global regulatory landscape for election integrity and platform regulation.
- Strong policy writing skills and ability to convert risk frameworks.
- Experience working with multidisciplinary teams, including Engineering and Legal.
- Excellent communication skills to explain complex policy rationales.
Preferred Qualifications
- Familiarity with political psychology and democratic resilience research.
- Understanding of persuasion theory, cognitive biases, and influence tactics.
- Experience working with EU institutions or regulatory bodies.
- Experience in adversarial testing, red teaming, or vulnerability assessments.
- Familiarity with generative AI and its application in personalized persuasion.
Compensation & Logistics
The annual salary range is $245,000 to $330,000 USD. A bachelor’s degree or equivalent experience is required. The role follows a location-based hybrid policy with a minimum 25% office attendance. Visa sponsorship is available under certain conditions.
How We're Different
Anthropic values big science in AI research, working as a cohesive team on high-impact projects that ensure trustworthy and steerable AI. Collaboration and clear communication are key in our research discussions and overall approach.
Key skills/competency
- Policy Development
- Trust & Safety
- Regulatory Compliance
- Risk Management
- Election Integrity
- Fraud Prevention
- Enforcement Guidelines
- Cross-functional Collaboration
- UX Policy Alignment
- Communication
How to Get Hired at Anthropic
- Research Anthropic's culture: Study their mission, values, and recent projects.
- Tailor your resume: Highlight policy development and regulatory experience.
- Showcase technical insights: Emphasize risk frameworks and AI safety skills.
- Prepare for interviews: Review real-world case studies and enforcement scenarios.
Frequently Asked Questions
Find answers to common questions about this job opportunity
Explore similar opportunities that match your background