Policy Manager, Harmful Persuasion
Anthropic(13 days ago)
About this role
The Safeguards Product Policy Manager for Harmful Persuasion at Anthropic will own policy development to prevent misuse of AI for influence operations, manipulation, and fraud. The role focuses on translating a harmful persuasion risk framework into clear, enforceable policy and shaping policy across areas like election integrity, information integrity, and fraud. The position sits on the Safeguards team and contributes to broader safety and governance efforts as AI capabilities evolve.
Required Skills
- Policy Development
- Policy Writing
- Election Integrity
- Information Integrity
- Fraud Prevention
- Adversarial Testing
- Regulatory Compliance
- Risk Assessment
- Model Safety
- Cross-Functional Collaboration
+1 more
Qualifications
- Bachelor's Degree
About Anthropic
anthropic.comAnthropic is an AI safety and research company focused on building reliable, interpretable, and steerable AI systems. It develops large language models (branded as Claude) and offers APIs and enterprise products that let organizations integrate conversational AI with safety-focused controls, moderation, and privacy features. The company prioritizes interpretability and alignment research, publishes technical work, and engages with policymakers to reduce risks from advanced AI. Customers choose Anthropic for its safety-first approach, controllability tools, and research-driven models.