Anthropic

Policy Manager, Harmful Persuasion

Anthropic(13 days ago)

HybridFull TimeSenior$245,000 - $330,000Safeguards
Apply Now

About this role

The Safeguards Product Policy Manager for Harmful Persuasion at Anthropic will own policy development to prevent misuse of AI for influence operations, manipulation, and fraud. The role focuses on translating a harmful persuasion risk framework into clear, enforceable policy and shaping policy across areas like election integrity, information integrity, and fraud. The position sits on the Safeguards team and contributes to broader safety and governance efforts as AI capabilities evolve.

View Original Listing

Required Skills

  • Policy Development
  • Policy Writing
  • Election Integrity
  • Information Integrity
  • Fraud Prevention
  • Adversarial Testing
  • Regulatory Compliance
  • Risk Assessment
  • Model Safety
  • Cross-Functional Collaboration

+1 more

Qualifications

  • Bachelor's Degree
Anthropic

About Anthropic

anthropic.com

Anthropic is an AI safety and research company focused on building reliable, interpretable, and steerable AI systems. It develops large language models (branded as Claude) and offers APIs and enterprise products that let organizations integrate conversational AI with safety-focused controls, moderation, and privacy features. The company prioritizes interpretability and alignment research, publishes technical work, and engages with policymakers to reduce risks from advanced AI. Customers choose Anthropic for its safety-first approach, controllability tools, and research-driven models.

ApplyBlast uses AI to match you with the right jobs, tailor your resume and cover letter, and apply automatically so you can land your dream job faster.

© All Rights Reserved. ApplyBlast.com