Anthropic

Offensive Security Research Engineer, Safeguards

Anthropic(2 hours ago)

HybridFull TimeSenior$320,000 - $405,000Security
Apply Now

About this role

Anthropic is seeking vulnerability researchers to analyze and mitigate risks associated with large language models (LLMs). The role involves researching how adversaries might misuse LLMs and developing strategies to defend against these threats, contributing to building safer and more trustworthy AI systems.

View Original Listing

Required Skills

  • Vulnerability Research
  • Penetration Testing
  • Security Exploitation
  • Reverse Engineering
  • Network Security
  • Software Engineering
  • AI Safety
  • Bug Bounty
  • Open Source
  • Threat Modeling
Anthropic

About Anthropic

anthropic.com

Anthropic is an AI safety and research company focused on building reliable, interpretable, and steerable AI systems. It develops large language models (branded as Claude) and offers APIs and enterprise products that let organizations integrate conversational AI with safety-focused controls, moderation, and privacy features. The company prioritizes interpretability and alignment research, publishes technical work, and engages with policymakers to reduce risks from advanced AI. Customers choose Anthropic for its safety-first approach, controllability tools, and research-driven models.

View more jobs at Anthropic

ApplyBlast uses AI to match you with the right jobs, tailor your resume and cover letter, and apply automatically so you can land your dream job faster.

© All Rights Reserved. ApplyBlast.com