VentureBeat August 8, 2024
Michael Nuñez

Anthropic, the artificial intelligence startup backed by Amazon, launched an expanded bug bounty program on Thursday, offering rewards up to $15,000 for identifying critical vulnerabilities in its AI systems. This initiative marks one of the most aggressive efforts yet by an AI company to crowdsource security testing of advanced language models.

The program targets “universal jailbreak” attacks — methods that could consistently bypass AI safety guardrails across high-risk domains like chemical, biological, radiological, and nuclear (CBRN) threats and cybersecurity. Anthropic will invite ethical hackers to probe its next-generation safety mitigation system before public deployment, aiming to preempt potential exploits that could lead to misuse of its AI models.

AI safety bounties: A new frontier in tech security

This move comes...

Today's Sponsors

LEK
ZeOmega

Today's Sponsor

LEK

 
Topics: AI (Artificial Intelligence), Healthcare System, Safety, Technology
Harnessing AI to reshape consumer experiences in healthcare
AI Ambient Scribes are poised to become indispensable tools for healthcare providers in 2025
Where AI Ambient Scribes Are Heading
AI agents’ momentum won’t stop in 2025
The cybersecurity provider’s next opportunity: Making AI safer

Share This Article