VentureBeat September 3, 2024
Louis Columbus

With weaponized large language models (LLMs) becoming lethal, stealthy by design and challenging to stop, Meta has created CyberSecEval 3, a new suite of security benchmarks for LLMs designed to benchmark AI models’ cybersecurity risks and capabilities.

“CyberSecEval 3 assesses eight different risks across two broad categories: risk to third parties and risk to application developers and end users. Compared to previous work, we add new areas focused on offensive security capabilities: automated social engineering, scaling manual offensive cyber operations, and autonomous offensive cyber operations,” write Meta researchers.

Meta’s CyberSecEval 3 team tested Llama 3 across core cybersecurity risks to highlight vulnerabilities, including automated phishing and offensive operations. All non-manual elements and guardrails, including CodeShield and LlamaGuard 3 mentioned...

Today's Sponsors

LEK
ZeOmega

Today's Sponsor

LEK

 
Topics: AI (Artificial Intelligence), Cybersecurity, Technology
Mobile Threats, AI Innovation, Third-Party Risks: Trends Shaping Healthcare In 2025
Healthcare Cybersecurity – 2025 Health IT Predictions
IoT and ransomware are big security risks, and health systems feel unprepared
Strengthening Cybersecurity: The Power Of Technical Partnerships For API Security
Cybersecurity Trends And Priorities To Watch For 2025

Share This Article