PYMNTS.com August 27, 2025

Artificial intelligence startups Anthropic and OpenAI said Wednesday (Aug. 27) that they evaluated each other’s public models, using their own safety and misalignment tests.

Sharing this news and the results in separate blog posts, the companies said they looked for problems like sycophancy, whistleblowing, self-preservation, supporting human misuse and capabilities that could undermine AI safety evaluations and oversight.

OpenAI wrote in its post that this collaboration was a “first-of-its-kind joint evaluation” and that it demonstrates how labs can work together on issues like these.

Anthropic wrote in its post that the joint evaluation exercise was meant to help mature the field of alignment evaluations and “establish production-ready best practices.”

Reporting the findings of its evaluations, Anthropic said OpenAI’s o3 and...

Today's Sponsors

Venturous
ZeOmega

Today's Sponsor

Venturous

 
Topics: AI (Artificial Intelligence), Healthcare System, Safety, Technology
The Download: OpenAI’s plans for science, and chatbot age verification
Around the nation: Amazon's One Medical launches new AI chatbot
The Medical Futurist’s 100 Digital Health And AI Companies Of 2026
Physician assistants say paperwork and AI training still lag
More Data Isn’t Always Better for AI Decisions

Share Article