PYMNTS.com August 27, 2025
Artificial intelligence startups Anthropic and OpenAI said Wednesday (Aug. 27) that they evaluated each other’s public models, using their own safety and misalignment tests.
Sharing this news and the results in separate blog posts, the companies said they looked for problems like sycophancy, whistleblowing, self-preservation, supporting human misuse and capabilities that could undermine AI safety evaluations and oversight.
OpenAI wrote in its post that this collaboration was a “first-of-its-kind joint evaluation” and that it demonstrates how labs can work together on issues like these.
Anthropic wrote in its post that the joint evaluation exercise was meant to help mature the field of alignment evaluations and “establish production-ready best practices.”
Reporting the findings of its evaluations, Anthropic said OpenAI’s o3 and...







