OpenAI and Anthropic conducted safety evaluations of each other’s AI systems

Most of the time, AI companies are locked in a race to the top, treating each other as rivals and competitors. Today, OpenAI and Anthropic revealed that they agreed to evaluate the alignment of each other’s publicly available systems and shared the results of their analyses. The full reports get pretty technical, but are worth a read for anyone who’s following the nuts and bolts of AI development. A broad summary showed some flaws with each company’s offerings, as well as revealing pointers for how to improve future safety tests.

Anthropic said it evaluated OpenAI models for “sycophancy, whistleblowing, self-preservation,

→ Continue reading at Engadget

Similar Articles

Advertisment

Most Popular