Anthropic’s latest feature for two of its Claude AI models could be the beginning of the end for the AI jailbreaking community. The company announced in a post on its website that the Claude Opus 4 and 4.1 models now have the power to end a conversation with users. According to Anthropic, this feature will only be used in “rare, extreme cases of persistently harmful or abusive user interactions.”
To clarify, Anthropic said those two Claude models could exit harmful conversations, like “requests from users for sexual
→ Continue reading at Engadget