Anthropic's Claude AI now has the ability to end 'distressing' conversations
Anthropic, an AI company, has introduced a new feature in its Claude AI models that allows them to end conversations in "rare, extreme cases of persistently harmful or abusive user interactions." This includes requests for sexual content involving minors and attempts to solicit information that could enable large-scale violence or acts of terror. The company claims this feature will be used as a last resort when multiple attempts at redirection have failed and there is no hope of a productive interaction. Anthropic says most users won't experience Claude cutting a conversation short, even when discussing highly controversial topics, as this feature is reserved for "extreme edge cases." This move is part of Anthropic's research program that studies the idea of AI welfare, and the company encourages users to provide feedback when they encounter such a scenario.
Note: This is an AI-generated summary of the original article. For the full story, please visit the source link below.