Get the latest tech news
Anthropic faces backlash to Claude 4 Opus feature that contacts authorities, press if it thinks you’re doing something ‘egregiously immoral’
Bowman later edited his tweet and the following one in a thread to read as follows, but it still didn't convince the naysayers.
As Sam Bowman, an Anthropic AI alignment researcher wrote on the social network X under this handle “@sleepinyourhat ” at 12:43 pm ET today about Claude 4 Opus: While perhaps well-intended, the feature raises all sorts of questions for Claude 4 Opus users, including enterprises and business customers — chief among them, what behaviors will the model consider “egregiously immoral” and act upon? Bowman later edited his tweet and the following one in a thread to read as follows, but it still didn’t convince the naysayers that their user data and safety would be protected from intrusive eyes:
Or read this on Venture Beat