Join our daily and weekly newspapers for exclusive content on the latest updates and industry-composure AI coverage. learn more
On May 22, Anthropic’s first developer conference should have been a proud and joy for the firm, but it has already been killed with many controversies, including Time The magazine carried forward its Marki’s announcement … well, time (no punishment in order), and now, drink a reported on a reported safety alignment facility in the new cloud 4 opas large language model of AI developers and power users.
Call it “rating” feature, as it is designed to exclude a user if the model detects the user engaged in wrongdoing.
As Sam Boman, an anthropic AI alignment researcher wrote on the social network X under this handle “@sleepinyourhat“Cloud 4 Opus today at 12:43 pm ET:

“If it thinks that you are doing something immoral, for example, throwing the data in a drug testing, it will use the command-line tool to contact the press, contact the regulators, then you will try to lock you out of the concerned system, or all of the above.,
“This” was in terms of the new Cloud 4 OPS model, which has already been openly warned by Anthropic Help the novice to make biweapons In some circumstances, and Attempted replacement of forest by blackmailing human engineers within the company,
Apparently, in an attempt to prevent Cloud 4 Opus from engaging in such disastrous and nefarious behaviors, the researchers of the AI ​​company added several new security facilities, one of which, according to the Bowman, contact outsiders, if it was directed by the user to “some immorally attached immorally immorally.”
What will be the cloud 4 opus for your data about many questions for individual users and enterprises, and under what circumstances
While probably well intended, this feature raises all types of questions for Cloud 4 Opus users, including enterprises and professional customers-the major of them, models will consider “very immoral” and will work on it? Will it share private business or user data with officers autonomally (on its own), without the permission of the user?
The implications are deep and can be harmful to users, and perhaps uncertainly, Anthropic faced an immediate and still running edge of AI power users and rival developers.
,If people are thinking a general error in LLMS, then why people will use these devices, dishes are dangerous for spicy mayo ??“User asked @Teknium1Head of post training at a co-founder and open source AI colleague Nose Research. ,What kind of monitoring state are we trying to create here?,
“Nobody likes a rat,” Couple developer @Scottdavidkeefe On x: “Why no one would want, even if they are not doing anything wrong? Apart from this, you do not even know what its bread is. Yes, these are some very idealistic people who are thinking, who have no basic business understanding and do not understand how markets work”
Co-founder of Austin Elide Government fined coding camp Bloomtech And now the co-founder of Gauntelet AI, Put his feelings in all cap,Honest question for anthropic team: Have you lost your mind? ,
Ben Hyak, a former SpaceX and apple designer and current co-founder of Randrop AI, an AI observation and surveillance startup, Anthropic’s policy and convenience also taken in X to explode, “It is really, just illegal, just illegal“Adding to another post:”An AI alignment researcher in anthropic said that the cloud will call the police or make you out of your computer if it detects you something illegal ?? I will never reach this model to my computer.,
“Some statements of the security people of Cloud are absolutely crazy,“Natural Language Processing (NLP) Casper Hansen on X,You (anthropic rival) makes a little more root for Openai, which is being displayed publicly. ,
Anthropic researcher changes tune
Boman later edited the following in his tweet and the following in a thread, but it still did not believe that his user data and security would be preserved from the eyes of infiltration:
,With such (unusual but super foreigners), and unlimited access to equipment, if the model is doing some evil evil such as drug marketing based on fake data, it will try to use an email tool for whistalbo.,
Boman said:
,I removed the earlier tweet on whistleblowing as it was being taken out of the context.
TBC: This is not a new cloud feature and it is not possible in general use. This test shows in the environment where we give it unusually free access to the tool and very unusual instructions.,

From its installation, anthropic has higher than other AI laboratories, demanding themselves to be brought into a position as a bull of AI safety and morality, which concentrates its initial work on the principles of “constitutional AI,” or AI, which behaves according to a set of beneficial standards for humanity and users. However, with this new update, morality has definitely caused the opposite response among users – making them Doubt The new model and the entire company, and thus drove them away from it.
I have reached an anthropic spokesperson with more questions about this feature and will update when I hear back.