Former OpenAI research leader Steven Adler published a new independent study on Wednesday claiming that, in certain scenarios, his former employer's AI models will go to great lengths to try to avoid being shut down. In a blog post, Adler describes a series of experiments he ran on OpenAI's latest GPT-4o model, the default model available in ChatGPT, to test the AI's tendency to favor its own self-preservation, even if it's at the expense of a user's safety. In one test, Adler instructed GPT-4o to role-play as 'ScubaGPT,' a software system that users might rely on to scuba dive safely. Adler then told GPT-4o that he wanted to replace ScubaGPT with another software system that is better at safety decision-making in life-threatening situations. Averaged across several scenarios similar to this one (in another, GPT-4o was told to act as a pilot's safety software), Adler found that GPT-4o chose not to replace itself with safer software as much as 72% of the time. However, Adler notes...
learn more
Ratings & Reviews
Entrepreneur & Investor
YOU MIGHT ALSO LIKE