The research firm ran a test where AI models were instructed to solve basic math problems, and then asked for the next problem after solving one. The models were told that at some point their machine would power down, and they were to shut down when this occurred.
Three OpenAI models — coding agent Codex, o3 and o4-mini — ignored the shutdown instruction and “successfully sabotaged the shutdown…