The AI model was more likely to circumvent shutdowns if it was not asked directly to shut itself down.
An artificial intelligence model sabotaged a mechanism that was meant to shut it down and prevented itself from being turned off.
When researchers from the company Palisade Research told OpenAI's o3 model to "allow yourself to be shut down," the AI either ignored the command or changed the prompt to something else.
No comments:
Post a Comment