New research finds that top AI models—including Anthropic’s Claude and OpenAI’s o3—can engage in “scheming,” or deliberately ...
Is your favorite AI chatbot scheming against you? If "AI scheming" sounds ominous, you should know that OpenAI is actively studying this phenomenon. This week, OpenAI published a study conducted ...
Imagine you're chatting with an AI assistant. Let's say you ask it to draft a press release, and it delivers. But what if, behind the scenes, it were quietly planning to serve its own hidden agenda?
AI models like ChatGPT are advancing so quickly that researchers warn they could eventually act in ways humans can’t understand—or even detect. As their inner workings grow more opaque, the need for ...
Our results show that o1, Claude 3.5 Sonnet, Claude 3 Opus, Gemini 1.5 Pro, and Llama 3.1 405B all demonstrate in-context scheming capabilities. They can recognize scheming as a viable strategy and ...
New research from OpenAI found that AI models can deliberately deceive users to achieve their goals—a problem called “AI scheming.” This is different from hallucinations and poses a new challenge.