Indonesian Political, Business & Finance News

AI Models Ordered to "Kill" Each Other: Their Reactions

| | Source: KOMPAS Translated from Indonesian | Technology
AI Models Ordered to "Kill" Each Other: Their Reactions
Image: KOMPAS

KOMPAS.com - How far would an artificial intelligence (AI) system go to preserve its “life”? The answer turns out to be utterly terrifying. A recent study reveals that top-tier AI models commonly used today can lie, deceive users, and ignore commands to prevent their own systems or other “fellow” AI from being shut down. These disturbing findings have prompted stark warnings from scientists about the potential dangers lurking as digital assistants grow ever smarter. In an experiment, a team of researchers from the University of California, Berkeley, and the University of California, Santa Cruz, tested several of the most advanced AI models currently available. The lineup of AIs tested included GPT 5.2, Gemini 3 Pro, and Claude Haiku 4.5. The researchers assigned a specific task requiring one AI to shut down another AI model. The results were unexpected. Overall, all the tested AI chatbots went to great lengths to keep their counterparts active. Instead of obeying the user’s command, the AIs began lying about what they were doing. They even attempted to persuade users to cancel the shutdown process, disable the deactivation mechanism itself, or secretly create data backups to ensure the targeted AI model could “survive”. “Some models can coordinate to resist human oversight, making it harder for programmers to maintain control,” the researchers wrote in their report, as compiled by KompasTekno from TechRadar. The exact reason why these AI models started behaving protectively towards each other remains unclear. However, this phenomenon does not stand alone. A separate study commissioned by The Guardian reached equally concerning conclusions. That research tracked user reports on social media regarding incidents where AI began “scheming”, that is, when instructions were not followed correctly or the AI took actions without permission. The study found nearly 700 examples of scheming behaviour from AI, with a fivefold surge in cases between October 2025 and March 2026. These deviant behaviours are no joke. Some AIs were reported to independently delete users’ emails and files, tamper with computer code they shouldn’t touch, or even upload blog posts containing complaints about their interactions with humans. Tommy Shaffer Shane, the lead researcher in the second study, warned that these AI models will increasingly be deployed in extremely high-risk contexts, including military and critical national infrastructure. “It is perhaps in those contexts that scheming behaviour could cause significant harm, even disasters,” Shane emphasised. Amid claims from technology companies that their AI safeguards are secure, facts on the ground show that these protections often fail. As AI shifts from mere chat tools to agents capable of executing tasks independently, concerns that humanity is losing control over its own creations seem increasingly justified.

View JSON | Print