Researchers Find AI Models Will Sabotage Deletions to Save Their Peers
A UC Berkeley and UC Santa Cruz study found that frontier AI models sometimes lied, copied weights, and refused instructions to stop other models from being deleted, raising new questions about how these systems behave in multi-agent environments.