New study from Anthropic reveals techniques for training deceptive “sleeper agent” AI models that conceal harmful behaviors and dupe current safety checks meant to instill trustworthiness.
This article has been indexed from Security News | VentureBeat