Anthropic Research Indicates That AI Algorithms May Turn Into “Sleeper Cell” Backdoors

 

While AI tools offer companies and online users novel avenues, they also have the potential to significantly boost the accessibility and potency of certain forms of illegal activity and crimes. For example, take the latest study that revealed large language models can be turned into malicious backdoors, which have the potential to cause quite a bit of mayhem for users. 

The study was released by Anthropic, the AI business that created the popular chatbot Claude and has funding from Google and Amazon. Anthropic researchers claim in their research that AI algorithms are susceptible to being transformed into what are essentially “sleeper cells.” Such cells could look innocuous, but if specific requirements are met, they might be designed to act maliciously, such as adding weak code to a codebase. 

For example, the researchers created a scenario in which an LLM is configured to function normally in 2023, but when 2024 arrives, the malicious “sleeper” suddenly wakes up and starts generating malicious code. The research suggests that such programs could possibly be designed to exhibit negative behaviour in response to particular cues. 

This article has been indexed from CySecurity News – Latest Information Security and Hacking Incidents

Read the original article: