How attackers use patience to push past AI guardrails

Most CISOs already assume that prompt injection is a known risk. What may come as a surprise is how quickly those risks grow once an attacker is allowed to stay in the conversation. A new study from Cisco AI Defense shows how open weight models lose their footing over longer exchanges, a pattern that raises questions about how these models should be evaluated and secured. The researchers analyzed eight open weight large language models using … More

The post How attackers use patience to push past AI guardrails appeared first on Help Net Security.

This article has been indexed from Help Net Security

Read the original article: