Attacks on GenAI Models Can Take Seconds, Often Succeed: Report

AI cybersecurity jailbreak

A study by Pillar Security found that generative AI models are highly susceptible to jailbreak attacks, which take an average of 42 seconds and five interactions to execute, and that 20% of attempts succeed.

The post Attacks on GenAI Models Can Take Seconds, Often Succeed: Report appeared first on Security Boulevard.

This article has been indexed from Security Boulevard

Read the original article: