GenAI models are easily compromised

95% of cybersecurity experts express low confidence in GenAI security measures while red team data shows anyone can easily hack GenAI models, according to Lakera. Attack methods specific to GenAI, or prompt attacks, are easily used by anyone to manipulate the applications, gain unauthorized access, steal confidential data and take unauthorized actions. Realizing this, only 5% of the 1,000 cybersecurity experts surveyed have confidence in the security measures protecting their GenAI applications, even though 90% … More

The post GenAI models are easily compromised appeared first on Help Net Security.

This article has been indexed from Help Net Security

Read the original article: