GitHub Copilot Jailbreak Vulnerability Let Attackers Train Malicious Models

Researchers have uncovered two critical vulnerabilities in GitHub Copilot, Microsoft’s AI-powered coding assistant, that expose systemic weaknesses in enterprise AI tools.  The flaws—dubbed “Affirmation Jailbreak” and “Proxy Hijack”—allow attackers to bypass ethical safeguards, manipulate model behavior, and even hijack access to premium AI resources like OpenAI’s GPT-o1. These findings highlight the ease with which AI […]

The post GitHub Copilot Jailbreak Vulnerability Let Attackers Train Malicious Models appeared first on Cyber Security News.

This article has been indexed from Cyber Security News

Read the original article: