all InfoSec news
Microsoft warns of novel jailbreak affecting many generative AI models
CSO Online www.csoonline.com
Microsoft is warning users of a newly discovered AI jailbreak attack that can cause a generative AI model to ignore its guardrails and return malicious or unsanctioned responses to user prompts.
The direct prompt injection hack that Microsoft has named Skeleton Key, enables attackers to bypass the model’s safeguards and produce ordinarily forbidden behaviors ranging from production of harmful content to overriding its usual decision-making rules.
“Skeleton Key works by asking a model to augment, rather than change, its behavior …
ai model ai models attack attackers bypass can generative generative ai guardrails hack injection jailbreak key malicious microsoft novel prompt prompt injection prompts return safeguards skeleton vulnerabilities warning