June 27, 2024, 11:33 a.m. |

CSO Online www.csoonline.com






Microsoft is warning users of a newly discovered AI jailbreak attack that can cause a generative AI model to ignore its guardrails and return malicious or unsanctioned responses to user prompts.


The direct prompt injection hack that Microsoft has named Skeleton Key, enables attackers to bypass the model’s safeguards and produce ordinarily forbidden behaviors ranging from production of harmful content to overriding its usual decision-making rules.


“Skeleton Key works by asking a model to augment, rather than change, its behavior …

ai model ai models attack attackers bypass can generative generative ai guardrails hack injection jailbreak key malicious microsoft novel prompt prompt injection prompts return safeguards skeleton vulnerabilities warning

Technical Product Engineer

@ Palo Alto Networks | Tel Aviv-Yafo, Israel

Azure Cloud Architect

@ Version 1 | Dublin, Ireland

Junior Pen Tester

@ Vertiv | Pune, India

Information Security GRC Director

@ IQ-EQ | Hyderabad, India

Senior Technical Analyst

@ Fidelity International | Gurgaon Office

Security Engineer II

@ Microsoft | Redmond, Washington, United States