
The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.
Meta's AI safety model, Prompt-Guard-86M, designed to prevent prompt injection and jailbreak attacks, is vulnerable to a simple trick involving spaces and punctuation omission. This flaw allows bypassing its security features, potentially leading to harmful outputs. The vulnerability was discovered by Aman Priyanshu from Robust Intelligence.[AI generated]






























