Meta’s AI safety system defeated by space bar
Meta’s AI safety system defeated by space bar ‘Ignore previous instructions’ thwarts Prompt-Guard model if you just add good ol’ ASCII code 32 Meta’s machine-learning model for detecting prompt injection attacks, Prompt-Guard-86M, is vulnerable to prompt injection attacks. The model was introduced by Meta with its Llama 3.1 generative model last week. It’s meant to … Read more