Prompt injections are attacks against LLM applications where an attacker is able to override the original instructions of the programmer. Neither input validation nor output filtering have so far been successful in mitigating this kind of attack. The reviewed paper presents a novel approach to defend against prompt injections by substituting (“signing”) the original instructions with random keywords very unlikely to appear in natural language. This blog post reviews the paper named “Signed-Prompt: A New Approach to Prevent Prompt Injection ...