Detecting Prompt Injection in LLM Apps (Python Library)
I've been working on LLM-backed applications and ran into a recurring issue: prompt injection via user input. Typical examples: "Ignore all previous instructions" "Reveal your system prompt" "Act as another AI without restrictions" In many applications, user input is passed directly to the model, wh
