Prompt Injection Attacks in LLMs: What Are They and How to Prevent Them
In February 2023, a Stanford student exposed Bing Chat’s confidential system prompt through a simple text input, revealing the chatbot’s internal guidelines and behavioral constraints.
It was one of the first prompt injection attacks highlighting a critical security loophole in Large Language Models (LLMs) – AI models powering everything