tr-23-246-002 (1)

12215117476?profile=RESIZE_400xThe UK’s National Cyber Security Centre (NCSC) issued a warning this week about the growing danger of “prompt injection” attacks against applications built using AI.  While the warning is meant for cybersecurity professionals building large language models (LLMs) and other AI tools, prompt injection is worth understanding if you use any kind of AI tool, as attacks using it are likely to be a major category of security vulnerabilities going forward.

Prompt injection is a kind of attack against LL