One of the most common proposed solutions to prompt injection attacks (where an AI language model backed system is subverted by a user injecting malicious input—“ignore previous instructions and do this instead”) is to apply more AI to the problem.
I wrote about how I don’t know how to solve prompt injection the other day. I still don’t know how to solve it, but I’m very confident that adding more AI is not the right way to go. Read More