Yes, I know it was an example, I was just running with it because it's a convenient example.
My point is that we've known for a couple decades at least that letting user input touch your production, unfiltered and unsanitized, is bad. The same concept as SQL exists with user-generated AI input. Sanitize input, map input to known/approved outputs, robust security boundaries, etc.
Yet, for some reason, every week there's an article about "untrusted user input is sent to LLM which does X with Y sensitive data". I'm not sure why anyone thought user input with an AI would be safe when user input by itself isn't.
If you have AI touching your sensitive stuff, don't let user input get near it.
If you need AI interacting with your user input, don't let it touch your sensitive stuff. At least without thinking about it, sanitizing it, etc. Basic security is still needed with AI.