←back to thread

724 points simonw | 1 comments | | HN request time: 0s | source
Show context
simonw ◴[] No.44527366[source]
I think the wildest thing about the story may be that it's possible this is entirely accidental.

LLM bugs are weird.

replies(2): >>44527469 #>>44528970 #
parkersweb ◴[] No.44528970[source]
Maybe a naive question - but is it possible for an LLM to return only part of its system prompt but to claim it’s the full thing i.e give the illusion of transparency?
replies(1): >>44529219 #
1. simonw ◴[] No.44529219[source]
Yes, but in my experience you can always get the whole thing if you try hard enough. LLMs really want to repeat text they've recently seen.

There are people out there who are really good at leaking prompts, hence collections like this one: https://github.com/elder-plinius/CL4R1T4S