←back to thread

645 points helloplanets | 3 comments | | HN request time: 0.338s | source
Show context
ec109685 ◴[] No.45005397[source]
It’s obviously fundamentally unsafe when Google, OpenAI and Anthropic haven’t released the same feature and instead use a locked down VM with no cookies to browse the web.

LLM within a browser that can view data across tabs is the ultimate “lethal trifecta”.

Earlier discussion: https://news.ycombinator.com/item?id=44847933

It’s interesting that in Brave’s post describing this exploit, they didn’t reach the fundamental conclusion this is a bad idea: https://brave.com/blog/comet-prompt-injection/

Instead they believe model alignment, trying to understand when a user is doing a dangerous task, etc. will be enough. The only good mitigation they mention is that the agent should drop privileges, but it’s just as easy to hit an attacker controlled image url to leak data as it is to send an email.

replies(7): >>45005444 #>>45005853 #>>45006130 #>>45006210 #>>45006263 #>>45006384 #>>45006571 #
ivape ◴[] No.45006384[source]
A smart performant local model will be the equivalent of having good anti-virus and firewall software. It will be the only thing between you and wrong prompts being sent every which way from which app.

We’re probably three or four years away from the hardware necessary for this (NPUs in every computer).

replies(1): >>45006570 #
ec109685 ◴[] No.45006570[source]
A local LLM wouldn’t have helped at all here.
replies(1): >>45006576 #
1. ivape ◴[] No.45006576[source]
You can’t imagine a MITM LLM that sits between you and the world?
replies(2): >>45007079 #>>45007850 #
2. QuadmasterXLII ◴[] No.45007079[source]
Local llms can get offline searched for vulnerabilities using gradient based attacks. they will always be very easy to prompt inject.
3. solid_fuel ◴[] No.45007850[source]
I can't imagine how such a thing would _help_, it seems like it would just be another injection target.