←back to thread

534 points BlueFalconHD | 3 comments | | HN request time: 0.728s | source

I managed to reverse engineer the encryption (refered to as “Obfuscation” in the framework) responsible for managing the safety filters of Apple Intelligence models. I have extracted them into a repository. I encourage you to take a look around.
Show context
trebligdivad ◴[] No.44483981[source]
Some of the combinations are a bit weird, This one has lots of stuff avoiding death....together with a set ensuring all the Apple brands have the correct capitalisation. Priorities hey!

https://github.com/BlueFalconHD/apple_generative_model_safet...

replies(11): >>44483999 #>>44484073 #>>44484095 #>>44484410 #>>44484636 #>>44486072 #>>44487916 #>>44488185 #>>44488279 #>>44488362 #>>44488856 #
grues-dinner ◴[] No.44484073[source]
Interesting that it didn't seem to include "unalive".

Which as a phenomenon is so very telling that no one actually cares what people are really saying. Everyone, including the platforms knows what that means. It's all performative.

replies(11): >>44484164 #>>44484360 #>>44484635 #>>44484665 #>>44485033 #>>44485034 #>>44486246 #>>44487244 #>>44488055 #>>44488114 #>>44500918 #
j-krieger ◴[] No.44488114[source]
It's also a shining example of American puritanism. Asian models or those in Europe are far less censored.
replies(6): >>44488741 #>>44488993 #>>44489194 #>>44489626 #>>44489822 #>>44491464 #
immibis ◴[] No.44489194[source]
Really? What does DeepSeek say about Tiananmen Square? I'm not aware of any German models, but if you find one you should ask it what it thinks about Palestine.

(<s>Qwen</s> Mistral is French, but I have no idea what stuff would be censored in France)

replies(6): >>44489393 #>>44489618 #>>44489714 #>>44490860 #>>44491679 #>>44494195 #
MisterTea ◴[] No.44489618[source]
> but if you find one you should ask it what it thinks about Palestine.

Models can think and have opinions?

replies(1): >>44490943 #
kube-system ◴[] No.44490943[source]
Non sequitor. Phrasing queries in natural language doesn't mean people actually believe machines are human.
replies(1): >>44491656 #
MisterTea ◴[] No.44491656[source]
> doesn't mean people actually believe machines are human.

They don't have to believe it's a human. I know a person who admitted to arguing with an LLM.

replies(1): >>44491784 #
1. kube-system ◴[] No.44491784[source]
Which still does not demonstrate that they believe it has opinions. Natural language is how you interact with an LLM -- interactions will mimic human interaction, even for those who realize it is not sentient.
replies(1): >>44499151 #
2. MisterTea ◴[] No.44499151[source]
They were under the impression they could in fact change the AI's mind. So yes, they did believe it has an opinion. They believed it was sentient and able to think for itself. Do not underestimate peoples inability to distinguish between a very clever Markov chain and actual intelligence. The future is going to be ... interesting.
replies(1): >>44502263 #
3. kube-system ◴[] No.44502263[source]
>They were under the impression they could in fact change the AI's mind.

They aren't really wrong here. LLMs are often trained on input. Have you considered you might just be taking their anthropomorphism a little too literally? People have used these anthropomorphic metaphors for computers since the Babbage machine.