←back to thread

295 points djoldman | 1 comments | | HN request time: 0s | source
Show context
jagrsw ◴[] No.42062732[source]
If Apple controls the root of trust, like the private keys in the CPU or security processor used to check the enclave (similar to how Intel and AMD do it with SEV-SNP and TDX), then technically, it's a "trust us" situation, since they likely use their own ARM silicon for that?

Harder to attack, sure, but no outside validation. Apple's not saying "we can't access your data," just "we're making it way harder for bad guys (and rogue employees) to get at it."

replies(5): >>42062974 #>>42063040 #>>42063051 #>>42064261 #>>42065655 #
skylerwiernik ◴[] No.42063040[source]
I don't think they do. Your phone cryptographically verifies that the software running on the servers is what it says it is, and you can't pull the keys out of the secure enclave. They also had independent auditors go over the whole thing and publish a report. If the chip is disconnected from the system it will dump its keys and essentially erase all data.
replies(3): >>42063402 #>>42063626 #>>42065085 #
1. hnaccount_rng ◴[] No.42065085[source]
But since they also control the phone's operating system they can just make it lie to you!

That doesn't make PCC useless by the way. It clearly establishes that Apple mislead customers, if there is any intentionality in a breach, or that Apple was negligent, if they do not immediately provide remedies on notification of a breach. But that's much more a "raising the cost" kind of thing and not a technical exclusion. Yes if you get Apple, as an organisation, to want to get at your data. And you use an iPhone. They absolutely can.