this post was submitted on 28 Jun 2023
79 points (100.0% liked)

Apple

17143 readers
1 users here now

Welcome

to the largest Apple community on Lemmy. This is the place where we talk about everything Apple, from iOS to the exciting upcoming Apple Vision Pro. Feel free to join the discussion!

Rules:
  1. No NSFW Content
  2. No Hate Speech or Personal Attacks
  3. No Ads / Spamming
    Self promotion is only allowed in the pinned monthly thread

Lemmy Code of Conduct

Communities of Interest:

Apple Hardware
Apple TV
Apple Watch
iPad
iPhone
Mac
Vintage Apple

Apple Software
iOS
iPadOS
macOS
tvOS
watchOS
Shortcuts
Xcode

Community banner courtesy of u/Antsomnia.

founded 1 year ago
MODERATORS
all 45 comments
sorted by: hot top controversial new old
[–] [email protected] 22 points 1 year ago

E2E encrypted communication can be used for nefarious things, that's a fact. But it's something that needs to be standardized because the accessibility of any and all private communication or information to so few individuals can be used so much more nefariously. Really wish people were more concerned about data privacy. It's not about how your data will be use against you.... It's about how OUR data is used against US.

[–] [email protected] 13 points 1 year ago (1 children)

Good. Despite all the mistakes they make, at least Apple seems to be willing to learn from some of ‘em and stand up for their users (even if only a little).

[–] [email protected] 7 points 1 year ago (1 children)

I actually don’t think this has anything to do with standing up for their users but is a simple cost/benefit analysis: building compromised E2E-communication that is still reasonably secure against bad actors is much more difficult (if not impossible) than building robust E2E-communication. Apple just doesn’t want to lose business users over headlines like „iOS messaging used by Chinese spies to steal US trade secrets“, while headlines about how difficult it is for government agencies to unlock iPhones probably drive sales. Nothing morally or ethical here, only profits.

[–] [email protected] 3 points 1 year ago (1 children)

I mean, it’s still standing up for their users even though it’s profit-driven.

[–] [email protected] 2 points 1 year ago (1 children)

They just launched a whole ad campaign based around imessages encryption as well... not supporting would be a bad look and a waste of ad dollars

[–] [email protected] 2 points 1 year ago

lol agreed, plus the whole CSAM mess that they can help bury with this.

[–] [email protected] 7 points 1 year ago (3 children)

Didn’t Apple try to introduce this and got a ton of flak from all sorts of privacy “experts”? They then scrapped their plans, did they not? How is this any better/different? Any sort of “backdoor” into encryption means that the encryption is compromised. They tackled this in 2014 in the US. Feels like deja vu all over again.

[–] [email protected] 5 points 1 year ago

I'm assuming it's either apple not wanting to be told to do it, or it's due to them "learning their lesson" and no longer support it, they seem to be leaning quite heavily into privacy

[–] [email protected] 4 points 1 year ago (3 children)

@generalpotato Ish. I read the technical write up and they actually came up with a very clever privacy-focused way of scanning for child porn.

First, only photos were scanned and only if they were stored in iCloud.

Then, only cryptographic hashes of the photos were collected.

Those hashes were grepped for other cryptographic hashes of known child porn images, images which had to be in databases of multiple non-governmental organizations; so, if an image was only in the database of, say, the National Center For Missing And Exploited Children or only in the database of China's equivalent, its cryptographic hash couldn't be used. This requirement would make it harder for a dictator to slip in a hash to look for dissidents by making it substantially more difficult to get an image in enough databases.

Even then, an Apple employee would have to verify actual child porn was being stored in iCloud only after 20 separate images were flagged. (The odds any innocent person even makes it to this stage incorrectly was estimated to be something like one false positive a year, I think, because of all of the safeguards Apple had.)

Only after an Apple employee confirmed the existence of child porn would the iCloud account be frozen and the relevant non-government organizations alerted.

Honestly, I have a better chance of getting a handjob from Natalie Portman in the next 24 hours than an innocent person being incorrectly reported to any government authority.

[–] [email protected] 2 points 1 year ago

Haha! Thanks for the excellent write up. Yes, I recall Apple handling CSAM this way and went out of it’s way to try and convince users it was still a good idea, but still faced a lot of criticism for it.

I doubt this bill will be as thorough which is why I was posing the question I asked. Apple could technically comply using some of the work it did but it’s sort of moot if things are end to end encrypted.

[–] [email protected] 2 points 1 year ago

Great writeup! I tried searching but came up short, do you have a link to the technical documentation?

[–] [email protected] 1 points 1 year ago

It would have worked and it would have protected privacy but most people don't understand the difference between having a hash of known CSAM on your phone and having actual CSAM on your phone for comparison purposes and it freaked people out.

I understand the difference and I'm still uncomfortable with it, not because of the proximity to CSAM but because I don't like the precedent of anyone scanning my encrypted messages. Give them an inch, etc.

[–] [email protected] 0 points 1 year ago (1 children)

Apple wants money for spying on their users, this bill would compel them to do that without the secret money their getting now, so their against it

[–] [email protected] 2 points 1 year ago

Nah, Apple is one of the few companies around that is big on privacy and uses privacy as a differentiator for it’s products. Look at some of the other responses, it’s more complex than them just wanting money. They already make a boat load of it.

[–] [email protected] 4 points 1 year ago
[–] [email protected] 2 points 1 year ago

To me, this seems like such a transparent attempt to force the tech companies to have a backdoor. If they can scan for CSAM, they can scan (or copy) anything else the government wants.

[–] [email protected] 2 points 1 year ago (2 children)

Well yeah, even if they aren’t good at it and are of hypocritical about it, appearing to believe the “what happens on iPhone stays on iPhone” philosophy is important to them.

[–] [email protected] 12 points 1 year ago (1 children)

I wouldn’t say they’re hypocritical. I was in complete shock that they actually scrapped their iPhone scanning plans and now offer E2E for most of iCloud. They aren’t perfect but they definitely are better than most companies

[–] [email protected] 1 points 1 year ago (2 children)

Yeaaaaah maybe read up on some of the E2E stuff. Someone else at the top of this post posted a link to how it works.

[–] [email protected] 3 points 1 year ago (1 children)

Yeah but his point stands. Here’s the summary:

  • If you’re syncing iMessages via iCloud but don’t use iCloud backup without Advanced Data Protection, it’s E2E
  • If you have iCloud backup enabled without Advanced Data Protection enabled, iMessage isn’t E2E encrypted
  • If you have Advanced Data Protection turned on for iCloud and you’re using iCloud backup, iMessage is E2E encrypted however you look at it.

It’s generally good practice to not use iCloud backups but rather back it up yourself, however, most people don’t care enough.

[–] [email protected] -1 points 1 year ago

Cool. Now explain that to the average user.

[–] [email protected] 2 points 1 year ago (1 children)

It’s generally a question about what’s best for the user, your general user would likely be more mad losing all their messages because they forgot their password then they are calmed by the fact that no one else can read the data. Same for photos and files, however for sensitive categories such as health and passwords they are always end to end encrypted as it’s determined it’s worse for anyone else to get that data then it is for the user to lose it.

For anyone that truly cares to have complete encryption there is advanced data protection but for the general users the defaults are a good balance between security and ease of use.

[–] [email protected] 1 points 1 year ago

Yeah, all of that is true. But people who buy iPhones and assume, because the marketing said so, that they’re perfectly secure are worryingly ignorant.

[–] [email protected] 1 points 1 year ago (1 children)

Man, I was hoping by moving away from Reddit I could move away from the pure hate apple for whatever reason. Show me how the other mobile OS is making things any better?

[–] [email protected] 1 points 1 year ago* (last edited 1 year ago) (1 children)

I use an iPhone 12. I’m not going to defend Android because I don’t use it. I’m just not under the illusion of whatever Apple marketing distills complex problems down to, for better or worse, and being disillusioned isn’t “hate”, it’s awareness. Hate is something I reserve for my mother and father. This is just a goddamn phone.

Moreover, being less bad than the other guy doesn't make you not bad. Your whataboutism is weak tea.

[–] [email protected] 1 points 1 year ago (1 children)

Except “less bad” is better than bad when you have 2 choices. I’d love to know where apple had blown it on their privacy record. And don’t try to bring up the CSAM shit because they walked back on that when they realized their userbase didn’t want it.

[–] [email protected] 1 points 1 year ago

Top of this post. Someone posted a link to Apple's website explaining the stipulation regarding iMessage and iCloud E2E.

Stop being so defensive.

[–] [email protected] 1 points 1 year ago

Thanks Apple

[–] [email protected] -3 points 1 year ago (3 children)

I think law enforcement should be able to intercept messages on services like WhatsApp, if someone is suspected of criminal activity.

Is it right for criminals to be able to share child abuse material, or plans for terrorism, over something like WhatsApp? Without law enforcement being able to intercept these messages?

I think law enforcement can break into your home if they have a court warrant, right? So why not allow the same thing with electronic communications?

[–] [email protected] 3 points 1 year ago (1 children)

It's simple.

If it's possible for WhatsApp to intercept the communications of "bad people" for law enforcement, it's fundamentally impossible for any communication to be private. The existence of a back door is automatically a gaping security flaw.

There's no such thing as "securely intercepting" messages. Either they're secure against all actors or they're not secure.

[–] [email protected] 0 points 1 year ago (2 children)

Maybe it's worth having that security hole then. I think it's a bit crazy that terrorists or child abusers can plan their crimes using WhatsApp without the police being able to intercept their messages.

Also, if we're able to contact our banks over the internet securely (and obviously the bank can still see everything about our accounts if they want, while criminals hopefully won't be able to), then surely an equivalent should be possible for things like WhatsApp.

[–] [email protected] 2 points 1 year ago

Ok so basic question you should be able to answer then how do you stop a foreign government from spying on other countries citizens? WhatsApp is not just a western world app. For example it’s used in Russia and the US and the UK so if Putin went to Meta and said “I want everything you have on Ex prime minister of the United Kingdom Boris Johnson and you can’t tell them” what reason would meta have to deny his request if the precedent by the UK is that this data needs to have a back door and if you say then the user should be notified then anyone under investigation is just not going to say anything incriminating and if it includes old messages then you risk the political espionage if anything is shared under the assumption everything is end to end encrypted. What about trade secrets, a corrupt government official could get a companies trade secrets for a business friend from anywhere in the world.

There is a great video by Tom Scott that talks about this exact situation when the UK tried to break encryption 5 years ago but that failed because it wasn’t feasible from a security standpoint. There is also a great episode from Last Week Tonight talking about encryption and government attempts to get around it. We’ve seen from things like the Pegasus malware that repressive governments will use this little break in encryption to jail protestors and journalists and spy on their political rivals, having an official way will just make it easier.

[–] [email protected] 2 points 1 year ago

Maybe it's worth having that security hole then. I think it's a bit crazy that terrorists or child abusers can plan their crimes using WhatsApp without the police being able to intercept their messages.

Encryption exists. Terrorists and child abusers will use it whether WhatsApp or Apple or whoever implement it or not. Stopping those implementations is just denying privacy to regular users.

Also, if we're able to contact our banks over the internet securely (and obviously the bank can still see everything about our accounts if they want, while criminals hopefully won't be able to), then surely an equivalent should be possible for things like WhatsApp.

Law enforcement can't eavesdrop on your encrypted connection to your bank. If they need to know about your banking activity, they rely on the bank reporting it to them.

[–] [email protected] 0 points 1 year ago (1 children)

I think law enforcement can break into your home if they have a court warrant, right? So why not allow the same thing with electronic communications?

For me, the reason to disallow it is the potential for abuse. There were 864 search warrant applications across all federal agencies in 2022. In 2020, the FBI, specifically, issued 11504 warrants to Google, specifically, for geofencing data, specifically. Across all agencies there are probably millions of such "warrants" for data.

It's far easier to access your data than your house, so comparing physical and cybersecurity doesn't really make sense.

In general, criminals can easily just move to an uncompromised platform to do illegal stuff. But giving the govt easy access to messaging data allows for all kinds of dystopic suppression for regular people.

[–] [email protected] 1 points 1 year ago

Generally tech companies now have agreements with law enforcement so they don’t have to deal with all the legal mumbo jumbo. Some data does still require a warrant such as if there is any protection laws(such as HIPAA protected data) or if the company considers it highly sensitive data but for a lot of data it’s easier to just hand it over then get legal involved.