Apple
Welcome
to the largest Apple community on Lemmy. This is the place where we talk about everything Apple, from iOS to the exciting upcoming Apple Vision Pro. Feel free to join the discussion!
Rules:
- No NSFW Content
- No Hate Speech or Personal Attacks
- No Ads / Spamming
Self promotion is only allowed in the pinned monthly thread
Communities of Interest:
Apple Hardware
Apple TV
Apple Watch
iPad
iPhone
Mac
Vintage Apple
Apple Software
iOS
iPadOS
macOS
tvOS
watchOS
Shortcuts
Xcode
Community banner courtesy of u/Antsomnia.
Remind me again why Federighi is so well-regarded.
He’s so terminally Applebrained that he thinks last year customers celebrated “the values we brought” to the AI discussion.
Apple is the only one running AI on-device first. Google just happily takes all your data.
Google has been quietly doing that for more than 10 years, only we didn't start really calling this stuff AI until 2022. Google had offline speech to text (and an always on local hotword detection for "hey Google") since the Moto X 2013, and added hardware support for image processing in the camera app, as images were captured.
The tasks they offloaded onto the Tensor chip starting in 2021 started opening up more image editing features (various algorithms for tuning and editing images), keyboard corrections and spelling/grammar recommendations that got better (and then worse), audio processing (better noise cancellation on calls, an always-on Shazam-like song recognition function that worked entirely offline), etc.
Apple went harder at trying to use those AI features into language processing locally and making it obvious, but personally I think that the tech industry as a whole has grossly overcorrected for trying to do flashy AI, pushed beyond the limits of what the tech can competently do, instead of the quiet background stuff that just worked, while using the specialized hardware functions that efficiently process tensor math.
Exactly this.