top | item 46593649

(no title)

Sevii | 1 month ago

Apple's goal is likely to run all inference locally. But models aren't good enough yet and there isn't enough RAM in an iPhone. They just need Gemini to buy time until those problems are resolved.

discuss

order

kennywinker|1 month ago

That was their goal, but in the past couple years they seem to have given up on client-side-only ai. Once they let that go, it became next to impossible to claw back to client only… because as client side ai gets better so does server side, and people’s expectations scale up with server side. And everybody who this was a dealbreaker for left the room already.

WorldMaker|1 month ago

Apple thinks they can get a best-of-both-worlds approach with Private Cloud Compute. They believe they can secure private servers specialized to specific client devices in a way that the cloud compute effort is still "client-side" from a trust standpoint, but still able to use extra server-side resources (under lock and key).

I don't know how close to that ideal they've achieved, but especially given this announcement is partly baked on an arrangement with Google that they are allowed to run Gemini on-device and in Private Cloud Compute, without using Google's more direct Gemini services/cloud, I'm excited that they are trying and I'm interested in how this plays out.

mr_toad|1 month ago

Phones will get upgrades, but then so will servers. The local models will always be behind the state of the art running on big iron. You can’t expect to stand still and keep up with the Red Queen.

O5vYtytb|1 month ago

Well DRAM prices aren't going down soon so I see this as quite the push away from local inference.