(no title)
mackopes | 6 months ago
You can't get a consumer-grade GPU with enough VRAM to run a large model, but you can do so with macbooks.
I wonder if doubling down on that and shipping devices that let you run third party AI models locally and privately will be their path.
If only they made their unified memory faster as that seems to be the biggest bottleneck regarding LLMs and their tk/s performance.
ChocolateGod|6 months ago
You can if you're willing to trust a modded GPU with leaked firmware from a Chinese backshop
Firerouge|6 months ago
gmays|6 months ago
We may care about running LLMs locally, but 99% of consumers don't. They want the easiest/cheapest path, which will always be the cloud models. Spending ~$6k (what my M4 Max cost) every N years since models/HW keep improving to be able to run a somewhat decent model locally just isn't a consumer thing. Nonviable for a consumer hardware business at Apple's scale.
unknown|6 months ago
[deleted]
csomar|6 months ago
Of course nobody knows how this will eventually play out. But people without inside information on what these big organizations have/possess, cannot make such predictions.
karmakaze|6 months ago
A pair of MaxSun/Intel Arc B60 48GB GPUs (dual 24GB B580's on one card) for $1200 each also outperforms the M4 Max.
tyleo|6 months ago
The tangible hardware you point out is $2,400 for two niche-specific components vs the Apple hardware which benefits more general use cases.
insane_dreamer|6 months ago
please point me to the laptop with these
orbifold|6 months ago
vonneumannstan|6 months ago
It is? I haven't seen anything about this.
billbrown|6 months ago
https://github.com/ml-explore/mlx/pull/1983
stefan_|6 months ago
mackopes|6 months ago