top | item 42559467

(no title)

timtom123 | 1 year ago

So much spam around this model. LocalLLaMA is stuffed with spam posts and even hacker news is getting spammed. Who has actually ran this model and verified performance? Does anyone know of a decent review from a trustworthy source?

discuss

order

starfezzy|1 year ago

Where’s the spam?

I scrolled dozens of posts without seeing a single mention of this—the biggest (certainly the most interesting) LLM news recently. When something big happens with Claude or ChatGPT there are more posts, but nobody calls that “spam”.

Anyways, if you were actually following locallama (a subreddit about running LLMs locally, where this is by far the biggest and most relevant news topic currently) you’d have seen this post https://www.reddit.com/r/LocalLLaMA/s/Yay5njt963 where a guy is working on running deepseek on llamacpp and demonstrates ~8tk/s using a cpu.

timtom123|1 year ago

I am not GPU poor and don't care about speed. I care about how good the model is which is much harder to measure and much harder to do. I have not seen many independent reviews. There are finally some coming out now but a lot of this is just marketing hype to drive attention. Every AI company does it.

x_may|1 year ago

The LMSYS leaderboards are crowdsourced and would be hard to fake, it showing a pretty strong performance in terms of human preference.

paxys|1 year ago

Crowdsourced data is the easiest to fake unless you can somehow ensure that you have a completely unbiased population (which is impossible). There's a reason why certain models do so well on upvote-based leaderboards but rank nowhere on objective tests.

feverzsj|1 year ago

I've tried it. It's average at best. Nothing comparable to ChatGPT.