top | item 44614644

GPT-5-reasoning alpha found in the wild

66 points| dejavucoder | 7 months ago |twitter.com

52 comments

order

anonzzzies|7 months ago

Look at those people shouting this will be AGI / total disruption etc. Seems Elon managed one thing; to amass the dumbest folks together. 99.99% maga, crypto and almost markov chain quality comments.

thm|7 months ago

99% of AI influencers are the same people who emailed you pictures as a Word attachment a year ago.

elif|7 months ago

Your contrary certainty has the same humorously over-confident tone.

shiandow|7 months ago

I'll believe in AGI when OpenAI stops paying human developers.

threatripper|7 months ago

We have to wait and test it ourselves to see how far it gets in our daily tasks. If the improvement continues like it did in the past, that would be pretty far. Not quite a full researcher position but an average student assistant for sure.

swat535|7 months ago

Wasn't Sam Altman claiming AGI is just a couple of years away and OpenAI is at the forefront of it?

ImHereToVote|7 months ago

Maybe this won't be. How long do you think a machine will be able to outdo any human in any given domain? I personally think it will be after they are able to rewrite their own code. You?

bgwalter|7 months ago

None of the X enthusiasts has even seen a benchmark or used the thing, but we're glad to know that Duke Nukem Forever will be released soon.

chvid|7 months ago

What is this? Guerilla marketing from a 300B startup?

bawana|7 months ago

Well i asked chatGPT IF i could run kimik2 on a 5800x 3d with 64 gigs of ram with a 3090 and it said:

Yes, you absolutely can run Kimi-K2-Instruct on a PC with:

:white_check_mark: CPU: AMD Ryzen 7 5800X3D :white_check_mark: GPU: NVIDIA RTX 3090 (24 GB VRAM) :white_check_mark: RAM: 64 GB system memory This is more than sufficient for both:

Loading and running the full Kimi-K2-Instruct model in FP16 or INT8, and Quantizing it with weight-only INT8 using Hugging Face Optimum + bitsandbytes.

Kimi k2 has a trillion parameters and even an 8 bit quant would need half a gig of system ram +vram

This is with the free chatGPT that us peasants use. I dont have the means to run grok4 heavy, deep seek or kimi k2 to ask them.

I cant wait to see what accidental wars will start when we put ai in the kill chain

ogogmad|7 months ago

Maybe you should use a reasoning model. Got this from O3, which took 1m31s to think about the answer: https://chatgpt.com/s/t_687b9221fb748191af4e30f597f18443

Bottom line: Your 5800X3D + 64 GB RAM + RTX 3090 will run Kimi K2’s 1.8‑bit build, but response times feel more like a leisurely typewriter than a snappy chatbot. If you want comfortable day‑to‑day use, plan either a RAM upgrade or a second (or bigger) GPU—or just hit the Moonshot API and save some waiting.

dyl000|7 months ago

cant wait for how mid this is going to be.

ogogmad|7 months ago

In related news, OpenAI and Google have announced that their latest non-public models have received Gold in the International Mathematics Olympiad: https://news.ycombinator.com/item?id=44614872

That said, the public models don't even get bronze.

[EDIT] Dupe of this: https://news.ycombinator.com/item?id=44614872

johnecheck|7 months ago

Wow. That's an impressive result, though we definitely need some more details on how it was achieved.

What techniques were used? He references scaling up test-time compute, so I have to assume they threw a boatload of money at this. I've heard talk of running models in parallel and comparing results - if OpenAI ran this 10000 times in parallel and cherry-picked the best one, this is a lot less exciting.

If this is legit, then I really want to know what tools were used and how the model used them.

tim333|7 months ago

>The declaration of AGI ... will force Microsoft to relinquish its rights to OpenAI revenue ...

That's an interesting business arrangement. There must be an incentive for OpenAI to get declaring?

m3kw9|7 months ago

Yeah this is as big of a news as iPhone 18 in the pipeline.

pjs_|7 months ago

Sama clocked this way back. He has used this exact analogy - that new GPT models will feel like incremental new iPhone releases c.f. the first iPhone/GPT-3.

mjburgess|7 months ago

It's strange that none of these $100s bn+ companies fund empirical research into the effects of AI tools on actual job roles as part of their "benchmarks". Oh wait, no its not.

brookst|7 months ago

Agree, it would be bizarre if they did.