top | item 44568114

(no title)

lsy | 7 months ago

I think two things can be true simultaneously:

1. LLMs are a new technology and it's hard to put the genie back in the bottle with that. It's difficult to imagine a future where they don't continue to exist in some form, with all the timesaving benefits and social issues that come with them.

2. Almost three years in, companies investing in LLMs have not yet discovered a business model that justifies the massive expenditure of training and hosting them, the majority of consumer usage is at the free tier, the industry is seeing the first signs of pulling back investments, and model capabilities are plateauing at a level where most people agree that the output is trite and unpleasant to consume.

There are many technologies that have seemed inevitable and seen retreats under the lack of commensurate business return (the supersonic jetliner), and several that seemed poised to displace both old tech and labor but have settled into specific use cases (the microwave oven). Given the lack of a sufficiently profitable business model, it feels as likely as not that LLMs settle somewhere a little less remarkable, and hopefully less annoying, than today's almost universally disliked attempts to cram it everywhere.

discuss

order

Some comments were deferred for faster rendering.

strange_quark|7 months ago

> There are many technologies that have seemed inevitable and seen retreats under the lack of commensurate business return (the supersonic jetliner)

I think this is a great analogy, not just to the current state of AI, but maybe even computers and the internet in general.

Supersonic transports must've seemed amazing, inevitable, and maybe even obvious to anyone alive at the time of their debut. But hiding under that amazing tech was a whole host of problems that were just not solvable with the technology of the era, let alone a profitable business model. I wonder if computers and the internet are following a similar trajectory to aerospace. Maybe we've basically peaked, and all that's left are optimizations around cost, efficiency, distribution, or convenience.

If you time traveled back to the 1970s and talked to most adults, they would have witnessed aerospace go from loud, smelly, and dangerous prop planes to the 707, 747 and Concorde. They would've witnessed the moon landings and were seeing the development of the Space Shuttle. I bet they would call you crazy if you told this person that 50 years later, in 2025, there would be no more supersonic commercial airliners, commercial aviation would basically look the same except more annoying, and also that we haven't been back to the moon. In the previous 50 years we went from the Wright Brothers to the 707! So maybe in 2075 we'll all be watching documentaries about LLMs (maybe even on our phones or laptops that look basically the same), and reminiscing about the mid-2020s and wondering why what seemed to be such a promising technology disappeared almost entirely.

kenjackson|7 months ago

I think this is both right and wrong. There was a good book that came out probably 15 years ago about how technology never stops in aggregate, but individual technologies tend to grow quickly and then stall. Airplane jets were one example in the book. The reason why I partially note this as wrong is that even in the 70s people recognized that supersonic travel had real concrete issues with no solution in sight. I don't think LLMs share that characteristic today.

A better example, also in the book, are skyscrapers. Each year they grew and new ones were taller than the ones last year. The ability to build them and traverse them increased each year with new technologies to support it. There wasn't a general consensus around issues that would stop growth (except at more extremes like air pressure). But the growth did stop. No one even has expectations of taller skyscrapers any more.

LLMs may fail to advance, but not because of any consensus reason that exists today. And it maybe that they serve their purpose to build something on top of them which ends up being far more revolutionary than LLMs. This is more like the path of electricity -- electricity in itself isn't that exciting nowadays, but almost every piece of technology built uses it.

I fundamentally find it odd that people seem so against AI. I get the potential dystopian future, which I also don't want. But the more mundane annoyance seems odd to me.

Earw0rm|7 months ago

From a system optimisation perspective, SSTs solved the wrong problem.

Want to save people time flying? Solve the grotesque inefficiency pit that is airport transit and check-in.

Like, I'm sorry, STILL no high speed, direct to terminal rail at JFK, LAX and a dozen other major international airports? And that's before we get to the absolute joke of "border security" and luggage check-in.

Sure, supersonic afterburning engines are dope. But it's like some 10GHz single-core CPU that pulls 1.2kW out of the wall. Like it or not, an iPhone 16 delivers far more compute utility in far more scenarios.

codebolt|7 months ago

The crucial point is that we simply do not know yet if there is an inherent limitation in the reasoning capabilities of LLMs, and if so whether we are currently near to pushing up against them. It seems clear that American firms are still going to increase the amount of compute by a lot more (with projects like the Stargate factory), so time will tell if that is the only bottleneck to further progress. There might also still be methodological innovations that can push capabilities further.

gniv|7 months ago

I don't think we're anywhere near peak capability for LLMs yet. It won't take 50 years but still it's been just 4 years.

hilux|7 months ago

> So maybe in 2075 we'll all be watching documentaries about LLMs (maybe even on our phones or laptops that look basically the same), and reminiscing about the mid-2020s and wondering why what seemed to be such a promising technology disappeared almost entirely.

It's hard for me to believe that anyone who works with technology in general, and LLMs in particular, could think this.

Lu2025|7 months ago

Oh no, LLMs won't disappear but they will be a lot less loud.

Progress is often an S shaped curve and we are nearing saturation.

yieldcrv|7 months ago

slower, no fast option, no smoking in the cabins, less leg room, but with TVs plastered on the back of every chair, sometimes

its actually kind of scary to think of a world where generative AI in the cloud goes away due to costs, in favor of some other lesser chimera version that can't currently be predicted

but good news is that locally run generative AI is still getting better and better with fewer and fewer resources consumed to use

SJC_Hacker|7 months ago

The problem with supersonic commercial jets was mainly one of marketing/politics. The so called "sonic boom" problem was vastly overhyped, as anyone who lives near an air force base can tell you.

The conspiracy theorist tells me the American aerospace manufacturers at the time (Boening, McDonnell-Douglas, etc.), did everything they could to kill the Concorde. With limited flyable routes (NYC and DC to Paris and London I think were the only ones), the financials didn't make sense. If overland routes were available, especially opening up LA, San Francisco and Chicago, it might have been a different story.

brokencode|7 months ago

> “most people agree that the output is trite and unpleasant to consume”

That is a such a wild claim. People like the output of LLMs so much that ChatGPT is the fastest growing app ever. It and other AI apps like Perplexity are now beginning to challenge Google’s search dominance.

Sure, probably not a lot of people would go out and buy a novel or collection of poetry written by ChatGPT. But that doesn’t mean the output is unpleasant to consume. It pretty undeniably produces clear and readable summaries and explanations.

pera|7 months ago

> People like the output of LLMs so much that ChatGPT is the fastest growing app ever

While people seem to love the output of their own queries they seem to hate the output of other people's queries, so maybe what people actually love is to interact with chatbots.

If people loved LLM outputs in general then Google, OpenAI and Anthropic would be in the business of producing and selling content.

underdeserver|7 months ago

> That is a such a wild claim. People like the output of LLMs so much that ChatGPT is the fastest growing app ever.

The people using ChatGPT like its output enough when they're the ones reading it.

The people reading ChatGPT output that other people asked for generally don't like it. Especially if it's not disclosed up front.

hattmall|7 months ago

I'm not really countering that ChatGPT is popular, it certainly is, but it's also sort of like "fastest growing tire brand" that came along with the adoption of vehicles. The amount of smartphone users is also growing at the fastest rate ever so whatever the new most popular app is has a good chance of being the fastest growing app ever.

ants_everywhere|7 months ago

> That is a such a wild claim.

Some people who hate LLMs are absolutely convinced everyone else hates them. I've talked with a few of them.

I think it's a form of filter bubble.

sejje|7 months ago

Maybe he's referencing how people don't like when other humans post LLM responses in the comments.

"Here's what chatGPT said about..."

I don't like that, either.

I love the LLM for answering my own questions, though.

xnx|7 months ago

> AI apps like Perplexity are now beginning to challenge Google’s search dominance

Now that is a wild claim. ChatGPT might be challenging Google's dominance, but Perplexity is nothing.

tikhonj|7 months ago

At some point, Groupon was the fastest growing company ever.

johnnyanmac|7 months ago

People "like" or people "suffice" with the output? This "rise of whatever" as one blog put it gives me feelings that people are instead lowering their standards and cutting corners. Letting them cut through to stuff they actually want to do.

satvikpendem|7 months ago

> People like the output of LLMs so much that ChatGPT is the fastest growing app ever

And how much of that is free usage, like the parent said? Even when users are paying, ChatGPT's costs are larger than their revenue.

JohnMakin|7 months ago

> That is a such a wild claim. People like the output of LLMs so much that ChatGPT is the fastest growing app ever.

And this kind of meaningless factoid was immediately usurped by the Threads app release, which IMO is kind of a pointless app. Maybe let's find a more meaningful metric before saying someone else's claim is wild.

shpongled|7 months ago

I would pay $5000 to never have to read another LLM-authored piece of text ever again.

Wowfunhappy|7 months ago

...I do wonder what percent of ChatGPT usage is just students cheating on their homework, though.

const_cast|7 months ago

> > “most people agree that the output is trite and unpleasant to consume”

> That is a such a wild claim.

I think when he said "consume" he meant in terms of content consumption. You know, media - the thing that makes Western society go round. Movies, TV, music, books.

Would I watch an AI generated movie? No. What about a TV show? Uh... no. What about AI music? I mean, Spotify is trying to be tricky with that one, but no. I'd rather listen to Remi Wolf's 2024 Album "Big Ideas", which I thought was, ironically, less inspired than "Juno" but easily one of the best albums of the year.

ChatGPT is a useful interface, sure, but it's not entertaining. It's not high-quality. It doesn't provoke thought or offer us some solace in times of sadness. It doesn't spark joy or make me want to get up and dance.

alonsonic|7 months ago

I'm confused with your second point. LLM companies are not making any money from current models? Openai generates 10b USD ARR and has 100M MAUs. Yes they are running at a loss right now but that's because they are racing to improve models. If they stopped today to focus on optimization of their current models to minimize operating cost and monetizing their massive user base you think they don't have a successful business model? People use this tools daily, this is inevitable.

lordnacho|7 months ago

Are you saying they'd be profitable if they didn't pour all the winnings into research?

From where I'm standing, the models are useful as is. If Claude stopped improving today, I would still find use for it. Well worth 4 figures a year IMO.

dvfjsdhgfv|7 months ago

> If they stopped today to focus on optimization of their current models to minimize operating cost and monetizing their user base you think they don't have a successful business model?

Actually, I'd be very curious to know this. Because we already have a few relatively capable models that I can run on my MBP with 128 GB of RAM (and a few less capable models I can run much faster on my 5090).

In order to break even they would have to minimize the operating costs (by throttling, maiming models etc.) and/or increase prices. This would be the reality check.

But the cynic in me feels they prefer to avoid this reality check and use the tried and tested Uber model of permanent money influx with the "profitability is just around the corner" justification but at an even bigger scale.

ehutch79|7 months ago

Revenue is _NOT_ Profit

dkdbejwi383|7 months ago

How many of those MAUs are crappy startups building a janky layer on top of the OpenAI API which will cease to exist in 2 years?

Forgeties79|7 months ago

> that's because they are racing improve models. If they stopped today to focus on optimization of their current models to minimize operating cost and monetizing their user base you think they don't have a successful business model?

I imagine they would’ve flicked that switch if they thought it would generate a profit, but as it is it seems like all AI companies are still happy to burn investor money trying to improve their models while I guess waiting for everyone else to stop first.

I also imagine it’s hard to go to investors with “while all of our competitors are improving their models and either closing the gap or surpassing us, we’re just going to stabilize and see if people will pay for our current product.”

bbor|7 months ago

It’s just the natural counterpart to dogmatic inevitabilism — dogmatic denialism. One denies the present, the other the (recent) past. It’s honestly an understandable PoV though when you consider A) most people understand “AI” and “chatbot” to be synonyms, and B) the blockchain hype cycle(s) bred some deep cynicism about software innovation.

Funny seeing that comment on this post in particular, tho. When OP says “I’m not sure it’s a world I want”, I really don’t think they’re thinking about corporate revenue opportunities… More like Rehoboam, if not Skynet.

mc32|7 months ago

Making money and operating at a loss contradict each other. Maybe someday they’ll make money —but not just yet. As many have said they’re hoping capturing market will position them nicely once things settle. Obviously we’re not there yet.

airstrike|7 months ago

No, because if they stop to focus on optimizing and minimizing operating costs, the next competitor over will leapfrog them with a better model in 6-12 months, making all those margin improvements an NPV negative endeavor.

827a|7 months ago

One thing we're seeing in the software engineering agent space right now is how many people are angry with Cursor [1], and now Claude Code [2] (just picked a couple examples; you can browse around these subreddits and see tons of complaints).

What's happening here is pretty clear to me: Its a form of enshittification. These companies are struggling to find a price point that supports both broad market adoption ($20? $30?) and the intelligence/scale to deliver good results ($200? $300?). So, they're nerfing cheap plans, prioritizing expensive ones, and pissing off customers in the process. Cursor even had to apologize for it [3].

There's a broad sense in the LLM industry right now that if we can't get to "it" (AGI, etc) by the end of this decade, it won't happen during this "AI Summer". The reason for that is two-fold: Intelligence scaling is logarithmic w.r.t compute. We simply cannot scale compute quick enough. And, interest in funding to pay for that exponential compute need will dry up, and previous super-cycles tell us that will happen on the order of ~5 years.

So here's my thesis: We have a deadline that even evangelists agree is a deadline. I would argue that we're further along in this supercycle than many people realize, because these companies have already reached the early enshitification phase for some niche use-cases (software development). We're also seeing Grok 4 Heavy release with a 50% price increase ($300/mo) yet offer single-digit percent improvement in capability. This is hallmark enshitification.

Enshitification is the final, terminal phase of hyperscale technology companies. Companies remain in that phase potentially forever, but its not a phase where significant research, innovation, and optimization can happen; instead, it is a phase of extraction. AI hyperscalers genuinely speedran this cycle thanks to their incredible funding and costs; but they're now showcasing very early signals of enshitifications.

(Google might actually escape this enshitification supercycle, to be clear, and that's why I'm so bullish on them and them alone. Their deep, multi-decade investment into TPUs, Cloud Infra, and high margin product deployments of AI might help them escape it).

[1] https://www.reddit.com/r/cursor/comments/1m0i6o3/cursor_qual...

[2] https://www.reddit.com/r/ClaudeAI/comments/1lzuy0j/claude_co...

[3] https://techcrunch.com/2025/07/07/cursor-apologizes-for-uncl...

erlend_sh|7 months ago

Exactly. This is basically the argument of “AI as Normal Technology”.

https://knightcolumbia.org/content/ai-as-normal-technology

https://news.ycombinator.com/item?id=43697717

highfrequency|7 months ago

Thanks for the link. The comparison to electricity is a good one, and this is a nice reflection on why it took time for electricity’s usefulness to show up in productivity stats:

> What eventually allowed gains to be realized was redesigning the entire layout of factories around the logic of production lines. In addition to changes to factory architecture, diffusion also required changes to workplace organization and process control, which could only be developed through experimentation across industries.

SirHumphrey|7 months ago

This seems like one the only sane arguments in this whole sea of articles.

api|7 months ago

My take since day one:

(1) Model capabilities will plateau as training data is exhausted. Some additional gains will be possible by better training, better architectures, more compute, longer context windows or "infinite" context architectures, etc., but there are limits here.

(2) Training on synthetic data beyond a very limited amount will result in overfitting because there is no new information. To some extent you could train models on each other, but that's just an indirect way to consolidate models. Beyond consolidation you'll plateau.

(3) There will be no "takeoff" scenario -- this is sci-fi (in the pejorative sense) because you can't exceed available information. There is no magic way that a brain in a vat can innovate beyond available training data. This includes for humans -- a brain in a vat would quickly go mad and then spiral into a coma-like state. The idea of AI running away is the information-theoretic equivalent of a perpetual motion machine and is impossible. Yudkowski and the rest of the people afraid of this are crackpots, and so are the hype-mongers betting on it.

So I agree that LLMs are real and useful, but the hype and bubble are starting to plateau. The bubble is predicated on the idea that you can just keep going forever.

ogogmad|7 months ago

The next step is clearly improved vision and everyday-physics models. These can also solve hallucinations.

ludicrousdispla|7 months ago

>> There are many technologies that have seemed inevitable and seen retreats under the lack of commensurate business return

120+ Cable TV channels must have seemed like a good idea at the time, but like LLMs the vast majority of the content was not something people were interested in.

strangescript|7 months ago

I think the difference between all previous technologies is scope. If you make a super sonic jet that gets people from place A to place B faster for more money, but the target consumer is like "yeah, I don't care that much about that at that price point", then your tech sort is of dead. You are also fully innovated on that product, like maybe you can make it more fuel efficient, sure, but your scope is narrow.

AI is the opposite. There are numerous things it can do and numerous ways to improve it (currently). There is lower upfront investment than say a supersonic jet and many more ways it can pivot if something doesn't work out.

sumeno|7 months ago

The number of things it can actually do is significantly lower than the number of things the hype men are claiming it can do.

digianarchist|7 months ago

It's not a great analogy. The only parallel with Concorde is energy consumption. I think a better analogy would have been VR.

peder|7 months ago

Most of the comments here feel like cope about AI TBH. There's never been an innovation like this ever, and it makes sense to get on board rather than be left behind.

dcow|7 months ago

The difference is that the future is now with LLMs. There is a microwave (some multiple) in almost every kitchen in the world. The Concord served a few hundred people a day. LLMs are already ingrained into hundreds of millions if not billions of people’s lives, directly and indirectly. My dad directly uses LLMs multiple times a week if not daily in an industry that still makes you rotate your password every 3 months. It’s not a question of whether the future will have them, it’s a question of whether the future will get tired of them.

jayd16|7 months ago

The huge leap that is getting pushback is the sentiment that LLMs will consume every use case and replace human labor. I don't think many are arguing LLMs will die off entirely.

eric-burel|7 months ago

Developers haven't even started extracting the value of LLMs with agent architectures yet. Using an LLM UI like open ai is like we just figured fire and you use it to warm you hands (still impressive when you think about it, but not worth the burns), while LLM development is about building car engines (here is you return on investment).

Jensson|7 months ago

> Developers haven't even started extracting the value of LLMs with agent architectures yet

There are thousands of startups doing exactly that right now, why do you think this will work when all evidence points towards it not working? Or why else would it not already have revolutionized everything a year or two ago when everyone started doing this?

clarinificator|7 months ago

Every booster argument is like this one. $trite_analogy triumphant smile

__loam|7 months ago

3 years into automating all white collar labor in 6 months.

camillomiller|7 months ago

>> Developers haven't even started extracting the value of LLMs with agent architectures yet.

What does this EVEN mean? Do words have any value still, or are we all just starting to treat them as the byproduct of probabilistic tokens?

"Agent architectures". Last time I checked an architecture needs predictability and constraints. Even in software engineering, a field for which the word "engineering" is already quite a stretch in comparison to construction, electronics, mechanics.

Yet we just spew the non-speak "Agentic architectures" as if the innate inability of LLMs in managing predictable quantitative operations is not an unsolved issue. As if putting more and more of these things together automagically will solves their fundamental and existential issue (hallucinations) and suddenly makes them viable for unchecked and automated integration.

pydry|7 months ago

Theyre doing it so much it's practically a cliche.

There are underserved areas of the economy but agentic startups is not one.

dvfjsdhgfv|7 months ago

> Developers haven't even started extracting the value of LLMs with agent architectures yet.

For sure there is a portion of developers who don't care about the future, are not interested in current developements and just live as before hoping nothing will change. But the rest already gave it a try and realized tools like Claude Code can give excellent results for small codebases to fail miserably at more complex tasks with the net result being negative as you get a codebase you don't understand, with many subtle bugs and inconsistencies created over a few days you will need weeks to discover and fix.

mns|7 months ago

>evelopers haven't even started extracting the value of LLMs with agent architectures yet.

Which is basically what? The infinite monkey theorem? Brute forcing solutions for problems at huge costs? Somehow people have been tricked to actually embrace and accept that now they have to pay subscriptions from 20$ to 300$ to freaking code? How insane is that, something that was a very low entry point and something that anyone could do, is now being turned into some sort of classist system where the future of code is subscriptions you pay for companies ran by sociopaths who don't care that the world burns around them, as long as their pockets are full.

UncleOxidant|7 months ago

Let's not ignore the technical aspects as well: LLMs are probably a local minima that we've gotten stuck in because of their rapid rise. Other areas in AI are being starved of investment because all of the capital is pouring into LLMs. We might have been better off in the long run if LLMs hadn't been so successful so fast.

Msurrow|7 months ago

> first signs of pulling back investments

I agree with you, but I’m curious; do you have link to one or two concrete examples of companies pulling back investments, or rolling back an AI push?

(Yes it’s just to fuel my confirmation bias, but it’s still feels nice:-) )

magic_hamster|7 months ago

There are pretty hidden assumption in this comment. First of all, not every business in the AI space is _training_ models, and the difference between training and inference is massive - i.e. most businesses can easily afford inference, perhaps depending on model, but they definitely can.

Another several unfounded claims were made here, but I just wanted to say LLMs with MCP are definitely good enough for almost every use case you can come up with as long as you can provide them with high quality context. LLMs are absolutely the future and they will take over massive parts of our workflow in many industries. Try MCP for yourself and see. There's just no going back.

ramoz|7 months ago

LLMs with tools*

MCP isn’t inherently special. A Claude Code with Bash() tool can do nearly anything a MCP server will give you - much more efficiently.

Computer Use agents are here and are only going to get better.

The conversation shouldn’t be about LLMs any longer. Providers will be providing agents.

dontlikeyoueith|7 months ago

> I just wanted to say LLMs with MCP are definitely good enough for almost every use case you can come up with as long as you can provide them with high quality context.

This just shows you lack imagination.

I have a lot of use cases that they are not good enough for.

nyarlathotep_|7 months ago

I do wonder where in the cycle this all is given that we've now seen yet another LLM/"Agentic" VSCode fork.

I'm genuinely surprised that Code forks and LLM cli things are seemingly the only use case that's approached viability. Even a year ago, I figured there'd be something else that's emerged by now.

alonsonic|7 months ago

But there are a ton of LLM powered products in the market.

I have a friend in finance that uses LLM powered products for financial analysis, he works in a big bank. Just now anthropic released a product to compete in this space.

Another friend in real estate uses LLM powered lead qualifications products, he runs marketing campaigns and the AI handles the initial interaction via email or phone and then ranks the lead in their crm.

I have a few friends that run small businesses and use LLM powered assistants to manage all their email comms and agendas.

I've also talked with startups in legal and marketing doing very well.

Coding is the theme that's talked about the most in HN but there are a ton of startups and big companies creating value with LLMs

philomath_mn|7 months ago

> most people agree that the output is trite and unpleasant to consume

This is likely a selection bias: you only notice the obviously bad outputs. I have created plenty of outputs myself that are good/passable -- you are likely surrounded by these types of outputs without noticing.

Not a panacea, but can be useful.

MonkeyIsNull|7 months ago

> 2. Almost three years in, companies investing in LLMs have not yet discovered a business model that justifies the massive expenditure of training and hosting them,

I always think back to how Bezos and Amazon were railed against for losing money for years. People thought that would never work. And then when he started selling stuff other than books? People I know were like: please, he's desperate.

Someone, somewhere will figure out how to make money off it - just not most people.

alexpotato|7 months ago

To use the Internet as a comparison:

Phase 1 - mid to late 1990s:

- "The Internet is going to change EVERYTHING!!!"

Phase 2 - late 1990s to early 2000s:

- "It's amazing and we are all making SO much money!"

- "Oh no! The bubble burst"

- "Of course everyone could see this coming: who is going to buy 40 lb bags of dogfood or their groceries over the Internet?!?!?"

Phase 3 - mid 2000s to 2020:

- "It is astounding the amount of money being by tech companies"

- "Who could have predicted that social media would change the ENTIRE landscape??"

gonzobonzo|7 months ago

My guess is that LLM's are bridge technology, the equivalent of cassette tapes. A big step forward, allowing things that we couldn't before. But before long they'll be surpassed by much better technology, and future generations will look back on them as primitive.

You have top scientists like LeCun arguing this position. I'd imagine all of these companies are desperately searching for the next big paradigm shift, but no one knows when that will be, and until then they need to squeeze everything they can out of LLMs.

moffkalast|7 months ago

ML models have the good property of only requiring investment once and can then be used till the end of history or until something better replaces them.

Granted the initial investment is immense, and the results are not guaranteed which makes it risky, but it's like building a dam or a bridge. Being in the age where bridge technology evolves massively on a weekly basis is a recipe for being wasteful if you keep starting a new megaproject every other month though. The R&D phase for just about anything always results in a lot of waste. The Apollo programme wasn't profitable either, but without it we wouldn't have the knowledge for modern launch vehicles to be either. Or to even exist.

I'm pretty sure one day we'll have an LLM/LMM/VLA/etc. that's so good that pretraining a new one will seem pointless, and that'll finally be the time we get to (as a society) reap the benefits of our collective investment in the tech. The profitability of a single technology demonstrator model (which is what all current models are) is immaterial from that standpoint.

wincy|7 months ago

Nah, if TSMC got exploded and there was a world war, in 20 years all the LLMs would bit rot.

dmix|7 months ago

> model capabilities are plateauing at a level where most people agree that the output is trite and unpleasant to consume.

What are you basing this on? Personal feelings?

fendy3002|7 months ago

LLMs need significant optimization or we get significant improvement on computing power while keeping the energy cost the same. It's similar with smartphone, when at the start it's not feasible because of computing power, and now we have one that can rival 2000s notebooks.

LLMs is too trivial to be expensive

EDIT: I presented the statement wrongly. What I mean is the use case for LLM are trivial things, it shouldn't be expensive to operate

killerstorm|7 months ago

LLM can give you thousands of lines of perfectly working code for less than 1 dollar. How is that trivial or expensive?

jsnell|7 months ago

But the thing is, LLMs are already incredibly cheap to operate compared to the alternatives. Both for trivial things and for complex things.

lblume|7 months ago

Imagine telling a person from five years ago that the programs that would basically solve NLP, perform better than experts at many tasks and are hard not to anthropomorphize accidentally are actually "trivial". Good luck with that.

trashchomper|7 months ago

Calling LLMs trivial is a new one. Yea just consume all of the information on the internet and encode it into a statistical model, trivial, child could do it /s

jittery41|7 months ago

Oh wow I forgot that the microwave oven was once marketed as the total replacement of cooking chores and in futuristic life people can just press a button and have a delicious good meal ( well you can now but microwave meals are often seen as worse than fastfood ).

JimmaDaRustla|7 months ago

Investments are mostly in model training. We have trained models now, we'll see a pullback in that regard as businesses will need to optimize to get the best model without spending billions in order to compete on price, but LLMs are here to stay.

giancarlostoro|7 months ago

> 2. Almost three years in, companies investing in LLMs have not yet discovered a business model that justifies the massive expenditure of training and hosting them, the majority of consumer usage is at the free tier, the industry is seeing the first signs of pulling back investments, and model capabilities are plateauing at a level where most people agree that the output is trite and unpleasant to consume.

You hit the nail on why I say to much hatred from "AI Bros" as I call them, when I say it will not take off truly until it runs on your phone effortlessly, because nobody wants to foot a trillion dollar cloud bill.

Give me a fully offline LLM that fits in 2GB of VRAM and lets refine that so it can plug into external APIs and see how much farther we can take things without resorting to burning billions of dollars' worth of GPU compute. I don't care that my answer arrives instantly, if I'm doing the research myself, I want to take my time to get the correct answer anyway.

saratogacx|7 months ago

We actually aren't too far off from that reality. There are several models you can run fully offline on your phone (phi-3, Gemma-3n-E2b-it, Qwen2.5-1.5b-instruct all run quite well on my Samsung S24 ultra). There are a few offline apps that also have tool calling (mostly for web search but I suspect this is extendable).

If you want to play around a bit and are on android there is PocketPal,ChatterUI, MyDeviceAI, SmolChat are good multi-model apps and Google's Edge gallery won't keep your chats but is a fun tech demo.

All are on github and can be installed using Obtainium if you don't want to

DSingularity|7 months ago

You aren’t extrapolating enough. Nearly the entire history of computing has been one that isolates between shared computing and personal computing. Give it time. These massive cloud bills are building the case for accelerators in phones. It’s going to happen just needs time.

xnx|7 months ago

> (the supersonic jetliner) ... (the microwave oven)

But have we ever had a general purpose technology (steam engine, electricity) that failed to change society?

blueflow|7 months ago

It wouldn't be general purpose if it fails to bring change. I'd take every previous iteration of "AI" as example, IBM Watson, that stuff

smrtinsert|7 months ago

They didn't really need the cloud either and yet...

Jach|7 months ago

I don't really buy your point 2. Just the other day Meta announced hundreds of billions of dollars investment into more AI datacenters. Companies are bringing back nuclear power plants to support this stuff. Earlier this year OpenAI and Oracle announced their $500bn AI datacenter project, but admittedly in favor of your point have run into funding snags, though that's supposedly from tariff fears with foreign investors, not lack of confidence in AI. Meta can just finance everything from their own capital and Zuck's decree, like they did with VR (and it may very well turn out similarly).

Since you brought up supersonic jetliners you're probably aware of the startup Boom in Colorado trying to bring it back. We'll see if they succeed. But yes, it would be a strange path, but a possible one, that LLMs kind of go away for a while and try to come back later.

You're going to have to cite some surveys for the "most people agree that the output is trite and unpleasant" and "almost universally disliked attempts to cram it everywhere" claims. There are some very vocal people against LLM flavors of AI, but I don't think they even represent the biggest minority, let alone a majority or near universal opinions. (I personally was bugged by earlier attempts at cramming non-LLM AI into a lot of places, e.g. Salesforce Einstein appeared I think in 2016, and that was mostly just being put off by the cutesy Einstein characterization. I generally don't have the same feelings with LLMs in particular, in some cases they're small improvements to an already annoying process, e.g. non-human customer support that was previously done by a crude chatbot front-end to an expert system or knowledge base, the LLM version of that tends to be slightly less annoying.)

Jach|7 months ago

Sort of a followup to myself if I come back searching this comment or someone sees this thread later... here's a study that just came out on AI attitudes: https://report2025.seismic.org/

I don't think it supports the bits I quoted, but it does include more negativity than I would have predicted before seeing it.