top | item 46131448

(no title)

blablabla123 | 3 months ago

Despite the flashy title that's the first "sober" analysis from a CEO I read about the technology. While not even really news, it's also worth mentioning that the energy requirements are impossible to fulfill

Also now using ChatGPT intensely since months for all kinds of tasks and having tried Claude etc. None of this is on par with a human. The code snippets are straight out of Stackoverflow...

discuss

order

mattlondon|2 months ago

Take this "sober" analysis with a big pinch of salt.

IBM have totally missed the AI boat, and a large chunk of their revenue comes from selling expensive consultants to clients who do not have the expertise to do IT work themselves - this business model is at a high risk of being disrupted by those clients just using AI agents instead of paying $2-5000/day for a team of 20 barely-qualified new-grads in some far-off country.

IBM have an incentive to try and pour water on the AI fire to try and sustain their business.

evanjrowley|2 months ago

Is this true in 2025?

Asking because the biggest IT consulting branch of IBM, Global Technology Services (GTS), was spun off into Kyndryl back in 2021[0]. Same goes for some premier software products (including one I consulted for) back in 2019[1]. Anecdotal evidence suggests the consulting part of IBM was already significantly smaller than in the past.

It's worth noting that IBM may view these AI companies as competitors to it's Watson AI tech[2]. It already existed before the GPU crunch and hyperscaler boom - runs on proprietary IBM hardware.

[0] https://en.wikipedia.org/wiki/Kyndryl

[1] https://www.prnewswire.com/news-releases/hcl-technologies-to...

[2] https://en.wikipedia.org/wiki/IBM_Watson

tw04|2 months ago

Missed the boat? Have you been living under a rock? Watson AI advertising has been everywhere for years.

It’s not that they aren't in the AI space, it’s that the CEO has a shockingly sober take on it. Probably because they’ve been doing AI for 30+ years combined with the fact they don’t have endless money with nowhere to invest it like Google.

belter|2 months ago

> IBM have an incentive to try and pour water on the AI fire to try and sustain their business.

IBM has faced multiple lawsuits over the years. From age discrimination cases to various tactics allegedly used to push employees out, such as requiring them to relocate to states with more employer friendly laws only to terminate them afterward.

IBM is one of the clearest examples of a company that, if given the opportunity to replace human workers with AI, would not hesitate to do so. Assume therefore, the AI does not work for such a purpose...

deepGem|2 months ago

I would any day take chatGPT/Claude over an IBM consultant. I worked at IBM.

OhMeadhbh|2 months ago

Are you suggesting IBM made up the numbers? Or that CAPEX is a pre-GAI measure and is useless in guiding decision making?

IBM may have a vested interest in calming (or even extinguishing) the AI fire, but they're not the first to point out the numbers look a little wobbly.

And why should I believe OpenAI or Alphabet/Gemini when they say AI will be the royal road to future value? Don't they have a vested interest in making AI investments look attractive?

Ragnarork|2 months ago

> a high risk of being disrupted by those clients just using AI agents instead of paying $2-5000/day for a team of 20 barely-qualified new-grads in some far-off country

Is there any concrete evidence of that risk being high? That doesn't come from people whose job is to sell AI?

ratelimitsteve|2 months ago

they have incentive but what's the sustainable, actually-pays-for-itself-and-generates-profit cost of AI? We have no idea. Everything is so heavily subsidized by burning investor capital for heat with the hope that they'll pull an amazon and make it impossible to do business on the internet without paying an AI firm. Maybe the 20 juniors will turn out to be cheaper. Maybe they'll turn out to be slightly better. Maybe they'll be loosely equivalent and the ability to automate mediocrity will drive down the cost of human mediocrity. We don't know and everyone seems to be betting heavily on the most optimistic case, so it makes an awful lot of sense to take the other side of that bet.

HacklesRaised|2 months ago

Isn't he one of the first ass clowns to start laying people off, replaced by AI.

Another empty suit.

TheCondor|2 months ago

How do you see the math working out?

The numbers are staggering.

vultour|2 months ago

Do you expect Sam Altman to come on stage and tell you the whole thing is a giant house of cards when the entire western economy seems to be propped up by AI? I wonder whose "sober" analysis you would accept, because surely the people that are making money hand over fist will never admit it.

Seems to me like any criticism of AI is always handwaved away with the same arguments. Either it's companies who missed the AI wave, or the models are improving incredibly quickly so if it's shit today you just have to wait one more year, or if you're not seeing 100x improvements in productivity you must be using it wrong.

diggyhole|2 months ago

Right. Just like Intel completely shitting the bed on GPUs then their CEO tweeted a prayer. Old tech companies are going to be left behind.

MDGeist|2 months ago

IBM was ahead of the boat! They had Watson on Jeopardy years ago! /s

I think you make a fair point about the potential disruption for their consulting business but didn't they try to de-risk a bit with the Kyndryl spinout?

infecto|2 months ago

I am a senior engineer, I use cursor a lot in my day to day. I find I can code longer and typically faster than without. Is it on par with human? It’s getting pretty darn close to be honest, I am sure the “10x” engineers of the world would disagree but it definitely has surpassed a junior engineer. We all have our anecdotes but I am inclined to believe on average there is net value.

boringg|2 months ago

I think surpassed is not the right word because it doesn't create/ideate. However it is incredibly resourceful. Maybe like having a jr engineer to do your bidding without thinking or growing.

lordnacho|2 months ago

Largely agree. Anything that is just a multi-file edit, like an interface change, it can do. Maybe not immediately, but you can have it iterate, and it doesn't eat up your attention.

It is without a doubt worth more than the 200 bucks a month I spend on it.

I will go as far as to say it has decent ideas. Vanilla ideas, but it has them. I've actually gotten it to come up with algorithms that I thought were industry secrets. Minor secrets, sure. But things that you don't just come across. I'm in the trading business, so you don't really expect a lot of public information to be in the dataset.

ratelimitsteve|2 months ago

i'm also a senior engineer and I use codex a lot. It has reduced many of the typical coding tasks to simply writing really good AC. I still have to write good AC, but I'm starting to see the velocity change from using good AI in a smart way.

enraged_camel|2 months ago

Senior engineer here as well. I would say Opus 4.5 is easily a mid-level engineer. It's a substantial improvement over Sonnet 4.5, which required a lot more hand-holding and interventions.

trgn|2 months ago

i think less. not sure if that's a good thing. but small little bugs and improvements get cleared so quickly now.

fvv|2 months ago

it surpassed 30 junior engineers

delaminator|3 months ago

Your assessment of Claude simply isn’t true.

Or Stackoverflow is really good.

I’m producing multiple projects per week that are weeks of work each.

bloppe|3 months ago

Would you mind sharing some of these projects?

I've found Claude's usefulness is highly variable, though somewhat predictable. It can write `jq` filters flawlessly every time, whereas I would normally spend 30 minutes scanning docs because nobody memorizes `jq` syntax. And it can comb through server logs in every pod of my k8s clusters extremely fast. But it often struggles making quality code changes in a large codebase, or writing good documentation that isn't just an English translation of the code it's documenting.

written-beyond|3 months ago

I'm just as much of an avid llm code generator fan as you may be but I do wonder about the practicality of spending time making projects anymore.

Why build them if other can just generate them too, where is the value of making so many projects?

If the value is in who can sell it the best to people who can't generate it, isn't it just a matter of time before someone else will generate one and they may become better than you at selling it?

blablabla123|3 months ago

Sure but these are likely just variations of existing things. And yet the quality is still behind the original

eschaton|3 months ago

I produce a lot of shit every week too, but I don’t brag about my digestive system on “Hacker” “News.”

baobabKoodaa|2 months ago

I'll do one better: I poop every day in the water closet!

tim333|2 months ago

An issue with the doom forecasts is most of the hypothetical $8tn hasn't happened yet. Current big tech capex is about $315bn this year, $250bn last against a pre AI level ~$100bn so ~$400bn has been spent so far on AI boom data centers. https://sherwood.news/business/amazon-plans-100-billion-spen...

The future spend is optional - AGI takeoff, you spend loads, not happening not so much.

Say it levels of at $800bn. The world's population is ~8bn so $100 a head so you'd need to be making $10 or $20 per head per year. Quite possibly doable.

trueismywork|2 months ago

65% of people in the world earn less than 3000 euros/year.

tempfile|2 months ago

Lol. If you ballpark numbers like that probably anything is doable!

mark_l_watson|2 months ago

I agree. re: energy and other resource use: the analogy I like is with driving cars: we use cars for transportation knowing the environmental costs so we don’t usually just go on two hour drives for the fun of it, rather we drive to get to work, go shopping. I use Gemini 3 but only in specific high value use cases. When I use commercial models I think a little about the societal costs.

In the USA we have lost the thread here: we don’t maximize the use of small tuned models throughout society and industry, instead we use the pursuit of advanced AI as a distraction to the reality that our economy and competitiveness are failing.

spider-mario|2 months ago

Most of the energy for AI does not go into chatbots. Using Gemini is not remotely close to driving a car for 2 hours. If a prompt is 0.3 Wh (https://cloud.google.com/blog/products/infrastructure/measur..., https://andymasley.substack.com/p/a-cheat-sheet-for-conversa...), each prompt is closer to using an e-bike for 50 metres.

You could have your morning shower 1°C less hot and save enough energy for about 200 prompts (assuming 50 litres per shower). (Or skip the shower altogether and save thousands of prompts.)

MisterTea|2 months ago

Yesterday I was talking to coworkers about AI I mentioned that a friend of mine used ChatGPT to help him move. So a coworker said I have to test this and asked ChatGPT if he could fit a set of the largest Magnepan speakers (the wide folding older room divider style) in his Infinity QX80. The results were hilarious. It had some of the dimensions right but it then decided the QX80 is as wide as a box truck (~8-8.5 feet/2.5 m) and to align the nearly 7 foot long speakers sideways between the wheel wells. It also posted hilariously incomprehensible ASCII diagrams.

TheOccasionalWr|2 months ago

I'm not sure what you mean with the "code snippets are straight out of Stackoverflow". That is factually incorrect just by how LLM works. By now there has been so much code ingested from all kinds of sources, including Stackoverflow LLM is able to help generate quite good code in many occasions. My point being it is extremly useful for super popular languages and many languages where resources are more scarce for developer but because they got the code from who knows where, it can definitely give you many useful ideas.

It's not human, which I'm not sure what is supposed to actually mean. Humans make mistakes, humans make good code. AI does also both. What it definitely needs is a good programmer still on top to know what he is getting and how to improve it.

I find AI (LLM) very useful as a very good code completion and light coder where you know exactly what to do because you did it a thousand times but it's wasteful to be typing it again. Especially a lot of boilerplate code or tests.

It's also useful for agentic use cases because some things you just couldn't do before because there was nothing to understand a human voice/text input and translate that to an actual command.

But that is all far from some AGI and it all costs a lot today an average company to say that this actually provided return on the money but it definitely speeds things up.

prewett|2 months ago

> I'm not sure what you mean with the "code snippets are straight out of Stackoverflow". That is factually incorrect just by how LLM works.

I'm not an AI lover, but I did try Gemini for a small, well-contained algorithm for a personal project that I didn't want to spend the time looking up, and it was straight-up a StackOverflow solution. I found out because I said "hm, there has to be a more elegant solution", and quickly found the StackOverflow solution that the AI regurgitated. Another 10 or 20 minutes of hunting uncovered another StackOverflow solution with the requisite elegance.

will4274|3 months ago

> While not even really news, it's also worth mentioning that the energy requirements are impossible to fulfill

If you believe this, you must also believe that global warming is unstoppable. OpenAI's energy costs are large compared to the current electricity market, but not so large compared to the current energy market. Environmentalists usually suggest that electrification - converting non-electrical energy to electrical energy - and then making that electrical energy clean - is the solution to global warming. OpenAI's energy needs are something like 10% of the current worldwide electricity market but less than 1% of the current worldwide energy market.

blablabla123|3 months ago

Google recently announced to double AI data center capacity every 6 month. While both unfortunately deal with exponential growth, we are talking about 1% growth CO2 which is bad enough vs 300% effectively per year according to Google

rvnx|3 months ago

Imagine how big pile of trash as the current generation of graphics cards used for LLM training will get outdated. It will crash the hardware market (which is a good news for gamers)

tikotus|3 months ago

I'd rather phrase it as "code is straight out of GitHub, but tailored to match your data structures"

That's at least how I use it. If I know there's a library that can solve the issue, I know an LLM can implement the same thing for me. Often much faster than integrating the library. And hey, now it's my code. Ethical? Probably not. Useful? Sometimes.

If I know there isn't a library available, and I'm not doing the most trivial UI or data processing, well, then it can be very tough to get anything usable out of an LLM.

guywithahat|2 months ago

> it's also worth mentioning that the energy requirements are impossible to fulfill

Maybe I'm misunderstanding you but they're definitely not impossible to fulfill, in fact I'd argue the energy requirements are some of the most straightforward to fulfill. Bringing a natural gas power plant online is not the hardest part in creating AGI

diggyhole|2 months ago

I've had decent results hackin', wackin' and smashin'.

lavezzi|2 months ago

> Despite the flashy title that's the first "sober" analysis from a CEO I read about the technology.

Didn't IBM just sign quite a big deal with Groq?

trgn|2 months ago

> Also now using ChatGPT intensely since months for all kinds of tasks and having tried Claude etc.

the facts though, read like an endorsement not a criticism