top | item 47010481

(no title)

Springtime | 17 days ago

Ars Technica being caught using LLMs that hallucinated quotes by the author and then publishing them in their coverage about this is quite ironic here.

Even on a forum where I saw the original article by this author posted someone used an LLM to summarize the piece without having read it fully themselves.

How many levels of outsourcing thinking is occurring to where it becomes a game of telephone.

discuss

order

sho_hn|16 days ago

Also ironic: When the same professionals advocating "don't look at the code anymore" and "it's just the next level of abstraction" respond with outrage to a journalist giving them an unchecked article.

Read through the comments here and mentally replace "journalist" with "developer" and wonder about the standards and expectations in play.

Food for thought on whether the users who rely on our software might feel similarly.

There's many places to take this line of thinking to, e.g. one argument would be "well, we pay journalists precisely because we expect them to check" or "in engineering we have test-suites and can test deterministically", but I'm not sure if any of them hold up. The "the market pays for the checking" might also be true for developers reviewing AI code at some point, and those test-suites increasingly get vibed and only checked empirically, too.

Super interesting to compare.

armchairhacker|16 days ago

- There’s a difference. Users don’t see code, only its output. Writing is “the output”.

- A rough equivalent here would be Windows shipping an update that bricks your PC or one of its basic features, which draws plenty of outrage. In both cases, the vendor shipped a critical flaw to production: factual correctness is crucial in journalism, and a quote is one of the worst things to get factually incorrect because it’s so unambiguous (inexcusable) and misrepresents who’s quoted (personal).

I’m 100% ok with journalists using AI as long as their articles are good, which at minimum requires factual correctness and not vacuous. Likewise, I’m 100% ok with developers using AI as long as their programs are good, which at minimum requires decent UX and no major bugs.

adamddev1|16 days ago

Excellent observation. I get so frustrated every time I hear the "we have test-suites and can test deterministically" argument. Have we learned absolutely nothing from the last 40 years of computer science? Testing does not prove the absence of bugs.

boothby|16 days ago

I look forward to a day when the internet is so uniformly fraudulent that we can set it aside and return to the physical plane.

anonymous908213|16 days ago

> When the same professionals advocating "don't look at the code anymore" and "it's just the next level of abstraction" respond with outrage to a journalist giving them an unchecked article.

I would expect there is literally zero overlap between the "professionals"[1] who say "don't look at the code" and the ones criticising the "journalists"[2]. The former group tend to be maximalists and would likely cheer on the usage of LLMs to replace the work of the latter group, consequences be damned.

[1] The people that say this are not professional software developers, by the way. I still have not seen a single case of any vibe coder who makes useful software suitable for deployment at scale. If they make money, it is by grifting and acting as an "AI influencer", for instance Yegge shilling his memecoin for hundreds of thousands of dollars before it was rugpulled.

[2] Somebody who prompts an LLM to produce an article and does not even so much as fact-check the quotations it produces can clearly not be described as a journalist, either.

ffsm8|16 days ago

While I don't subscribe to the idea that you shouldn't look at the code - it's a lot more plausible for devs because you do actually have ways to validate the code without looking at it.

E.g you technically don't need to look at the code if it's frontend code and part of the product is a e2e test which produces a video of the correct/full behavior via playwright or similar.

Same with backend implementations which have instrumentation which expose enough tracing information to determine if the expected modules were encountered etc

I wouldn't want to work with coworkers which actually think that's a good idea though

rsynnott|14 days ago

> When the same professionals advocating "don't look at the code anymore" and "it's just the next level of abstraction" respond with outrage to a journalist giving them an unchecked article.

I doubt, by and large, that it's the same people. Just as this LLM misquoting is journalistic malpractice, "don't look at the code anymore" is engineering malpractice.

ChrisMarshallNY|16 days ago

I’ve been saying the same kind of thing (and I have been far from alone), for years, about dependaholism.

Nothing new here, in software. What is new, is that AI is allowing dependency hell to be experienced by many other vocations.

Dylan16807|15 days ago

I haven't seen a single person advocate not looking at the code.

I'm sure that person exists but they're not representative of HN as a whole.

mattgreenrocks|16 days ago

So much projection these days in so many areas of life.

tliltocatl|15 days ago

> the same professionals

Same forum, not necessary same people.

usefulposter|16 days ago

Incredible. When Ars pull an article and its comments, they wipe the public XenForo forum thread too, but Scott's post there was archived. Username scottshambaugh:

https://web.archive.org/web/20260213211721/https://arstechni...

>Scott Shambaugh here. None of the quotes you attribute to me in the second half of the article are accurate, and do not exist at the source you link. It appears that they themselves are AI hallucinations. The irony here is fantastic.

Instead of cross-checking the fake quotes against the source material, some proud Ars Subscriptors proceed to defend Condé Nast by accusing Scott of being a bot and/or fake account.

EDIT: Page 2 of the forum thread is archived too. This poster spoke too soon:

>Obviously this is massive breach of trust if true and I will likely end my pro sub if this isnt handled well but to the credit of ARS, having this comment section at all is what allows something like this to surface. So kudos on keeping this chat around.

bombcar|16 days ago

This is just one of the reasons archiving is so important in the digital era; it's key to keeping people honest.

asddubs|16 days ago

I read the forum thread, and most people seem to be critical of ars. One person said scott is a bot, but this read to me as a joke about the situation

vor_|16 days ago

The comment calling him a bot is sarcasm.

sphars|16 days ago

Aurich Lawson (creative director at Ars) posted a comment[0] in response to a thread about what happened, the article has been pulled and they'll follow-up next week.

[0]: https://arstechnica.com/civis/threads/journalistic-standards...

_HMCB_|16 days ago

It’s funny they say the article “may have” run afoul of their journalistic standards. May have is carrying a lot of weight there.

epistasis|16 days ago

Yikes I subscribed to them last year on the strength of their reporting in a time where it's hard to find good information.

Printing hallucinated quotes is a huge shock to their credibility, AI or not. Their credibility was already building up after one of their long time contributors, a complete troll of a person that was a poison on their forums, went to prison for either pedophilia or soliciting sex from a minor.

Some serious poor character judgement is going on over there. With all their fantastic reporters I hope the editors explain this carefully.

singpolyma3|16 days ago

TBF even journalists who interview people for real and take notes routinely quite them saying things they didn't say. The LLMs make it worse, but it's hardly surprising behaviour from them

justinclift|16 days ago

> Their credibility was already building up ...

Don't you mean diminishing or disappearing instead of building up?

Building up sounds like the exact opposite of what I think you're meaning. ;)

trollbridge|17 days ago

The amount of effort to click an LLM’s sources is, what, 20 seconds? Was a human in the loop for sourcing that article at all?

phire|17 days ago

Humans aren't very diligent in the long term. If an LLM does something correctly enough times in a row (or close enough), humans are likely to stop checking its work throughly enough.

This isn't exactly a new problem we do it with any bit of new software/hardware, not just LLMs. We check its work when it's new, and then tend to trust it over time as it proves itself.

But it seems to be hitting us worse with LLMs, as they are less consistent than previous software. And LLM hallucinations are partially dangerous, because they are often plausible enough to pass the sniff test. We just aren't used to handling something this unpredictable.

prussia|16 days ago

The kind of people to use LLM to write news article for them tend not to be the people who care about mundane things like reading sources or ensuring what they write has any resemblance to the truth.

kortilla|16 days ago

The source would just be the article, which the Ars author used an LLM to avoid reading in the first place.

adamddev1|16 days ago

The problem is that the LLM's sources can be LLM generated. I was looking up some health question and tried clicking to see the source for one of the LLMs claim. The source was a blog post that contained an obvious hallucination or false elaboration.

kmeisthax|15 days ago

If a human had enough time to check all the sources they wouldn't have been using an LLM to write for them.

seanhunter|16 days ago

It’s fascinating that on the one hand Ars Technica didn’t think the article was worth writing (so got an LLM to do it) but expect us to think it’s worth reading. Then some people don’t think it’s worth reading (so get an LLM to do it) but think somehow we will think it’s not worth reading the article but is worth reading the llm summary. Feel like you can carry on that process ad infinitum always going for a smaller and smaller audience who are somehow willing to spend less and less effort (but not zero).

0xbadcafebee|16 days ago

> How many levels of outsourcing thinking is occurring to where it becomes a game of telephone

How do you know quantum physics is real? Or radio waves? Or just health advice? We don't. We outsource our thinking around it to someone we trust, because thinking about everything to its root source would leave us paralyzed.

Most people seem to have never thought about the nature of truth and reality, and AI is giving them a wake-up call. Not to worry though. In 10 years everyone will take all this for granted, the way they take all the rest of the insanity of reality for granted.

Lerc|16 days ago

Has it been shown or admitted that the quotes were hallucinations, or is it the presumption that all made up content is a hallucination now?

vor_|16 days ago

Another red flag is that the article used repetitive phrases in an AI-like way:

"...it illustrates exactly the kind of unsupervised output that makes open source maintainers wary."

followed later on by

"[It] illustrates exactly the kind of unsupervised behavior that makes open source maintainers wary of AI contributions in the first place."

joquarky|16 days ago

Gen AI only produces hallucinations (confabulations).

The utility is that the infrenced output tends to be right much more often than wrong for mainstream knowledge.

Pay08|16 days ago

You could read the original blog post...

DonHopkins|16 days ago

[deleted]

moomin|16 days ago

Ironically, if you actually know what you’re doing with an LLM, getting a separate process to check the quotations are accurate isn’t even that hard. Not 100% foolproof, because LLM, but way better than the current process of asking ChatGPT to write something for you and then never reading it before publication.

Springtime|16 days ago

The wrinkle in this case is the author blocked AI bots from their site (doesn't seem to be a mere robots.txt exclusion from what I can tell), so if any such bot were trying to do this it may have not been able to read the page to verify, so instead made up the quotes.

This is what the author actually speculated may have occurred with Ars. Clearly something was lacking in the editorial process though that such things weren't human verified either way.

giobox|16 days ago

More than ironic, it's truly outrageous, especially given the site's recent propensity for negativity towards AI. They've been caught red-handed here doing the very things they routinely criticize others for.

The right thing to do would be a mea-culpa style post and explain what went wrong, but I suspect the article will simply remain taken down and Ars will pretend this never happened.

I loved Ars in the early years, but I'd argue since the Conde Nast acquisition in 2008 the site has been a shadow of its former self for a long time, trading on a formerly trusted brand name that recent iterations simply don't live up to anymore.

khannn|16 days ago

Is there anything like a replacement? The three biggest tech sites that I traditionally love are ArsTechnica, AnandTech(rip), and Phoronix. One is dead man walking mode, the second is ded dead, and the last is still going strong.

I'm basically getting tech news from social media sites now and I don't like that.

jandrewrogers|16 days ago

Conde Nast are the same people wearing Wired magazine like a skin suit, publishing cringe content that would have brought mortal shame upon the old Wired.

antod|16 days ago

While their audience (and the odd staff member) is overwhelming anti AI in the comments, the site itself overall editorially doesn't seem to be.

emmelaich|16 days ago

Outrageous, but more precisely malpractice and unethical to not double check the result.

netsharc|16 days ago

Probably "one bad apple", soon to be fired, tarred and feathered...

llbbdd|16 days ago

Honestly frustrating that Scott chose not to name and shame the authors. Liability is the only thing that's going to stop this kind of ugly shit.

rectang|16 days ago

There is no need to rush to judgment on the internet instant-gratification timescale. If consequences are coming for journalist or publication, they are inevitable.

We’ll know more in only a couple days — how about we wait that long before administering punishment?

arduanika|16 days ago

I mean, I'm even more frustrated by this in Scott's original post:

> If you are the person who deployed this agent, please reach out. It’s important for us to understand this failure mode, and to that end we need to know what model this was running on and what was in the soul document. I’m not upset and you can contact me anonymously if you’d like.

I can see where he's coming from, and I suppose he's being the bigger man in the situation, but at some point one of these reckless moltbrain kiddies is going to have to pay. Libel and extortion should carry penalties no matter whether you do it directly, or via code that you wrote, or via code that you deployed without reading it.

The AI's hit piece on Scott was pretty minor, so if we want to wait around for a more serious injury that's fine, just as long as we're standing ready to prosecute when (not 'if') it happens.

asddubs|16 days ago

I mean, he linked the archived article. You're one click away from the information if you really want to know.

JPKab|16 days ago

I just wish people would remember how awful and unprofessional and lazy most "journalists" are in 2026.

It's a slop job now.

Ars Technica, a supposedly reputable institution, has no editorial review. No checks. Just a lazy slop cannon journalist prompting an LLM to research and write articles for her.

Ask yourself if you think it's much different at other publications.

joquarky|16 days ago

I would assume that most who were journalists 10 years ago have now either gone independent or changed careers

The ones that remain are probably at some extreme on one or more attributes (e.g. overworked, underpaid) and are leaning on genAI out of desperation.

troyvit|16 days ago

I work with the journalists at a local (state-wide) public media organization. It's night and day different from what is described at ars. These are people who are paid a third (or less) of what a sales engineer at meta makes. We have editorial review and ban LLMs for any editorial work except maybe alt-text if I can convince them to use it. They're over-worked, underpaid, and doing what very few people here (including me) have the dedication to do. But hey, if people didn't hate journalists they wouldn't be doing their job.

neya|16 days ago

Ars Technica has always trash even before LLMs and is mostly an advertisement hub for the highest bidder