(no title)
Springtime | 17 days ago
Even on a forum where I saw the original article by this author posted someone used an LLM to summarize the piece without having read it fully themselves.
How many levels of outsourcing thinking is occurring to where it becomes a game of telephone.
sho_hn|16 days ago
Read through the comments here and mentally replace "journalist" with "developer" and wonder about the standards and expectations in play.
Food for thought on whether the users who rely on our software might feel similarly.
There's many places to take this line of thinking to, e.g. one argument would be "well, we pay journalists precisely because we expect them to check" or "in engineering we have test-suites and can test deterministically", but I'm not sure if any of them hold up. The "the market pays for the checking" might also be true for developers reviewing AI code at some point, and those test-suites increasingly get vibed and only checked empirically, too.
Super interesting to compare.
armchairhacker|16 days ago
- A rough equivalent here would be Windows shipping an update that bricks your PC or one of its basic features, which draws plenty of outrage. In both cases, the vendor shipped a critical flaw to production: factual correctness is crucial in journalism, and a quote is one of the worst things to get factually incorrect because it’s so unambiguous (inexcusable) and misrepresents who’s quoted (personal).
I’m 100% ok with journalists using AI as long as their articles are good, which at minimum requires factual correctness and not vacuous. Likewise, I’m 100% ok with developers using AI as long as their programs are good, which at minimum requires decent UX and no major bugs.
adamddev1|16 days ago
boothby|16 days ago
anonymous908213|16 days ago
I would expect there is literally zero overlap between the "professionals"[1] who say "don't look at the code" and the ones criticising the "journalists"[2]. The former group tend to be maximalists and would likely cheer on the usage of LLMs to replace the work of the latter group, consequences be damned.
[1] The people that say this are not professional software developers, by the way. I still have not seen a single case of any vibe coder who makes useful software suitable for deployment at scale. If they make money, it is by grifting and acting as an "AI influencer", for instance Yegge shilling his memecoin for hundreds of thousands of dollars before it was rugpulled.
[2] Somebody who prompts an LLM to produce an article and does not even so much as fact-check the quotations it produces can clearly not be described as a journalist, either.
ffsm8|16 days ago
E.g you technically don't need to look at the code if it's frontend code and part of the product is a e2e test which produces a video of the correct/full behavior via playwright or similar.
Same with backend implementations which have instrumentation which expose enough tracing information to determine if the expected modules were encountered etc
I wouldn't want to work with coworkers which actually think that's a good idea though
rsynnott|14 days ago
I doubt, by and large, that it's the same people. Just as this LLM misquoting is journalistic malpractice, "don't look at the code anymore" is engineering malpractice.
ChrisMarshallNY|16 days ago
Nothing new here, in software. What is new, is that AI is allowing dependency hell to be experienced by many other vocations.
Dylan16807|15 days ago
I'm sure that person exists but they're not representative of HN as a whole.
mattgreenrocks|16 days ago
tliltocatl|15 days ago
Same forum, not necessary same people.
usefulposter|16 days ago
https://web.archive.org/web/20260213211721/https://arstechni...
>Scott Shambaugh here. None of the quotes you attribute to me in the second half of the article are accurate, and do not exist at the source you link. It appears that they themselves are AI hallucinations. The irony here is fantastic.
Instead of cross-checking the fake quotes against the source material, some proud Ars Subscriptors proceed to defend Condé Nast by accusing Scott of being a bot and/or fake account.
EDIT: Page 2 of the forum thread is archived too. This poster spoke too soon:
>Obviously this is massive breach of trust if true and I will likely end my pro sub if this isnt handled well but to the credit of ARS, having this comment section at all is what allows something like this to surface. So kudos on keeping this chat around.
bombcar|16 days ago
asddubs|16 days ago
vor_|16 days ago
sphars|16 days ago
[0]: https://arstechnica.com/civis/threads/journalistic-standards...
_HMCB_|16 days ago
usefulposter|16 days ago
All threads have since been locked:
https://arstechnica.com/civis/threads/journalistic-standards...
https://arstechnica.com/civis/threads/is-there-going-to-be-a...
https://arstechnica.com/civis/threads/um-what-happened-to-th...
epistasis|16 days ago
Printing hallucinated quotes is a huge shock to their credibility, AI or not. Their credibility was already building up after one of their long time contributors, a complete troll of a person that was a poison on their forums, went to prison for either pedophilia or soliciting sex from a minor.
Some serious poor character judgement is going on over there. With all their fantastic reporters I hope the editors explain this carefully.
singpolyma3|16 days ago
justinclift|16 days ago
Don't you mean diminishing or disappearing instead of building up?
Building up sounds like the exact opposite of what I think you're meaning. ;)
trollbridge|17 days ago
phire|17 days ago
This isn't exactly a new problem we do it with any bit of new software/hardware, not just LLMs. We check its work when it's new, and then tend to trust it over time as it proves itself.
But it seems to be hitting us worse with LLMs, as they are less consistent than previous software. And LLM hallucinations are partially dangerous, because they are often plausible enough to pass the sniff test. We just aren't used to handling something this unpredictable.
prussia|16 days ago
kortilla|16 days ago
adamddev1|16 days ago
kmeisthax|15 days ago
seanhunter|16 days ago
0xbadcafebee|16 days ago
How do you know quantum physics is real? Or radio waves? Or just health advice? We don't. We outsource our thinking around it to someone we trust, because thinking about everything to its root source would leave us paralyzed.
Most people seem to have never thought about the nature of truth and reality, and AI is giving them a wake-up call. Not to worry though. In 10 years everyone will take all this for granted, the way they take all the rest of the insanity of reality for granted.
DonHopkins|16 days ago
[deleted]
Lerc|16 days ago
vor_|16 days ago
"...it illustrates exactly the kind of unsupervised output that makes open source maintainers wary."
followed later on by
"[It] illustrates exactly the kind of unsupervised behavior that makes open source maintainers wary of AI contributions in the first place."
joquarky|16 days ago
The utility is that the infrenced output tends to be right much more often than wrong for mainstream knowledge.
Pay08|16 days ago
DonHopkins|16 days ago
[deleted]
moomin|16 days ago
Springtime|16 days ago
This is what the author actually speculated may have occurred with Ars. Clearly something was lacking in the editorial process though that such things weren't human verified either way.
giobox|16 days ago
The right thing to do would be a mea-culpa style post and explain what went wrong, but I suspect the article will simply remain taken down and Ars will pretend this never happened.
I loved Ars in the early years, but I'd argue since the Conde Nast acquisition in 2008 the site has been a shadow of its former self for a long time, trading on a formerly trusted brand name that recent iterations simply don't live up to anymore.
khannn|16 days ago
I'm basically getting tech news from social media sites now and I don't like that.
jandrewrogers|16 days ago
antod|16 days ago
emmelaich|16 days ago
netsharc|16 days ago
unknown|16 days ago
[deleted]
llbbdd|16 days ago
rectang|16 days ago
We’ll know more in only a couple days — how about we wait that long before administering punishment?
arduanika|16 days ago
> If you are the person who deployed this agent, please reach out. It’s important for us to understand this failure mode, and to that end we need to know what model this was running on and what was in the soul document. I’m not upset and you can contact me anonymously if you’d like.
I can see where he's coming from, and I suppose he's being the bigger man in the situation, but at some point one of these reckless moltbrain kiddies is going to have to pay. Libel and extortion should carry penalties no matter whether you do it directly, or via code that you wrote, or via code that you deployed without reading it.
The AI's hit piece on Scott was pretty minor, so if we want to wait around for a more serious injury that's fine, just as long as we're standing ready to prosecute when (not 'if') it happens.
asddubs|16 days ago
JPKab|16 days ago
It's a slop job now.
Ars Technica, a supposedly reputable institution, has no editorial review. No checks. Just a lazy slop cannon journalist prompting an LLM to research and write articles for her.
Ask yourself if you think it's much different at other publications.
joquarky|16 days ago
The ones that remain are probably at some extreme on one or more attributes (e.g. overworked, underpaid) and are leaning on genAI out of desperation.
troyvit|16 days ago
neya|16 days ago