top | item 46113990

The consumption of AI-generated content at scale

32 points| ivansavz | 3 months ago |sh-reya.com

31 comments

order

chemotaxis|2 months ago

The best part is that this article is almost certainly AI-generated or heavily AI-assisted too.

Before people get angry with me... there's plenty of small tells, starting with section headings, a lot of linguistic choices, and low information density... but more importantly, the author openly says she writes using LLMs: https://www.sh-reya.com/blog/ai-writing/#how-i-write-with-ll...

DarkNova6|2 months ago

Most likely. I got turned off instantly reading it but then realized that this is part of the joke.

phainopepla2|2 months ago

I would think a decent LLM would know the difference between a metaphor and simile, unlike the author

absoluteunit1|2 months ago

Was thinking this as well.

Just skimming throught the first two paragraphs felt like I as reading a ChatGPT response. That and the fact that there's multiple em dashes in the intro alone.

krupan|2 months ago

"Now, with LLM-generated content, it’s hard to even build mental models for what might go wrong, because there’s such a long tail of possible errors. An LLM-generated literature review might cite the right people but hallucinate the paper titles. Or the titles and venues might look right, but the authors are wrong."

This is insidious and if humans were doing it they would be fired and/or cancelled on the spot. Yet we continue to rave about how amazing LLMs are!

It's actually a complete reversal on self driving car AI. Humans crash cars and hurt people all the time. AI cars are already much safer drivers than humans. However, we all go nuts when a Waymo runs over a cat, but ignore the fact that humans do that on a daily basis!

Something is really broken in our collective morals and reasoning

carbarjartar|2 months ago

> AI cars are already much safer drivers than humans.

I feel this statement should come with a hefty caveat.

"But look at this statistic" you might retort, but I feel the statistics people pose are weighted heavily in the autonomous service's favor.

The frontrunner in autonomous taxis only runs in very specific cities for very specific reasons.

I avoid using them out of a feeble attempt to 'do my part', but I was recently talking to a friend and was surprised that they avoid using these autonomous services because they drive, what would be to a human driver, very strange routes.

I wondered if these unconventional, often longer, routes were also taken in order to stick to well trodden and predictable paths.

"X deaths/injuries per mile" is a useless metric when the autonomous vehicles only drive in specific places and conditions.

To get the true statistic you'd have to filter the human driver statistics to match the autonomous services' data. Things like weather, cities, number of and location of people in the vehicle, and even which streets.

These service providers could do this, they have the data, compute, and engineering to do so, though they are disincentivized to do so as long as everyone keeps parroting their marketing speak for them.

watwut|2 months ago

> AI cars are already much safer drivers than humans.

Nothing like that was shown. We have a bunch of very "motivated reasoning" kind of studies and best you can conclude from them is that "some circumstances where ai cars are safer drivers exist". The common trick is to compare overall human records with ai car record in super tailored circumstances.

They have potential to be safer drivers one day, if they will be produced by companies that are forced to care about safety by regulations.

pessimizer|2 months ago

I'm pretty sure that the reason everything seems like AI is that AI produces stupid, pointless content at scale, and our "writers" have become people who generate stupid, pointless content at scale.

There's no reason for most things to have been written. Whatever point is being made is pointless. It's not really entertaining, it's meant to be identified with; it's not a call to any specific action; it doesn't create some new fertile interpretation of past events or ideas; it's not even a cry for help. It's just pointless fluff to surround advertising. From a high concept likely dictated by somebody's boss.

AI has no passion and no point. It is not trying to convince anyone of anything, because it does not care. If AI were trying to be convincing, it would try to conceal its own style. But it doesn't mean anything for an AI to try. It's just running through the motions of filling out an idea to a certain length. It's whatever the opposite of compression is.

A generation of writers raised on fanfiction and prestige tv who grew up to write Buzzfeed articles at the rate of five a day are indistinguishable from AI.

Why This Matters

FarmerPotato|2 months ago

God help us if we give the bag of words a reason to live for. It might try to be convincing.

nancyminusone|2 months ago

In my opinion, this is the biggest (current) problem with AI. It is really good at that thing you used to do when you had to hit a word count in a school essay. How long until the world's hard drive space is filled up with filler words and paragraphs of text that goes nowhere, and how could you possibly search and find anything in such conditions?

Havoc|2 months ago

Noticing this most in visual content rather than LLMs. That era of anyone young and perpetually online can spot AI via uncanny valley was remarkably shortlived. [0]

>In the pre-LLM era, I could build mental models, rely on heuristics, or spot-check information strategically.

I wonder if this will be an enduring advantage of the current generation - building your formative world model in a pre-AI era. It seems plausible to me that anyone who built the foundations there has a much higher chance of having instincts that are more grounded even if post-AI experiences are layered on later

[0] https://www.reddit.com/r/antiai/comments/1p8z6y6/nano_banana...

SpaceManNabs|2 months ago

> If something seems off, I can just regenerate and hope the next version is better. But that’s not the same as actually checking. It feels like a slot machine—pull the lever again, see if you get a better result—substitutes for the slower, harder work of understanding whether the output is correct.

What a great point. In some work loops I feel like I get addicted to seeing what pops in the next generation.

One of the things i Learned from moderating internet usage is not fall prey to recommendation systems. As in, when I am on the web, I only consume what I explicitly looked for, and not what the algorithm thinks i should consume next.

sites like reddit and HN make this tricky.

furyofantares|2 months ago

Scroll through and read only the section headers. I would be shocked if this wasn't at the very least run through an LLM itself. For sure the section headers are, I'll skip the rest unless someone posts that it's worth a read for some reason.

It doesn't appear to be section headings glued together with bullet lists so maybe the content really does retain the author's perspective but at this point I'd rather skip stuff I know has been run through an LLM and miss a few gems rather than get slopped daily.

SunshineTheCat|2 months ago

What's crazy is you're starting to see an overreaction to this fact as well.

The other day I posted a short showcasing some artwork I made for a TCG I'm in the process of creating.

Comments poured in saying it was "doomed to fail" because it was just "AI slop"

In the video itself I explained how I made them, in Adobe Illustrator (even showing some of the layers, elements, etc).

Next I'm actually posting a recording of me making a character from start to finish, a timelapse.

Will be interesting if I get any more "AI slop" comments, but it's becoming increasingly difficult to share anything drawn now because people immediately assume it's generated.

raincole|2 months ago

I feel you, but people nowadays go extreme lengths to present AI-generated artworks as hand-drawn.

It's not even funny. You can google "asamiarts tracing over AI" and read the whole drama. They have not only timelapse, but real world footage as 'evidence.' And they are not the only case.

It's not the fight you can win. Either ignore the comments calling you AI or just use AI.

phainopepla2|2 months ago

I have seen this as well. Any nicely formatted medium to long text without obvious errors immediately comes under suspicion, even without the obvious tells

p_l|2 months ago

The people commenting about AI Slop, at least considerable portion, do so because it allows them to feel morally superior at little effort.

Do not expect them to retract or stop if there's a way to not see the making of :P

nh23423fefe|2 months ago

gpt is eternal september for normies

tensegrist|2 months ago

> There’s a frustration I can’t quite shake when consuming content now—

perhaps even a frustration you can't quite name

bryanrasmussen|3 months ago

yeah everything sounds like AI, and why is that? Well it might be because everything is AI but I think that writing style is more LinkedIn than LLM, the style of people who might get slapped down if they wrote something individual.

Much of the world has agreed to sound like machines.

Another thing I've noticed is that weird stuff that is perhaps off in some way, also gets accused of being LLMs because it doesn't feel right.

If you sound unique and weird you get accused of being a bad LLM that can't falsify humanity well enough, and if you sounds boring and bland and boosterist, you get accused of being a good LLM.

You can't write like no one else, but you also can't write like everybody else.

1bpp|2 months ago

Text feeling awkward or not flowing very well has ironically become a very strong signal for human-written text for me, and usually makes me pay more attention now

FarmerPotato|2 months ago

When I encounter this LLM generated Mad Lib:

"We embody <adjective> <noun> through <adjective> <noun>, <adjective> <noun>, and <adjective> <noun>. "

my uncanny warning blares--so I test if it becomes more intelligible with the adjectives stripped out. These padded-out pabulums are the tells.

I hope Elements of Style is rediscovered.