LLMs are the eternal September for software, in that the sort of people who couldn’t make it through a bootcamp can now be “programming thought leaders”. There’s no longer a reliable way to filter signal from noise.
Those 3000 early adopters who are bookmarking a trivial markdown file largely overlap with the sort of people who breathlessly announce that “the last six months of model development have changed everything!”, while simultaneously exhibiting little understanding of what has actually changed.
There’s utility in these tools, but 99% of the content creators in AI are one intellectual step above banging rocks together, and their judgement of progress is not to be trusted.
I'm sure i'll piss off a lot of people with this one but I don't care any more. I'm calling it what it is.
LLMs empower those without the domain knowledge or experience to identify if the output actually solves the problem. I have seen multiple colleagues deliver a lot of stuff that looks fancy but doesn't actually solve the prescribed problem at all. It's mostly just furniture around the problem. And the retort when I have to evaluate what they have done is "but it's so powerful". I stopped listening. It's a pure faith argument without any critical reasoning. It's the new "but it's got electrolytes!".
The second major problem is corrupting reasoning outright. I see people approaching LLMs as an exploratory process and let the LLM guide the reasoning. That doesn't really work. If you have a defined problem, it is very difficult to keep an LLM inside the rails. I believe that a lot of "success" with LLMs is because the users have little interest in purity or the problem they are supposed to be solving and are quite happy to deliver anything if it is demonstrable to someone else. That would suggest they are doing it to be conspicuous.
So we have a unique combination of self-imposed intellectual dishonesty, mixed with irrational faith which is ultimately self-aggrandizing. Just what society needs in difficult times: more of that! :(
> LLMs are the eternal September for software, in that the sort of people who couldn’t make it through a bootcamp can now be “programming thought leaders”
Do you ever filter signal from noise by the quality of the code? The code written by the Google founders was eventually rewritten by others, and it was likely worse than what a fresh grad produces today. Still, that initial search engine is the most influential thing they ever built, and it's something the modern Bay Area will probably never create again.
The Markdown file looks like it’s written for people who either haven’t discovered Plan mode, or who can’t be bothered to read a generated plan before running with it.
That might be because AI is being pushed largely by leaders that do not have the experience you’re referring to. Determinism is grossly undervalued - look at how low the knowledge of formal verification is, let alone its deployment in the real world!
I've got a few clients using a Wordpress setup. They like being able to edit their own blogs, add new job listings, and rearrange the leadership page when people are promoted. All of that can happen without bugging the developer (me) and all the changes are fully auditable and reversible.
and there is a crap-ton of "apps" that repackage the entire world^W^W excuse me, Chromium, hog RAM, and destroy any semblance of native feel - all to write "production-ready" [sic] cross-platform code in JavaScript, a language more absurd than C++[0] but so easy to start with.
My probably incorrect, uninformed hunch is that users convinced of how AI should act actually end up nerfing its capabilities with their prompts. Essentially dumbing it down to their level, losing out on the wisdom it's gained through training.
I've experienced in times of gpt 3, and 3.5 that existence of any, even 1-word system message changed output drastically in the worse direction. I did not verify this behaviour with recentl models.
Since then I don't impose any system prompts on users of my tg bot. This is so unusual and wild in relation to what others do that very few actually appreciate it. I'm happy I don't need to make money for living with this project thus I and can keep it ideologically clean: user's control over system prompts, temperature, top_p, giving selection of the top barebones LLMs.
I often wonder this as well, things are moving so quickly that unless you want to keep chasing the next best prompt/etc then you are better running as close to vanilla as you can IMHO.
Similar for MCP/Skills/Prompts, I’m not saying they can’t/don’t help but I think you can shoot yourself in the foot very easily and spend all your time trying to maintain those things and/or try to force the agent to use your Skill/MCP. That or having your context eaten up with bad MCP/Skills.
I read a comment the other day about sometime talking about Claude Code getting dumber then they went on to explain switching would be hard due to their MCP/Skills/Skill router setup. My dude, maybe _that’s_ the problem?
So we reached a point where the quality of a `piece` of software is decided based on stars on GitHub.
The exact same thing happened with xClaw where people where going "look at this app that got thousands of stars on GitHub in only a few days!".
How is that different than the followers/likes counts on the usual social networks?
Given how much good it did to give power to strangers based on those counts, it's hard not to think that we're going in the completely wrong direction.
All good advice in general. Could add others, like x-y problems etc.
This feels like a handbook for a senior engineer becoming a first level manager talking to junior devs. Which is exactly what it should be.
However, this will go horribly wrong if junior devs are thus “promoted “ to eng managers without having cut their teeth on real projects first. And that’s likely to happen. A lot.
Bro science is rampant in the AI world. Every new model that comes out is the best there ever was, every trick you can think of is the one that makes all the other users unsophisticated, "bro, you are still writing prompts as text? You have to put them into images so the AI can understand them visually as well as textually".
It isn't strange that this is the case, because you'd be equally hard pressed to compare developers at different companies. Great to have you on the team Paul, but wouldn't it be better if we had Harry instead? What if we just tell you to think before you code, would that make a difference?
You know, it's good old prompt/context engineering. To be fair, markdowns actually can be useful because of LLM's (Transformer's) gullible/susceptible nature... At least that's what I discovered developing a prompting framework.
Of course it's hilarious a single markdown got 4000 starts, but it looks like just another example of how people chase a buzzing x post in tech space.
If exatly such markdown was written by some Joe from the internet no one would notice it. So these stars exist not because of quality or utility of the text.
Funny. The author is not sure if his (and original) extension improves Claude output, but since the original project has 4k+ start on Github, "surely, 4,000 developers can’t be wrong". So, "Please try for yourself! Install the extension, don’t forget to star my repository and see the results". No matter if it's good, just star my repo.
I have been thinking a lot about the use of AI and how to use it. Part of my process has been watching others, namely the people who I thought were incompetent at their job before AI.
I have found the following, but I suspect as AI gets better this will change.
1) those who where incompetent before still are, but AI hides it.
2) those who were competent before AI do vastly more with AI. They seem to apply it in away that simply overshadows what the incompetent are doing.
3) the incompetent seem to be fascinated with things like skills, pre prompts and, setting policies and guidelines, and workshops. The competent seem to need none of this, are not going to workshops, already have their own and simply are more productive.
Why are you surprised that it takes only 65 lines of Markdown to create the next Linus Torvalds or Donald Knuth!
I’m learning Haskell by Thinking before Coding, without working through the difficult exercises. I finally understand Monads because of Thinking first.
Maybe I'm just really lucky but reading those instructions it's basically how I find Claude Code behaves. That repo with 4k stars is only 2 weeks old as well, so it's obviously not from a much less competent model.
Same here, I find most of these skills/prompts a bit redundant. Some people argue that in including these in the conversation, one is doing latent space management of sorts and bringing the model closer to where one wants it to be.
I wonder what will happen with new LLMs that contain all of these in their training data.
Next inevitable step is LLM alchemy. People will be writing crazy-ass prompts which ununderstandable text which somehow get system work better than the straight-human-text prompts.
1. Between Jan 27th and Feb 3rd stars grew quickly to 3K, project was released at that time.
2. People star it to be on top of NEW changes, people wanted to learn more about what's coming - but it didn't come. Doesn't mean people are dumb.
3. If OP synthesized the Markdown into a single line: "Think before coding" - why did he went through this VS Code extension publishing? Why can't they just share learnings and tell the world, "Add 'Think before coding' before your prompt and Please try for yourself!"
PS: no I haven't starred this project, I didn't know about it. But I disagree with the authors "assumptions" about stars and correlating it to some kind of insight revelation
I just packaged the extension for the fun of it! And I do want people to try for themselves, that is the point. About people that are not dumb; surely many people are not dumb; many people are very smart indeed. But that does not prove there are no dumb or gullible people!
These hopeful incantations are a kind of cargo cult… But when applied to programming, the wild thing is that the cult natives actually built the airports and the airplanes but they don’t know what makes them fly and where the cargo comes from.
I find the whole premise of writing some vague instructions, feeding them to a stochastic parrot and expecting a solid engineering process to materialize out of the blue quite ridiculous.
Any sufficiently advanced "AI" technology is indistinguishable from bullshit.
This is so incredibly depressing. As of the time of me posting this comment this has 73 upvotes. I'm sorry, but this is absurd. People put real effort in real posts that don't see half this many votes but AI Slop on top of AI Slop? Upvote!
This post is about a prompt that has 4K+ stars (that matters to people?) so they wrapped it up in an extension for Cursor and they ask you to try it out and star their repo (Please clap?).
And this "Sensation"?
> Was the result better? I’m not really sure.
I cannot even...
A Slop article about a Slop prompt that a bunch of Slop people starred, "I don't know if it helps but here is an extension that you should use", just the laziest of everything.
I'd bet that this prompt, if it is even helpful at all, will stop be helpful or be redundant in a couple weeks at most time.
I'm _far_ from anti-LLM, I use them quite a bit daily, I run multiple Claude Code instances (which I review, gasp!), but this is reaching a fever pitch and this article was the straw that broke this camel's back.
timr|17 days ago
Those 3000 early adopters who are bookmarking a trivial markdown file largely overlap with the sort of people who breathlessly announce that “the last six months of model development have changed everything!”, while simultaneously exhibiting little understanding of what has actually changed.
There’s utility in these tools, but 99% of the content creators in AI are one intellectual step above banging rocks together, and their judgement of progress is not to be trusted.
adjfasn47573|17 days ago
So I wouldn’t give anything on 3k stars at all.
dgxyz|17 days ago
I'm sure i'll piss off a lot of people with this one but I don't care any more. I'm calling it what it is.
LLMs empower those without the domain knowledge or experience to identify if the output actually solves the problem. I have seen multiple colleagues deliver a lot of stuff that looks fancy but doesn't actually solve the prescribed problem at all. It's mostly just furniture around the problem. And the retort when I have to evaluate what they have done is "but it's so powerful". I stopped listening. It's a pure faith argument without any critical reasoning. It's the new "but it's got electrolytes!".
The second major problem is corrupting reasoning outright. I see people approaching LLMs as an exploratory process and let the LLM guide the reasoning. That doesn't really work. If you have a defined problem, it is very difficult to keep an LLM inside the rails. I believe that a lot of "success" with LLMs is because the users have little interest in purity or the problem they are supposed to be solving and are quite happy to deliver anything if it is demonstrable to someone else. That would suggest they are doing it to be conspicuous.
So we have a unique combination of self-imposed intellectual dishonesty, mixed with irrational faith which is ultimately self-aggrandizing. Just what society needs in difficult times: more of that! :(
OJFord|17 days ago
sph|17 days ago
The democratization of programming (derogatory)
pikachu0625|16 days ago
EdNutting|17 days ago
unknown|17 days ago
[deleted]
squidbeak|17 days ago
[deleted]
pyrale|17 days ago
The absence of means to measure outcomes of these prompt documents makes me feel like the profession is regressing further into cargo culting.
bandrami|17 days ago
1. Reproducibility
2. Chain of custody/SBOM
3. Verification of artifacts of CI
All three of which are not simply difficult but in fact by nature impossible when using an LLM
EdNutting|17 days ago
unknown|17 days ago
[deleted]
XenophileJKO|17 days ago
onion2k|17 days ago
Apparently almost half of all the websites on the internet run on WordPress, so it's entirely possible for developers to be wrong at scale.
joenot443|17 days ago
What FOSS solution would you recommend instead?
bananaflag|17 days ago
bpavuk|17 days ago
[0]: https://jsdate.wtf
john_owl|17 days ago
crousto|17 days ago
croes|17 days ago
>Think Before Coding
xlbuttplug2|17 days ago
podgorniy|17 days ago
I've experienced in times of gpt 3, and 3.5 that existence of any, even 1-word system message changed output drastically in the worse direction. I did not verify this behaviour with recentl models.
Since then I don't impose any system prompts on users of my tg bot. This is so unusual and wild in relation to what others do that very few actually appreciate it. I'm happy I don't need to make money for living with this project thus I and can keep it ideologically clean: user's control over system prompts, temperature, top_p, giving selection of the top barebones LLMs.
joshstrange|17 days ago
Similar for MCP/Skills/Prompts, I’m not saying they can’t/don’t help but I think you can shoot yourself in the foot very easily and spend all your time trying to maintain those things and/or try to force the agent to use your Skill/MCP. That or having your context eaten up with bad MCP/Skills.
I read a comment the other day about sometime talking about Claude Code getting dumber then they went on to explain switching would be hard due to their MCP/Skills/Skill router setup. My dude, maybe _that’s_ the problem?
wiether|17 days ago
The exact same thing happened with xClaw where people where going "look at this app that got thousands of stars on GitHub in only a few days!".
How is that different than the followers/likes counts on the usual social networks?
Given how much good it did to give power to strangers based on those counts, it's hard not to think that we're going in the completely wrong direction.
golem14|17 days ago
This feels like a handbook for a senior engineer becoming a first level manager talking to junior devs. Which is exactly what it should be.
However, this will go horribly wrong if junior devs are thus “promoted “ to eng managers without having cut their teeth on real projects first. And that’s likely to happen. A lot.
mosselman|17 days ago
It isn't strange that this is the case, because you'd be equally hard pressed to compare developers at different companies. Great to have you on the team Paul, but wouldn't it be better if we had Harry instead? What if we just tell you to think before you code, would that make a difference?
ccozan|17 days ago
That would be game changing!
arjie|17 days ago
3371|17 days ago
Of course it's hilarious a single markdown got 4000 starts, but it looks like just another example of how people chase a buzzing x post in tech space.
podgorniy|17 days ago
quiet35|17 days ago
mbrumlow|16 days ago
I have found the following, but I suspect as AI gets better this will change.
1) those who where incompetent before still are, but AI hides it.
2) those who were competent before AI do vastly more with AI. They seem to apply it in away that simply overshadows what the incompetent are doing.
3) the incompetent seem to be fascinated with things like skills, pre prompts and, setting policies and guidelines, and workshops. The competent seem to need none of this, are not going to workshops, already have their own and simply are more productive.
sathish316|16 days ago
Why are you surprised that it takes only 65 lines of Markdown to create the next Linus Torvalds or Donald Knuth!
I’m learning Haskell by Thinking before Coding, without working through the difficult exercises. I finally understand Monads because of Thinking first.
mcintyre1994|17 days ago
cobolexpert|17 days ago
I wonder what will happen with new LLMs that contain all of these in their training data.
brador|17 days ago
podgorniy|17 days ago
throwaw12|17 days ago
This is such a negative messaging!
Let's check star history: https://www.star-history.com/#forrestchang/andrej-karpathy-s...
1. Between Jan 27th and Feb 3rd stars grew quickly to 3K, project was released at that time.
2. People star it to be on top of NEW changes, people wanted to learn more about what's coming - but it didn't come. Doesn't mean people are dumb.
3. If OP synthesized the Markdown into a single line: "Think before coding" - why did he went through this VS Code extension publishing? Why can't they just share learnings and tell the world, "Add 'Think before coding' before your prompt and Please try for yourself!"
PS: no I haven't starred this project, I didn't know about it. But I disagree with the authors "assumptions" about stars and correlating it to some kind of insight revelation
roywashere|17 days ago
pavlov|17 days ago
ciconia|17 days ago
https://github.com/forrestchang/andrej-karpathy-skills/blob/...
I find the whole premise of writing some vague instructions, feeding them to a stochastic parrot and expecting a solid engineering process to materialize out of the blue quite ridiculous.
Any sufficiently advanced "AI" technology is indistinguishable from bullshit.
unknown|17 days ago
[deleted]
joshstrange|17 days ago
This post is about a prompt that has 4K+ stars (that matters to people?) so they wrapped it up in an extension for Cursor and they ask you to try it out and star their repo (Please clap?).
And this "Sensation"?
> Was the result better? I’m not really sure.
I cannot even...
A Slop article about a Slop prompt that a bunch of Slop people starred, "I don't know if it helps but here is an extension that you should use", just the laziest of everything.
I'd bet that this prompt, if it is even helpful at all, will stop be helpful or be redundant in a couple weeks at most time.
I'm _far_ from anti-LLM, I use them quite a bit daily, I run multiple Claude Code instances (which I review, gasp!), but this is reaching a fever pitch and this article was the straw that broke this camel's back.
csomar|17 days ago
rootlocus|17 days ago