(no title)
sockgrant | 3 months ago
IMHO the bleeding edge of what’s working well with LLMs is within software engineering because we’re building for ourselves, first.
Claude code is incredible. Where I work, there are an incredible number of custom agents that integrate with our internal tooling. Many make me very productive and are worthwhile.
I find it hard to buy in to opinions of non-SWE on the uselessness of AI solely because I think the innovation is lagging in other areas. I don’t doubt they don’t yet have compelling AI tooling.
SoftTalker|3 months ago
01100011|3 months ago
No, it's not going to write all your code for you. Yes your skills are still needed to design, debug, perform teamwork(selling your designs, building consensus, etc), etc.. But it's time to get on the train.
noduerme|3 months ago
sanmon3186|3 months ago
code51|3 months ago
Know this: someone is coming after this already.
One day someone from management will hear about a cost-saving story at a dinner table, the words GPT, Cursor, Antigravity, reasoning, AGI will cause a buzzing in her ear. Waking up with tinnitus the next morning, they'll instantly schedule a 1:1 to discuss "the degree of AI use and automation"
gherkinnn|3 months ago
Feed an LLM stack traces or ask it to ask you questions about a topic you're unfamiliar about. Give it a rough hypothesis and demand it poke holes in it. These things it does well. I use Kagi's auto summariser to distil search results in to a hand full of paragraphs and then read through the citations it gives me.
Know that LLMs will suck up to you and confirm your ideas and make up bonkers things a third of the time.
reactordev|3 months ago
You are doing yourself a huge disservice.
devjam|3 months ago
I'm a long time SWE and in the last week, I've made and shipped production changes across around 6 different repos/monorepos, ranging from Python to Golang, to Kotlin to TS to Java. I'd consider myself "expert" in maybe one or two of those codebases and only having a passing knowledge of the others.
I'm using AI, not to fire-and-forget changes, but to explain and document where I can find certain functionality, generate snippets and boilerplate, and produce test cases for the changes I need. I read, review and consider that every line of code I commit has my name against it, and treat it as such.
Without these tools I'd estimate being around 25% as effective when it comes to getting up to speed on unfamiliar code and service. For that alone, AI tooling is utterly invaluable.
dcre|3 months ago
tmikaeld|3 months ago
jatora|3 months ago
bostik|3 months ago
AI and LLMs are tools. The best tools tend to be highly focused in their application. I expect AI to eventually find its way to various specific tool uses, but I have no crystal ball to predict what those tools might be or where they will surface. Although I have to say that I have seen, earlier this week, the first genuinely interesting use-case for AI-powered code generation.
A very senior engineer (think: ~40 years of hands-on experience) had joined a company and was frustrated by lack of integration tests. Unit tests, yes. E2E test suite, yes. Nothing in between. So he wrote a piece of machinery to automatically test integration between a selected number of interacting components, and eventually was happy with the result. But since that was only a small portion of the stack, he would have had to then replicate that body of work for a whole lot of other pieces - and thought "I could make AI repeat this chore".
The end result is a set of very specific prompts, constraints, requirements, instructions, and sequences of logical steps that tell one advanced model what to do. One of the instructions is along the lines of "use this body of work I wrote $OVER_THERE as a reference". That the model is building iteratively a set of tests that self-validate the progress certainly helps. The curious insight in the workflow is that once the model has finished, he then points the generated body of work to another advanced model from a different vendor, and has that do an automated code review, again using his original work as a reference material. And then feeds that back to the first model to fix things.
That means that he still has to do the final review of the results, and tweak/polish parts where the two-headed AI went off the rails. But overall the approach saves quite a lot of time and actually scales pretty much linearly to the size of the codebase and stack depth. To quote his presentation note, "this way AI works as a highly productive junior that can follow good instructions, not as a misguided oracle that comes up with inventive reinterpretations."
He made modern AI repeat his effort, but crucially he had had to do the work at least once to know precisely what constraints would apply. I suspect that eventually we'll be seeing more of these increasingly sophisticated but very narrowly tailored tooling use cases to pop up. The best tools are after all focused, even surgical.
ß: Who could have predicted in 1900 that radioactive compounds would change fields ranging from medicine to food storage?
perfmode|3 months ago
mrweasel|3 months ago
Or, more correctly, they don't work well for my problems or usage. They can at best answer basic questions, stuff you could lookup using a search engine, if you knew what to look for. They can also generate code for inspiration, but you'll end up rewriting all of it before you're done. What they can't do it solve your problem start to end. They really do need a RTFM mode, where they will just insult you if you're approach or design is plain wrong or at least just let you know that it will now stop helping as you're clearly of the rails.
We need to bubble to pop, it'll be a year or two, the finance bros aren't done extracting value from the stonks. Once it does, we can focus on what's working and what isn't and refine the good stuff.
Right now the LLMs are the product, and they can't be, it makes no sense. They need to be embedded within product, either as a built in feature, e.g. CoPilot in Visual Studio, or as plugins, like LSPs.
Clearly others are having more luck with LLMs than I do, and do amazing projects, but that sort of illustrates the point, their aren't ready and we don't have a solution for them to be universally useful (and here I'm even restricting myself to thinking about coding).
Aeolun|3 months ago
alfiedotwtf|3 months ago
You’re deliberately disadvantaging yourself by a mile. Give it a go
… the first one’s free ;)
monkaiju|3 months ago
whstl|3 months ago
In a 1-year company, the only tech person that's been there for more than 3-4 months (the CTO), only really understands a tiny fraction of the codebase and infrastructure, and can't review code anymore. Application size has blown up tremendously despite being quite simple. Turnover is crazy and people rarely stay for more than a couple months. The team works nights and weekends, and sales is CONSTANTLY complaining about small bugs that take weeks to solve.
The funny thing is that this is an AI company, but I see the CTO constantly asking developers "how much of that code is AI?". Paranoia has set in for him.
overfeed|3 months ago
lumost|3 months ago
They have not necessarily changed the rate at which I produce valuable outputs (yet).
awinter-py|3 months ago
gniv|3 months ago
bgwalter|3 months ago
It's a code laundering machine. Software engineering has a higher number of people who have never created anything by themselves and have no issues with copyright infringement. Other professions still tend to take a broader view. Even unproductive people in other professions may have compunctions about stealing other people's work.
a_bonobo|3 months ago
Kind of how for the longest time, Google used to be best at finding solutions to programming problems and programming documentation: say, a Google built by librarians would have a totally different slant.
Perhaps that's why designers don't see it yet, no designers have built Claude's 'world-view'.
ihaveajob|3 months ago
verdverm|3 months ago
With subagents and A2A generally, you should be able to hook any of them into your preferred agentic interface
hagbarth|3 months ago
NitpickLawyer|3 months ago
AGI is a lot of things, a lot of ever moving targets, but it's never (under any sane definition) "infinite growth". That's already ASI territory / singularity and all that stuff. I see more and more people mixing the two, and arguing against ASI being a thing, when talking about AGI. "Human level competences" is AGI. Super-human, ever improving, infinite growth - that's ASI.
If and when we reach AGI is left for everyone to decide. I sometimes like to think about it this way: how many decades would you have to go back, and ask people from that time if what we have today is "AGI".
hollowturtle|3 months ago
redorb|3 months ago
signa11|3 months ago
the jury is still out on that...
morshu9001|3 months ago
buu700|3 months ago
Of course it's a collaborative process — you can't just tell it to document the code with no other information and expect it to one-shot exactly what you wanted — but I find that documentation is actually a major strength of LLMs.
toenail|3 months ago
dr_dshiv|3 months ago
voidhorse|3 months ago
This ignorance or failure to address these aspects of the issue and solely focus on its utility in a vacuum is precisely the blinkered perspective that will enable the consolidations of power the essay is worried about...the people pushing this stuff are overjoyed that so few people seem to be paying any attention to the more significant shifts they are enacting (as the article states, land purchase, political/capital power accumulation, reduction of workforces and operating costs and labor power... the list goes on)
pigpop|3 months ago
insane_dreamer|3 months ago
The mistake is to project the same level of productivity provided by LLMs in coding to all other areas of work.
The point of TFA is that LLMs are an excellent tool for particular aspects of work (coding being one of them), not a general intelligence tool that improves all aspects (as we're being sold).
smoody07|3 months ago
If LLMs only disrupt software engineering and content slop, the economy is going to undergo rapid changes. Every car wash will have a forward deployed engineer maintaining their mobile app, website, backend, and LLM-augmented customer service. That happens even if LLMs plateau in six months.
elevatortrim|3 months ago
However, while I like using AI for software development, as also a middle-manager, it increased my output A TON because AI works better for virtually anything that's not software development.
Examples: Update Jira issues in bulk, write difficult responses and incident reports, understand a tool or system I'm not familiar with, analyse 30 projects to understand which of them have this particular problem, review tickets in bulk to see if they have anything missing that was mentioned in the solution design, and so on ... All sorts tasks that used to take hours, now take minutes.
This is in line with what I'm hearing from other people: My CFO is complaining daily about running out of tokens. My technical sales relative says it is now taking him minutes to create tech specs from requirements of his customers, while it used to take hours.
While devs are rightfully "meh" because they truly need to review every single line generated by AI and type-writing the code is not their bottleneck anyway. It is harder to realise the gains for them.
lomase|3 months ago
The only reply I have got to this question was: it created a sap script.
unknown|3 months ago
[deleted]
muldvarp|3 months ago
How are we building _for_ ourselves when we literally automate away our jobs? This is probably one of the _worst_ things someone could do to me.
DennisP|3 months ago
Maybe that will continue with AI, or maybe our long-standing habit will finally turn against us.
jstanley|3 months ago
If you're just in it to collect a salary, then yeah, maybe you do benefit from delivering the minimum possible productivity that won't get you fired.
But if you like making computers do things, and you get joy from making computers do more and new things, then LLMs that can write programs are a fantastic gift.