These blog posts are fascinating to read. I don't have a personal blog, but if I did I'm sure I would've written a very similar post as I've been wrestling with similar thoughts over the last few weeks. I have the distinct sense that I will look back on February 2026 as an inflection point, where AI crossed over from being an interesting parlor trick to something that fundamentally and irreversibly altered what I do day-to-day. It's bittersweet, for sure - it feels inevitable that the craft of software development that I've loved for years will be seen as an archaic relic at some point in the not too distant future. It may be several years yet before the impact is broadly felt (the full impact of today's frontier models has yet to be felt by the general public - to say nothing of models that will be released in the next few years) but this train doesn't seem to be slowing down anytime soon. This post was a helpful reminder that who I am is not defined by the code I write (or don't write) - there's so much more to life than code.
DonThomasitos|1 day ago
bityard|1 day ago
But: programming languages, libraries, and abstractions are not going away. It is still possible (and might always be possible) to get deep into the weeds of Python or Rust or whatever to understand how those work and really harness them to their full potential, or develop them further. It just won't be _compulsary_ (in most industries) if your only goal is to trade lines of code for dollars in your bank account.
pvillano|18 hours ago
melagonster|11 hours ago
chasd00|7 hours ago
UncleOxidant|1 day ago
We definitely saw some kind of non-linear step function jump in quality around the beginning of the year - it's hard to express how good Claude opus/sonnet 4.6 is now. However, I wonder if we're going to see the same kind of improvement from here? It's kind of like we got to the 80% point but the next 20% is going to be a lot harder/take longer than that first 80% (pareto principle). Also, as more and more code out there is AI generated it's going to be like the snake eating it's own tail. Training models on AI generated code doesn't seem like it will lead to improvements.
fragmede|15 hours ago
jimbokun|18 hours ago
throwaway1492|9 hours ago
sublinear|6 hours ago
Your job has always been to be accountable for the technical decisions made. Sometimes I read comments like yours and wonder if the almost apologetic tone is an admission of guilt that you never really liked this job. You can speak for yourself, but I didn't get into this to just noodle around or work at unhinged startups where nobody gives a rat's ass about code quality.