(no title)
lo_fye | 10 months ago
What if every time you had an Aha! moment, you blogged about it in detail. Many people do. AI ingests those blog posts. It uses what they say when writing new code, or assessing existing code. It does use hard-won knowledge; it just wasn't hard-won by AI itself.
ath3nd|10 months ago
The current crop of LLMs has a lot of knowledge, but severely lacks on the "intelligence" part. Sure, it can "guess" how to write a unit test consistent with your codebase (based on examples), but for those 1% where you need to make a deviation from the rule, it's completely clueless how to do it properly. Guessing is not intelligence, although it might appear masked as such.
Don't get me wrong, the "guessing" part is sometimes uncannily good, but it just can't replace real reasoning.
DontchaKnowit|10 months ago