(no title)
nafizh
|
2 years ago
My use of ChatGPT has just organically gone down 90%. It's unable to do any sort of task of non-trivial complexity e.g. complex coding tasks, writing complex prose that conforms precisely to what's been asked etc. Also I hate the fact that it has to answer everything in bullet points, even when it's not needed, clearly rlhf-ed. At this point, my question types have become what you would ask a tool like perplexity.
Kranar|2 years ago
ChatGPT writes excellent API documentation and can also document snippets of code to explain what they do, it does 80% of the work for unit tests, it can fill in simple methods like getters/setters, initialize constructors, I've even had it write a script to perform some substantial code refactoring.
Use ChatGPT for grunt work and focus on the more advanced stuff yourself.
hirvi74|2 years ago
For example, I was looking up Epipens (Epinephrine), and I happened to notice the side-effects were similar to how overdosing on stimulants would manifest.
So, I asked it, "if someone was having a severe allergic reaction and no Epipen was available, then could Crystal Methamphetamine be used instead?"
GPT answered the question well, but the answer is no. Apparently, stimulants lack the targeted action on alpha and beta-adrenergic receptors that makes epinephrine effective for treating anaphylaxis.
I do not know why I ask these questions because I am not severely allergic to anything, nor anyone else that I know of, and I do not have nor wish to have access to Crystal Meth.
I've been using GPT for helping prepare for dev technical interviews, and it's been pretty damn great. I also do not have access to a true senior dev at work either, so I tend to use GPT to kind of pair program. Honestly, it's been life changing. I have also not encountered any hallucinations that weren't easy to catch, but I mainly only ask it more project architectural, design questions, and a documentation search engine than using it to write code for me.
Like you, I think not using GPT for overly complex tasks is best for now. I use it make life easier, but not easy.
ekms|2 years ago
OJFord|2 years ago
Similar I think to what you're calling 'rlhf-ed', though I think useful for code, it definitely seems to kind of scratchpad itself, and stub out how it intends to solve a problem before filling in the implementation. Where this becomes really useful though is in asking for a small change it doesn't (it seems) recompute the whole thing, but just 'knows' to change one function from what it already has.
They also seem to have it somehow set up to 'test' itself and occasionally it just says 'error' and tries again. I don't really understand how that works.
Perplexity's great for finding information with citations, but (I've only used the free version) IME it's 'just' a better search engine (for difficult to find information, obviously it's slower), it suffers a lot more from the 'the information needs to be already written somewhere, it's not new knowledge' dismissal.
nafizh|2 years ago
txutxu|2 years ago
This.
People talks about prompt engineering, but then it fails on really simple details, like "on lowercase", "composed by max two words", etc... and when you point at the failure, apologizes, and composes something else that forgets the other 95% of the original prompt.
Or worse, apologizes and makes again the very same mistake.
skywhopper|2 years ago
vonwoodson|2 years ago
dr_kiszonka|2 years ago
alecco|2 years ago