top | item 41787167

(no title)

hayden_k | 1 year ago

hey! sorry about that, it’s still not perfect but shows that using CoT prompt does improve llm responses. compared with its base model, you can clearly see some difference. If you like, please email me at contact@pixelverse.tech with some prompts you provided that t1 failed to respond correctly and I can take a look.

discuss

order

latexr|1 year ago

> but shows that using CoT prompt does improve llm responses.

A wrong answer is a wrong answer. In one of the questions it failed exactly in the same manner that GPT-4o did when I asked, so it’s not clear at all this is better. I could even see the chain and identify exactly where it made the mistake, but that’s not really a consolation.

hayden_k|1 year ago

As I said - it’s not perfect at answering every question right. What I am saying is that CoT promoting does have an effect on the quality of LLM responses. Ask how many r in strawberry or a similar question to t1 and llama 3.1 and you will see that CoT strategy has some effect.

hayden_k|1 year ago

Also to be clear - I never claimed that t1 is better than gpt 4o and o1, but thank you for trying it and providing feedback :)