top | item 39699147

(no title)

andher | 1 year ago

Finally! I've been using the assistants api in building an ai mock interviewer (https://comp.lol) but the responses were painfully slow when using the latest iterations of the gpt-4 model. This will make things so much more responsive

discuss

order

cosmotic|1 year ago

I'd still want to see the entire response all at once. Having it stream in while I read it would be very distracting and make it difficult for me to read.

qwertox|1 year ago

It's a request the front-end developer should be confronted with, not OpenAI.

The website could as well buffer the incoming stream until the used clicks an area to request the display of the next block of the response, once he has finished reading the initial sentences.

TowerTall|1 year ago

yes, it like surfing porn in the early internet year using a dialup modem. One line a the time until you finally can see enough of the picture (reply) to realize that is was not the reply you were looking for.

LLM streaming must be a cost saving feature to prevent you from overloading the servers by asking to many questions with in a short time frame. Annoying feature IMHO

pieterhg|1 year ago

Same it was super slow and unusable when I tried. 10 seconds for a reply or smth. GPT4 API itself was way faster