(no title)
andher
|
1 year ago
Finally! I've been using the assistants api in building an ai mock interviewer (https://comp.lol) but the responses were painfully slow when using the latest iterations of the gpt-4 model. This will make things so much more responsive
cosmotic|1 year ago
qwertox|1 year ago
The website could as well buffer the incoming stream until the used clicks an area to request the display of the next block of the response, once he has finished reading the initial sentences.
TowerTall|1 year ago
LLM streaming must be a cost saving feature to prevent you from overloading the servers by asking to many questions with in a short time frame. Annoying feature IMHO
unknown|1 year ago
[deleted]
pieterhg|1 year ago