(no title)
podnami | 6 months ago
On ChatGPT.com o3 thought for for 13 seconds, on OpenRouter GPT OSS 120B thought for 0.7 seconds - and they both had the correct answer.
podnami | 6 months ago
On ChatGPT.com o3 thought for for 13 seconds, on OpenRouter GPT OSS 120B thought for 0.7 seconds - and they both had the correct answer.
swores|6 months ago
Workaccount2|6 months ago
nisegami|6 months ago
golergka|6 months ago
Imustaskforhelp|6 months ago
I am not kidding but such progress from a technological point of view is just fascinating!
xpe|6 months ago
What is being measured here? For end-to-end time, one model is:
t_total = t_network + t_queue + t_batch_wait + t_inference + t_service_overhead