top | item 37030858

(no title)

petulla | 2 years ago

What's the inference time without gpu?

discuss

order

lm2s|2 years ago

It might the time mentioned at the bottom of the page since the author isn't sure that the GPU is being used:

>How to speed this up—right now my Llama prompts often take 20+ seconds to complete.