top | item 44715342

(no title)

Jabrov | 7 months ago

Not exactly true ... KV and prompt caching is a thing

discuss

order

yahoozoo|7 months ago

Assuming you include the same prompts in the new request that were cached in the previous ones.

throw310822|7 months ago

As far as I understand, the entire chat is the prompt. So at the each round, the previous chat up to that point could already be cached. If I'm not wrong, Claude APIs require an explicit request to cache the prompt, while OpenAI's handle this automatically.

littlestymaar|7 months ago

I don't understand how you are downvoted…