You're probably calling openai from a backend server. If you're making the call because a user asked a question -> your endpoint received it -> did some processing -> called openai -> returned the response,
why would you cache the openai call instead of the endpoint that's receiving the user call?
No comments yet.