top | item 39614434

(no title)

valstu | 2 years ago

I assume you need to split the data to suitable sized database rows matching your model max length? Or does it do some chunking magic automatically?

discuss

order

chuckhend|2 years ago

There is no chunking built into the postgres extension yet, but we are working on it.

It does check the context length of the request against the limits of the chat model before sending the request, and optionally allows you to auto-trim the least relevant documents out of the request so that it fits the model's context window. IMO its worth spending time getting chunks prepared, sized, tuned for your use case though. There are some good conversations above discussing methods around this, such as using a summarization model to create the chunks.