top | item 38471034

(no title)

kaelinl | 2 years ago

How many tokens was that before it diverged from your original request? If it filled the context window, the original prompt would no longer be available to the model and it would start predicting just based on the repeated word. That certainly looks like what's happened here, although I haven't checked what model context size you used or how long the response is.

discuss

order

warkanlock|2 years ago

None. This was the original request.

kaelinl|2 years ago

The question is not how many prior messages there were in the conversation. That's irrelevant for context window purposes. The question is how many tokens there are between the prompt and the point where the reply breaks down.