top | item 41482245

(no title)

Zondartul | 1 year ago

My hunch is that since LLMs are trained on a per word basis (okay, per-token), vacuus verbosity is overrepresented.

If you have one normal sentence and one overly verbose, the latter will have more tokens and therefore more weight.

discuss

order

No comments yet.