top | item 38388039 (no title) nvrmnd | 2 years ago I don't think that's accurate, it generates novel outputs that were not observed in the training data. discuss order hn newest thomastjeffery|2 years ago It doesn't generate new tokens.Train an LLM on text that only uses lowercase, and it will never output an uppercase letter. nvrmnd|2 years ago So the model is limited to using words and characters that already exist. I agree with you but I don't see why is a limitation worth pointing out. load replies (1) pixl97|2 years ago But can it put two tokens together10 01 = 1001?
thomastjeffery|2 years ago It doesn't generate new tokens.Train an LLM on text that only uses lowercase, and it will never output an uppercase letter. nvrmnd|2 years ago So the model is limited to using words and characters that already exist. I agree with you but I don't see why is a limitation worth pointing out. load replies (1) pixl97|2 years ago But can it put two tokens together10 01 = 1001?
nvrmnd|2 years ago So the model is limited to using words and characters that already exist. I agree with you but I don't see why is a limitation worth pointing out. load replies (1)
thomastjeffery|2 years ago
Train an LLM on text that only uses lowercase, and it will never output an uppercase letter.
nvrmnd|2 years ago
pixl97|2 years ago
10 01 = 1001?