top | item 45645557

(no title)

rafaelero | 4 months ago

The problem with this approach to text generation is that it's still not flexible enough. If during inference the model changes its mind and wants to output something considerably different it can't because there are too many tokens already in place.

discuss

order

nodja|4 months ago

That's not true, you could just have looked at the first gif animation in the OP and seen that tokens disappear, the only part that stays untouched is the prompt, adding noise is part of the diffusion process and the code that does it is even posted in the article (ctrl+f "def diffusion_collator").

rafaelero|4 months ago

Looks like you are correct.

didibus|4 months ago

Could maybe be solved by reintroducing noise steps in between denoising step?

oezi|4 months ago

Didn't anybody add backspace to an LLM's output token set yet?