top | item 37551232

(no title)

borzunov | 2 years ago

Hi, a Petals dev here. </s> means "end of sequence" for LLMs. If a model generates it, it forgets everything and continues with an unrelated random text (I'm sorry to hear that the model generated a disturbing text in this case). Still, I doubt that malicious actors are involved here.

Apparently, the Colab code snippet is just too simplified and does not handle </s> correctly. This is not the case with the full chatbot app at https://chat.petals.dev - you can try it out instead.

discuss

order

quickthrower2|2 years ago

Thanks for the reply. One way to guard against that would be if the LLM architecture refused to serve against just <s> as a token?