LLMLingua uses a well-trained small language model after alignment, such as GPT2-small or LLaMA-7B, to detect the unimportant tokens in the prompt and enable inference with the compressed prompt in black-box LLMs, achieving up to 20x compression with minimal performance loss.
cyanydeez|2 years ago
arthurcolle|2 years ago
Linked for the culture: https://www.youtube.com/watch?v=bctjSvn-OC8&t=4s
Sleep big last night
unknown|2 years ago
[deleted]
sroussey|2 years ago
behnamoh|2 years ago
TarqDirtyToMe|2 years ago
I don’t think this models use of alignment implies any sort of censorship, it’s just being tuned to accomplish the task of outputting only important tokens for the target llm
nathan_compton|2 years ago
I agree that like "tone" alignment is silly and pointless for models in the public domain, but if I were a big company who wanted to keep customers I'd align my models this way. It isn't censorship, its marketing.