top | item 47108707 (no title) serendip-ml | 8 days ago The compression analogy is interesting. Another way of looking at it could be fine-tuning as "knowing what to leave out" - a 3B model for example tuned for a narrow task doesn't need the capacity that makes 70B good at many things. discuss order hn newest No comments yet.
No comments yet.