Does anyone have real life experience (preferably verified in production environment) of fine-tuning actually adding new knowledge to the existing LLM in a reliable and consistent manner? I've seen claims that fine-tuning only adapt the "forms" but can't adding new knowledge, while some claim otherwise. I couldn't convince myself either way with my limited adhoc/anecdotal experiments.
ozr|1 year ago
I have no idea where the myth of ‘can’t add new knowledge via fine-tuning’ came from. It’s a sticky meme that makes no sense.
Pretraining obviously adds knowledge to a model. The difference between pretraining and fine-tuning is the number of tokens and learning rate. That’s it.
mvkel|1 year ago
Aren't rag and fine tuning fundamentally flawed, because they only play at the surface of the model? Like sprinkles on the top of the cake, expecting them to completely change the flavor. I know LoRA is supposed to appropriately weight the data, but the results say that's not the solution.
Also anecdotal, but way less work!
lmeyerov|1 year ago
RAG is effectively prompt context optimization, so categorically rejecting doing that doesn't make sense to me. Maybe if models internalized that or scaled... But they don't.
ozr|1 year ago
fredliu|1 year ago
unreal6|1 year ago
simonw|1 year ago
lmeyerov|1 year ago
Nexusflow probably too, as it also does function calling and would need to bake in, or explicit fine-tuning for RAG use, which I don't recall seeing
I haven't look recently, but there is also a cool category of models that provide GIS inferencing via LLM
unknown|1 year ago
[deleted]
coder543|1 year ago
fredliu|1 year ago
unknown|1 year ago
[deleted]
mmoskal|1 year ago
theendisney|1 year ago
emersonrsantos|1 year ago
andy99|1 year ago