They generate text based on quite a large context, including hidden prompts we don’t see and their weights are distorted heavily by training. So I think there’s a lot more than a simple probability of word x coming next. That makes ‘predict next word’ a reductive summary IMO.
I do not personally feel it resembles thinking or reasoning though and really object to that framing because it is misleading many people.
grey-area|3 days ago
I do not personally feel it resembles thinking or reasoning though and really object to that framing because it is misleading many people.
karamanolev|3 days ago
What does that even mean? Their weights are essentially created by training. There aren't some magic golden weights that are then distorted.