Solving the strawberry problem will probably require a model that just works with bytes of text. There have been a few attempts at building this [1] but it just does not work as well as models that consume pre-tokenized strings.[1]: https://arxiv.org/abs/2106.12672
randomdata|1 year ago
jeroenhd|1 year ago
Even if these models did have a concept of the letters that make up their tokens, the problem still exists. We catch these mistakes and we can work around them by altering the question until they answer correctly because we can easily see how wrong the output is, but if we fix that particular problem, we don't know if these models are correct in the more complex use cases.
In scenarios where people use these models for actual useful work, we don't alter our queries to make sure we get the correct answer. If they can't answer the question when asked normally, the models can't be trusted.
mistercow|1 year ago
One of the things that makes that problem interesting is that during training, “what the model is good at” is a moving target.
eithed|1 year ago
viraptor|1 year ago
karterk|1 year ago