I would assume Google search is using a cheaper, flakier model. But it could also be that some contractor spent 30 minutes teaching Gemini that Kenya starts with a K. This specific example is a well-known LLM mistake and it seems plausible that Gemini would specifically be trained to avoid it.
The basic problem with commercial LLMs from Big Tech is that they have the resources to "patch over" errors in reasoning with human refinement, making it seem like the reasoning error is fixed when it is only fixed for a narrow category of questions. If Gemini knows about Africa and K, does it know Asia and O? (Oman) Or some other simple variation.
nicklecompte|1 year ago
The basic problem with commercial LLMs from Big Tech is that they have the resources to "patch over" errors in reasoning with human refinement, making it seem like the reasoning error is fixed when it is only fixed for a narrow category of questions. If Gemini knows about Africa and K, does it know Asia and O? (Oman) Or some other simple variation.
worewood|1 year ago