(no title)
Kronopath | 1 year ago
It’s very irresponsible for this article to advocate and provide a pathway to immediate superintelligence (regardless of whether or not it actually works) without even discussing the question of how you figure out what you’re searching for, and how you’ll prevent that superintelligence from being evil.
nullc|1 year ago
The obvious way to incorporate good search is to have extremely fast models that are being used in the search interior loop. Such models would be inherently less general, and likely trained on the specific problem or at least domain-- just for performance sake. The lesson in this article was that a tiny superspecialized model inside a powerful transitional search framework significantly outperformed a much larger more general model.
Use of explicit external search should make the optimization system's behavior and objective more transparent and tractable than just sampling the output of an auto-regressive model alone. If nothing else you can at least look at the branches it did and didn't explore. It's also a design that's more easy to bolt in varrious kinds of regularizes, code to steer it away from parts of the search space you don't want it operating in.
The irony of all the AI scaremongering is that if there is ever some evil AI with some LLM as an important part of its reasoning process if it is evil it may well be so because being evil is a big part of the narrative it was trained on. :D
coldtea|1 year ago
drdeca|1 year ago
aidan_mclau|1 year ago
>The cool thing about using modern LLMs as an eval/policy model is that their RLHF propagates throughout the search.
>Moreover, if search techniques work on the token level (likely), their thoughts are perfectly interpretable.
I suspect a search world is substantially more alignment-friendly than a large model world. Let me know your thoughts!
Tepix|1 year ago
Mobile Safari, phone set to french.