(no title)
z7 | 7 months ago
In 2021 Paul Christiano wrote he would update from 30% to "50% chance of hard takeoff" if we saw an IMO gold by 2025.
He thought there was an 8% chance of this happening.
Eliezer Yudkowsky said "at least 16%".
Source:
https://www.lesswrong.com/posts/sWLLdG6DWJEy3CH7n/imo-challe...
sigmoid10|7 months ago
AlphaAndOmega0|7 months ago
zeroonetwothree|7 months ago
grillitoazul|7 months ago
The more prior predictive power of human agents imply the more a posterior acceleration of progress in LLMs (math capability).
Here we are supposing that the increase in training data is not the main explanatory factor.
This example is the gem of a general framework for assessing acceleration in LLM progress, and I think its application to many data points could give us valuable information.
tunesmith|7 months ago
fxwin|7 months ago
davidclark|7 months ago
exegeist|7 months ago
We may certainly hope Eliezer's other predictions don't prove so well-calibrated.
rafaelero|7 months ago
causal|7 months ago
dcre|7 months ago
shuckles|7 months ago
andrepd|7 months ago
sailingparrot|7 months ago
You definitely should assume they are. They are rationalists, the modus operandi is to pull stuff out of thin air and slap a single digit precision percentage prediction in front to make it seems grounded in science and well thought out.
c1ccccc1|7 months ago
The point of giving such estimates is mostly an exercise in getting better at understanding the world, and a way to keep yourself honest by making predictions in advance. If someone else consistently gives higher probabilities to events that ended up happening than you did, then that's an indication that there's space for you to improve your prediction ability. (The quantitative way to compare these things is to see who has lower log loss [1].)
[1] https://en.wikipedia.org/wiki/Cross-entropy
ohdeargodno|7 months ago
Clowns, mostly. Yudkowski in particular, whose only job today seems to be making awful predictions and letting lesswrong eat it up when one out of a hundred ends up coming true, solidifying his position as AI-will-destroy-the-world messiah. They make money from these outlandish takes, and more money when you keep talking about them.
It's kind of like listening to the local drunkard at the bar that once in a while ends up predicting which team is going to win in football inbetween drunken and nonsensical rants, except that for some reason posting the predictions on the internet makes him a celebrity, instead of just a drunk curiosity.
meindnoch|7 months ago
Be glad you don't know anything about them. Seriously.
Maxious|7 months ago
empiricus|7 months ago
Xenoamorphous|7 months ago
Workaccount2|7 months ago
On the other hand, I think human hubris naturally makes us dramatically overestimate how special brains are.
UltraSane|7 months ago
sailingparrot|7 months ago
tedsanders|7 months ago
Take a look at this paper: https://scholar.harvard.edu/files/rzeckhauser/files/value_of...
They took high-precision forecasts from a forecasting tournament and rounded them to coarser buckets (nearest 5%, nearest 10%, nearest 33%), to see if the precision was actually conveying any real information. What they found is that if you rounded the forecasts of expert forecasters, Brier scores got consistently worse, suggesting that expert forecast precision at the 5% level is still conveying useful, if noisy, information. They also found that less expert forecasters took less of a hit from rounding their forecasts, which makes sense.
It's a really interesting paper, and they recommend that foreign policy analysts try to increase precision rather than retreating to lumpy buckets like "likely" or "unlikely".
Based on this, it seems totally reasonable for a rationalist to make guesses with single digit precision, and I don't think it's really worth criticizing.
c1ccccc1|7 months ago
If the variance (uncertainty) in a number is large, correct thing to do is to just also report the variance, not to round the mean to a whole number.
Also, in log odds, the difference between 5% and 10% is about the same as the difference between 40% and 60%. So using an intermediate value like 8% is less crazy than you'd think.
People writing comments in their own little forum where they happen not to use sig-figs to communicate uncertainty is probably not a sinister attempt to convince "everyone" that their predictions are somehow scientific. For one thing, I doubt most people are dumb enough to be convinced by that, even if it were the goal. For another, the expected audience for these comments was not "everyone", it was specifically people who are likely to interpret those probabilities in a Bayesian way (i.e. as subjective probabilities).
danlitt|7 months ago
ben_w|7 months ago
To add to what tedsanders wrote: there's also research that shows verbal descriptions, like those, mean wildly different things from one person to the next: https://lettersremain.com/perceptions-of-probability-and-num...
jdmoreira|7 months ago
https://en.wikipedia.org/wiki/Brier_score
also:
https://en.m.wikipedia.org/wiki/Superforecaster
mewpmewp2|7 months ago
baxtr|7 months ago
And since we’re at it: why not give confidence intervals too?
meindnoch|7 months ago
The rest of the sentence is not necessary. No, you're not the only one.
unknown|7 months ago
[deleted]
jere|7 months ago
unknown|7 months ago
[deleted]
Veedrac|7 months ago
drexlspivey|7 months ago