fela | 6 years ago | on: Statisticians want to abandon science’s standard measure of ‘significance’
fela's comments
fela | 6 years ago | on: I've stopped flying to conferences for climate change reasons
fela | 7 years ago | on: U.S. to ground Boeing 737 Max 8
That is different then stating the probability of it being as safe as the average airplane, which you can't do as easily without additional modelling/priors and bayesian statistics.
fela | 7 years ago | on: Bayes’ Theorem in the 21st Century (2013) [pdf]
P(Hypothesis|Data) = P(Hypothesis) * evidence_factor
P(Hypothesis) is the prior probability of the Hypothesis being true, in other words the probability we gave to the Hypothesis before seeing any of the data we are using in the theorem. When new data is observed, we use Bayes' theorem to update our believe in the hypothesis, which in practice means multiplying our prior probability by a number that depends on how well the new data fits our hypothesis. More precisely:
evidence_factor = P(Data|Hypothesis)/P(Data)
So it is the ratio of how likely our data is if our hypothesis is true, compared to (divided by) how likely it is in general. If it is more likely to occur in our Hypothesis, our probability of it being true increases, if it is more likely in general (and thus also more likely in case our hypothesis is not true, you can prove mathematically that those two statements are the same), then our believe in the hypothesis decreases.
TLDR: Prob(Hypothesis after I have seen new data) = Prob(Hypothesis before I saw the new data) * (how likely I am to see the data if my hypothesis is true, compared to in general)
fela | 7 years ago | on: Carlo Rovelli on the ‘greatest remaining mystery’: The nature of time
fela | 8 years ago | on: How Did Anyone Do Math in Roman Numerals?
fela | 9 years ago | on: The Whale
fela | 9 years ago | on: Ask HN: Do you still use browser bookmarks?
1. Autocompletion: for any website I use regularly I just write a substring of the url or Title (Firefox does this especially well). This covers probably 70% of my browsing.
2. Google. This might take slightly longer in case I want to find a specific article I had read some time ago, but it still seems less effort that having to bother with bookmarks, in my experience: either you have a very long list of unsorted bookmarks, in witch it's hard to search, or you have to spend time sorting them into sub-folders.
Now that I think of it, the following would be a very useful Google feature: +1 an url so that it becomes much more likely to bubble to the top in future searches.
fela | 9 years ago | on: Robot Baristas Serve Up the Future of Coffee at Cafe X
fela | 9 years ago | on: Statistical Mistakes and How to Avoid Them
fela | 9 years ago | on: Statistical Mistakes and How to Avoid Them
This is wrong. It’s telling you that there’s at most an alpha chance that a difference like that (or more) would have arisen from random chance if the quantities are actually equal. And if the quantities are equal 95 out of 100 parallel universes would not be able to reject the null hypothesis.
Is he saying that he would take the xkcd bet[0] on the frequentist side?
fela | 9 years ago | on: Statistical Mistakes and How to Avoid Them
If you say it like this it will very easily be misinterpreted. Once your results are in there are two cases: (1) either the null hypothesis is true and you got those results due to chance, or (2) the null hypothesis is false and there was some actual effect outside of the null hypothesis that helped you get the results.
Due to this it is very easy to interpret you statement as referring to the probability of (1).
Two two following definitions of p-values sound similar but are not:
[Correct] The probability of getting the results by chance if the null hypothesis is true P(Results|H0)
[Wrong] The probability that you got the results by chance and thus the null hypothesis was actually true P(H0|Results)
I'm not saying you didn't get it, but somebody reading what you wrote can very easily be fooled. And there are a lot of dead wrong definitions on the web[0][1][2][3].
[0] https://www.americannursetoday.com/the-p-value-what-it-reall...
[1] https://practice.sph.umich.edu/micphp/epicentral/p_value.php
[2] http://natajournals.org/doi/full/10.4085/1062-6050-51.1.04
[3] http://www.cdc.gov/des/consumers/research/understanding_scie...
fela | 9 years ago | on: Statistical Mistakes and How to Avoid Them
It's the odds of having that results due to chance, if the null hypothesis is true[0]. That latter part might sound pedantic, but the whole point is that we don't know how likely the null hypothesis is. If I test wheather the sun has just died[1] and get a p-value of 0.01 it's still very likely that this result is due to change (surely more than 1%)! We need a prior probability (i.e. bayesian statistics) to calculate the probability that the result was due to chance, that is why that partial definition is incomplete and actually very misleading. This point is subtle, but very important to really understand p-values.
Another way to look at it is: if we knew the probability that the result was due to chance we could also just take 1-p and have to probability of there actually being some effect, a probability that hypothesis testing cannot give us.
There is one nice property that hypothesis testing does have (and why presumably it's so widely used): if the idea you are testing is wrong (which actually means "null hypothesis true") you will most likely (1-p) not find any positive results. This is good, this means that if the sun in fact did not die, and use 0.01 as your threshold, 99% of the experiments will conclude that there is no reason to believe the sun has died. So hypothesis testing does limit the number of false positive findings. The xkcd comic is a bit misleading it this regard, yes it does highlight the limitations of frequentist hypothesis testing, but the scenario depicted is a very unlikely one, in 99% of the cases there would have been a boring and reasonable "No, the sun hasn't died".
For an incredibly interesting article about the difficulty of concluding anything definitive from scientific results I highly recommend "The Control Group is out of Control" at slatestarcodex[2].
[0] To be even more pedantic you would have to add "equal or more extreme", and "under a given model", but "if the null hypothesis is true" is by far the most important piece often missing.
[2] http://slatestarcodex.com/2014/04/28/the-control-group-is-ou...
fela | 9 years ago | on: Moral Machine
fela | 9 years ago | on: Moral Machine
fela | 9 years ago | on: Moral Machine
And as autonomous vehicles will have to make decisions that have moral implications, they better do so in a way that humans will be happy with. I think this is an important area of research. This won't mean a machines will have morals of his own, whatever that means, but that they should do what (most?) humans would consider morally right. And what do humans consider morally right? Well that is exactly what we should try to find out.
fela | 9 years ago | on: Moral Machine
fela | 9 years ago | on: Moral Machine
The hope must be that if people consistently prefer saving the life of young people in this made up scenario they will have similar preferences in a more realistic scenario. Of course weather such a generalization holds will have to be confirmed by further studies. But this seems like a good first step to explore moral decisions more.
fela | 9 years ago | on: Meet DevBot, a self-driving electric racing car
http://www.theverge.com/2016/8/22/12592938/roborace-self-dri...
fela | 9 years ago | on: “I Want to Know What Code Is Running Inside My Body”
It is much harder to formalise how hard it is for an attacker to find out what algorithm you use, so it is risky relying too much on him not being able to do so.