I believe the entire point of the ergodicity question here is "If you apply this process n times, with n approaching infinity, obviously the result may depend on what point in the n-times iterated distribution you sample, but if you choose a volume of vanishingly small measure to exclude, can you make a single concrete statement about what the process is doing without taking an expected value over the different outcomes"And the answer is yes - with probability approaching 1 as n increases (ie excluding a portion of the distribution whose measure decreases to 0), the random process matches a deterministic process which is described by "you lose 5% each round".
jakell|2 years ago
bzax|2 years ago
I asked a friend who is an econ professor "Why does this Peters guy explain this so poorly" and his response was more or less, yes, all of economics has been wondering that too since he first published his Nature Physics paper on this a decade ago.