I did a deep dive on entropy a couple years ago. I found the concept to be much harder to understand than I expected! Specifically, it was confusing to shift from the intuitive but wrong “entropy is disorder” to “entropy is about the number of possible microstates in a macrostate” (Boltzmann Entropy) https://en.wikipedia.org/wiki/Boltzmann%27s_entropy_formula
I was extra confused when I discovered that a spread out cloud of hydrogen is lower entropy than the same cloud gravitationally bound together in a star. So entropy isn’t just about “spreading out,” either.
I found that Legos provide a really nice example to illustrate entropy, so I’ll share that here.
Consider a big pile of Legos, the detritus of many past projects. Intuitively, a pile of Legos is high entropy because it is disordered—but if we are trying to move beyond order/disorder, we need to relate it to micro states and macro states.
Therefore, a pile of Legos is high entropy because if you randomly swap positions of the pieces it will all be the same macrostate—ie a big pile of Legos. Nevertheless, each of the Lego pieces is still in a very specific position— and if we could clearly snapshot all those positions, that would be the specific microstate. That means that the macrostate of the pile has an astronomical number of possible microstates — there are many ways to reorganize the pieces that still look like a pile.
On the other hand, consider a freshly built Lego Death Star. This is clearly low entropy. But to understand why in terms of microstates, it is because very few Legos can be swapped or moved without it not really being a Death Star anymore. The low entropy is because there are very few microstates (specific Lego positions) that correspond to the given macro state (being a Death Star).
This specific case helped me grok Boltzmann entropy. To extend it, consider a box with a small ice crystal in it: this has many fewer possible microstates than the same box filled with steam. In the steam, molecules can pretty much we swapped and moved anywhere and the macrostate is the same. With the crystal, if you start randomly swapping molecules to different microstates, it stops being an ice crystal quickly. So an ice crystal is low entropy.
Now, the definition of what counts as a macrostate is very important in this… but this comment is long enough and I still haven’t gotten to the gym…
I appreciate the explanation, but the very first example doesn't sit well with me. Water forming into ice cubes spontaneously looks weird simply because we’re not used to seeing it. Consider a time-lapse of an icicle forming as a sort of counter-example: https://m.youtube.com/watch?v=mmHQft7-iSU
(Not refuting entropy as the order of time at all, just noting a visual example is not great evidence.)
Ah, but the icicle isn’t really equivalent to water undripping and refreezing back into a nice cube-shaped object, at presumed room temperature (since our point of reference is the first gif). That would be weird to see IRL always, no matter what. You could watch that gif a million times and you’d still shit your pants if that happened to the glass of water on your desk.
So if I get this right, there is an infinitely small possibility that a cracked egg returns to its initial state.
Imagine that happening and being put on video. We'd all believe we're living in a simulation and witnessed a glitch.
No-one would believe the scientists explaining that although highly improbable, the uncracked egg does make scientific sense.
It's not really whether it makes scientific sense or not, it's just that it's so very highly improbable (really, really improbable) that other explanations make more sense: the video's a fake, it's mass hysteria, or even that we're living in a simulation.
I think your question mainly demonstrates how much trouble most people have about reasoning about exponentials. Not intending any personal insult, but unless you do it regularly (ie, probably are a physicist), you will use a term like "highly improbable" when referring to quantities that can only be expressed in scientific notation.
In other words, most humans have bad intuition about large numbers. And I'm not talking about "small" large numbers like "how many Teslas could Elon buy". I mean "how many atoms are in a chicken egg" (and what are their statistical properties at room temperature)
Now imagine that this was in gradeschool curricula!
I really think education is mostly about providing higher-level intuitions - making correct thought habitual and thus easy.
Part of what's so attractive about this particular article is how it would mesh with related fields (chemistry, statistics, politics, evolution, astophysics, climate science, etc)
> entropy is just a fancy word for ‘number of possible arrangements’
It isn’t though.
Entropy is a fancy word for potential distribution over negative potential. Negative potential is the “surface area” over which potential may distribute. The “number of possible arrangements” casually fits into this, yet misses some unintuitive possibilities, like the resistive variance or other characteristics not preempted by who ever constructed the intellectual model.
Idealists insist entropy is a scalar state resolve of delta probability in their model. They are self deceived. Entropy is the existential tendency for potential to distribute toward equilibrium.
As long as boffins can throw away results that do not correlate, they can insist it is anything they like.
>Entropy is a fancy word for potential distribution over negative potential. Negative potential is the “surface area” over which potential may distribute.
Maro|1 year ago
Almost all of them have Python code to illustrate concepts.
-
1. Entropy of a fair coin toss - https://bytepawn.com/what-is-the-entropy-of-a-fair-coin-toss...
2. Cross entropy, joint entropy, conditional entropy and relative entropy - https://bytepawn.com/cross-entropy-joint-entropy-conditional...
3. Entropy in Data Science - https://bytepawn.com/entropy-in-data-science.html
4. Entropy of a [monoatomic] ideal gas with coarse-graining - https://bytepawn.com/entropy-of-an-ideal-gas-with-coarse-gra...
5. All entropy related posts - https://bytepawn.com/tag/entropy.html
dr_dshiv|1 year ago
dr_dshiv|1 year ago
I was extra confused when I discovered that a spread out cloud of hydrogen is lower entropy than the same cloud gravitationally bound together in a star. So entropy isn’t just about “spreading out,” either.
I found that Legos provide a really nice example to illustrate entropy, so I’ll share that here.
Consider a big pile of Legos, the detritus of many past projects. Intuitively, a pile of Legos is high entropy because it is disordered—but if we are trying to move beyond order/disorder, we need to relate it to micro states and macro states.
Therefore, a pile of Legos is high entropy because if you randomly swap positions of the pieces it will all be the same macrostate—ie a big pile of Legos. Nevertheless, each of the Lego pieces is still in a very specific position— and if we could clearly snapshot all those positions, that would be the specific microstate. That means that the macrostate of the pile has an astronomical number of possible microstates — there are many ways to reorganize the pieces that still look like a pile.
On the other hand, consider a freshly built Lego Death Star. This is clearly low entropy. But to understand why in terms of microstates, it is because very few Legos can be swapped or moved without it not really being a Death Star anymore. The low entropy is because there are very few microstates (specific Lego positions) that correspond to the given macro state (being a Death Star).
This specific case helped me grok Boltzmann entropy. To extend it, consider a box with a small ice crystal in it: this has many fewer possible microstates than the same box filled with steam. In the steam, molecules can pretty much we swapped and moved anywhere and the macrostate is the same. With the crystal, if you start randomly swapping molecules to different microstates, it stops being an ice crystal quickly. So an ice crystal is low entropy.
Now, the definition of what counts as a macrostate is very important in this… but this comment is long enough and I still haven’t gotten to the gym…
kitd|1 year ago
kaonwarb|1 year ago
(Not refuting entropy as the order of time at all, just noting a visual example is not great evidence.)
throwuxiytayq|1 year ago
xtiansimon|1 year ago
“The most misunderstood concept in physics", by Veritasium (YouTube, 2023) (https://youtu.be/DxL2HoqLbyA?si=5a_4lCnuv85lRb57)
LaundroMat|1 year ago
No-one would believe the scientists explaining that although highly improbable, the uncracked egg does make scientific sense.
speakeron|1 year ago
markhahn|1 year ago
In other words, most humans have bad intuition about large numbers. And I'm not talking about "small" large numbers like "how many Teslas could Elon buy". I mean "how many atoms are in a chicken egg" (and what are their statistical properties at room temperature)
markhahn|1 year ago
I really think education is mostly about providing higher-level intuitions - making correct thought habitual and thus easy.
Part of what's so attractive about this particular article is how it would mesh with related fields (chemistry, statistics, politics, evolution, astophysics, climate science, etc)
NewsaHackO|1 year ago
ruthmarx|1 year ago
If you want to reference a relevant sci-fi, I'd say Asimov's The Last Question is a better fit.
xtrapol8|1 year ago
It isn’t though.
Entropy is a fancy word for potential distribution over negative potential. Negative potential is the “surface area” over which potential may distribute. The “number of possible arrangements” casually fits into this, yet misses some unintuitive possibilities, like the resistive variance or other characteristics not preempted by who ever constructed the intellectual model.
Idealists insist entropy is a scalar state resolve of delta probability in their model. They are self deceived. Entropy is the existential tendency for potential to distribute toward equilibrium.
As long as boffins can throw away results that do not correlate, they can insist it is anything they like.
ninetyninenine|1 year ago
I don't understand this. Please elucidate.