From the linked Andrew Molitor item:
Why Extropic insists on talking about thermodynamics at all is a mystery, especially since “thermodynamic computing” is an established term that means something quite different from what Extropic is trying to do. This is one of several red flags.
I have a feeling this is related to wanking about physics in the e/acc holy gospels. They invoke thermodynamics the way that people trying to sell you healing crystals for your chakras invoke quantum mechanics.
The existence of a Wikipedia page for dinosaur erotica must prove that back in the days when humans co-existed with stegosaurs, the ones who fucked them lived better.
Erin go Bleagh.
"Consider it from the perspective of someone who does not exist and therefore has no preferences. Who would they pick?"
The idea that formalist experimentation and deliberately pushing the boundaries of a medium are only one of several goals to which art can strive is apparently too sophisticated for Scott Adderall. He also takes a leap from "influential" to "meaningful", an elision so hackish it's trite.
It's a very "steampunk (derogatory)" picture, like something I would have found on sale in the artist room of the science-fiction convention that convinced me I don't like science-fiction conventions very much.
I think that in this particular instance, it's OK to kinkshame
"So, professor sir, are you OK with psychologically torturing Black people, or do you just not care?"
Superficially, it looks like he's making a testable prediction. But that "prediction" is a number from a bullshit calculation (or maybe two or three different, mutually inconsistent bullshit calculations — it's hard to be sure). So if someone wasted their time and did the experiment, he'd handwave away the null result by fiddling the input bullshit.
I will try to have some more comments about the physics when I have time and energy. In the meanwhile:
Entropy in thermodynamics is not actually a hard concept. It's the ratio of the size of a heat flow to the temperature at which that flow is happening. (So, joules per kelvin, if you're using SI units.) See episodes 46 and 47 of The Mechanical Universe for the old-school PBS treatment of the story. The last time I taught thermodynamics for undergraduates, we used Finn's Thermal Physics, for the sophisticated reason that the previous professor used Finn's Thermal Physics.
Entropy in information theory is also not actually that hard of a concept. It's a numerical measure of how spread-out a probability distribution is.
It's relating the two meanings that is tricky and subtle. The big picture is something like this: A microstate is a complete specification of the positions and momenta of all the pieces of a system. We can consider a probability distribution over all the possible microstates, and then do information theory to that. This bridges the two definitions, if we are very careful about it. One thing that trips people up (particularly if they got poisoned by pop-science oversimplifications about "disorder" first) is forgetting the momentum part. We have to consider probabilities, not just for where the pieces are, but also for how they are moving. I suspect that this is among Vopson's many problems. Either he doesn't get it, or he's not capable of writing clearly enough to explain it.
So these two were published in American Institute of Physics Advances, which looks like a serious journal about physics. Does anyone know about it? It occupies a space where I can’t easily find any obvious issues, but I also can’t find anyone saying “ye this is legit”. It claims to be peer-reviewed, and at least isn’t just a place where you dump a PDF and get a DOI in return.
I have never heard of anything important being published there. I think it's the kind of journal where one submits a paper after it has been rejected by one's first and second (and possibly third) choices.
However, after skimming, I can at least say that it doesn’t seem outlandish?
Oh, it's worse than "outlandish". It's nonsensical. He's basically operating at a level of "there's an E in this formula and an E in this other formula, so I will set them equal and declare it revolutionary new physics".
Here's a passage from the second paragraph of the 2023 paper:
The physical entropy of a given system is a measure of all its possible physical microstates compatible with the macrostate, S~Phys~. This is a characteristic of the non-information bearing microstates within the system. Assuming the same system, and assuming that one is able to create N information states within the same physical system (for example, by writing digital bits in it), the effect of creating a number of N information states is to form N additional information microstates superimposed onto the existing physical microstates. These additional microstates are information bearing states, and the additional entropy associated with them is called the entropy of information, S~Info~. We can now define the total entropy of the system as the sum of the initial physical entropy and the newly created entropy of information, S~tot~ = S~Phys~ + S~Info~, showing that the information creation increases the entropy of a given system.
wat
Storing a message in a system doesn't make new microstates. How could it? You're just rearranging the pieces to spell out a message — selecting those microstates that are consistent with that message. Choosing from a list of available options doesn't magically add new options to the list.