In statistical mechanics, entropy is defined as a fundamental quantity. However, its unit, J/K, involves that of temperature, which is only subsequently defined—and defined in terms of entropy. This circularity arises with the introduction of Boltzmann's constant into the very expression of entropy. The J/K carried by the constant prevents entropy from finding a unit of its own while simultaneously obfuscating its informational nature. Following the precepts of information theory, we argue that entropy is well measured in bits and coincides with information capacity at thermodynamic equilibrium. Consequently, not only is the temperature of a system in equilibrium expressed in J/bit, but it acquires a clear meaning: It is the cost in energy to increase its information capacity by 1 bit. Viewing temperature as joules per bit uncovers the strong duality exhibited by Gibbs long ago between available capacity and free energy. It also simplifies Landauer's cost and clarifies that it is a cost of displacement, not of erasure. Replacing the kelvin with the bit as an SI unit would remove Boltzmann's constant from the seven defining constants.

Entropy is a fundamental quantity in statistical mechanics, and its SI unit is J/K. Yet, in statistical mechanical terms, what does a J/K represent? Tentative answers must confront the circularity mentioned in the abstract, where entropy is defined first yet incorporates a logically posterior temperature unit. Articulating entropy in terms of an as-yet-undefined unit demands an explanation that necessarily breaks with entropy's otherwise simple definition. It must foreshadow a logically posterior definition, like T=(S/E)1, or theorem, like equipartition. Moreover, to think in joules per kelvin introduces into our view of entropy all the parochial dependencies encompassed in the development of the Kelvin scale. For instance, this includes the aggregate behavior of water and the number 100—and with it, our physiology: the number of our fingers.

The flipside of the aforementioned circularity is that, upon defining temperature from energy and entropy, the kelvin conspicuously crops up alone, as if it were the unit of a fundamental quantity. For instance, Schroeder1 wrote: “Thanks to the factor of Boltzmann's constant in the definition of entropy, the slope S/E of a system's entropy vs. energy graph has the units of (J/K)/J=1/K. If we take the reciprocal of this slope, we get something with units of kelvins, just what we want for temperature.”

In Sec. II, we first explain how the circular inconsistency regarding the entropy units arises in statistical mechanics as a result of its historical ties to phenomenological thermodynamics. The significant role of information in thermodynamics is emphasized in Sec. III, which motivates the information-theoretic account of entropy. As explained in Sec. IV, at thermodynamic equilibrium, entropy amounts to information capacity. In Sec. V, we show how temperature arises as a bridge between energy and information capacity and how it acquires a clear meaning when formulated in units of joules per bit. Not only does it uncover the notion of available information capacity completely analogous to free energy, but it also sheds light on Landauer's erasure. We also discuss how our proposal alters the current SI.

“Careful, that's hot!” Temperature is one of the most intuitive physical concepts, as we can vividly feel it through our skin. Arguably, temperature is more intuitive than energy; and definitely more than entropy. Perhaps unsurprisingly, the historical development of the three concepts has also followed that order.

The first thermometers were built in the 17th century, by the end of which some precursor of kinetic energy, the so-called vis viva, was introduced by Leibniz.2 As Young3 coined the term energy at the beginning of the 19th century, its existence in various forms was quickly realized. Notably, in 1824, Carnot4 studied engines that could use energy of the caloric type to produce energy of the work type. Clausius5 then characterized the form of energy that is inevitably lost during a thermodynamical cycle in such engines. Only then did he introduce entropy, a monotonically increasing quantity that assured the irreversibility of some processes.

The formalization of these concepts at a macroscopic level, rooted in empirical observations, established the first theory of heat: thermodynamics.*

The first theory of heat is a theory of thermal processes relating macroscopic quantities. Its starting point is temperature. Indeed, the zeroth law entails grouping systems into equivalence classes based on thermal equilibrium. Temperature then serves as a real-valued label for these equivalence classes, endowing them with an ordering that dictates the direction of possible heat flow. Clausius' formula for entropy is ΔS=Q/T, where Q is the amount of heat entering or escaping the system and T is the absolute temperature of the system.

The second theory of heat, statistical mechanics, departs from the first by taking into consideration the underlying atomic structure. Statistical mechanics depends on microstates, and it takes their count Ω, the number of possible configurations, as more fundamental than temperature. Following this precept, Boltzmann's entropy could be, as it was in Boltzmann's writings,6 simply without constant:

Boltzmann's entropy in this raw form is a combinatorial object, a log-count, a very different quantity than the entropy of Clausius. Had lnΩ been given a life of its own, perhaps Boltzmann's entropy had eventually found its own unit—in hindsight the natural unit of information (nat).

However, it did not. Planck prefactored lnΩ with what is now known as Boltzmann's constant, kB. The constant harmonizes Boltzmann's statistical mechanical entropy with Clausius's thermodynamic entropy, both in J/K, while also assuring a statistical mechanical temperature in kelvins.

We suggest that this introduction of the constant kB into Boltzmann's entropy, as important as it was for the early development of statistical mechanics, tied the infant theory of statistical mechanics to its predecessor, thermodynamics. However, the theory of thermodynamics should be obtained as a limiting case of statistical mechanics and not be used as an overarching framework for statistical mechanics. This is because statistical mechanics supersedes thermodynamics; it claims more explanatory power. For instance, it yields fluctuation results, it accommodates quantum statistics, it explains phase transitions from first principles, and it better predicts low temperature behaviors. The chronological discovery of ideas does not, therefore, amount to their logical priority.

Crucial insights and results followed the introduction of information-processing considerations into the picture of thermodynamics. For modern reviews, see Refs. 7 and 8. A serious challenge to the second law was proposed by Maxwell9–11 with his famous demon who sorts a system via measurements and manipulations, thereby reducing its entropy. Szilard's engine12 has become the canonical system on which a demon is imagined to perform its puzzling manipulations. It consists of a one-molecule gas container that can be bisected by a partition, which thereupon acts as a bidirectional piston. After inserting the partition, the demon observes on which side the molecule is trapped and then expands adiabatically the piston in the opposite direction. This extracts kBTln2 of free energy from the molecule. After thermalization with the environment, the process can be repeated if the demon has kept his ability to measure and act on the system. Szilard suggested that the second law should be saved by the act of measurement of the demon, which, he thought, should unavoidably create kBTln2 of heat, namely, just enough to compensate. The idea to tie a thermodynamic cost to the act of measurement per se has been followed by many,13–15 and was likely stimulated by the confusion that arose with measurement in quantum theory.

In 1961, Landauer16 correctly identified information erasure—and not measurement—as the precise thermodynamically irreversible step that needs to be compensated by heat dissipation. The erasure of one bit of information must be accompanied by an amount of kBTln2 of free energy lost to the environment or to the non-information-bearing degrees of freedom of a computer. The reason for heat dissipation is purely physical: Information cannot be processed independently of real devices, or as Landauer would have it, “Information is physical.”

The fact that erasure has a thermodynamic cost does not a priori preclude measurement from also having such a cost. This concern was part of the larger misconception that held that computing devices should unavoidably involve logical irreversibility and, with it, heat dissipation. Upon demonstrating that universal computation can be done via logically reversible steps, Bennett17 paved the way for developing thermodynamically reversible models of computation, the most spectacular of which is perhaps Fredkin and Toffoli's ballistic computer.18 In this light, the status of measurement, a very special kind of computation, has been clarified:19 An apparatus initialized in a ready state can measure non-dissipatively. This yields a detailed and satisfactory resolution of Maxwell's problem: With an initialized memory, the demon can measure the system and act on it to reduce its entropy. However, this is no paradox, as it merely displaces the entropy of the system onto its memory. To operate in a cycle, the demon needs to reset the memory to its initial state, that is, to get rid of the information stored—an erasure that, by Landauer's bound, dissipates a quantity of heat greater or equal to the entropy reduction of the system times the environmental temperature.

The logical reversibility of dynamical laws applies to all physical systems, including those with the ability to store information and act based on it. Information-processing agents cannot avoid the second law. Yet, the point of view of information processing has offered limits to what can or cannot be done, thermodynamically. In particular, Landauer's erasure cost can be considered one of the many expressions of the second law, which, as eloquently stated by Schumacher,20 mandates that “No physical process has as its sole result the erasure of information.”

The resolution of Maxwell's demon has been highly influential in promoting the role of information theory in thermodynamics. The scientific literature on the topic has been booming and ramifying in many ways. The advent of Shannon's mathematical theory of communication21 generated insights22,23 (and debates24,25) on the nature of entropy. Algorithmic information theory26–28 permits a quantification of information based on individual objects, yielding more sophisticated notions of entropy29,30 and macrostates.31 The significance of quantum information was realized32,33 and incorporated into various thermodynamical analyses.34 Axiomatic reconstructions of thermodynamics have been suggested.35–37 

Energy is a fundamental physical quantity whose conservation principle has permitted significant theoretical advancements and has been proven to be repeatedly consistent with experimental tests. In a conversation about thermodynamics, we shall be satisfied with its SI unit, the joule (J), itself expressed in terms of the kilogram, the meter, and the second.

However, we expressed the problematic circularity with the SI unit of entropy as measured in J/K (Sec. I) and advocated instead for the information-theoretic view of entropy (Sec. III). Many such entropy measures have been proposed, most of which are tied to bits. For instance, Shannon entropy H(X) is the expected number of bits required to communicate the outcome of a random variable X in an optimal prefix code.21 More convoluted measures were proposed; yet, for our purpose, we step back on what is perhaps the simplest: information capacity.

The information capacity S(E) (or simply the capacity) of a system at a given energy E is quantified by
where Ω(E) is the number of possible configurations a system at energy E can adopt, and a bit is the information capacity of a 2-level system. The logarithm of a microstate count that corresponds to a given constraint is the gist of Boltzmann's entropy. By using the base 2 instead of the natural logarithm, and prefactoring it by 1bit instead of kB, we obtain a measure of information capacity in bits, corresponding to a given energy E. Information capacity is a counterfactual property of the system, namely, an assessment of all the configurations in which the system could be, and it does not depend on the system's actual configuration. Nor does it depend on any probabilistic assessment, thereby avoiding the question of subjectivity and agent dependency.

Like a distance in space measured by the number of meters that can fit in that space, the information capacity of a system is measured by the number of bits that can fit in that system. In this light, S(E) is the number of bits required to label all distinguishable states at energy E, or equivalently, the number of bits that can be encoded in the system if it is seen as a storage resource. At the macroscopic scale, physical systems have a very large number of microstates, which translates to a large value for information capacity. The divide between a two-level system and a macroscopic system can be bridged, for instance, by measuring information capacity in yottabit (Ybit), which corresponds to 1024bits. This unit assesses more conveniently the information capacity of macroscopic systems.

In modern days, it has become second nature to quantify the storage capacity of devices in bits. In a similar way, any physical system can be viewed as an information-storage system. Compared to the memory of a computing device that has been engineered to be stable in some relevant set of environments, a generic physical system has microstates which, for all practical purposes, cannot be prepared nor maintained in any chosen configuration. Despite this technological (but not fundamental) impracticality for user interplay, it remains that information can be encoded in physical systems, a quantity which is bounded by its information capacity. The assumption of thermodynamic equilibrium amounts to viewing a system's entropy as maximal and, therefore, reaching its full capacity.§

In this section, we develop the logical implication of taking entropy as fundamental and assigning it its own unit: temperature should have the units of J/bit. The information capacity S is a function of its internal energy E. The temperature is obtained as the reciprocal of the slope between information capacity and energy, T=(S/E)1, or, when the slope is well-defined, directly as E/S. Therefore, with entropy measured in bits, temperature is in J/bit.

Non-exotic systems are of positive temperature and positive heat capacity, yielding both positive first and positive second derivatives of E with respect to S, as displayed in Fig. 1(a).** In thermodynamic contexts, where large systems are concerned, 1bit is negligible compared to the capacity of the system, so the slope E/S is well approximated by a finite difference. Thus, temperature can be interpreted as the increase in internal energy (in J) required to increase the information capacity by 1bit.

Fig. 1.

Graphs of energy vs capacity. Systems usually have a concave up curve (Fig. 1(a)). A heat bath is a linear idealization (Fig. 1(b)).

Fig. 1.

Graphs of energy vs capacity. Systems usually have a concave up curve (Fig. 1(a)). A heat bath is a linear idealization (Fig. 1(b)).

Close modal

The statistical mechanical definition of temperature has many advantages, one of which is the possibility to make sense of negative temperature.38 In joules per bit, it is to be interpreted as the amount of energy that needs to be extracted from the system in order to increase its capacity by one bit. Furthermore, temperature as such is independent of an a priori notion of heat baths, which instead can be understood in terms of systems whose E(S) function is of constant slope, namely, of constant temperature, as displayed in Fig. 1(b). Such systems can be thought of as idealizations of very large systems or, as close-ups of some E(S) function, which then appears to be linear. The energy cost to increase the information capacity of a heat bath at temperature T (in J/bit) by X (in bits) is XT (in J). However, for systems that are not well approximated by a heat bath, the energy cost for an additional xbits need not be linear: The energy cost must be integrated over the interval of increased information capacity.

In 1873, Gibbs39 found out that systems whose entropy is not maximal have available energy—now known as free energy—as the system can be used to produce work. This energy is extracted from the system by transforming it into a state of lower energy, while conserving its entropy, until the capacity curve is reached. It is thus the amount of energy that can be extracted from the system with no need to store an excess of entropy. In the same breath, Gibbs presents what he calls the capacity for entropy, or as we like to view it, available capacity.†† It is the number of bits that can be encoded in the system at no extra energy cost. Available capacity quantifies the amount of structure in the system, like blanks on a tape, or the empty registers in the memory of Maxwell's demon. The view advocated here, in which energy and entropy are more fundamental than temperature, highlights well the duality captured by Gibbs between available energy and available capacity. Figure 2 illustrates these quantities in the energy vs capacity plane of an isolated system.

Fig. 2.

The point's coordinates are given by the system's entropy and energy. It represents a nonequilibrium state, as its entropy is not maximal, i.e., the point is not on the capacity curve. As an example, it could correspond to the state of an isolated gas in a box, where all the particles occupy a smaller region than that of the whole box. The vertical red dashed line represents the available energy while the horizontal blue dashed line represents the available capacity.

Fig. 2.

The point's coordinates are given by the system's entropy and energy. It represents a nonequilibrium state, as its entropy is not maximal, i.e., the point is not on the capacity curve. As an example, it could correspond to the state of an isolated gas in a box, where all the particles occupy a smaller region than that of the whole box. The vertical red dashed line represents the available energy while the horizontal blue dashed line represents the available capacity.

Close modal

Moreover, that duality is also captured by idealized systems that are efficient to store either energy or entropy. On one end, a battery is a system designed to keep the energy available; namely, the internal energy of the system can be changed with no significant changes in entropy. As an example, consider a weight in a gravitational field. The (potential) energy of the weight can be easily changed without affecting the information that is encoded in it. When changing the position of the weight, except for the energy, all the intrinsic properties of the system are unchanged. On the other end, some systems are efficient for storing information (or entropy) with no energy change—we call them tapes. Degenerate ground states and ideal hard disk drives are instances of those tapes.

Be it classical or quantum, information processing follows the logical reversibility of the physical laws of motion. As a consequence, the information (i.e., entropy) of an isolated system does not decrease spontaneously. When considering the entire universe as one isolated system, the conclusion is straightforward: “No information is ever lost.” This statement has been recognized as a formulation of the second law of thermodynamics.42,43

The term “information erasure” is, therefore, a misnomer: Information is never erased as if free energy had the power of fundamental erasure. Rather, information is displaced. When information leaves the relevant degrees of freedom of an information-storage device, it moves to nearby systems, its environment. The environment, therefore, should be seen as a memory which is itself described by an energy vs information capacity curve of the sort shown in Fig. 1. Landauer's principle then reflects a clear application of temperature as joules per bit: To accommodate one additional bit of information, the environment's capacity must be expanded via a precise quantity of energy. By definition, this quantity is the temperature T of the environment times one bit. No unnecessary kB. No unnecessary ln2.

Relatedly, Schumacher advocated20 that environmental temperature and erasure cost are interchangeable. Namely, one can define temperature as the erasure cost.

Adopting the primacy of the bit over the kelvin has implications. Taken seriously, our proposal leads to the elimination of one of the seven defining constants of the SI, namely, kB.

Entropy can be measured in bits, it can be measured in nats, or, as it has been done historically, it can be measured in J/K. A system with an entropy of 1nat has lnΩ=1 and, therefore, that same system also has entropy kB. Thus,
and the above-mentioned expressions relate to the bit via 1bit=ln2nat. Alternatively, the constant kB also plays the role of a conversion factor between the Kelvin scale and the J/bit scale,

Seen in this light, Boltzmann's constant is not a fundamental constant, but a conversion factor between those different standards.‡‡ As pointed out by Callen:44 “The constant prefactor [in Boltzmann's entropy] merely determines the scale of [Boltzmann's entropy]; it is chosen to obtain agreement with the Kelvin scale of temperature.” Therefore, Boltzmann's constant embodies the same parochial dependencies as those of the kelvin.

Should the SI adopt the bit instead of the kelvin, the other base units would be left unchanged, as can be readily seen by inspecting Fig. 3. Indeed, no other SI unit depends on the current definition of the kelvin. Moreover, the connection between entropy, energy, and temperature is made obvious in the diagram of Fig. 3(b): Temperature makes the bridge between energy (in J, defined by the arrows coming from the meter, the second, and the kilogram) and entropy (in bits).

Fig. 3.

The result of altering the International System of Units (SI) to recognize the primacy of the bit and, with it, temperature as J/bit. Figure 3(a) displays the current defining constants and SI units, and their relationships to one another. Figure 3(b) displays our modification. While Boltzmann's constant kB has been removed, and the Kelvin replaced, no other SI constant or unit is affected by the modification. Note that the units of temperature are no longer a SI base unit. Those of entropy are.

Fig. 3.

The result of altering the International System of Units (SI) to recognize the primacy of the bit and, with it, temperature as J/bit. Figure 3(a) displays the current defining constants and SI units, and their relationships to one another. Figure 3(b) displays our modification. While Boltzmann's constant kB has been removed, and the Kelvin replaced, no other SI constant or unit is affected by the modification. Note that the units of temperature are no longer a SI base unit. Those of entropy are.

Close modal

Obviously, we do not expect the J/bit to feature in everyday discussions about temperature. After all, even if the kelvin is the scientific unit, the common usage of the Fahrenheit and the Celsius scales is still widespread. Still, we could describe our preferred pool temperature (80°F) as 2 869J/Ybit. As another example, let us consider the entropy of one mole of helium at room temperature and atmospheric pressure. As computed from the Sackur–Tetrode equation (Ref. 1, Chapter II), its value is 126  J/K. This corresponds to approximately 13.17Ybit.

By taking seriously the well-accepted logical priority of entropy over temperature in statistical mechanics, we suggest that entropy ought to find a unit of its own, the bit. This offers an interdisciplinary interpretation of thermodynamics centered on the duality between energy and information. Temperature arises as the nexus between them, linking joules to bits—and hence measured in J/bit. When a system's entropy is maximal, its information capacity is a useful proxy for its entropy, as they coincide. The system's temperature is then given by the energy cost to increase the information capacity by one bit. When the system's entropy is not maximal, it has available energy, or available capacity, like batteries and tapes do. Viewing the environment as a storehouse of information explains Landauer's “displacement”; namely, since a bit is never erased, the environment's capacity needs to be increased by one bit to accommodate it, which is achieved at the energy cost (in joules) of T×1 bit, where T is the temperature of the environment in joules per bit.

The ideas in this paper were developed under the assumption of classical physical laws. However, this limitation does not detract from the significance of our proposal. We conjecture that the interpretation of temperature as energy per unit of information remains valid, regardless of whether the laws of physics are quantum or of any other form yet to be discovered.

Like energy, which takes very different forms across the many domains of physics, information might also be characterized in different ways. Investigating them and interlinking information-theoretic notions with many of the traditional concepts of thermodynamics shall be fruitful.

The authors are grateful to Charles H. Bennett, Gilles Brassard, David Deutsch, Paul Erker, Hlér Kristjánsson, Richard MacKenzie, Bryan W. Roberts, Tommaso Toffoli, and Maria Violaris for fruitful discussions and comments on earlier versions of this work. The authors also thank the Jude the Obscure Pub for its enlightening atmosphere. Additionally, the authors acknowledge Gilles Brassard for generously enabling the publication of this work in open access. Figure 3 was generated from an adaptation of the Mathematica notebook of Emilio Pisanty (dated 2016–2018); made available under the CC BY-SA 4.0 license. X.C.-R. thanks the Fonds National Suisse (FNS) for financial support through the Postdoc Mobility fellowship program. X.C.-R. acknowledges funding from the BMW endowment fund. C.A.B.'s work is supported by the Fonds de recherche du Québec—Nature et technologie, the FNS, and the Hasler Foundation. S.W. acknowledges support from FNS through Project No. 214808.

The authors have no conflicts to disclose.

*

While in this work we use the term “thermodynamics” in its general sense encompassing also statistical mechanics, the present section assumes “thermodynamics” to specifically denote the phenomenological and macroscopic framework that remains agnostic to any underlying microscopic physics, i.e., not statistical mechanics.

A classic interpretation of Ω is the cardinality of a macrostate, which here is solely determined by the system's energy.

A valid defense of the natural logarithm is its convenience when the machinery of differential calculus is used to elaborate the theory. While true, this is easily resolved by the conversion log2Ω(E)bits=lnΩ(E)nat.

§

The fact that a system at maximal entropy saturates its capacity can be easily recognized when quantifying entropy in the probabilistic setting: In this case, the Gibbs–Shannon entropy S(ρ)=i=1Ω(E)pilog2pi of a distribution ρ over a configuration space of Ω(E) possible values is maximal for the uniform distribution, and is then equal to the capacity.

**

In most contexts, S(E) can be inverted into a function E(S). We opted for the graph of the latter, so as to have the slope directly equal to the temperature (instead of the inverse temperature).

††

Available capacity is also known by the names negative entropy40 and negentropy.41 

‡‡

It is often said that the Boltzmann's constant links the average kinetic energy with the temperature, but this contingency in the equipartition theorem is explained by the exponential relation between S and E that arises when we consider degrees of freedom that are quadratic in the energy.

1.
Daniel V.
Schroeder
,
An Introduction to Thermal Physics
(
Addison Wesley
,
2000
).
2.
Gottfried W.
Leibniz
, “
Specimen dynamicum pro admirandis Naturae legibus circa corporum vires et mutuas actiones detegendis, et ad suas causas revocandis
,”
Acta Eruditorum
4
,
145
157
(
1695
).
3.
Thomas
Young
,
A Course of Lectures on Natural Philosophy and the Mechanical Arts: in Two Volumes
(
Johnson
,
1807
), Vol.
2
.
4.
Sadi
Carnot
,
Réflexions Sur la Puissance Motrice du Feu et Sur Les Machines Propres à Développer Cette Puissance
(
Bachelier
,
Paris
,
1824
).
5.
Rudolf
Clausius
, “
Über die bewegende Kraft der Wärme und die Gesetze, welche sich daraus für die Wärmelehre selbst ableiten lassen
,”
Ann. Phys.
155
(
3
),
368
397
(
1850
).
6.
Ankita
Anirban
, “
A constant by any other name
,”
Nat. Phys.
17
(
5
),
660
660
(
2021
).
7.
Koji
Maruyama
,
Franco
Nori
, and
Vlatko
Vedral
, “
Colloquium: The physics of Maxwell's demon and information
,”
Rev. Mod. Phys.
81
(
1
),
1
23
(
2009
).
8.
Juan M. R.
Parrondo
,
Jordan M.
Horowitz
, and
Takahiro
Sagawa
, “
Thermodynamics of information
,”
Nat. Phys.
11
(
2
),
131
139
(
2015
).
9.
James C.
Maxwell
,
Theory of Heat
(
Cambridge U. P.
,
1871
).
10.
Harvey S.
Leff
and
Andrew F.
Rex
,
Maxwell's Demon: Entropy, Information, Computing
(
Princeton U. P.
,
1990
).
11.
Harvey
Leff
and
Andrew F.
Rex
,
Maxwell's Demon 2 Entropy, Classical and Quantum Information, Computing
(
CRC Press
,
2002
).
12.
Leo
Szilard
, “
Über die Entropieverminderung in einem thermodynamischen system bei Eingriffen intelligenter Wesen
,”
Z Phys.
53
(
11–12
),
840
856
(
1929
).
13.
John
Von Neumann
,
Mathematical Foundations of Quantum Mechanics
(
Princeton U. P.
,
1955
).
14.
Denis
Gabor
IV
, “
Light and information
,” in
Progress in Optics
(
Elsevier
,
1961
), Vol.
1
, pp.
109
153
.
15.
Leon
Brillouin
,
Science and Information Theory
, 2nd ed. (
Academic Press
,
1962
).
16.
Rolf
Landauer
, “
Irreversibility and heat generation in the computing process
,”
IBM J. Res. Dev.
5
(
3
),
183
191
(
1961
).
17.
Charles H.
Bennett
, “
Logical reversibility of computation
,”
IBM J. Res. Dev.
17
(
6
),
525
532
(
1973
).
18.
Edward
Fredkin
and
Tommaso
Toffoli
, “
Conservative logic
,”
Int. J. Theor. Phys.
21
(
3–4
),
219
253
(
1982
).
19.
Charles H.
Bennett
, “
The thermodynamics of computation—A review
,”
Int. J. Theor. Phys.
21
(
12
),
905
940
(
1982
).
20.
Benjamin
Schumacher
, in
Quantum Information and Foundations of Thermodynamics
(
ETH Zürich
,
2011
) (accessed November 11, 2023). <https://www.video.ethz.ch/conferences/2011/qiftw11/f9bb6f3d-7461-4971-8213-2e3ceb48c0ba.html >
21.
Claude E.
Shannon
, “
A mathematical theory of communication
,”
Bell Syst. Tech. J.
27
(
3
),
379
423
(
1948
).
22.
Edwin T.
Jaynes
, “
Information theory and statistical mechanics
,”
Phys. Rev.
106
(
4
),
620
630
(
1957
).
23.
Edwin T.
Jaynes
, “
Information theory and statistical mechanics. II
,”
Phys. Rev.
108
(
2
),
171
190
(
1957
).
24.
Abner
Shimony
, “
The status of the principle of maximum entropy
,”
Synthese
63
(
1
),
35
53
(
1985
).
25.
Tommaso
Toffoli
, “
Entropy? Honest!
,”
Entropy
18
(
7
),
247
(
2016
).
26.
Ray J.
Solomonoff
, “
A formal theory of inductive inference. Part I
,”
Inf. Control
7
(
1
),
1
22
(
1964
).
27.
Andreï N.
Kolmogorov
, “
Three approaches to the quantitative definition of information
,”
Probl. Peredachi Inf.
1
(
1
),
3
11
(
1965
).
28.
Gregory J.
Chaitin
, “
On the length of programs for computing finite binary sequences
,”
J. ACM
13
(
4
),
547
569
(
1966
).
29.
Wojciech H.
Zurek
, “
Algorithmic randomness and physical entropy
,”
Phys. Rev. A
40
(
8
),
4731
4751
(
1989
).
30.
Murray
Gell-Mann
and
Seth
Lloyd
, “
Effective complexity
,” in
Murray Gell-Mann: Selected Papers
(
World Scientific
,
2010
), pp.
391
402
.
31.
Ämin
Baumeler
and
Stefan
Wolf
, “
Causality–complexity–consistency: Can space-time be based on logic and computation?
,” in
Time in Physics
(
Springer
,
2017
), pp.
69
101
.
32.
Alexander S.
Holevo
, “
Some estimates for the amount of information transmittable by a quantum communication channel (in Russian)
,”
Probl. Predachi Inf.
9
,
3
11
(
1973
).
33.
Vlatko
Vedral
,
Introduction to Quantum Information Science.
(
Oxford U. P.
,
2006
).
34.
John
Goold
,
Marcus
Huber
,
Arnau
Riera
,
Lídia
Del Rio
, and
Paul
Skrzypczyk
, “
The role of quantum information in thermodynamics—a topical review
,”
J. Phys. A: Math. Theor.
49
(
14
),
143001
(
2016
).
35.
Elliott H.
Lieb
and
Jakob
Yngvason
, “
The physics and mathematics of the second law of thermodynamics
,”
Phys. Rep.
310
(
1
),
1
96
(
1999
).
36.
Chiara
Marletto
, “
Constructor theory of thermodynamics
,” arXiv:1608.02625 (
2016
).
37.
Austin
Hulse
,
Benjamin
Schumacher
, and
Michael D.
Westmoreland
, “
Axiomatic information thermodynamics
,”
Entropy
20
(
4
),
237
(
2018
).
38.
Ämin
Baumeler
,
Carla
Rieger
, and
Stefan
Wolf
, “
Thermodynamics as combinatorics: A toy theory
,” in
IEEE Information Theory Workshop (ITW)
(
IEEE
,
2022
), pp.
362
367
.
39.
Josiah Willard
Gibbs
, “
A method of geometrical representation of the thermodynamic properties by means of surfaces
,”
The Collected Works of J. Willard Gibbs
(Yale U. P.,
1957
), pp.
33
54
.
40.
Erwin
Schrödinger
,
What is Life?: The Physical Aspect of the Living Cell
(
Cambridge U. P.
,
1944
).
41.
Leon
Brillouin
, “
The negentropy principle of information
,”
J. Appl. Phys.
24
(
9
),
1152
1163
(
1953
).
42.
Stefan
Wolf
, “
Second thoughts on the second law
,” in
Adventures Between Lower Bounds and Higher Altitudes
(
Springer
,
2018
), pp.
463
476
.
43.
Ämin
Baumeler
and
Stefan
Wolf
, “
Free energy of a general computation
,”
Phys. Rev. E
100
(
5
),
052115
(
2019
).
44.
Herbert B.
Callen
,
Thermodynamics and an Introduction to Thermostatistics
(
John Wiley & Sons
,
1991
).