The information capacity of the universe has been a topic of great debate since the 1970s and continues to stimulate multiple branches of physics research. Here, we used Shannon’s information theory to estimate the amount of encoded information in all the visible matter in the universe. We achieved this by deriving a detailed formula estimating the total number of particles in the observable universe, known as the Eddington number, and by estimating the amount of information stored by each particle about itself. We determined that each particle in the observable universe contains 1.509 bits of information and there are ∼6 × 1080 bits of information stored in all the matter particles of the observable universe.

The issue of information content associated with the physical measurement, the matter and, by extension, the information content of the universe, has been a topic of scientific debate since the late 1920s. Szilard analyzed the relationship of information to physical processes in 1929,1 demonstrating that information about a system dictates its possible ways of evolution and behavior and offering an elegant solution to Maxwell’s famous paradox, known as Maxwell’s demon.2 

With the emergence of digital computers, digital technologies, and digital data storage, the topic of information physics entered a new era. In 1961, Landauer first proposed a link between thermodynamics and information, demonstrating that digital information is not just an abstract mathematical entity, but it is in fact physical.3 Later, Landauer speculated that the universe could be a giant computer simulating itself,4 an idea further developed in more recent studies by Lloyd.5–7 Wheeler considered the universe made up of particles, fields, and information, proposing that everything emanates from the information inherent within it, i.e., “It from Bit.”8 Landauer’s principle and the physical nature of information has been recently demonstrated,9–12 and it also stimulated several new studies and novel concepts including the black hole information paradox and the holographic principle,13 the emergent gravity,14,15 and the quantized inertia.16,17 Other studies extrapolated the physical nature of information to estimate the mass of information,18–21 culminating with the recent publication of the mass-energy-information equivalence principle,22 and Vopson postulated that “information” is the fifth form of dominant matter in the universe along solid, liquid, gas, and plasma.23,24

These radical theories are based on the principle that information is physical, the information is registered by physical systems, and all physical systems can register information.25 Accordingly, there is a given amount of information stored in the universe, regardless whether it is observed or not. The proposed existence of this information imposes some fundamental questions about it: “Why is there information stored in the universe and where is it?” and “How much information is stored in the universe?” Let us deal with these questions in detail.

To answer the first question, let us imagine an observer tracking and analyzing a random elementary particle. Let us assume that this particle is a free electron moving in the vacuum of space, but the observer has no prior knowledge of the particle and its properties. Upon tracking the particle and commencing the studies, the observer will determine, via meticulous measurements, that the particle has a mass of 9.109 × 10–31 kg, charge of −1.602 × 10–19 C, and a spin of 1/2. If the examined particle was already known or theoretically predicted, then the observer would be able to match its properties to an electron, in this case, and to confirm that what was observed/detected was indeed an electron. The key aspect here is the fact that by undertaking the observations and performing the measurements, the observer did not create any information. The three degrees of freedom that describe the electron, any electron anywhere in the universe, or any elementary particle, were already embedded somewhere, most likely in the particle itself. This is equivalent to saying that particles and elementary particles store information about themselves, or by extrapolation, there is an information content stored in the matter of the universe. Due to the mass-energy-information equivalence principle,22 we postulate that information can only be stored in particles that are stable and have a non-zero rest mass, while interaction/force carrier bosons can only transfer information via waveform. Hence, in this work, we are only examining the information content stored in the matter particles that make up the observable universe, but it is important to mention that information could also be stored in other forms, including on the surface of the space–time fabric itself, according to the holographic principle.13 

The second question is: “How much information content is there in the observable universe?”

This question has been addressed in several studies going back as far as the late 1970s. Using the Bekenstein–Hawking formula for the black-hole entropy,26,27 Davies calculated the information content of the universe by applying the black-hole entropy formula to the whole universe,28 
(1)
where G is the gravitational constant, Mu is the mass of the universe enclosed within its horizon, h is Planck’s constant, and c is the speed of light. In fact, Davies produced a time-dependent information formula that allows studying the time evolution of the information in the universe, as discussed in detail by Treumann.29 
Wheeler estimated the number of bits in the present universe at T = 2.735 K from entropy considerations, resulting in 8 × 1088 bits content.8 Lloyd took a similar approach and estimated the total information capacity of the universe as6 
(2)
where S is the total entropy of the matter dominated universe, kB is the Boltzmann constant, ρ is the matter density of the universe, t is the age of the universe at the present time, c is the speed of light, and ħ is the reduced Planck constant.

Applying Landauer’s principle, Gough developed an information equation of state and calculated that 1087 bits of intrinsic universe information content could account for all the dark energy.30 His estimates are for a cosmic time coinciding with stars formation, when most of the universe bit information content is represented by high temperature baryons giving an average bit energy value of 120 eV.30 Using the mass-energy-information equivalence principle,22 Vopson estimated that around 52 × 1093 bits would be enough to account for all the missing dark matter in the observable universe.22,24 However, this is most likely overestimated, as he assumed that all the information bits are stored at T = 2.735 K. This is inaccurate because a significant amount of the baryonic matter is contained in stars, intergalactic gas, and dust, which all have temperatures larger than the cosmic microwave background (CMB).

The prime objective of this study is to estimate the bits of information contained in the matter particles of the observable universe via a new approach, using Shannon’s information theory. This approach is therefore independent of the dynamics of the universe expansion, or the internal dynamics in the universe, giving an accurate bit content at any given time of observation.

The amount of information extracted from observing the occurrence of an event can be estimated using Shannon’s information theory formalism published in 1948.31 Ignoring any particular features of the event, the observer, or the observation method, Shannon developed his theory using an axiomatic approach in which he defined information (I) extracted from observing an event as a function of the probability (p) of the event to occur or not, I(p), and he identified that the only function satisfying the imposed axiomatic properties is a logarithmic function. Hence, for a single event whose probability of occurrence is p, the information extracted from observing the event is
(3)
where b is an arbitrary base, which gives the units of information, i.e., for binary bits of information, b = 2. Let us assume a set of n independent and distinctive events X=x1,x2,,xn having a probability distribution P=p1,p2,,pn on X, so that each event xj has a probability of occurring pj = p(xj), where pj ≥ 0 and j=1npj=1. According to Shannon, the average information per event, or the number of bits of information per event, one can extract when observing the set X once is
(4)
H(X) is called the information entropy function, and it is maximum when the events xj have equal probabilities, pj = 1/n, so Hmax(x) = logbn. When observing N sets of events X, or equivalently observing N times the set of events X, the number of bits of information extracted from the observation is N · H(X). To exemplify this, let us assume a two-state system corresponding to a digital bit 0 or 1. Hence, n = 2 and X = {0, 1}, and assuming no biasing or external work on the system, then the two states have equal probabilities, pj = 1/n = 1/2 and p = {p1, p2} = {1/2, 1/2}. Using (4), the information entropy of this system is
(5)

H(X) = 1 means that 1 bit of information is encoded in this process, or conversely, observing the above event generates 1 bit of information.

We will use the same approach to estimate the information content of matter in the universe. The observable universe is made up of building blocks called elementary particles. According to the Standard Model, there are 30 elementary particles and antiparticles known today. These are six quarks and their corresponding anti-particles, six leptons and their corresponding anti-particles, five vector bosons, and one scalar boson. Out of all these particles, only a fraction of them are participants of the observable universe, as observed today. Most of the other particles or anti-particles are unstable and have lifetimes extremely short, so their observation is only possible via artificially created experimental conditions or theoretically. Therefore, their participation to the observable universe is negligible and, by extrapolation, their capacity to register information is also negligible. As already postulated in Sec. I, all force carrier particles/bosons and particles with zero rest mass cannot store/register information and they can only transfer information, so they are not considered in our estimates.

Each elementary particle is fully described by at least three degrees of freedom: mass, charge, and spin. As we already discussed, these degrees of freedom are embedded within each elementary particle, like an intrinsic label, or “particle DNA.” We regard this as the information content registered in each particle about itself, measured in bits. Excluding the bosons, the anti-particles and all the unstable elementary particles, out of all 30 elementary particles, only protons, neutrons, and electrons are actual participants to the observable universe today. Of course, protons and neutrons are in fact not elementary particles, as they are made up of three quarks each, which would need to be accounted for. We begin, however, our estimate accounting only for protons, neutrons, and electrons, and later we introduce the quarks.

The key aspect is to regard each elementary particle in the universe as an event in Shannon’s information theory formalism. In order to calculate the information content per elementary particle in the observable universe, we use the percentage weight abundance of elements in the universe, which are about 75% hydrogen, 23% helium, and 2% heavy elements (A > 4) (see the supplementary material). Each hydrogen atom contains one proton (p+) and one electron (e). Each helium atom contains 2p+, 2e, and two neutrons (n0). These are summarized as follows:

The x and y values must be known to perform an exact calculation. Since the heavy elements represent only 2% of the observable universe, one option is to use average x and y values representing all the heavy elements. However, in this study, in order to get an accurate picture, we will perform the exact calculation for all the known elements contained in the universe, including the huge range of heavy elements. The data table in the supplementary material shows the percentage weight abundance of all the elements in the observable universe. The key aspect here is to realize that the percentage weight abundance must be converted into the percentage number of atoms abundance, if one needs to estimate the total number of elementary particles in the universe and their information content. This is done using the molar mass values (units of mol/g) for each element, to work out the moles (number of atoms) of each kind. This calculation shows that the percentage weight abundance of elements in the universe is very different to the percentage of atoms abundance. For example, % weight values for hydrogen and helium are 75% and 23%, respectively, while their % numbers of atoms are 92.68% and 7.15%, respectively (see the supplementary material for full list of values).

Taking into account the internal structure of the atoms by using the exact number of p+, e, and n0 in each kind of atom, and using the % abundance of atoms determined, we can work out the total number of p+, e, and n0 corresponding to each kind of atom. Summing up all p+, e, and n0, we obtain: 108.27 electrons, 108.27 protons, and 15.6 neutrons, where the fractional values reflect the probabilistic nature of the estimates. The total statistical number of atomic constituents representative for the matter in the observable universe is then a total of 232.14 particles (p+, e, and n0). This allows us to estimate the probabilities of observing a p+, e, and n0 in the observable universe as Pp+ = Pe = 108.27/232.14 = 0.466 and Pn0 = 15.6/232.14 = 0.067. We can now regard this baryonic distribution of matter in the observable universe as a set of three possible states/events, {p+, e, n0}, with the occurrence probabilities {Pp+, Pe, Pn0} = {0.466, 0.466, 0.067}. Shannon’s classical information theory gives us an elegant mechanism to estimate the bits information content per event. In the case of a three-state system, the maximum information (I) encoded per event is I = log23 = 1.585 bits, corresponding to the case when the three events/states are equally probable. This is not the case here, so the average bits information content per event, meaning per any element of the set {p+, e, n0}, is given by
(6)

Introducing the numerical values in (6), we obtain 1.288 bits of information encoded per elementary particle. The meaning of this result is that the most effective way that information can be compressed into a p+, e, or n0, or the average amount of information content stored in each of these particles, is 1.288 bits per particle.

To find out the total information stored in all the matter particles in the universe, one needs to estimate the total number of protons, electrons, and neutrons in the observable universe, and to multiply that number by the bits information per elementary particle, i.e., 1.288.

Let us remember that the information content estimated above is in fact inaccurate as the protons and neutrons are not elementary particles and they have an internal structure containing three quarks each. According to the standard model, a proton contains two up quarks and one down quark, and a neutron contains one up quark and two down quarks. Hence, we can estimate the number of up/down quarks per each kind of atom, and using the % abundance of atoms, we can estimate how many up and down quarks are in a sample of 100 atoms. Summing up, we obtain a total of 216.55 up quarks and 108.27 down quarks contained in protons, and 15.6 up quarks and 31.21 down quarks contained in neutrons. This results in a total of 232.15 up quarks and 139.48 down quarks. Adding up the total number of electrons (108.27) to the total number of up quarks and down quarks, we obtain 479.9, which is the total number of elementary particles contained in a statistical sample of 100 random atoms. Using the acronyms uq for up quarks and dq for down quarks, we can now regard, again, these elementary particles as a set of three possible states/events, where the three possible states are {uq, dq, e}, and their occurrence probabilities are {Puq, Pdq, Pe}. The occurrence probabilities are obtained as Puq = 232.15/479.9, Pdq = 139.48/479.9 and Pe = 108.27/479.9, resulting in {Puq, Pdq, Pe} = {0.483, 0.29, 0.225}.

Shannon’s information entropy formula for this set of events and probabilities is
(7)

Introducing the numerical values in (7), we obtain 1.509 bits of information encoded per elementary particle. If the information content per elementary particle in the observable universe is known, the total information stored in the baryonic matter in the universe could be worked out if one knows how many particles, Nb, are in the universe.

The total number of particles in the universe, protons more precisely, is known as the Eddington number, NEdd, and it was first calculated in 1938.32 Eddington estimated that the universe contains NEdd = 136 × 2256, or about 1.57 × 1079 protons.

Here, we provide an analytical formula for the total number of baryons in the universe, which allows a more detailed estimation of the number of particles in the universe. Using the Friedmann equation33 for a homogeneous, isotropic, and flat universe, we determine the well-known critical density, ρc, of the universe as
(8)
where H = 2.2 × 10−18 s−1 is the Hubble parameter today and G = 6.674 × 10−11 m3/kg s2 is the gravitational constant. The critical density is then ρc = 9.6 × 10−27 kg/m3. The fraction of the energy density stored in baryons is Ωb = 0.0485,34 which allows us to estimate the baryons density as ρb = Ωb · ρc. The radius of the observable universe is estimated to be about 46.5 × 109 light-years,35 or Lu = 44 × 1025 m. The volume of the observable universe is then Vu = 4/3 · π · L3 = 4 × 1080 m3. The total mass of ordinary matter in the universe, or the mass of all baryons in the observable universe, Mb, can be calculated as
(9)
Considering the probabilities of occurrence of each p+, e, and n0, {Pp+, Pe, Pn0} = {0.466, 0.466, 0.067}, the effective mass is defined as follows:
(10)
me = 9.109 × 10–31 kg, mp+ = 1.672 × 10–27 kg, and mn0 = 1.675 × 10–27 kg are the rest masses of the electron, proton, and neutron, respectively. The number of baryons in the universe is then obtained by dividing the mass of all the baryons in the universe by the effective mass,
(11)
Performing the numerical calculations, we obtain Nb = 1.93 × 1080 baryons in the universe (more precisely the total number of particles in the universe as it includes electrons, which are not baryons). This can be used to determine the exact number of electrons (Ne), protons (Np+), and neutrons (Nn0) as
(12)
The numerical values are Ne = Np+ = 9.027 × 1079 and Nn0 = 1.298 × 1079. Since each neutron and proton contains three quarks, the total number of particles in the observable universe is Ntot = Ne + 3(Np+ + Nn0) or
(13)

The calculations indicate that the observable universe contains Ntot = 4 × 1080 particles, which is in good agreement with Eddington’s number giving only the number of protons.

Since we calculated in Sec. II that each particle stores Huqdqe = 1.509 bits of information, the total information content of the baryonic matter in the observable universe is Nbits = Ntot × Huqdqe = 6.036 × 1080 bits. This value is lower than the previous estimates.6,8,22,24,28–30 However, our calculation refers only to the information stored in particles, such as electrons, protons, neutrons, and their constituent quarks. The previous estimates refer to the total information content of the universe. Moreover, our calculation takes also a different approach, using the information theory, which gives the most effective information compression. The discrepancy in the information content values could therefore be explained by the fact that our estimate is for the most effective compression mechanism, according to the information theory. Other possible reasons for the information content discrepancy could perhaps suggest that the universe most likely contains more information, which is stored in other elementary particles, or other media, not accounted for in this study. For example, we have considered all bosons to be force/interaction particles responsible for the transfer of information, rather than storage of information. We also ignored all the anti-particles, as well as all neutrinos. Although there is scope for further refinements of this work, here it is the first time that the information theory has been used to estimate the information content of the observable universe, and this formalism most certainly would stimulate future studies. Moreover, the current approach offers a unique tool for estimating the information content per elementary particle, which is very useful for designing practical experiments to test these predictions.

The supplementary material contains the known percentage weight abundance of all the elements in the observable universe36 and their estimated abundance by the number of atoms. The supplementary material also contains the number of electrons, protons, and neutrons per each element, which allows estimating their occurrence probabilities.

The author acknowledges the financial support received to undertake this research from the School of Mathematics and Physics, University of Portsmouth.

1.
L.
Szilard
, “
Uber die enfropieuerminderung in einem thermodynamischen system bei eingrifen intelligenter wesen
,”
Z. Phys.
53
,
840
856
(
1929
).
2.
W.
Thomson
, “
The sorting demon of Maxwell
,”
Nature
20
,
126
(
1879
).
3.
R.
Landauer
, “
Irreversibility and heat generation in the computing process
,”
IBM J. Res. Dev.
5
(
3
),
183
191
(
1961
).
4.
R.
Landauer
, “
Wanted: A physically possible theory of physics
,”
IEEE Spectrum
4
(
9
),
105
109
(
1967
).
5.
S.
Lloyd
,
Programming the Universe: A Quantum Computer Scientist Takes on the Cosmos
(
Alfred A. Knopf
,
2006
), eISBN-13: 978-0-307-26471-8.
6.
S.
Lloyd
, “
Computational capacity of the universe
,”
Phys. Rev. Lett.
88
,
237901
(
2002
).
7.
S.
Lloyd
, “
Ultimate physical limits to computation
,”
Nature
406
,
1047
1054
(
2000
).
8.
J. A.
Wheeler
, “
Information, physics, quantum: The search for links
,” in
Complexity, Entropy and the Physics of Information
, edited by
W. H.
Zurek
(
Addison-Wesley
,
Redwood City, CA
,
1990
), p.
354
.
9.
J.
Hong
,
B.
Lambson
,
S.
Dhuey
, and
J.
Bokor
, “
Experimental test of Landauer’s principle in single-bit operations on nanomagnetic memory bits
,”
Sci. Adv.
2
(
3
),
e1501492
(
2016
).
10.
R.
Gaudenzi
,
E.
Burzurí
,
S.
Maegawa
,
H. S. J.
van der Zant
, and
F.
Luis
, “
Quantum Landauer erasure with a molecular nanomagnet
,”
Nat. Phys.
14
,
565
568
(
2018
).
11.
A.
Bérut
,
A.
Arakelyan
,
A.
Petrosyan
,
S.
Ciliberto
,
R.
Dillenschneider
, and
E.
Lutz
, “
Experimental verification of Landauer’s principle linking information and thermodynamics
,”
Nature
483
,
187
189
(
2012
).
12.
Y.
Jun
,
M.
Gavrilov
, and
J.
Bechhoefer
, “
High-precision test of Landauer’s principle in a feedback trap
,”
Phys. Rev. Lett.
113
(
19
),
190601
(
2014
).
13.
J. D.
Bekenstein
, “
Black holes and information theory
,”
Contemp. Phys.
45
(
1
),
31
43
(
2004
).
14.
E.
Verlinde
, “
Emergent gravity and the dark universe
,”
SciPost Phys.
2
,
016
(
2017
).
15.
E.
Verlinde
, “
On the origin of gravity and the laws of Newton
,”
J. High Energy Phys.
2011
,
29
.
16.
M. E.
McCulloch
, “
Modelling the Pioneer anomaly as modified inertia
,”
Mon. Not. R. Astron. Soc.
376
(
1
),
338
342
(
2007
).
17.
M. E.
McCulloch
and
J.
Giné
, “
Modified inertial mass from information loss
,”
Mod. Phys. Lett. A
32
(
28
),
1750148
(
2017
).
18.
L. B.
Kish
, “
Gravitational mass of information?
,”
Fluctuation Noise Lett.
7
,
C51
C68
(
2007
).
19.
L. B.
Kish
and
C. G.
Granqvist
, “
Does information have mass?
,”
Proc. IEEE
101
(
9
),
1895
1899
(
2013
).
20.
L.
Herrera
, “
The mass of a bit of information and the Brillouin’s principle
,”
Fluctuation Noise Lett.
13
(
1
),
1450002
(
2014
).
21.
E.
Bormashenko
, “
The Landauer principle: Re-formulation of the second thermodynamics law or a step to great unification
,”
Entropy
21
,
918
(
2019
).
22.
M. M.
Vopson
, “
The mass-energy-information equivalence principle
,”
AIP Adv.
9
,
095206
(
2019
).
23.
M. M.
Vopson
, “
The information catastrophe
,”
AIP Adv.
10
,
085014
(
2020
).
24.
M. M.
Vopson
, “
The information content of the universe and the implications for the missing dark matter
.” (
2019
), https://doi.org/10.13140/RG.2.2.19933.46560.
25.
R.
Landauer
, “
Dissipation and noise immunity in computation and communication
,”
Nature
335
,
779
784
(
1988
).
26.
J. D.
Bekenstein
, “
Black holes and entropy
,”
Phys. Rev. D
7
(
8
),
2333
(
1973
).
27.
S. W.
Hawking
, “
Particle creation by black holes
,”
Commun. Math. Phys.
43
,
199
(
1975
).
28.
P. C. W.
Davies
, “
Why is the physical world so comprehensible?
,” in
Complexity, Entropy and the Physics of Information
, edited by
W. H.
Zurek
(
Addison-Wesley
,
Redwood City, CA
,
1990
), p.
61
.
29.
R. A.
Treumann
, “
Evolution of the information in the universe
,”
Astrophys. Space Sci.
201
,
135
147
(
1993
).
30.
M. P.
Gough
, “
Information equation of state
,”
Entropy
10
,
150
159
(
2008
).
31.
C. E.
Shannon
, “
A mathematical theory of communication
,”
Bell Syst. Tech. J.
27
,
379
423
(
1948
).
32.
E.
Whittaker
, “
Eddington’s theory of the constants of nature
,”
Math. Gaz.
29
(
286
),
137
144
(
1945
).
33.
A.
Friedman
, “
On the curvature of space
,”
Gen. Relativ. Gravitation
31
,
1991
2000
(
1999
).
34.
S.
Eidelman
et al (
Particle Data Group
), “
The review of particle physics
,”
Phys. Lett. B
592
,
1
(
2004
).
35.
J. R.
Gott
III
,
M.
Jurić
,
D.
Schlegel
,
F.
Hoyle
,
M.
Vogeley
,
M.
Tegmark
,
N.
Bahcall
, and
J.
Brinkmann
, “
A map of the universe
,”
Astrophys. J.
624
(
2
),
463
484
(
2005
).
36.
See https://periodictable.com/Properties/A/UniverseAbundance.v.log.html for the abundance of elements in the universe by weight percentage provided by Mathematica’s Element Data function from Wolfram Research, Inc.

Supplementary Material