The relativistic Hopfield model constitutes a generalization of the standard Hopfield model that is derived by the formal analogy between the statistical-mechanic framework embedding neural networks and the Lagrangian mechanics describing a fictitious single-particle motion in the space of the tuneable parameters of the network itself. In this analogy, the cost-function of the Hopfield model plays as the standard kinetic-energy term and its related Mattis overlap (naturally bounded by one) plays as the velocity. The Hamiltonian of the relativisitc model, once Taylor-expanded, results in a p-spin series with alternate signs: the attractive contributions enhance the information-storage capabilities of the network, while the repulsive contributions allow for an easier unlearning of spurious states, conferring overall more robustness to the system as a whole. Here, we do not deepen the information processing skills of this generalized Hopfield network, rather we focus on its statistical mechanical foundation. In particular, relying on Guerra’s interpolation techniques, we prove the existence of the infinite-volume limit for the model free-energy and we give its explicit expression in terms of the Mattis overlaps. By extremizing the free energy over the latter, we get the generalized self-consistent equations for these overlaps as well as a picture of criticality that is further corroborated by a fluctuation analysis. These findings are in full agreement with the available previous results.

1.
E.
Agliari
,
A.
Barra
,
A.
Galluzzi
,
F.
Guerra
, and
F.
Moauro
, “
Multitasking associative networks
,”
Phys. Rev. Lett.
109
,
268101
(
2012
).
2.
E.
Agliari
,
A.
Barra
,
A.
De Antoni
, and
A.
Galluzzi
, “
Parallel retrieval of correlated patterns: From Hopfield networks to Boltzmann machines
,”
Neural Networks
38
,
52
(
2013
).
3.
E.
Agliari
,
A.
Barra
,
A.
Galluzzi
,
F.
Guerra
,
D.
Tantari
, and
F.
Tavani
, “
Retrieval capabilities of hierarchical networks: From Dyson to Hopfield
,”
Phys. Rev. Lett.
114
,
028103
(
2015
).
4.
E.
Agliari
,
A.
Barra
,
C.
Longo
, and
D.
Tantari
, “
Neural networks retrieving binary patterns in a sea of real ones
,”
J. Stat. Phys.
168
,
1085
(
2017
).
5.
E.
Agliari
,
D.
Migliozzi
, and
D.
Tantari
, “
Non-convex multi-species Hopfield models
,”
J. Stat. Phys.
172
,
1247
(
2018
).
6.
E.
Agliari
,
A.
Barra
,
G.
Landolfi
,
S.
Murciano
, and
S.
Perrone
, “
Complex reaction kinetics in chemistry: A unified picture suggested by mechanics in physics
,”
Complexity
2018
,
7423297
.
7.
D. J.
Amit
,
Modeling Brain Functions
(
Cambridge University Press
,
1989
).
8.
A.
Barra
, “
The mean field Ising model trough interpolating techniques
,”
J. Stat. Phys.
132
(
5
),
787
(
2008
).
9.
A.
Barra
, “
Notes on ferromagnetic p-spin and REM
,”
Math. Methods Appl. Sci.
32
,
783
(
2008
).
10.
A.
Barra
,
M.
Beccaria
, and
A.
Fachechi
, “
A new mechanical approach to handle generalized Hopfield neural networks
,”
Neural Networks
106
,
205
(
2018
).
11.
A.
Barra
,
G.
Genovese
, and
F.
Guerra
, “
Equilibrium statistical mechanics of bipartite spin systems
,”
J. Phys. A: Math. Theor.
44
(
24
),
245002
(
2011
).
12.
A.
Barra
,
G.
Genovese
, and
F.
Guerra
, “
The replica symmetric approximation of the analogical neural network
,”
J. Stat. Phys.
140
(
4
),
784
(
2010
).
13.
A.
Barra
,
F.
Guerra
,
G.
Genovese
, and
D.
Tantari
, “
How glassy are neural networks?
,”
J. Stat. Mech.: Theory Exp.
2012
,
P07009
.
14.
A.
Bovier
and
V.
Gayrard
, “
Hopfield models as generalized random mean field models
,” in
Mathematical Aspects of Spin Glasses and Neural Networks
(
Birkhauser Press
,
Boston
,
1998
).
15.
A.
Bovier
,
V.
Gayrard
, and
P.
Picco
, “
Gibbs states of the Hopfield model in the regime of perfect memory
,”
Probab. Theory Relat. Fields
100
(
3
),
329
(
1994
).
16.
A. C. C.
Coolen
,
R.
Kuhn
, and
P.
Sollich
,
Theory of Neural Information Processing Systems
(
Oxford Press
,
2005
).
17.
V.
Dotsenko
,
An Introduction to the Theory of Spin Glasses and Neural Networks
(
World Scientific
,
1995
).
18.
A.
Fachechi
,
E.
Agliari
, and
A.
Barra
, “
Dreaming neural networks: Forgetting spurious memories and reinforcing pure ones
,”
Neural Networks
112
,
24
(
2019
).
19.
I.
Goodfellow
,
Y.
Bengio
, and
A.
Courville
,
Deep Learning
(
MIT Press
,
2017
).
20.
F.
Guerra
, “
Sum rules for the free energy in the mean field spin glass model
,”
Math. Phys. Math. Phys.
30
,
161
(
2001
).
21.
F.
Guerra
and
F. L.
Toninelli
, “
The thermodynamic limit in mean field spin glass models
,”
Commun. Math. Phys.
230
(
1
),
71
79
(
2002
).
22.
F.
Guerra
and
F. L.
Toninelli
, “
The infinite volume limit in generalized mean field disordered models
,”
Markov Proc. Rel. Fields
9
(
2
),
195
207
(
2003
).
23.
J. J.
Hopfield
, “
Neural networks and physical systems with emergent collective computational abilities
,”
Proc. Natl. Acad. Sci. U. S. A.
79
(
8
),
2554
2558
(
1982
).
24.
J. J.
Hopfield
,
D. I.
Feinstein
, and
R. G.
Palmer
, “
Unlearning has a stabilizing effect in collective memories
,”
Nat. Lett.
304
,
158
(
1983
).
25.
J. A.
Horas
and
P. M.
Pasinetti
, “
On the unlearning procedure yielding a high-performance associative memory neural network
,”
J. Phys. A: Math. Gen.
31
,
L463
L471
(
1998
).
26.
D.
Krotov
and
J. J.
Hopfield
, “
Dense associative memory is robust to adversarial inputs
,”
Neural Comput.
30
,
3151
(
2018
).
27.
Y.
Le Cun
,
Y.
Bengio
, and
G.
Hinton
, “
Deep learning
,”
Nature
521
,
436
444
(
2015
).
28.
K.
Nokura
, “
Spin glass states of the anti-Hopfield model
,”
J. Phys. A: Math. Gen.
31
,
7447
(
1998
).
29.
K.
Nokura
, “
Paramagnetic unlearning in neural network models
,”
Phys. Rev. E
54
(
5
),
5571
, (
1996
).
30.
K.-S.
Oh
and
J.
Keechul
, “
GPU implementation of neural networks
,”
Pattern Recognit.
37
(
6
),
1311
(
2004
).
31.
D.
Ruelle
,
Statistical Mechanics: Rigorous Results
(
World Scientific
,
1999
).
32.
M.
Talagrand
, “
Rigorous results for the Hopfield model with many patterns
,”
Probab. Theory Relat. Fields
110
(
2
),
177
(
1998
).
33.
M.
Talagrand
, “
Exponential inequalities and convergence of moments in the replica-symmetric regime of the Hopfield model
,”
Ann. Probab.
28
,
1393
(
2000
).
34.
J.
Tubiana
and
R.
Monasson
, “
Emergence of compositional representations in restricted Boltzmann machines
,”
Phys. Rev. Lett.
118
(
13
),
138301
(
2017
).
35.
S.
Wimbauer
and
J. L.
van Hemmen
, “
Hebbian unlearning
,” in
Analysis of Dynamical and Cognitive Systems
(
Springer
,
Berlin
,
1995
).
36.

Notice that, once the existence of the thermodynamic limit is proved for the free energy, it holds, in a cascade fashion, also for other various quantities of interest, as entropy and internal energy.22 

37.

It is worth mentioning that biased patterns (e.g., where positive entries are more likely), beyond being less informative, may also lead to technical issues as spurious configurations would present large overlap with pure memories, so the attracting power of the latter is strongly downsized, see also Ref. 7.

38.

For instance, information can be supplied to the network as an external field to keep a physical jargon.

39.

Actually, we are committing a small abuse of notation as the “standard” free energy α̃(β) is defined as α̃(β)=β1α(β), but, as from the mathematical side α̃(β) and α(β) are (a constant apart) equivalent, we will keep our choice as it makes calculations more transparent, obviously, without affecting the results.

40.

This assumption crucially allows us to use the Wick Theorem, namely, given a Gaussian variable z we can write E(zf(z))=E(z2)E(zf(z)).

You do not currently have access to this content.