Next Article in Journal
Exponential Families with External Parameters
Previous Article in Journal
Dynamics of Entropy Production Rate in Two Coupled Bosonic Modes Interacting with a Thermal Reservoir
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A New Look at the Spin Glass Problem from a Deep Learning Perspective

by
Petr Andriushchenko
1,*,†,
Dmitrii Kapitan
1,2,3,† and
Vitalii Kapitan
2,3,†
1
National Center for Cognitive Research, ITMO University, bldg. A, Kronverksky Pr. 49, 197101 Saint Petersburg, Russia
2
Department of Theoretical Physics and Smart Technologies, Far Eastern Federal University, Russky Island, 10 Ajax Bay, 690922 Vladivostok, Russia
3
Institute of Applied Mathematics, Far Eastern Branch, Russian Academy of Science, 7 Radio, 690041 Vladivostok, Russia
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Entropy 2022, 24(5), 697; https://doi.org/10.3390/e24050697
Submission received: 3 May 2022 / Revised: 10 May 2022 / Accepted: 10 May 2022 / Published: 14 May 2022
(This article belongs to the Section Statistical Physics)

Abstract

:
Spin glass is the simplest disordered system that preserves the full range of complex collective behavior of interacting frustrating elements. In the paper, we propose a novel approach for calculating the values of thermodynamic averages of the frustrated spin glass model using custom deep neural networks. The spin glass system was considered as a specific weighted graph whose spatial distribution of the edges values determines the fundamental characteristics of the system. Special neural network architectures that mimic the structure of spin lattices have been proposed, which has increased the speed of learning and the accuracy of the predictions compared to the basic solution of fully connected neural networks. At the same time, the use of trained neural networks can reduce simulation time by orders of magnitude compared to other classical methods. The validity of the results is confirmed by comparison with numerical simulation with the replica-exchange Monte Carlo method.

Graphical Abstract

1. Introduction

Spin glasses fundamentally differ from other lattice models by the presence of frustrations—strong competition of magnetic interactions, and disorder—freezing of atoms upon cooling. Due to these key features, spin glasses have long relaxation times, a rough energy landscape, and macroscopic degeneracy of ground states. This leads to the fact that the numerical simulation and even more so the analytical description of such systems becomes a challenging task. The first attempts at theoretical description of the spin glass model [1,2] encountered several difficulties. One of the difficulties was the lack of translational invariance in the spin glass systems. Another problem was the non-ergodicity of the spin glass phases due to the presence of many local energy minima, which are separated by high potential barriers. Hence, the problem of configurational averaging and the calculation of residual entropy follows. The problem is that the entropy of the system being in the true thermodynamic equilibrium should be equal to zero. However, this is not applicable to spin glasses, since the energy of the system may depend not only on the temperature, but also on the history of states of the sample. Thus, at the lowest temperatures, the so-called residual entropy can be observed [3]. It can be calculated through the degeneracy of the ground states. This parameter is one of the key parameters of systems with competing interactions.
Many processes that occur in spin glasses cannot be described within the framework of the classical theory of phase transitions and require new approaches. We would also like to note the relevance of such a problem since the approaches being developed to describe spin glasses lead to entirely new results far beyond theoretical physics. Such results include, for example, contributions to solving multiparameter optimization problems or problems in the area of associative and distributed memory [4,5,6].
Since the class of spin models on lattices described by Ising-like Hamiltonians can be solved analytically only in extremely rare cases [7], numerical probabilistic methods, such as different variations of the Monte Carlo method, are now most commonly used to describe the physics of spin glass [8,9,10,11]. However, properties of spin glasses, such as long relaxation time, rough energy landscape, macroscopic degeneracy of ground states, and the effects of critical slowing down, significantly decrease the efficiency of the Monte Carlo algorithms. The motion of the system in phase space is very slow, so it requires generation of an extremely large number of states to move to an equilibrium state. On the other hand, there has been exponential growth in computational power and the rapid development of Monte Carlo methods, leading to the use of these methods in almost all fields of physics [12,13,14,15]. It allows to partially offset the increasing complexity of calculations with the increasing size of the spin glass systems [16,17,18].
Simultaneously with the development of numerical Monte Carlo methods, the exponential growth of computational power led to the second revolution in the area of neural networks and the emergence of completely new architectures and approaches to neural network training—convolutional neural networks, autoencoders, constrained Boltzmann machines, etc. [19]. All of these approaches have dramatically reduced training time and increased the dimensionality of the tasks to be solved. This revolution has led to an unprecedented expansion of machine learning methods into all areas of life, business and science. In particular, machine learning methods have begun to be applied to statistical physics [20,21,22].
Several approaches have been proposed to study magnetic systems using machine learning. The first one is to use supervised machine learning algorithms to solve the problem of classifying the states of magnetic systems (with nearest-neighbors interaction on a square lattice) into thermodynamic phases. The input of such a model is the configuration of the magnetic system (states of all spins) obtained by Monte Carlo simulation, and the output of the model predicts the most probable phase where the configuration could appear [23,24,25]. The advantage of this approach is the ability to use convolutional neural networks, which are commonly used for image recognition tasks. It can be explained by the fact that such architectures naturally convey the fundamental properties of a square lattice—the spatial arrangement of spins and the interactions with the nearest four neighbors.
The second approach is to use supervised machine learning algorithms to solve the regression problem of searching for the lowest energy configurations (ground states) of spin systems with complex interactions, such as spin glass or spin ice. The problem of finding the lowest energy states of spin glasses is a key problem since such states are highly degenerate, often asymmetric, and separated by high energy barriers. However, such states have the highest probability at a lower temperature. Consequently, they are the ones that contribute the most to the statistical sum and thus to all the thermodynamic averages [26,27].
In this paper, we propose to solve a more general problem: using machine learning methods to solve the problem of regression of the basic thermodynamic characteristics E , M (average energy and magnetization) or any other system characteristics, as a function of temperature T for spin glasses on a square lattice. Widely known, there is Cybenko’s theorem on the universal approximator, proving that any continuous function of many variables can be approximated with a given accuracy by a feedforward neural network with one hidden layer [28]. For this purpose, we consider the spin glass as a weighted graph, in which the architecture of the graph corresponds to the lattice and the values of the edges correspond to the values of the exchange interaction. Thus, using a neural network, we are looking for a functional dependence between the spatial distribution of the exchange integral on the square lattice of the spin glass J k = f J ( x k , y k ) and the main average thermodynamic characteristics of the system E , M . Here, f J is the function of the spatial distribution of spin glass bonds values, J k is the bond value, x k , y k is the bond coordinates for bond k. To solve this class of problems, it is relevant to find an architecture of a neural network that will be able to approximate this function most effectively using the spatial regularities of the lattice.

2. Model and Data

2.1. Spin Glass Model

In this paper, we consider spin glass models with periodic boundary conditions on a square lattice N = L × L , in which each spin is an Ising spin, i.e., it has two states S i = ± 1 , and has four nearest neighbors, the interactions with which are determined by the exchange integral J k = ± 1 (see Figure 1). The standard Hamiltonian of such a system has the form:
H = < i j > J k S i S j ,
where S i , S j are the interacting spins that are nearest neighbors on the square lattice, < i , j > is the summation that occurs only by nearest neighbors, J k is the value of the exchange interaction between spins S i and S j , the index k = k ( i , j ) is a function of i , j .
Mean energy E of the spin glass at temperature T is calculated by (2), and the mean magnetization M by (3):
E T = 1 N H T ,
M T = 1 N i S i T .
Figure 1. Example of the spin glass model on a square lattice 6 × 6 of Ising spins with a periodical boundaries condition. S i = ± 1 —spins of the lattice, J k = ± 1 —exchange integral.
Figure 1. Example of the spin glass model on a square lattice 6 × 6 of Ising spins with a periodical boundaries condition. S i = ± 1 —spins of the lattice, J k = ± 1 —exchange integral.
Entropy 24 00697 g001
The spatial distribution of the exchange integral J k = f J ( x k , y k ) determines all macroscopic characteristics of spin glass. Such a description of the spin glass model, at first glance, may seem extremely simple. Since it is a classical spin model, it seems that numerical simulations can be performed quite easily with classical numerical Monte Carlo methods [29]. However, this is a misleading simplicity. There are many examples of extremely simple systems that present unpredictable behavior. Examples of such systems are given in Stephen Wolfram’s book [30], in which he described remarkably simple models that gave complex, non-predictable behavior that cannot be described by analytical approaches. Models of spin glasses possess all of the previously mentioned characteristics: long relaxation times, a rough energy landscape, macroscopic degeneracy of ground states, and critical slowing down effect. All of these features together make it extremely difficult to study such systems.

2.2. Data

For the lattice N = L × L , there are 2 2 N possible distributions of the exchange integral, starting with the distribution where all exchange interactions are antiferromagnetic J k = 1 , i k = 1 2 N J k = 2 N , passing through all possible combinations { J 1 , J 2 , J 3 , J 2 N } of exchange integrals and ending with the classical ferromagnetic model in which all interactions J k = 1 , i k = 1 2 N J k = 2 N . To find the dependence between the spatial distribution function of the exchange integral and the main average thermodynamic characteristics of the system, it is necessary to train the neural network for detecting different patterns of the mutual location of the exchange interaction values on the lattice and their influence on the macroscopic parameters of the system. For this purpose, the considered configuration of the spin glass with the given values { J 1 , J 2 , J 3 , J 2 N } and the temperature T should be fed to the input of the neural network. Then, it can be trained to predict the output mean energy E and magnetization M of the considered configuration of the spin glass. We would like to note that in the same way, it is possible to train a neural network to predict the probability density of states, residual entropy, heat capacity, susceptibility and other parameters characterizing the considered spin glass.
In order to train the neural networks, it was necessary to prepare datasets for training, validation and testing of neural network models. To this end, 60 temperatures from 0.1 to 6 in step 0.1 were calculated for each considered configuration of the spin glass. Simulations were performed using a parallel replica-exchange Monte Carlo (MC) method. There were 10,000 equilibration MC steps, then the energy and magnetization of the system were calculated and averaged over the next 100,000 MC steps according to Equations (2) and (3). To overcome the effects of critical slowing down and getting stuck in local minima, the system was simulated in parallel at 60 temperatures and the system configurations were exchanged every 1000 MC steps with a probability dependent on the system energy:
p ( X X ) = 1 if Δ 0 exp ( Δ ) if Δ > 0 ,
where Δ = ( 1 / T 1 / T ) ( E E ) , E and E are the energies corresponding to X and X configurations, respectively.
In total, two datasets each were calculated for two system sizes. The datasets contained 2 N values of all interactions J, one value of temperature T and two output values of mean energy E and mean magnetization M , therefore, a total of 2 N + 3 values. For the spin glass model with N = 6 × 6 number of spins, the small dataset consisted of 834 configurations (with dimension of 50,040 × 75 since each spin glass configuration was calculated at 60 different temperature values) and the large one consisted of 41,405 configurations (2,484,300 × 75). For the model with N = 10 × 10 number of spins, the small dataset consisted of 10,302 configurations (618,120 × 203) and the large of 43,596 configurations (2,615,760 × 203).
All configurations were randomly generated, with different constraints on the total sum of all interactions k = 1 2 N J k . In the small datasets, the spin glass configurations were presented almost uniformly with respect to all possible values of the sum of all interactions k = 1 2 N J k from 2 N to 2 N in steps of 2. In large datasets, the distribution of spin glass configurations over the sum of all interactions tended to the corresponding values of binomial coefficients k = 1 2 N J k = 2 N 2 j j 2 N , where j is the number of negative interactions J k = 1 . All datasets were divided into train, validation and test subsets in proportions of 0.8:0.15:0.05.

2.3. Deep Neural Networks

A deep neural network (DNN) is an artificial neural network (ANN) of forwarding propagation, i.e., multilayer perceptron, with more than one hidden layer. Similar to biological neurons with axons and dendrites, a DNN represents layers of artificial neurons with a given activation function, which are interconnected by trainable coefficients (see Figure 2). The first layer is called the input layer, the last one is the output layer, and all layers between them are called hidden layers. At the initial stage, the neural network is untrained, i.e., the linkage weights are set randomly and not optimized for the certain problem. The training of a neural network involves the adaptation of the network to the solution of a particular problem, carried out by adjusting the weight coefficients of each layer to minimize a given loss function L [31,32]. In this work, we use the mean squared error (MSE) as the loss function.
ANN training is performed using the error backpropagation method in two stages. During forward propagation, the input data are fed to the input of the ANN and then propagated through all the hidden layers to the output layer. The neurons of each layer receive data from the neurons of the previous layer, and their values are calculated using a matrix of weights W, bias b (5) and activation function h (6). The resulting values are transmitted to the next layer, i.e., the output of layer l 1 becomes the input of layer l, and so on, to the output layer of the network. A parametric rectified linear unit (PReLU) was used as the activation function h, which solved the problem of the so-called “dying ReLU”, when some neurons were simply turned off from training (7).
y [ l ] = W [ l ] h [ l 1 ] + b [ l ] ,
h [ l ] = f ( y [ l ] ) ,
f ( y ) = y if y > 0 a y if y 0 ,
where a is a learnable parameter controlling the slope of the negative part of the function.
In the second step, all ANN weights are updated so as to minimize the loss function on a given dataset. For this purpose, the gradients for the variable parameters are calculated according to (8)–(10):
L y [ l 1 ] = [ W [ l ] ] T L y [ l ] f ( y [ l 1 ] ) ,
L W [ l ] = L y [ l ] [ h [ l 1 ] ] T ,
L b [ l ] = L y [ l ] .
Figure 2. Deep Neural Network architecture FC4 (Fully Connected) with three hidden layers h 1 = 2 N + 1 , h 2 = N , h 3 = N .
Figure 2. Deep Neural Network architecture FC4 (Fully Connected) with three hidden layers h 1 = 2 N + 1 , h 2 = N , h 3 = N .
Entropy 24 00697 g002
Then, the weights values and the bias are updated according to the calculated gradients:
W [ l ] : = W [ l ] α L W [ l ] ,
b [ l ] : = b [ l ] α L b [ l ] ,
where α is a learning rate parameter. This is the way the ANN is trained, during which with each step we descend to the global minimum of the convex loss function L with speed α .

3. Results and Discussion

Fully connected (FC) network architectures with different numbers and sizes of hidden layers were proposed as a baseline (see Figure 2). The following FC network architectures with one, two, three and four hidden layers were tested on datasets of the model with 6 × 6 size:
  • FC1: h 1 = 73 ;
  • FC2: h 1 = 36 , h 2 = 36 ;
  • FC3: h 1 = 73 , h 2 = 36 ;
  • FC4: h 1 = 73 , h 2 = 36 , h 3 = 36 ;
  • FC5: h 1 = 73 , h 2 = 36 , h 3 = 36 , h 4 = 36 .
The effect of the number and size of DNN’s hidden layers h on the quality and speed of learning was investigated. Five different fully connected neural network architectures (FC1–FC5) were trained on small and large datasets. The average learning time per epoch in seconds was calculated for them, and the root mean squared errors (RMSE) of the mean energy and magnetization was calculated with the DNN from the initial values obtained with the replica-exchange Monte Carlo (see Table 1). For the big dataset, the number of epochs was 500, and for the small dataset it was 1000.
The architecture FC4 with three hidden layers, shown in Figure 2, showed optimal results, yielding only slightly in accuracy on a small dataset of a network with four hidden layers (FC5). Further results will be compared with these two architectures.
Table 1. Comparison of average single epoch learning rate and RMSE of mean energy E and magnetization M for s m a l l and b i g datasets of FC1–FC5 fully connected neural networks.
Table 1. Comparison of average single epoch learning rate and RMSE of mean energy E and magnetization M for s m a l l and b i g datasets of FC1–FC5 fully connected neural networks.
FC1FC2FC3FC4FC5
RMSE s m a l l E 2.4832.3782.3702.3232.321
RMSE s m a l l M 0.1330.1030.1010.0880.088
Time s m a l l , c33334
RMSE b i g E 8.9982.6552.3401.9381.938
RMSE b i g M 0.2710.1790.0930.0650.065
Time b i g , c2628313545
To improve the speed and accuracy of the calculations, we investigated DNNs whose architectures would transmit information about the spatial arrangement of the connections on the square lattice. We proposed to replace fully connected hidden layers with layers in which neurons would be connected similarly to spins on a square lattice. The speed improvement is achieved by having fewer training weights in the layers compared to fully connected architectures.
Two architectures of DNN with two levels of spin lattice abstraction were considered CC1 and CC2 (CC—Custom Connected). The first architecture CC1 proposes to consider the first hidden layer h 1 as virtual bonds, and the second layer h 2 as virtual spins. In such a network, all neurons of layer h 1 , except the temperature neuron, are connected to the corresponding neurons of layer h 2 in the same way that bonds in a square lattice are connected to spins. For example, spin S 1 has four bonds: J 1 , J 2 , J L × L + 1 = 37 , J L × ( L + 1 ) + 1 = 43 (Figure 1), so in the neural network, neurons 1 , 2 , 37 and 43 of layer h 1 will be connected to the first neuron of layer h 2 , see Figure 3. Thus, each neuron of the h 1 layer is connected to two corresponding neurons of the h 2 layer, except for the temperature neuron of the h 1 layer, which is connected to all neurons of the h 2 layer. In such an architecture, the connection between layers h 2 and h 3 , as well as h 3 and the output layer, was fully connected.
The second architecture of CC2 has the same approach as in CC1, except for the connections between layers h 2 and h 3 . In CC2, it was proposed to consider the h 3 layer also as virtual spins, and to connect the h 2 layer with h 3 in a manner similar to the neighboring spins in a square lattice (Figure 1). For example, if spin S 2 is a neighbor of spins S 1 , S 3 , S L + 2 = 8 , S ( L × L 1 ) + 2 = 32 , it means that the second neuron of layer h 3 will be connected with neurons 1 , 3 , 8 , 32 and also with neuron 2 (See Figure 3). Hidden layer h 3 is fully connected with the output layer. The connection between layers h 3 and the output layer was fully connected.
To study the accuracy of the proposed DNN architectures with a baseline solution, the neural networks were trained and tested on big datasets for spin glass systems N = 6 × 6 and N = 10 × 10 . To control the overfitting of the neural networks during training, the loss function L was calculated on a validation sample that was not involved in the training. The graph of the loss function value L as a function of the number of training epochs is shown in Figure 4. This figure shows that for 500 epochs, overfitting does not occur for any of the architectures considered. However, there is a large difference in the speed and accuracy of learning.
The results of the work of neural networks were scored by root mean squared error (RMSE) of the average energy E and magnetization M . The resulting RMSE values depending on DNN architecture and system size are presented in Table 2. The table shows that the CC1 architecture was the most accurate, reducing the average energy calculation error by one-and-a-half times compared to the fully connected architectures. Figure 5 also shows the dependence of the root mean squared error (RMSE) on temperature for the average energy E and magnetization M for neural networks of different architectures. It is well noticeable that the computational error increases with decreasing temperature. This is due to the complexity of calculating the ground states of the spin glass models. To reduce the error, it is possible to use, for example, the approach described in [26], which allows using a restricted Boltzmann machine to calculate the ground states of the spin glass systems. It is also clear that the CC1 architecture reduces the error, this difference can be seen especially at small temperatures.
Figure 4. Dependence of the loss function value L during validation on epoch number for the FC4 and the proposed CC1 and CC2 architectures, trained on a large dataset for the model N = 6 × 6 (a); N = 10 × 10 (b).
Figure 4. Dependence of the loss function value L during validation on epoch number for the FC4 and the proposed CC1 and CC2 architectures, trained on a large dataset for the model N = 6 × 6 (a); N = 10 × 10 (b).
Entropy 24 00697 g004
Table 2. Comparison of the results of the root mean squared error (RMSE), calculated by different architectures of DNN, of the mean energy E and magnetization M for spin glasses of sizes 6 × 6 and 10 × 10 .
Table 2. Comparison of the results of the root mean squared error (RMSE), calculated by different architectures of DNN, of the mean energy E and magnetization M for spin glasses of sizes 6 × 6 and 10 × 10 .
ArchitecturesRMSE E RMSE M
6 × 6 10 × 10 6 × 6 10 × 10
FC41.99913.76600.06010.0491
FC52.00453.81680.06880.0492
CC11.48542.60710.06420.0443
CC21.76743.01730.06730.0581
Figure 5. Dependence of the root mean squared error (RMSE) on the temperature for the average energy E (a) and magnetization M (b) for neural networks of different architectures (FC4, CC1 and CC2).
Figure 5. Dependence of the root mean squared error (RMSE) on the temperature for the average energy E (a) and magnetization M (b) for neural networks of different architectures (FC4, CC1 and CC2).
Entropy 24 00697 g005
Figure 6 shows an example of mean energy calculation using a replica-exchange MC and DNNs of different architectures (FC4, CC1 and CC2). The configuration of the calculated spin glass is shown in the corner of the figure. It can be seen that the calculation result of the network with the CC1 architecture is almost identical to that obtained with the replica-exchange MC, while the results of the networks with the FC4 and CC2 architectures have some deviations at low and medium temperatures.

4. Conclusions

In this paper, we presented a method for calculating the values of thermodynamic averages of a frustrated spin glass model using deep neural networks. The influence of the neural network architecture on the speed and accuracy of calculations for spin glass models N = 6 × 6 and 10 × 10 with different distributions of the exchange integral J was studied. Specific neural network architectures have been proposed to increase the accuracy and reduce the error compared to fully connected models. The use of trained neural networks can significantly reduce the time compared to classical numerical approaches when modeling spin glass systems. This approach, after appropriate training, allows modeling of any global characteristics of spin glass including probability density of states, residual entropy, heat capacity, susceptibility, and so on.
With the help of deep neural networks, it has become possible to calculate the global characteristics of the system quite accurately based on the microarchitecture (a certain distribution of edges values) of a particular configuration of the spin glass. Based on the obtained results, we can conclude that neural network architectures that simulate the structure of spin lattices are better adapted to the calculation of spin glass models. In further development of this topic of using neural networks in the modeling of complex magnetic systems, it is interesting to consider convolutional models of neural networks on lattices. As convolution should be performed not by nodes, but by bonds, the recently proposed graph neural networks (GCN) [33], which allow to move away from clearly fixed lattice sizes and calculate systems of any size, are interesting. This will make it possible to train such networks on small lattices computed by various exact methods [34], and to extend the obtained functional regularities to large systems.

Author Contributions

Conceptualization, P.A.; Data curation, D.K.; Funding acquisition, P.A.; Investigation, P.A., D.K. and V.K.; Methodology, P.A.; Project administration, P.A. and V.K.; Software, D.K.; Supervision, P.A. and V.K.; Visualization, P.A.; Writing—original draft, P.A. and D.K.; Writing—review & editing, P.A. All authors have read and agreed to the published version of the manuscript.

Funding

Implementation and research by the custom neural network architecture were carried out within the frame of the grant of the Russian Science Foundation No. 21-72-00157. Implementation and research by the replica-exchange Monte Carlo were made within the frame of the state task of the Ministry of Science and Higher Education of the Russian Federation No. 0657-2020-0005.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data of the current study are available from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Edwards, S.F.; Anderson, P.W. Theory of spin glasses. J. Phys. Met. Phys. 1975, 5, 965. [Google Scholar] [CrossRef]
  2. Kirkpatrick, S.; Sherrington, D. Infinite-ranged models of spin-glasses. Phys. Rev. B 1978, 17, 4384. [Google Scholar] [CrossRef]
  3. Langer, S.A.; Sethna, J.P.; Grannan, E.R. Nonequilibrium entropy and entropy distributions. Phys. Rev. B 1990, 41, 2261. [Google Scholar] [CrossRef] [PubMed]
  4. Zarinelli, E. Spin-Glass Models and Interdisciplinary Applications. Ph.D. Thesis, École Doctorale Physique de la réGion Parisienne, Paris, France, 2012. [Google Scholar]
  5. Venkataraman, G.; Athithan, G. Spin glass, the travelling salesman problem, neural networks and all that. Pramana 1991, 36, 1–77. [Google Scholar] [CrossRef]
  6. Amit, D.J.; Gutfreund, H.; Sompolinsky, H. Storing infinite numbers of patterns in a spin-glass model of neural networks. Phys. Rev. Lett. 1985, 55, 1530. [Google Scholar] [CrossRef] [PubMed]
  7. Barzegar, A.; Pattison, C.; Wang, W.; Katzgraber, H.G. Optimization of population annealing Monte Carlo for large-scale spin-glass simulations. Phys. Rev. E 2018, 98, 053308. [Google Scholar] [CrossRef] [Green Version]
  8. Swendsen, R.H.; Wang, J.S. Replica Monte Carlo simulation of spin-glasses. Phys. Rev. Lett. 1986, 57, 2607. [Google Scholar] [CrossRef]
  9. Andriushchenko, P. Influence of cutoff dipole interaction radius and dilution on phase transition in kagome artificial spin ice. J. Magn. Magn. Mater. 2019, 476, 284–288. [Google Scholar] [CrossRef]
  10. Andriushchenko, P.; Soldatov, K.; Peretyatko, A.; Shevchenko, Y.; Nefedev, K.; Otsuka, H.; Okabe, Y. Large peaks in the entropy of the diluted nearest-neighbor spin-ice model on the pyrochlore lattice in a [111] magnetic field. Phys. Rev. E 2019, 99, 022138. [Google Scholar] [CrossRef]
  11. Soldatov, K.; Peretyatko, A.; Andriushchenko, P.; Nefedev, K.; Okabe, Y. Comparison of diluted antiferromagnetic Ising models on frustrated lattices in a magnetic field. Phys. Lett. A 2019, 383, 1229–1234. [Google Scholar] [CrossRef] [Green Version]
  12. Kovtanyuk, A.; Nefedev, K.; Prokhorov, I. Advanced computing method for solving of the polarized-radiation transfer equation. In Russia-Taiwan Symposium on Methods and Tools of Parallel Processing; Springer: Berlin/Heidelberg, Germany, 2010; pp. 268–276. [Google Scholar]
  13. Shevchenko, Y.A.; Makarov, A.; Andriushchenko, P.; Nefedev, K. Multicanonical sampling of the space of states of H(2, n)-vector models. J. Exp. Theor. Phys. 2017, 124, 982–993. [Google Scholar] [CrossRef] [Green Version]
  14. Chepak, A.K.; Afremov, L.L.; Mironenko, A.Y. Concentration phase transition in a two-dimensional ferromagnet. Solid State Phenomena. Trans. Tech. Publ. 2020, 312, 244–250. [Google Scholar]
  15. Vasil’ev, E.V.; Perzhu, A.V.; Korol, A.O.; Kapitan, D.Y.; Rybin, A.E.; Soldatov, K.S.; Kapitan, V.Y. Numerical simulation of two-dimensional magnetic skyrmion structures. Comput. Res. Model. 2020, 12, 1051–1061. [Google Scholar] [CrossRef]
  16. Landau, D.; Binder, K. A Guide to Monte Carlo Simulations in Statistical Physics; Cambridge University Press: Cambridge, UK, 2021. [Google Scholar]
  17. Soldatov, K.; Nefedev, K.; Kapitan, V.Y.; Andriushchenko, P. Approaches to numerical solution of 2D Ising model. J. Phys. Conf. Ser. 2016, 741, 012199. [Google Scholar] [CrossRef] [Green Version]
  18. Makarov, A.G.; Makarova, K.; Shevchenko, Y.A.; Andriushchenko, P.D.; Kapitan, V.Y.; Soldatov, K.S.; Perzhu, A.V.; Rybin, A.; Kapitan, D.Y.; Vasil’ev, E.; et al. On the numerical calculation of frustrations in the Ising model. JETP Lett. 2019, 110, 702–706. [Google Scholar] [CrossRef]
  19. Dean, J.; Patterson, D.; Young, C. A new golden age in computer architecture: Empowering the machine-learning revolution. IEEE Micro 2018, 38, 21–29. [Google Scholar] [CrossRef]
  20. Carleo, G.; Cirac, I.; Cranmer, K.; Daudet, L.; Schuld, M.; Tishby, N.; Vogt-Maranto, L.; Zdeborová, L. Machine learning and the physical sciences. Rev. Mod. Phys. 2019, 91, 045002. [Google Scholar] [CrossRef] [Green Version]
  21. Butler, K.T.; Davies, D.W.; Cartwright, H.; Isayev, O.; Walsh, A. Machine learning for molecular and materials science. Nature 2018, 559, 547–555. [Google Scholar] [CrossRef]
  22. Suchsland, P.; Wessel, S. Parameter diagnostics of phases and phase transition learning by neural networks. Phys. Rev. B 2018, 97, 174435. [Google Scholar] [CrossRef] [Green Version]
  23. Carrasquilla, J.; Melko, R.G. Machine learning phases of matter. Nat. Phys. 2017, 13, 431–434. [Google Scholar] [CrossRef] [Green Version]
  24. Shiina, K.; Mori, H.; Okabe, Y.; Lee, H.K. Machine-learning studies on spin models. Sci. Rep. 2020, 10, 2177. [Google Scholar] [CrossRef] [PubMed]
  25. Kapitan, V.; Vasiliev, E.; Perzhu, A.; Kapitan, D.; Rybin, A.; Korol, A.; Soldatov, K.; Shevchenko, Y. Numerical simulation of magnetic skyrmions on flat lattices. AIP Adv. 2021, 11, 015041. [Google Scholar] [CrossRef]
  26. Korol, A.; Kapitan, V.; Perzhu, A.; Padalko, M.; Kapitan, D.; Volotovskii, R.; Vasiliev, E.; Rybin, A.; Ovchinnikov, P.; Andriushchenko, P.; et al. Calculation of the ground states of spin glasses using the Restricted Boltzmann Machine. JETP Lett. 2022, 115, 500. [Google Scholar]
  27. Bukov, M.; Schmitt, M.; Dupont, M. Learning the ground state of a non-stoquastic quantum Hamiltonian in a rugged neural network landscape. SciPost Phys. 2021, 10, 147. [Google Scholar] [CrossRef]
  28. Cybenko, G. Approximation by superpositions of a sigmoidal function. Math. Control. Signals Syst. 1989, 2, 303–314. [Google Scholar] [CrossRef]
  29. Newman, M.E.; Barkema, G.T. Monte Carlo Methods in Statistical Physics; Clarendon Press: Tokyo, Japan, 1999. [Google Scholar]
  30. Wolfram, S. A New Kind of Science; Wolfram Media: Champaign, IL, USA, 2002; Volume 5. [Google Scholar]
  31. Janocha, K.; Czarnecki, W.M. On loss functions for deep neural networks in classification. arXiv 2017, arXiv:1702.05659. [Google Scholar] [CrossRef]
  32. El-Amir, H.; Hamdy, M. Deep Learning Fundamentals. In Deep Learning Pipeline; Springer: Berlin/Heidelberg, Germany, 2020; pp. 279–343. [Google Scholar]
  33. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  34. Padalko, M.A.; Shevchenko, Y.A.; Kapitan, V.Y.; Nefedev, K.V. Parallel Computing of Edwards-Anderson Model. Algorithms 2022, 15, 13. [Google Scholar] [CrossRef]
Figure 3. The proposed architecture of the CC2 deep neural network with three hidden layers h 1 = 2 N + 1 , h 2 = N , h 3 = N . The first hidden layer h 1 connected with the layer h 2 in the same way as bonds J k connect spins S j on a square lattice (neurons of the layer h 1 have 2 outputs, and h 2 have 5 inputs). The second hidden layer h 2 is connected with the third h 3 just like the neighboring spins are connected with each other on the square lattice (neurons of the layer h 2 have 5 outputs, and h 3 5 inputs, see Figure 1).
Figure 3. The proposed architecture of the CC2 deep neural network with three hidden layers h 1 = 2 N + 1 , h 2 = N , h 3 = N . The first hidden layer h 1 connected with the layer h 2 in the same way as bonds J k connect spins S j on a square lattice (neurons of the layer h 1 have 2 outputs, and h 2 have 5 inputs). The second hidden layer h 2 is connected with the third h 3 just like the neighboring spins are connected with each other on the square lattice (neurons of the layer h 2 have 5 outputs, and h 3 5 inputs, see Figure 1).
Entropy 24 00697 g003
Figure 6. Example of average energy calculation using replica-exchange MC as well as neural networks of different architectures (FC4, CC1 and CC2). The configuration of the calculated spin glass is shown in the lower right corner of the figure. Black circles denote spins, red rhombuses denote J = 1 bonds, and blue J = 1 bonds.
Figure 6. Example of average energy calculation using replica-exchange MC as well as neural networks of different architectures (FC4, CC1 and CC2). The configuration of the calculated spin glass is shown in the lower right corner of the figure. Black circles denote spins, red rhombuses denote J = 1 bonds, and blue J = 1 bonds.
Entropy 24 00697 g006
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Andriushchenko, P.; Kapitan, D.; Kapitan, V. A New Look at the Spin Glass Problem from a Deep Learning Perspective. Entropy 2022, 24, 697. https://doi.org/10.3390/e24050697

AMA Style

Andriushchenko P, Kapitan D, Kapitan V. A New Look at the Spin Glass Problem from a Deep Learning Perspective. Entropy. 2022; 24(5):697. https://doi.org/10.3390/e24050697

Chicago/Turabian Style

Andriushchenko, Petr, Dmitrii Kapitan, and Vitalii Kapitan. 2022. "A New Look at the Spin Glass Problem from a Deep Learning Perspective" Entropy 24, no. 5: 697. https://doi.org/10.3390/e24050697

APA Style

Andriushchenko, P., Kapitan, D., & Kapitan, V. (2022). A New Look at the Spin Glass Problem from a Deep Learning Perspective. Entropy, 24(5), 697. https://doi.org/10.3390/e24050697

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop