Welcome to visit Communications in Theoretical Physics,
Quantum Physics and Quantum Information

Learning topological defects formation with neural networks in a quantum phase transition

  • Han-Qing Shi , 1 ,
  • Hai-Qing Zhang , 1, 2,
Expand
  • 1Center for Gravitational Physics, Department of Space Science, Beihang University, Beijing 100191, China
  • 2Peng Huanwu Collaborative Center for Research and Education, Beihang University, Beijing 100191, China

Author to which any correspondence should be addressed.

Received date: 2023-12-13

  Revised date: 2024-01-12

  Accepted date: 2024-03-11

  Online published: 2024-04-17

Copyright

© 2024 Institute of Theoretical Physics CAS, Chinese Physical Society and IOP Publishing

Abstract

Neural networks possess formidable representational power, rendering them invaluable in solving complex quantum many-body systems. While they excel at analyzing static solutions, nonequilibrium processes, including critical dynamics during a quantum phase transition, pose a greater challenge for neural networks. To address this, we utilize neural networks and machine learning algorithms to investigate time evolutions, universal statistics, and correlations of topological defects in a one-dimensional transverse-field quantum Ising model. Specifically, our analysis involves computing the energy of the system during a quantum phase transition following a linear quench of the transverse magnetic field strength. The excitation energies satisfy a power-law relation to the quench rate, indicating a proportional relationship between the excitation energy and the kink numbers. Moreover, we establish a universal power-law relationship between the first three cumulants of the kink numbers and the quench rate, indicating a binomial distribution of the kinks. Finally, the normalized kink-kink correlations are also investigated and it is found that the numerical values are consistent with the analytic formula.

Cite this article

Han-Qing Shi , Hai-Qing Zhang . Learning topological defects formation with neural networks in a quantum phase transition[J]. Communications in Theoretical Physics, 2024 , 76(5) : 055101 . DOI: 10.1088/1572-9494/ad3227

1. Introduction

One of the most challenging problems in modern physics is the so-called many-body problem. In its quantum version—quantum many-body physics, the exponential complexity of the states in the Hilbert space makes the strongly correlated systems difficult to deal with [1]. Only limited analytical solutions are amenable to a few simple models [2]. Therefore, resorting to a powerful algorithm by using suitable parameters to generalize the physical states becomes an intriguing direction. In this way, many numerical methods are proposed, such as the density matrix renormalization group [3] and quantum Monte Carlo [4, 5]. Although these methods work well for some specific problems, they lack universalities.
Fortunately, neural network methods have more universalities. The same neural network can be used to represent the states or to study the dynamical processes for various systems, such as those with different dimensions or with different interactions. The recent state-of-the-art neural networks have been shown to provide high efficient representations of such complex states, making the overwhelming complexity computationally tractable [6, 7]. Except for the success in the industrial applications, such as the image and speech recognitions [8], the autonomous driving, and the game of Go [9], neural networks have been widely adopted to study a broad spectrum of areas in physics, ranging from statistical and quantum physics to high energy and cosmology [1014].
Among these successful applications in physical sciences, the more challenging task is to use neural networks to study nonequilibrium problems. Recently, an algorithm of artificial neural networks was proposed to solve the unitary time evolutions in a quantum many-body system [15]. Later developments in this direction can be found in [1627]. At the same time, matrix product states (MPS) is used to represent one-dimensional quantum systems [28], while tensor networks can represent higher-dimensional systems [29, 30]. By using the density matrix renormalization group (DMRG) and some delicate techniques, these methods can simulate quantum systems with high accuracy while requiring less computational complexity. Moreover, tensor networks can represent both finite and infinite systems, while neural networks typically can only represent systems that are not too large. Although MPS and tensor networks have also been developed for handling quantum many-body problems, neural networks still have their advantages. Restricted Boltzmann Machines (RBM) have been proven to accurately represent states with massive entanglement [31] and quantum topological states [32] which cannot be represented efficiently in terms of MPS or tensor-network states. In nonequilibrium dynamics, a common issue is the emergence of universalities during the critical dynamics of a phase transition. One of the most well-known challenges in this regard is the formation of topological defects. It is stated that topological defects will arise in the course of a phase transition with symmetry breaking of the system due to the celebrated Kibble–Zurek mechanism (KZM) [33, 34]. Number density of topological defects was found to satisfy a universal power-law with respect to the quench rate. The formation of topological defects and the KZM have been widely examined in various numerical simulations and experiments, including in quantum phase transition [35, 36], in quantum field theory with matrix product states [37], in AdS/CFT correspondence [3841], in programmable quantum simulators [42, 43] and in D-wave devices [44], to name some relevant references.
In [15], the machine learning methods was merely applied in the unitary dynamics without phase transitions. Critical dynamics, i.e., the dynamics across the critical point of a phase transition is more complex and has richer phenomena [45]. Critical slowing down near the phase transition point may invalidate the applicability of this method. In this paper, we extend the machine learning methods introduced in [15] to study the nonequilibrium process of critical dynamics in a one-dimensional transverse field quantum Ising model (TFQIM). TFQIM is a widely used model to study the phase transitions of one-dimensional spin chain and has been extensively studied analytically or experimentally such as in [4648]. Therefore, TFQIM is a very suitable testbed to check the representing accuracy of neural networks and the robustness of machine learning methods. Specifically, we study the time evolutions of the energy, universal statistics and correlations of the topological defects formed in the TFQIM after quantum phase transition induced by a quench. In particular, we quench the strength of the transverse magnetic field to drive the system from a paramagnetic state into a ferromagnetic state, during which the topological defects, i.e. the kinks where the polarization of the spins changes their directions, will form due to the KZM. In the machine learning we introduce the RBM as a representation of the quantum state for TFQIM. The RBM is a kind of neural networks with two layers of neurons, i.e. visible layer and hidden layer (see figure 1). In order to solve the ground state and the time evolution of the system, the stochastic reconfiguration (SR) method and time-dependent variational Monte Carlo (VMC) approach [49] are utilized, respectively. We find that time evolutions of the energy expectation value from the neural networks are perfectly consistent with the results reported in [46]. After the quench, the excited energy of the system is found to satisfy a power-law relation against the quench rate, which reveals the proportional relationship between the excitation energy and the kink numbers. Besides, the counting statistics of the kink numbers satisfy the Poisson binomial distributions introduced previously in [36, 50]. By computing the first three cumulants of the kink pair numbers, we find that they satisfy universal power-law scalings to the quench rate consistent with the theoretical predictions. Additionally, we compute the kink-kink correlations at the end of the quench. The numerical data match the analytic formula presented in [47] very well. Therefore, our results show a very high accuracy of neural networks to investigate the critical dynamics of TFQIM.
Figure 1. Map of the structure of the RBM. RBM has a hidden layer with hi (i = 1, ⋯ M) as hidden neurons and a visible layer with sj (j = 1, ⋯ N) as visible neurons. The lines linking the hidden points and visible points represent the interactions. There is no intralayer interactions in the hidden layer or visible layer themselves.

2. Results

2.1. Quantum quench of TFQIM

We study the formation of topological defects in one-dimensional TFQIM, with the Hamiltonian of a spin chain of N sites in a transverse magnetic field [46],
$\begin{eqnarray}H=-J\displaystyle \sum _{i=1}^{N}\left({\sigma }_{i}^{z}{\sigma }_{i+1}^{z}+h{\sigma }_{i}^{x}\right),\end{eqnarray}$
where ${\sigma }_{i}^{z}$ and ${\sigma }_{i}^{x}$ are the Pauli matrices at the site i in the z and x directions, while J and h respectively denote the coupling strengths between the nearest-neighbor sites and transverse magnetic field strength. We consider the periodic boundary conditions (PBC) for this spin chain by imposing ${\vec{\sigma }}_{N+1}={\vec{\sigma }}_{1}$ with even N for simplicity. There exists a quantum phase transition at the critical point ∣hc∣ = 1. Without loss of generality, we will only focus on the regime h ≥ 0. As h ≫ 1, the ground state is in a paramagnetic state. On the other hand, while h ≪ 1 the ground state is in the two-degenerated ferromagnetic states with spins up or down along the z-direction. In the limit of N → ∞ the energy gap at the critical point hc=1 tends to zero, which resembles the critical slowing down. Therefore, if the system goes from the regime h > 1 to the regime h < 1, it is impossible to cross the critical point without exciting the system. As a result, the system will end up in the configurations that spins will point up or point down in some finite domains. Consequently, the kinks, which are a kind of topological defects in one dimension, form.
Conventionally, one evolves the system by linearly quenching the transverse magnetic field as,
$\begin{eqnarray}h(t)=-\displaystyle \frac{t}{{\tau }_{Q}},\qquad t\in [-T,0]\end{eqnarray}$
in which τQ is the quench rate. In the initial time T/τQ ≫ 1, we prepare the ground state in a strong transverse magnetic field, thus the spins will all point up along the transverse field direction. The system will evolve according to the quench profile (2), cross the critical point and then end at h(t = 0) = 0. During this phase transition, kinks for the polarization directions of spins will form according to KZM. The analytic solutions to the dynamics of the Hamiltonian (1) were previously carried out in [46]. In the limit of N → ∞ , the average number density of kinks $\langle \hat{{ \mathcal N }}\rangle $ (where $\hat{{ \mathcal N }}\,=\tfrac{1}{2}{\sum }_{i=1}^{N}(1-{\sigma }_{i}^{z}{\sigma }_{i+1}^{z})$ is the kink number operator) are found to satisfy a universal power-law to the quench rate as $\langle \hat{{ \mathcal N }}\rangle \propto {\tau }_{Q}^{-1/2}$, consistent with the KZM.

2.2. Time evolution of energy expectation value

Utilizing the machine learning method (see section 4 ‘Materials and Methods’), we consider the time evolution of a one-dimensional TFQIM (1) under a linear quench (2) through the critical point with various quench rates. We set PBC to the spin chains, thus satisfying the even parity [46]. The coupling strengths between the nearest-neighbor site J and the lattice spacing were set to be unit. The number of lattice points we take is N = 100 and the time period is t ∈ [ − 2τQ, 0], corresponding to the strength of transverse magnetic field h from 2 to 0. The hidden site we take is M = 4N. The time step we use is Δt = 1/1000. Then, the system will evolve under the time-dependent VMC method. In the left plot of figure 2, we show the evolution of the energy expectation value for three different quench rates τQ = 1, 4 and 8. We can see that the solutions from the machine learning methods (solid lines) and the solutions from the analytic methods in [46] (black dashed lines) match each other very well. In particular, we find that earlier on, the energies satisfy a linear function as E(t)/(JN) ≈ t/τQ, which can be readily derived from the Hamiltonian (1) and the quench profile (2), since earlier on, the term $h{\sigma }_{i}^{x}$ makes a dominant contribution to the Hamiltonian. Then, roughly in the intermediate time t/τQ ∈ [ − 1, − 0.5], the faster quench (smaller τQ) increases the energy faster, leading to a higher final energy as the quench ends (t/τQ = 0). The relative errors for both methods at the end of the quench are relatively small, specifically the errors are (1.41%, 0.51%, 0.02%) for τQ = (1, 4, 8), respectively, as the inset plot shows in the left plot of figure 2.
Figure 2. (Left) Time evolution of the energy expectation value E(t)/J per lattice point with respect to the reduced time t/τQ. The solid lines represent the evolution of the energy for three different quench rates from machine learning methods, while the black dashed lines are the comparing results from the methods in [46]. The inset plot exhibits the relative errors for both methods, from which we see that they match each other very well. (Right) The excitation energy density ΔE with respect to τQ at the end of the quench. The fitting line has a power-law scaling as ${\rm{\Delta }}E\approx 0.245\times {\tau }_{Q}^{-0.56}$, which indicates that the excitation energy is proportional to the kink numbers.
Since each kink has the same energy at h = 0, we can determine the number of kinks by analyzing their excitation energy. In the right plot of figure 2, we show the excitation energy density ΔE at t = 0 with respect to the quench rate τQ. The excitation energy density is defined as
$\begin{eqnarray}{\rm{\Delta }}E=(E-{E}_{0})/({JN})\end{eqnarray}$
where E0 is the ground state energy as h = 0 in the equation (1). The fitting line is roughly ${\rm{\Delta }}E\approx 0.245\times {\tau }_{Q}^{-0.56}$. We see that the power-law ${\tau }_{Q}^{-0.56}$ is close to the power-law predicted by KZM, and that the mean kink number density is proportional to the quench rate as $\langle \hat{{ \mathcal N }}\rangle \propto {\tau }_{Q}^{-0.5}$. We will investigate this in the following in detail. This consistency demonstrates our assumption that the excitation energy is proportional to the kink numbers.

2.3. Statistics of kink numbers beyond KZM

The topological defects, i.e., kinks form in the course of the quantum phase transitions due to the KZM. It predicts that the power-law scalings of the mean value of the kink numbers to the quench rate is proportional to ${\tau }_{Q}^{-d\nu /(1+\nu z)}$, where d = 1 for the one-dimensional spin chain; ν and z are respectively the static and dynamic critical exponents. For the TFQIM, they are ν = z = 1. Therefore, the theoretical prediction of the power-law scaling between the mean kink number and the quench rate is $\langle \hat{{ \mathcal N }}\rangle \propto {\tau }_{Q}^{-1/2}$. However, the KZM predicts only the power-law relation between the average number of the kinks and the quench rate. From [36] the distribution of kinks in the TFQIM is assumed to satisfy the Poisson binomial distributions. Consequently, the fluctuations away from the mean of the kink numbers, such as the variance, satisfy a universal power-law scaling to the quench rate as well. Therefore, it will be vital to study these universal power-laws beyond KZM by virtue of the neural networks. These universal power-laws beyond KZM can be achieved by computing the higher order cumulants of the kink numbers. Since we adopt the PBC in the spin chain, the outcomes of the kink numbers are all even [51, 52]. Thus, in practice we compute the higher cumulants for the numbers of kink pairs, i.e., ${\hat{{ \mathcal N }}}_{P}=\hat{{ \mathcal N }}/2$. According to [36, 50], all the cumulants should be proportional to the mean, i.e., ${\kappa }_{q}\propto \langle {\hat{{ \mathcal N }}}_{P}\rangle $ where q are positive integers. Due to the cost of time and the limited resources of the computer, we only compute the first three cumulants, i.e. κq with q = (1, 2, 3) of the kink pair numbers.3(3For the sites number N = 100 and the neural network size α = M/N = 4, it would take approximately 48 hours to run the code once with τQ = 16. If increasing the size of N or α, or bigger τQ, the consumption of time will be much expensive. Therefore, considering the accuracy of the results and the cost of time, we set N = 100, α = 4 and the greatest quench rate is τQ = 16.) Specifically, these cumulants can be expressed as
$\begin{eqnarray}{\kappa }_{1}=\langle {\hat{{ \mathcal N }}}_{P}\rangle ,\end{eqnarray}$
$\begin{eqnarray}{\kappa }_{2}=\langle {\hat{{ \mathcal N }}}_{P}^{2}\rangle -\langle {\hat{{ \mathcal N }}}_{P}{\rangle }^{2},\end{eqnarray}$
$\begin{eqnarray}{\kappa }_{3}=\langle {({\hat{{ \mathcal N }}}_{P}-\langle {\hat{{ \mathcal N }}}_{P}\rangle )}^{3}\rangle .\end{eqnarray}$
In other words, κ1 is the mean value of the kink pair numbers; κ2 is the variance of the kink pair numbers while κ3 is related to the skewness of the kink pair numbers as ${\kappa }_{3}=\mathrm{Skew}({\hat{{ \mathcal N }}}_{P}){\kappa }_{2}^{3/2}$.
In figure 3, we show the first three cumulants of the kink pair number distribution κq, (q = 1, 2, 3) as a function of the quench rate τQ. The reference line (dashed line) represents the theoretical scaling ${\kappa }_{q}\propto {\tau }_{Q}^{-0.5}$. The numerical data from the neural network methods are shown in the circles (κ1), triangles (κ2) and squares (κ3). The error bars indicate the standard errors in the statistics. The solid lines are obtained from the analytic methods used in [46]. In order to not cause confusion, we stress that the solid lines are not the fitting lines for the numerical data from the neural networks method. By fitting these cumulants, the power-law scaling ${\tau }_{Q}^{\beta }$ for κ1,2,3 have the powers β = (−0.58, −0.61, −0.64), respectively. The fitting range is taken from τQ ≥ 3 for κ1,2 and 3 ≤ τQ ≤ 8 for κ3. For fast quench (relatively small quench rates), i.e. τQ ≲ 3, the scaling relations will deviate from the KZM power-laws because of the finite size effect [36, 50].
Figure 3. Double logarithmic plots for the cumulants κq (q = 1, 2, 3) of the kink pair number distributions with respect to quench rate τQ. The circles, triangles and squares are the data from the neural network methods, while the solid lines are from the analytic method introduced in [46]. The dashed line is the reference line with theoretical power-law ${\tau }_{Q}^{-0.5}$. The error bars represent the standard errors.
From figure 3, we can see that the results of the power β, in particular for κ1 and κ2, is a little bit away from the theoretical power −1/2, since we are using the finite size N = 100. In the limit of large N, this power-law would tend to the theoretical predictions as [36, 46, 50] demonstrated. Nevertheless, the power ${\beta }_{{\kappa }_{1}}=-0.58$ for N = 100 in our case is perfectly consistent with the numerical results in [35], where the authors also adopted N = 100 sites. It should be noted that κ3 deviates relatively largely compared to the first two cumulants. This is because κ3 is more sensitive to the sites number N. Similar results in experiments, numerical simulations and holography were reported previously in [5053]. In appendix A, we compare the numerical results for the cumulants κ1,2,3 with different N's. And we find that as N increases, the power-law behavior will improve and get closer to the theoretical predictions.
Moreover, from figure 3 we notice that the value of the ratio κ2/κ1 is roughly κ2/κ1 ≈ 0.30 from the fitting, which is very close to the theoretical predictions ${\kappa }_{2}/{\kappa }_{1}\,=(2-\sqrt{2})/2\approx 0.29$ in [36, 50]. Besides, the ratio κ3/κ1 ≈ 0.042 in our case is a little bit away from the predictions ${\kappa }_{3}/{\kappa }_{1}=(1-3/\sqrt{2}+2/\sqrt{3})\approx 0.033$ in [36, 50]. The reason is that κ3 is more sensitive to the sites number N as we stated above. It is expected that increasing the sites number would improve this problem.

2.4. Kink-kink correlations

Correlations between kinks are the delicate quantum features for the Ising model [47], and it has been successfully tested on a one-dimensional transverse field Ising chain using programmable quantum annealer [48]. Therefore, it will be very important to check the accuracy of the neural network methods by studying the kink-kink correlation CrKK. Following [48], we define the normalized connected kink-kink correlation as
$\begin{eqnarray}{C}_{r}^{{KK}}=\displaystyle \frac{1}{N}\displaystyle \sum _{i=1}^{N}\displaystyle \frac{\left(\langle {K}_{i}{K}_{i+r}\rangle -{\bar{n}}^{2}\right)}{{\bar{n}}^{2}},\end{eqnarray}$
where ${K}_{i}=\tfrac{1}{2}(1-{\sigma }_{i}^{z}{\sigma }_{i+1}^{z})$ is the kink number operator at the site i, $\bar{n}=\langle \hat{{ \mathcal N }}\rangle /N$ is the average kink number density. The analytic formula for the normalized kink-kink correlation can be read out from [47] as,
$\begin{eqnarray}{C}_{r}^{{KK}}=\alpha \displaystyle \frac{\hat{\xi }}{l}{\left(\displaystyle \frac{r}{l}\right)}^{2}{{\rm{e}}}^{-3\pi {(r/l)}^{2}}-{{\rm{e}}}^{-2\pi {(r/\hat{\xi })}^{2}},\end{eqnarray}$
in which $\alpha =\tfrac{9747\pi }{3200}\approx 9.57$ is a numerical factor, $\hat{\xi }$ is the typical KZ correlation length that indicates the average distance between kinks, i.e., $\hat{\xi }=N/\langle \hat{{ \mathcal N }}\rangle =1/\bar{n}$ and $l\,=\hat{\xi }\sqrt{1+{\left(3\mathrm{log}{\tau }_{Q}/(4\pi )\right)}^{2}}$ is the correlation range. In figure 4, we show the numerical results of CrKK against the normalized distance $r/\hat{\xi }$ for various quenches τQ = 3, 4 and 5. The green line is from the analytic formula equation (8) with τQ = 5. For τQ = 3, 4 and 5, the differences between the analytic formula equation (8) are tiny. In order to distinguish the figure clearly, we only plot the analytic lines for τQ = 5. We see that the numerical data will collapse together and match the analytic line very well. In particular, as $r/\hat{\xi }\to 0$ the correlation goes to −1 and there is a peak at around $r/\hat{\xi }\approx 0.5$, which were already demonstrated in [47, 48]. The correlation tends to zero as the distance is very large, which satisfies the physical intuitions. Therefore, we see that the neural network methods can also uncover the delicate quantum properties of the Ising model.
Figure 4. The normalized kink-kink correlations CrKK against the normalized distances $r/\hat{\xi }$. The green line is from the analytic formula in equation (8) with τQ = 5. The error bars in the numerical data represent the standard error.

3. Discussions

We have realized the time evolutions of the energy expectation value, the universal statistics of the topological defects numbers and the kink-kink correlations in a quantum phase transition of a TFQIM by virtue of the neural networks. The results were found to satisfy theoretical predictions. Thus, it numerically verifies that the neural network methods not only can extend but also work well for the critical dynamics in a quantum phase transition.
In our paper, we used the network size α = M/N = 4. However, we find that even with smaller α, the numerical results will still be consistent with theoretical predictions. In appendix B, we compare the effects of network size α = 2 and α = 4 on the kink number cumulants and the energy expectation values. We see that for α = 2 the numerical results are in agreement with the those for α = 4. All of these suggested that the neural networks method perform very well and showed high accuracy for the quantum phase transition. We expect that the neural networks and machine learning methods may shed light on further complex dynamics in quantum many-body physics.

4. Materials and methods

4.1. Machine learning with neural networks

The states of the quantum system can be characterized by wave functions. One can approximate the wave functions with neural network quantum states (NQS) as ${{\rm{\Psi }}}_{{NN}}(s,{ \mathcal W })$, where s = (s1, s2...,sN) denotes the spin configurations at each site directed in the z-direction basis, and ${ \mathcal W }$ is the neural network parameter. Different ${ \mathcal W }$'s correspond to different quantum states of the system. Therefore, suitable values of ${ \mathcal W }$ can make the NQS describe the states with topological defects after the phase transition from paramagnetic state to ferromagnetic state. Our goal is find out this parameter ${ \mathcal W }$ to simulate the process of phase transition and quantify the statistics and correlations of the kinks. In this work, we adopt the RBM as neural networks which work well for the quantum spin chain [15]. It consists of two layers, one is the visible layers with N visible neurons sj and the other is the hidden layers with M hidden neurons hi, refer to figure 1. There are interactions between hidden and visible layers, but no intralayer interactions between themselves. According to the structure of the neural networks, the quantum state can be described as
$\begin{eqnarray}\begin{array}{l}{{\rm{\Psi }}}_{{NN}}(s,{ \mathcal W })=\displaystyle \sum _{\{h\}}\exp \left[\displaystyle \sum _{j}{a}_{j}{s}_{j}\right.\\ \,\left.+\,\displaystyle \sum _{i}{b}_{i}{h}_{i}+\displaystyle \sum _{i,j}{w}_{{ij}}{h}_{i}{s}_{j}\right],\end{array}\end{eqnarray}$
where s = {sj} denotes the spin configurations and ${ \mathcal W }=\{a,b,w\}$ are the neural network parameters. When we set ${ \mathcal W }$ to be complex, the wave function $\Psi$NN can represent the amplitudes and phases of the states. The parameter hi = {−1, 1} are the hidden variables. Since there is no intralayer interactions in the hidden layers, the hidden variables can be traced out explicitly in the first step. Therefore, the wave function becomes
$\begin{eqnarray}\begin{array}{l}{{\rm{\Psi }}}_{{NN}}(s,{ \mathcal W })\\ =\,{{\rm{e}}}^{\displaystyle \sum _{j}{a}_{j}{s}_{j}}\displaystyle \prod _{i=1}^{M}2\cosh [{b}_{i}+\displaystyle \sum _{j}{w}_{{ij}}{s}_{j}].\end{array}\end{eqnarray}$
In the machine learning, the number of independent parameters contributes to the computational complexity of the wave functions $\Psi$NN. Fortunately, we can reduce the number of parameters from the symmetry of the system or some physical constraints. When a system has a symmetry under some operation $\hat{T}$, we can impose a constraint for the neural network ${{\rm{\Psi }}}_{{NN}}(s,{ \mathcal W })={{\rm{\Psi }}}_{{NN}}(\hat{T}s,{ \mathcal W })$. This constraint will largely reduce the number of independent network parameters and facilitate the computations. In our case, lattice translational symmetry is used to reduce the parameters from the order of magnitude ${ \mathcal O }(M\times N)$ to ${ \mathcal O }(M)$. In this paper, we need to evolve the system from an initial ground state with strong transverse magnetic field to a state without any transverse magnetic field. In order to obtain this initial state, we have to train a beginning wave function whose parameters are random complex numbers with machine learning algorithms, in which this learning is realized through minimizing the expectation value of the energy ⟨E⟩ = ⟨$\Psi$NNH∣$\Psi$NN⟩/⟨$\Psi$NN∣$\Psi$NN⟩ with SR method.
After preparing the initial ground state, the system will evolve according to the quench profile (2). Therefore, the wave function should also depend on time. To this end, we render the neural network parameters to be functions of time. Then, the parameters will be computed at every time step with the time-dependent VMC method, by minimizing the distances δ between the exact time evolution and the approximate variational evolution
$\begin{eqnarray}\begin{array}{l}\delta =\mathrm{dist}\left[{\partial }_{t}{{\rm{\Psi }}}_{{NN}}\left({ \mathcal W }(t)\right),-{\rm{i}}{H}{{\rm{\Psi }}}_{{NN}}\left({ \mathcal W }(t)\right)\right],\end{array}\end{eqnarray}$
where,
$\begin{eqnarray}\begin{array}{l}\mathrm{dist}\left[{\rm{\Psi }}^{\prime} ,{\rm{\Psi }}\right]\equiv \arccos \sqrt{\displaystyle \frac{\langle {\rm{\Psi }}^{\prime} | {\rm{\Psi }}\rangle \langle {\rm{\Psi }}| {\rm{\Psi }}^{\prime} \rangle }{\langle {\rm{\Psi }}^{\prime} | {\rm{\Psi }}^{\prime} \rangle \langle {\rm{\Psi }}| {\rm{\Psi }}\rangle }}.\end{array}\end{eqnarray}$
For more detailed information, we refer readers to [15]. In order to decrease the influence of the noise from the Monte Carlo method, some regularization methods are needed. We utilize the singular value decomposition (SVD) regularization method introduced in [18]. This method can eliminate the effect from Monte Carlo sampling noise and make the program more stable. More specifically, we decompose the matrix with SVD, and remove the singular value less than a tolerence λ ∼ 10−8. As the approximate wave function is obtained, the average number of topological defects $\langle {{\rm{\Psi }}}_{{NN}}| \hat{{ \mathcal N }}| {{\rm{\Psi }}}_{{NN}}\rangle $ can be computed through the kink number operator $\hat{{ \mathcal N }}$.

Conflict of interest

The authors declare that they have no conflict of interest.

Appendix A. Finite size effects on the cumulants

In [46], it has been analytically confirmed that as the site number N goes to infinity, the scaling of the mean kink number to the quench rate will be proportional to ${\tau }_{Q}^{-0.5}$ which satisfies the KZM’s prediction. Later, this scaling law was extended to higher cumulants of the kink numbers [36, 50], and it was found that all cumulants should be parallel to each other in the large N limit.
In figure A1, we study the effects of N on the first three cumulants κ1,2,3 of the kink pair numbers. Specifically, we show four plots with N = 40, 60, 80 and 100, respectively. From the figure, we can see that as N = 40, the first cumulant κ1 is already very straight in the range τQ ≥ 3, except the last point deviates a little bit away from the line. However, as N grows we see that the last point of κ1 will align in a straight line which is roughly parallel to the theoretical predictions ${\tau }_{Q}^{-0.5}$. For κ2, we see that as N = 40, κ2 deviates from the straight line ${\tau }_{Q}^{-0.5}$ greatly as τQ is relatively big. However, for N = 60, 80 and 100, the behavior of κ2 improves and tends to be parallel with the theoretical prediction. For κ3, as N = 40 it even disappears for greater τQ since in this case κ3 was computed to be negative. However, as N increases we see that more and more data of κ3 will appear and tend to align in the straight line ${\tau }_{Q}^{-0.5}$. But as we stated in the main text, κ3 needs much more data to get a better behavior. N = 100 is still not big enough for a perfect κ3. However, as we have shown in figure A1, we can vividly see that as N increases, the behavior of κ1,2,3 will certainly improve and tend to be parallel with the line ${\tau }_{Q}^{-0.5}$.
Figure A1. The first three cumulants against the quench rate for various sites numbers N. The dashed lines are the theoretical predictions with the power-law scaling ${\tau }_{Q}^{-0.5}$. The numerical data are from the neural network methods.

Appendix B. Effects of neural network size on the cumulants and the energy expectation value

In this part, we will show the effects of the neural network size on the cumulants and the energy expectation value. We define the neural network size α as the ratio between the number of hidden neurons and the number of visible neurons, i.e, α = M/N. In figure B1, we show the effects of α on the cumulants. Specifically, we take α = 2 and 4. From figure B1, we see that in the range of α ∈ [2, 4], it has minor effects to the cumulants κ1 and κ2. However, α has strong effect on κ3. It is clear that as α = 4, the behavior of κ3 is much better than that of α = 2. Considering the cost of time and the resources of the computer, we take α = 4 in the main text.
Figure B1. The first three cumulants against the quench rate for various neural network sizes α. The asterisks represent the cumulants for α = 2 while the circles for α = 4. We take N = 100 in this figure.
Figure B2 shows the time evolutions of the relative errors of the energy for α = 2 and 4. The relative error is defined as ε = (Eneural networkEexact)/Eexact, where Eexact is from the analytic methods in [46]. From figure B2 we see that for α ∈ [2, 4], there is no much distinctions between the relative error of the energy. Thus, from both figure B1 and figure B2, it suggests that the neural network size α = 4 we take is precise enough for the TFQIM.
Figure B2. Time evolutions of the relative errors of the energies ε for α = 2 (left) and α = 4 (right).

We appreciate the illuminating discussions with Dr. Peng-Zhang He. This work was partially supported by the National Natural Science Foundation of China (Grants No.11875095 and 12175008).

1
Fetter A L, Walecka J D 2012 Quantum Theory of Many-Particle Systems North Chelmsford, MA Courier Corporation

2
Albeverio S, Gesztesy F, Hoegh-Krohn R, Holden H 2012 Solvable Models in Quantum Mechanics Berlin Springer

3
White S R 1992 Density matrix formulation for quantum renormalization groups Phys. Rev. Lett. 69 2863

DOI

4
Ceperley D, Alder B 1986 Quantum monte carlo Science 231 555

DOI

5
Troyer M, Wiese U-J 2005 Computational complexity and fundamental limitations to fermionic quantum monte carlo simulations Phys. Rev. Lett. 94 170201

DOI

6
Carleo G, Cirac I, Cranmer K, Daudet L, Schuld M, Tishby N, Vogt-Maranto L, Zdeborová L 2019 Machine learning and the physical sciences Rev. Mod. Phys. 91 045002

DOI

7
Karagiorgi G, Kasieczka G, Kravitz S, Nachman B, Shih D 2022 Machine learning in the search for new fundamental physics Nat. Rev. Phys. 4 399 412

DOI

8
LeCun Y, Bengio Y, Hinton G 2015 Deep learning Nature 521 436

DOI

9
Silver D 2016 Mastering the game of go with deep neural networks and tree search Nature 529 484

DOI

10
Lam J, You Y-Z 2021 Machine learning statistical gravity from multi-region entanglement entropy Phys. Rev. Res. 3 043199

DOI

11
Raissi M, Perdikaris P, Karniadakis G E 2019 Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations J. Comput. Phys. 378 686

DOI

12
Carrasquilla J, Melko R G 2017 Machine learning phases of matter Nat. Phys. 13 431

DOI

13
Shi H-Q, Sun X-Y, Zeng D-F 2019 Neural-network quantum state of transverse-field ising model Commun. Theor. Phys. 71 1379

DOI

14
Park C-Y, Kastoryano M J 2020 Geometry of learning neural quantum states Physical Review Research 2 023232

DOI

15
Carleo G, Troyer M 2017 Solving the quantum many-body problem with artificial neural networks Science 355 602

DOI

16
Schmitt M, Heyl M 2020 Quantum many-body dynamics in two dimensions with artificial neural networks Phys. Rev. Lett. 125 100503

DOI

17
Hartmann M J, Carleo G 2019 Neural-network approach to dissipative quantum many-body dynamics Phys. Rev. Lett. 122 250502

DOI

18
Hofmann D, Fabiani G, Mentink J H, Carleo G, Sentef M A 2022 Role of stochastic noise and generalization error in the time propagation of neural-network quantum states SciPost Phys. 12 165

DOI

19
Schmitt M, Heyl M 2018 Quantum dynamics in transverse-field ising models from classical networks SciPost Physics 4 013

DOI

20
Schmitt M, Rams M M, Dziarmaga J, Heyl M, Zurek W H 2022 Quantum phase transition dynamics in the two-dimensional transverse-field Ising model Sci. Adv. 8 abl6850

DOI

21
Czischek S, Gärttner M, Gasenzer T 2018 Quenches near ising quantum criticality as a challenge for artificial neural networks Phys. Rev. B 98 024311

DOI

22
Fabiani G, Mentink J 2019 Investigating ultrafast quantum magnetism with machine learning SciPost Physics 7 004

DOI

23
Luo D, Chen Z, Carrasquilla J, Clark B K 2022 Autoregressive neural network for simulating open quantum systems via a probabilistic formulation Phys. Rev. Lett. 128 090501

DOI

24
Reh M, Schmitt M, Gärttner M 2021 Time-dependent variational principle for open quantum systems with artificial neural networks Phys. Rev. Lett. 127 230501

DOI

25
Donatella K, Denis Z, Boité A L, Ciuti C 2022 Dynamics with autoregressive neural quantum states: application to critical quench dynamics arXiv:2209.03241

26
Gutiérrez I L, Mendl C B 2022 Real time evolution with neural-network quantum states Quantum 6 627

DOI

27
Yuan D, Wang H-R, Wang Z, Deng D-L 2021 Solving the liouvillian gap with artificial neural networks Phys. Rev. Lett. 126 160401

DOI

28
Schollwöck U 2011 The density-matrix renormalization group in the age of matrix product states Ann. Phys. 326 96

DOI

29
Verstraete F, Murg V, Cirac J I 2008 Matrix product states, projected entangled pair states, and variational renormalization group methods for quantum spin systems Adv. Phys. 57 143

DOI

30
Gu Z-C, Wen X-G 2009 Tensor-entanglement-filtering renormalization approach and symmetry-protected topological order Phys. Rev. B 80 155131

DOI

31
Deng D-L, Li X, Das Sarma S 2017 Quantum entanglement in neural network states Phys. Rev. X 7 021021

DOI

32
Deng D-L, Li X, Das Sarma S 2017 Machine learning topological states Phys. Rev. B 96 195145

DOI

33
Kibble T W 1976 Topology of cosmic domains and strings J. Phys. A: Math. Gen. 9 1387

DOI

34
Zurek W H 1985 Cosmological experiments in superfluid helium? Nature 317 505

DOI

35
Zurek W H, Dorner U, Zoller P 2005 Dynamics of a quantum phase transition Phys. Rev. Lett. 95 105701

DOI

36
Campo A Del 2018 Universal statistics of topological defects formed in a quantum phase transition Phys. Rev. Lett. 121 200601

DOI

37
Gillman E, Rajantie A 2018 Kibble zurek mechanism of topological defect formation in quantum field theory with matrix product states Phys. Rev. D 97 094505

DOI

38
Sonner J, del Campo A, Zurek W H 2015 Universal far-from-equilibrium dynamics of a holographic superconductor Nature Commun. 6 7406

DOI

39
Chesler P M, Garcia-Garcia A M, Liu H 2015 Defect formation beyond Kibble-Zurek mechanism and holography Phys. Rev. X 5 021015

DOI

40
Zeng H-B, Xia C-Y, Zhang H-Q 2021 Topological defects as relics of spontaneous symmetry breaking from black hole physics J. High Energy Phys. JHEP03(2021)136

DOI

41
Li Z-H, Shi H-Q, Zhang H-Q 2022 Holographic topological defects in a ring: role of diverse boundary conditions J. High Energy Phys. JHEP05(2022)056

DOI

42
Keesling A 2019 Quantum kibble-zurek mechanism and critical dynamics on a programmable rydberg simulator Nature 568 207

DOI

43
Ebadi S 2021 Quantum phases of matter on a 256-atom programmable quantum simulator Nature 595 227

DOI

44
Weinberg P, Tylutki M, Rönkkö J M, Westerholm J, Åström J A, Manninen P, Törmä P, Sandvik A W 2020 Scaling and diabatic effects in quantum annealing with a d-wave device Phys. Rev. Lett. 124 090502

DOI

45
Sachdev S 2011 Quantum Phase Transitions Cambridge Cambridge University Press

46
Dziarmaga J 2005 Dynamics of a quantum phase transition: Exact solution of the quantum ising model Phys. Rev. Lett. 95 245701

DOI

47
Nowak R J, Dziarmaga J 2021 Quantum kibble-zurek mechanism: Kink correlations after a quench in the quantum ising chain Phys. Rev. B 104 075448

DOI

48
King A D 2022 Coherent quantum annealing in a programmable 2,000 qubit Ising chain Nature Phys. 18 1324

DOI

49
Sorella S 2001 Generalized lanczos algorithm for variational quantum monte carlo Phys. Rev. B 64 024512

DOI

50
Gómez-Ruiz F J, Mayo J J, Campo A Del 2020 Full counting statistics of topological defects after crossing a phase transition Phys. Rev. Lett. 124 240602

DOI

51
Cui J-M, Gómez-Ruiz F J, Huang Y-F, Li C-F, Guo G-C, del Campo A 2020 Experimentally testing quantum critical dynamics beyond the kibble-zurek mechanism Communications Physics 3 1

DOI

52
Bando Y, Susa Y, Oshiyama H, Shibata N, Ohzeki M, Gómez-Ruiz F J, Lidar D A, Suzuki S, Del Campo A, Nishimori H 2020 Probing the universality of topological defect formation in a quantum annealer: Kibble-zurek mechanism and beyond Phys. Rev. Res. 2 033369

DOI

53
del Campo A, Gómez-Ruiz F J, Li Z-H, Xia C-Y, Zeng H-B, Zhang H-Q 2021 Universal statistics of vortices in a newborn holographic superconductor: beyond the Kibble-Zurek mechanism J. High Energy Phys. JHEP06(2021)061

DOI

Outlines

/