This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Stability of energy landscape for Ising models

Bruno Hideki Fukushima-Kimura111Faculty of Science, Hokkaido University, Japan.   Akira Sakai11footnotemark: 1 222https://orcid.org/0000-0003-0943-7842    Hisayoshi Toyokawa    Yuki Ueda Institute of Mathematics for Industry, Kyushu University, Japan.Department of Mathematics, Hokkaido University of Education, Japan.
Abstract

In this paper, we explore the stability of the energy landscape of an Ising Hamiltonian when subjected to two kinds of perturbations: a perturbation on the coupling coefficients and external fields, and a perturbation on the underlying graph structure. We give sufficient conditions so that the ground states of a given Hamiltonian are stable under perturbations of the first kind in terms of order preservation. Here by order preservation we mean that the ordering of energy corresponding to two spin configurations in a perturbed Hamiltonian will be preserved in the original Hamiltonian up to a given error margin. We also estimate the probability that the energy gap between ground states for the original Hamiltonian and the perturbed Hamiltonian is bounded by a given error margin when the coupling coefficients and local external magnetic fields of the original Hamiltonian are i.i.d. Gaussian random variables. In the end we show a concrete example of a system which is stable under perturbations of the second kind.

1 Introduction

Finding optimal solutions for combinatorial optimization problems, some of which are known to be NP-hard, is a very important problem. Among many possible approaches to such problems, the application of Ising models to solve real social problems has been getting attention due to its versatility (see [1]). More precisely, a given social combinatorial optimization problem can be mapped into a Hamiltonian HH on a graph G=(V,E)G=(V,E), whose expression is given by

H(σ)=b={x,y}EJbσxσyxVhxσx\displaystyle H(\sigma)=-\sum_{b=\{x,y\}\in E}J_{b}\sigma_{x}\sigma_{y}-\sum_{x\in V}h_{x}\sigma_{x}

for every Ising spin configuration σ{1,1}V\sigma\in\{-1,1\}^{V}, where {Jb}bE\{J_{b}\}_{b\in E} are coupling coefficients and {hx}xV\{h_{x}\}_{x\in V} are local magnetic fields. In that approach, an optimal solution for the intended combinatorial problem corresponds to a ground state (or global minimum) σG\sigma_{G} of HH, that is, σGargminH\sigma_{G}\in\mathrm{arg\,min}\,H. There are some well-known methods that can be applied to obtain a ground state. Implementing a Markov chain Monte Carlo (such as Glauber dynamics and stochastic cellular automata) is known as a way to find an approximation for the Gibbs distribution whose highest peaks correspond to the ground states of HH. We refer for details to [2, 3, 4, 5] and also [6].

However, as long as we use Ising machines or any computer to perform numerical simulations to find a ground state, we cannot avoid the error occurring due to the analog nature or the difficulty of representing real numbers (see [7]). Because of these reasons, we should incorporate the error coming from the coupling coefficients and local magnetic fields by introducing a perturbed Hamiltonian. Hence, our original Hamiltonian HH will be perturbed, originating a perturbed Hamiltonian HδH_{\delta} whose coupling coefficients and local magnetic fields have a maximal error δ\delta. Then, the following natural questions arise:

  1. (1)

    For any pair of configurations which are ordered in terms of energy with respect to the perturbed Hamiltonian HδH_{\delta}, is that ordering preserved in the original Hamiltonian HH, up to a given error margin?

  2. (2)

    Given a Hamiltonian HH with coupling coefficients and local magnetic fields distributed as i.i.d. Gaussian random variables, what is the probability that the energy gap in HH between two ground states respectively for HH and the corresponding perturbed Hamiltonian HδH_{\delta} is sufficiently small?

In addition to the above questions (1) and (2), the following problem is also important when using Ising machines and computers. It may be somewhat a waste of resources taking all coupling coefficients and local magnetic fields into account. It may be useful to “eliminate” vertices of a given graph whose contribution to the total energy is relatively small, in order to save memories of computers. Hence, we also have the following natural question:

  1. (3)

    Can we find a subset of a given graph such that for an arbitrary choice of configuration outside of that region, the energy variation can be controlled?

In this paper, we investigate the stability of energy landscape of a given Hamiltonian under perturbations from the view point of order preservation, aiming at answering the questions we addressed above. Thanks to the order preservation property, we can obtain better estimates for the success probability of finding a ground state compared to the result given in [7].

This paper is organized as follows. In Section 2, we provide a precise formulation for the questions we just posed and raise them again. In Section 3, we answer the questions (1’) and (2’) from Section 2. In Section 4, we provide an example together with a sufficient condition that guarantees a positive answer for question (3’).

2 Setting and the main questions

In this section, we introduce some necessary definitions and terminologies for discussing the stability of energy landscape. Further we also introduce the notion of order preservation for a perturbed system, which plays a central role in this paper. Here, order preservation means, roughly speaking, if we take a ground state for a perturbed Hamiltonian (implemented by a device) then it should be close to the ground state for an original Hamiltonian (intended mathematical problem) in energy, up to a given error margin.

Let us begin by introducing the precise setting. Let G=(V,E)G=(V,E) be a finite simple graph with the vertex set VV and the edge set EE. The so-called original Hamiltonian HH with coupling coefficients {Jb}bE\{J_{b}\}_{b\in E} and external magnetic fields {hx}xV\{h_{x}\}_{x\in V} on GG is defined by

H(σ)=b={x,y}EJbσxσyxVhxσx\displaystyle H(\sigma)=-\sum_{b=\{x,y\}\in E}J_{b}\sigma_{x}\sigma_{y}-\sum_{x\in V}h_{x}\sigma_{x} (2.1)

for each σ={σx}xV{1,1}V\sigma=\{\sigma_{x}\}_{x\in V}\in\{-1,1\}^{V}. Such a function HH can be regarded as a cost function of an intended problem. Given δ>0\delta>0, we denote by HδH_{\delta} the perturbed Hamiltonian with the coupling coefficients {Jb}bE\{J^{\prime}_{b}\}_{b\in E} and external fields {hx}xV\{h^{\prime}_{x}\}_{x\in V}, i.e.,

Hδ(σ)=b={x,y}EJbσxσyxVhxσx\displaystyle H_{\delta}(\sigma)=-\sum_{b=\{x,y\}\in E}J^{\prime}_{b}\sigma_{x}\sigma_{y}-\sum_{x\in V}h^{\prime}_{x}\sigma_{x} (2.2)

where the JbJ^{\prime}_{b}’s and hxh^{\prime}_{x}’s satisfy the bounds supb|JbJb|δ\sup_{b}\lvert J_{b}-J^{\prime}_{b}\rvert\leq\delta and supx|hxhx|δ\sup_{x}\lvert h_{x}-h^{\prime}_{x}\rvert\leq\delta. This perturbation will be often interpreted as a round-off in the following way. Let (Jb(1)Jb(2))(J_{b}^{(1)}J_{b}^{(2)}\dots) and (hx(1)hx(2))(h_{x}^{(1)}h_{x}^{(2)}\dots) be the binary expansions of the fractional parts of JbJ_{b} and hxh_{x}, i.e.,

Jb=Jb(0)+i1Jb(i)2i,hx=hx(0)+i1hx(i)2i\displaystyle J_{b}=J_{b}^{(0)}+\sum_{i\geq 1}\frac{J_{b}^{(i)}}{2^{i}},\quad h_{x}=h_{x}^{(0)}+\sum_{i\geq 1}\frac{h_{x}^{(i)}}{2^{i}} (2.3)

where Jb(0),hx(0)J_{b}^{(0)},h_{x}^{(0)}\in\mathbb{Z} and Jb(i),hx(i){0,1}J_{b}^{(i)},h_{x}^{(i)}\in\{0,1\} for i1i\geq 1. If we set Jb=Jb(0)+i=1NJb(i)2iJ_{b}^{\prime}=J_{b}^{(0)}+\sum_{i=1}^{N}\frac{J_{b}^{(i)}}{2^{i}} and hx=hx(0)+i=1Nhx(i)2ih_{x}^{\prime}=h_{x}^{(0)}+\sum_{i=1}^{N}\frac{h_{x}^{(i)}}{2^{i}} in the equation (2.2), then the error δ\delta can be taken as 2N2^{-N}. It means that the perturbed Hamiltonian HδH_{\delta} is obtained by rounding off the given parameters JbJ_{b}’s and hxh_{x}’s uniformly from the (N+1)(N+1)-th digit of their binary expansions.

The main purpose of this paper is to clarify the stability of the ground states for a given Hamiltonian under a perturbation in terms of order preservation. In this paper, we will answer the following questions:

  1. (1’)

    Find a δ>0\delta>0 corresponding to a given ε>0\varepsilon>0, so that, for any pair (σ,τ)(\sigma,\tau) that satisfies Hδ(σ)Hδ(τ)H_{\delta}(\sigma)\geq H_{\delta}(\tau), the ordering is preserved in HH up to the error margin εsupξ,η|H(ξ)H(η)|\varepsilon\sup_{\xi,\eta}\left\lvert H(\xi)-H(\eta)\right\rvert, i.e.,

    H(σ)H(τ)εsupξ,η|H(ξ)H(η)|.\displaystyle H(\sigma)\geq H(\tau)-\varepsilon\sup_{\xi,\eta}\left\lvert H(\xi)-H(\eta)\right\rvert. (2.4)

    Here, supξ,η|H(ξ)H(η)|\sup_{\xi,\eta}\left\lvert H(\xi)-H(\eta)\right\rvert is the total margin of the original Hamiltonian.

  2. (2’)

    Let (Ω,,)(\Omega,\mathcal{F},\mathbb{P}) be a probability space and let {Jb}bE\{J_{b}\}_{b\in E} and {hx}xV\{h_{x}\}_{x\in V} be mutually independent standard Gaussian random variables on this probability space. Estimate the probability that the energy gap in HH between ground states for HH and HδH_{\delta}, say σG\sigma_{G} and σ~G\tilde{\sigma}_{G}, respectively, is controled by the given error margin, explicitly,

    (0H(σ~G)H(σG)εsupξ,η|H(ξ)H(η)|).\displaystyle\mathbb{P}\left(0\leq H(\tilde{\sigma}_{G})-H(\sigma_{G})\leq\varepsilon\sup_{\xi,\eta}\left\lvert H(\xi)-H(\eta)\right\rvert\right). (2.5)

A different aspect of stability of a given system is to find a nontrivial subsystem so that the energy gap between any two spin configurations whose spins restricted to the subgraph coincide is bounded above by a given error margin. Also, at the same time, we require that the number of vertices that can be disregarded is at least of order NαN^{\alpha}, where N=|V|N=\lvert V\rvert and α[0,1)\alpha\in{[}0,1), so that such a number can go to infinity as NN\to\infty. In the later part of this paper, we answer the following question for a particular case:

  1. (3’)

    Let |V|=N\lvert V\rvert=N, and let {Jb}bE\{J_{b}\}_{b\in E} and {hx}xV\{h_{x}\}_{x\in V} be mutually independent standard Gaussian random variables. Find a subset V0VV_{0}\subset V for a given ε>0\varepsilon>0 and α[0,1)\alpha\in{[}0,1) such that

    (supσ,τ{1,1}N|H(σ)H(σV0,τVV0)|<εsupξ,η|H(ξ)H(η)|&CNα|VV0|<N)\displaystyle{\mathbb{P}}\left(\sup_{\sigma,\tau\in\{-1,1\}^{N}}\left\lvert H(\sigma)-H(\sigma_{V_{0}},\tau_{V\setminus V_{0}})\right\rvert<\varepsilon\sup_{\xi,\eta}\left\lvert H(\xi)-H(\eta)\right\rvert\;\&\;CN^{\alpha}\leq\lvert V\setminus V_{0}\rvert<N\right)

    is close to 11, where σV0{1,1}V0\sigma_{V_{0}}\in\{-1,1\}^{V_{0}} is the spin configuration σ\sigma restricted to V0V_{0} and τVV0{1,1}VV0\tau_{V\setminus V_{0}}\in\{-1,1\}^{V\setminus V_{0}} is the restriction of the spin configuration τ\tau to VV0V\setminus V_{0}.

Questions (1’), (2’) and (3’) above correspond to questions (1), (2) and (3) from Section 1, respectively. In Section 3, we investigate the first two questions above, where for the second one we adopt two different approaches. We obtain answers for question (2’) by means of a method involving the LL^{\infty}-distance and a graph’s structure approach, and we compare these two methods for three different graphs. Specifically, we consider sufficient conditions on the perturbation δ\delta to satisfy order preservation, and calculate the probability that such a sufficient condition holds. In Section 4, we obtain an answer for the question (3’) when the graph is a one-dimensional torus /N\mathbb{Z}/N\mathbb{Z} without external fields.

3 Stability under a Hamiltonian perturbation

This part is dedicated to provide solutions for questions (1’) and (2’) just posed in the end of the previous section. Before we proceed to the next sections, let us introduce the quantity RHR_{H} defined by

RHmaxξ,η|H(ξ)H(η)|,R_{H}\coloneqq\max_{\xi,\eta}\lvert H(\xi)-H(\eta)\rvert, (3.1)

which is defined whenever a Hamiltonian HH is given. Moreover, if G=(V,E)G=(V,E) is a finite simple graph, then we define kGk_{G} by

kG:=|E|+|V|.k_{G}:=|E|+|V|. (3.2)

Keeping in mind the mathematical setting introduced in the beginning of Section 2, let us start by showing that the order preservation property holds, that is, let us first answer the question (1’), which consists in finding a δ>0\delta>0 corresponding to a given ε>0\varepsilon>0 such that Hδ(σ)Hδ(τ)H_{\delta}(\sigma)\geq H_{\delta}(\tau) implies H(σ)H(τ)εRHH(\sigma)\geq H(\tau)-\varepsilon R_{H}; and later on, assuming some randomness on the spin-spin couplings and external fields, we adopt two different approaches to answer question (2’) and estimate the probability that the condition H(σ~G)H(σG)εRHH(\tilde{\sigma}_{G})-H(\sigma_{G})\leq\varepsilon R_{H} is satisfied.

In order to solve the second problem, we will adopt two distinct approaches: a method that relies on uniform estimates and a method where combinatorial estimates are considered instead, which will be presented in Sections 3.2 and 3.3, respectively. In the last part of this section, we compare these two methods and conclude that depending on the underlying graph structure of the problem, one of them will give us a better lower bound for the probability from equation (2.5).

3.1 Order preservation of energy

The answer for question (1’) from Section 2 is provided by Theorem 3.2, however, let us show first a preliminary result.

In [5], we have already established lower bounds for the total margin RHR_{H} of the Hamiltonian HH, but for the reader’s convenience we include its proof in the present paper.

Lemma 3.1 (See [5]).

Let us consider a finite simple graph G=(V,E)G=(V,E) and a Hamiltonian HH written in the form

H(σ)=b={x,y}EJbσxσyxVhxσx\displaystyle H(\sigma)=-\sum_{b=\{x,y\}\in E}J_{b}\sigma_{x}\sigma_{y}-\sum_{x\in V}h_{x}\sigma_{x}

for each σ{1,1}V\sigma\in\{-1,1\}^{V}. Then, we have

RHvH,where vHbJb2+xhx2.\displaystyle R_{H}\geq\sqrt{v_{H}},\qquad\text{where }v_{H}\coloneqq\sum_{b}J_{b}^{2}+\sum_{x}h_{x}^{2}.
Proof.

For any probability measure μ\mu on the configuration space {1,1}V\{-1,1\}^{V}, we have

RH(𝔼μ[H2]𝔼μ[H]2)1/2,\displaystyle R_{H}\geq\left(\mathbb{E}_{\mu}[H^{2}]-\mathbb{E}_{\mu}[H]^{2}\right)^{1/2},

where 𝔼μ\mathbb{E}_{\mu} stands for the expectation with respect to the probability measure μ\mu. If μ\mu is particularly chosen as the uniform distribution on {1,1}V\{-1,1\}^{V}, then we have

𝔼μ[H]\displaystyle\mathbb{E}_{\mu}[H] 12|V|σH(σ)=0,\displaystyle\coloneqq\frac{1}{2^{\lvert V\rvert}}\sum_{\sigma}H(\sigma)=0,{}

and

𝔼μ[H2]\displaystyle\mathbb{E}_{\mu}[H^{2}] 12|V|σH(σ)2\displaystyle\coloneqq\frac{1}{2^{\lvert V\rvert}}\sum_{\sigma}H(\sigma)^{2}
=12|V|σ(b={x,y}EJbσxσyxVhxσx)2\displaystyle=\frac{1}{2^{\lvert V\rvert}}\sum_{\sigma}\left(-\sum_{b=\{x,y\}\in E}J_{b}\sigma_{x}\sigma_{y}-\sum_{x\in V}h_{x}\sigma_{x}\right)^{2}
=12|V|σ(b,bEJbJbσxσyσxσy+x,xVhxhxσxσx)\displaystyle=\frac{1}{2^{\lvert V\rvert}}\sum_{\sigma}\left(\sum_{b,b^{\prime}\in E}J_{b}J_{b^{\prime}}\sigma_{x}\sigma_{y}\sigma_{x^{\prime}}\sigma_{y^{\prime}}+\sum_{x,x^{\prime}\in V}h_{x}h_{x^{\prime}}\sigma_{x}\sigma_{x^{\prime}}\right)
=12|V|σ(bEJb2+xVhx2)=vH.\displaystyle=\frac{1}{2^{\lvert V\rvert}}\sum_{\sigma}\left(\sum_{b\in E}J_{b}^{2}+\sum_{x\in V}h_{x}^{2}\right)={v_{H}}.

Therefore, RH(𝔼μ[H2]𝔼μ[H]2)1/2=vHR_{H}\geq\left(\mathbb{E}_{\mu}[H^{2}]-\mathbb{E}_{\mu}[H]^{2}\right)^{1/2}=\sqrt{v_{H}}. ∎

In order to prove the next result, it is convenient to consider the following notation introduced by [7]. For any Ising spin configurations σ\sigma and τ\tau, we consider the sets Dσ,τD_{\sigma,\tau} and Wσ,τW_{\sigma,\tau} defined by

Dσ,τ{xV:σxτx=1}D_{\sigma,\tau}\coloneqq\{x\in V:\sigma_{x}\tau_{x}=-1\}

and

Wσ,τ{{x,y}E:σxσyτxτy=1},W_{\sigma,\tau}\coloneqq\{\{x,y\}\in E:\sigma_{x}\sigma_{y}\tau_{x}\tau_{y}=-1\},

where the products σxτx\sigma_{x}\tau_{x} and σxσyτxτy\sigma_{x}\sigma_{y}\tau_{x}\tau_{y} are called the spin overlap and the link overlap, respectively.

Theorem 3.2.

Given ε>0\varepsilon>0 and configurations σ\sigma and τ\tau, if the condition

0<δkG12εvH\displaystyle 0<\delta k_{G}\leq\frac{1}{2}\varepsilon\sqrt{v_{H}}

is satisfied, then Hδ(σ)Hδ(τ)H_{\delta}(\sigma)\geq H_{\delta}(\tau) implies H(σ)H(τ)εRHH(\sigma)\geq H(\tau)-\varepsilon R_{H}.

Proof.

If we suppose that Hδ(σ)Hδ(τ)H_{\delta}(\sigma)\geq H_{\delta}(\tau), then, we have

H(τ)H(σ)\displaystyle H(\tau)-H(\sigma) =(H(τ)Hδ(τ))+Hδ(τ)Hδ(σ)+(Hδ(σ)H(σ))\displaystyle=\left(H(\tau)-H_{\delta}(\tau)\right)+H_{\delta}(\tau)-H_{\delta}(\sigma)+\left(H_{\delta}(\sigma)-H(\sigma)\right)
(Hδ(σ)H(σ))(Hδ(τ)H(τ))\displaystyle\leq\left(H_{\delta}(\sigma)-H(\sigma)\right)-\left(H_{\delta}(\tau)-H(\tau)\right)
b={x,y}E|JbJb||σxσyτxτy|+xV|hxhx||σxτx|\displaystyle\leq\sum_{b=\{x,y\}\in E}\left\lvert J_{b}-J_{b}^{\prime}\right\rvert\left\lvert\sigma_{x}\sigma_{y}-\tau_{x}\tau_{y}\right\rvert+\sum_{x\in V}\left\lvert h_{x}-h_{x}^{\prime}\right\rvert\left\lvert\sigma_{x}-\tau_{x}\right\rvert
=2bWσ,τ|JbJb|+2xDσ,τ|hxhx|\displaystyle=2\sum_{b\in W_{\sigma,\tau}}\left\lvert J_{b}-J_{b}^{\prime}\right\rvert+2\sum_{x\in D_{\sigma,\tau}}\left\lvert h_{x}-h_{x}^{\prime}\right\rvert
2δ(|Wσ,τ|+|Dσ,τ|).\displaystyle\leq 2\delta(\lvert W_{\sigma,\tau}\rvert+\lvert D_{\sigma,\tau}\rvert).

Since |Wσ,τ||E|\lvert W_{\sigma,\tau}\rvert\leq|E|, |Dσ,τ||V||D_{\sigma,\tau}\rvert\leq|V|, kG=|E|+|V|k_{G}=|E|+|V|, and RHvHR_{H}\geq\sqrt{v_{H}}, then, by our assumption, we obtain

H(τ)H(σ)2δkGεvHεRH.\displaystyle H(\tau)-H(\sigma)\leq 2\delta k_{G}\leq\varepsilon\sqrt{v_{H}}\leq\varepsilon R_{H}.

Therefore, the conclusion of this theorem follows. ∎

3.2 Stability of ground states: first approach

In the previous subsection, we did not assume any randomness on the spin-spin couplings JbJ_{b}’s and local external fields hxh_{x}’s. In this subsection, let us consider the same setting as stated in question (2’) from Section 2. Precisely speaking, we assume that {Jb}bE\{J_{b}\}_{b\in E} and {hx}xV\{h_{x}\}_{x\in V} are mutually independent random variables distributed according to a standard Gaussian distribution.

Under such assumptions, let us estimate the probability that the inequality H(σ~G)H(σG)εRHH(\tilde{\sigma}_{G})-H(\sigma_{G})\leq\varepsilon R_{H} holds, where ε\varepsilon is a given positive constant, by using a method that relies on uniform bounds with respect to certain spin configurations. In the following lemma, we provide an upper bound for the difference H(σ~G)H(σG)H(\tilde{\sigma}_{G})-H(\sigma_{G}).

Lemma 3.3.

Given δ>0\delta>0, if σG\sigma_{G} and σ~G\tilde{\sigma}_{G} are ground states for HH and HδH_{\delta}, respectively, then, we have

H(σ~G)H(σG)2δkG.H(\tilde{\sigma}_{G})-H(\sigma_{G})\leq 2\delta k_{G}. (3.3)
Proof.

It follows from the definition of a ground state that Hδ(σ~G)Hδ(σG)0H_{\delta}(\tilde{\sigma}_{G})-H_{\delta}(\sigma_{G})\leq 0, then

H(σ~G)H(σG)\displaystyle H(\tilde{\sigma}_{G})-H(\sigma_{G}) =(H(σ~G)Hδ(σ~G))+Hδ(σ~G)Hδ(σG)+(Hδ(σG)H(σG))\displaystyle=\left(H(\tilde{\sigma}_{G})-H_{\delta}(\tilde{\sigma}_{G})\right)+H_{\delta}(\tilde{\sigma}_{G})-H_{\delta}(\sigma_{G})+\left(H_{\delta}(\sigma_{G})-H(\sigma_{G})\right)
2HδH,\displaystyle\leq 2\lVert H_{\delta}-H\rVert_{\infty},

where \|\cdot\|_{\infty} stands for the uniform norm, as usual. Furthermore, for any spin configuration σ\sigma, we have

|Hδ(σ)H(σ)|\displaystyle\lvert H_{\delta}(\sigma)-H(\sigma)\rvert =|b={x,y}E(JbJb)σxσy+xV(hxhx)σx|\displaystyle=\left\lvert\sum_{b=\{x,y\}\in E}(J_{b}-J_{b}^{\prime})\sigma_{x}\sigma_{y}+\sum_{x\in V}(h_{x}-h_{x}^{\prime})\sigma_{x}\right\rvert
bE|JbJb|+xV|hxhx|\displaystyle\leq\sum_{b\in E}\left\lvert J_{b}-J_{b}^{\prime}\right\lvert+\sum_{x\in V}\left\rvert h_{x}-h_{x}^{\prime}\right\rvert
δ(|E|+|V|)=δkG.\displaystyle\leq\delta(\lvert E\rvert+\lvert V\rvert)=\delta k_{G}.

Then, HδHδkG\lVert H_{\delta}-H\rVert_{\infty}\leq\delta k_{G}, therefore, we conclude the proof. ∎

By the lemma above , it follows that

(H(σ~G)H(σG)εRH)(δkG12εRH),\displaystyle\mathbb{P}\left(H(\tilde{\sigma}_{G})-H(\sigma_{G})\leq\varepsilon R_{H}\right)\geq\mathbb{P}\left(\delta k_{G}\leq\frac{1}{2}\varepsilon R_{H}\right),

and by using the fact that RHvHR_{H}\geq\sqrt{v_{H}} (see Lemma 3.1), we conclude that

(H(σ~G)H(σG)εRH)(δkG12εvH).\mathbb{P}\left(H(\tilde{\sigma}_{G})-H(\sigma_{G})\leq\varepsilon R_{H}\right)\geq\mathbb{P}\left(\delta k_{G}\leq\frac{1}{2}\varepsilon\sqrt{v_{H}}\right). (3.4)

Finally, we have the following estimation for the probability that H(σ~G)H(σG)εRHH(\tilde{\sigma}_{G})-H(\sigma_{G})\leq\varepsilon R_{H} holds, which consists of one of the answers for the question (2’).

Theorem 3.4.

Let {Jb}bE\{J_{b}\}_{b\in E} and {hx}xV\{h_{x}\}_{x\in V} be mutually independent standard Gaussian random variables. It follows that

(H(σ~G)H(σG)εRH)1γ(kG;(2δkGε)2),\mathbb{P}\left(H(\tilde{\sigma}_{G})-H(\sigma_{G})\leq\varepsilon R_{H}\right)\geq 1-\gamma\left(k_{G};\left(\frac{2\delta k_{G}}{\varepsilon}\right)^{2}\right), (3.5)

where γ(s;x)\gamma(s;x) is the distribution function of the chi-square distribution with s>0s>0 degrees of freedom, that is,

γ(s;x)12s/2Γ(s/2)0xts/21et/2𝑑t\displaystyle\gamma(s;x)\coloneqq\frac{1}{2^{s/2}\Gamma(s/2)}\int_{0}^{x}t^{s/2-1}e^{-t/2}dt

for x0x\geq 0, and γ(s;x)0\gamma(s;x)\coloneqq 0 for x<0x<0.

Proof.

It follows from the above discussion that we have

(H(σ~G)H(σG)εRH)\displaystyle\mathbb{P}\left(H(\tilde{\sigma}_{G})-H(\sigma_{G})\leq\varepsilon R_{H}\right) (δkG12εvH)\displaystyle\geq\mathbb{P}\left(\delta k_{G}\leq\frac{1}{2}\varepsilon\sqrt{v_{H}}\right)
=(vH(2δkGε)2)\displaystyle=\mathbb{P}\left(v_{H}\geq\left(\frac{2\delta k_{G}}{\varepsilon}\right)^{2}\right)
=1(vH<(2δkGε)2).\displaystyle=1-\mathbb{P}\left(v_{H}<\left(\frac{2\delta k_{G}}{\varepsilon}\right)^{2}\right).

Since {Jb}bE\{J_{b}\}_{b\in E} and {hx}xV\{h_{x}\}_{x\in V} are mutually independent random variables distributed according to a standard Gaussian distribution, then the random variable vHv_{H} is distributed as the chi-square distribution with kGk_{G} degrees of freedom. Therefore,

(vH<(2δkGε)2)=γ(kG;(2δkGε)2).\displaystyle\mathbb{P}\left(v_{H}<\left(\frac{2\delta k_{G}}{\varepsilon}\right)^{2}\right)=\gamma\left(k_{G};\left(\frac{2\delta k_{G}}{\varepsilon}\right)^{2}\right).

Thus, we obtain the lower bound of the target probability. ∎

3.3 Stability of ground states: second approach

Before we proceed, let us point out the fundamental difference between the uniform approach and the current approach to solve question (2’). Note that, if we use the same computations as considered in the proof of Theorem 3.2 in the particular case where τ=σ~G\tau=\tilde{\sigma}_{G} and σ=σG\sigma=\sigma_{G} and use the fact that Hδ(σG)Hδ(σ~G)H_{\delta}(\sigma_{G})\geq H_{\delta}(\tilde{\sigma}_{G}), then it follows that

H(σ~G)H(σG)2δ(|WσG,σ~G|+|DσG,σ~G|).H(\tilde{\sigma}_{G})-H(\sigma_{G})\leq 2\delta(\lvert W_{\sigma_{G},\tilde{\sigma}_{G}}\rvert+\lvert D_{\sigma_{G},\tilde{\sigma}_{G}}\rvert). (3.6)

Recall that the proof of Theorem 3.4 fundamentally relied on the fact that, by using the LL^{\infty}-distance estimates, the left-hand side of equation (3.6) could be bounded above by 2δkG2\delta k_{G}. Note that the right-hand side of equation (3.6) is also bounded above by 2δkG2\delta k_{G}, therefore, let us explore the geometry of the underlying graph GG in order to see whether it is possible to obtain better bounds.

The value of |WσG,σ~G|+|DσG,σ~G|\lvert W_{\sigma_{G},\tilde{\sigma}_{G}}\rvert+\lvert D_{\sigma_{G},\tilde{\sigma}_{G}}\rvert depends on the underlying graph structure and the relationship between the ground states σG\sigma_{G} and σ~G\tilde{\sigma}_{G}. Therefore, we should check the value of |WσG,σ~G|+|DσG,σ~G|\lvert W_{\sigma_{G},\tilde{\sigma}_{G}}\rvert+\lvert D_{\sigma_{G},\tilde{\sigma}_{G}}\rvert for the intended problem. In general, we look for a uniform estimation for the value of |Wσ,τ|+|Dσ,τ|\lvert W_{\sigma,\tau}\rvert+\lvert D_{\sigma,\tau}\rvert for any σ\sigma and τ\tau since the ground states σG\sigma_{G} and σ~G\tilde{\sigma}_{G} in practice are unknown. First, let us show the following lemma.

Lemma 3.5.

For any two configurations σ\sigma and τ\tau, we have

|Wσ,τ|(degG)min{|Dσ,τ|,|VDσ,τ|},\displaystyle\lvert W_{\sigma,\tau}\rvert\leq(\deg G)\cdot\min\left\{\lvert D_{\sigma,\tau}\rvert,\lvert V\setminus D_{\sigma,\tau}\rvert\right\},

where degG\deg G stands for the maximum degree of GG.

Proof.

Let us assume that |Dσ,τ|=s\lvert D_{\sigma,\tau}\rvert=s, for some ss such that 0s|V|0\leq s\leq\lvert V\rvert. Then, let us enumerate Dσ,τD_{\sigma,\tau} as Dσ,τ={x1,,xs}VD_{\sigma,\tau}=\{x_{1},\dots,x_{s}\}\subset V, where xiVx_{i}\in V for each i=1,,si=1,\dots,s. Moreover, we have |VDσ,τ|=|V|s\lvert V\setminus D_{\sigma,\tau}\rvert=\lvert V\rvert-s, and therefore we can write VDσ,τ={y1,,y|V|s}VV\setminus D_{\sigma,\tau}=\{y_{1},\dots,y_{\lvert V\rvert-s}\}\subset V, where yiVy_{i}\in V for each i=1,,|V|si=1,\dots,\lvert V\rvert-s. By the definition of Dσ,τD_{\sigma,\tau}, we have σxiτxi=1\sigma_{x_{i}}\tau_{x_{i}}=-1 for every i=1,,si=1,\dots,s and σyjτyj=1\sigma_{y_{j}}\tau_{y_{j}}=1 for all j=1,,|V|sj=1,\dots,\lvert V\rvert-s. If {xi,xj}E\{x_{i},x_{j}\}\in E for distinct ii and jj in {1,,s}\{1,\dots,s\}, then

σxiσxjτxiτxj=(σxiτxi)(σxjτxj)=(1)2=1.\displaystyle\sigma_{x_{i}}\sigma_{x_{j}}\tau_{x_{i}}\tau_{x_{j}}=(\sigma_{x_{i}}\tau_{x_{i}})(\sigma_{x_{j}}\tau_{x_{j}})=(-1)^{2}=1.

Thus, {xi,xj}Wσ,τ\{x_{i},x_{j}\}\notin W_{\sigma,\tau}. In a similar way, we conclude that in case {yi,yj}E\{y_{i},y_{j}\}\in E for distinct ii and jj in {1,,|V|s}\{1,\dots,\lvert V\rvert-s\}, it follows that {yi,yj}Wσ,τ\{y_{i},y_{j}\}\notin W_{\sigma,\tau}. If {xi,yj}E\{x_{i},y_{j}\}\in E for some i{1,,s}i\in\{1,\dots,s\} and some j{1,,|V|s}j\in\{1,\cdots,\lvert V\rvert-s\}, then

σxiσyjτxiτyj=(σxiτxi)(σyjτyj)=(1)×1=1.\displaystyle\sigma_{x_{i}}\sigma_{y_{j}}\tau_{x_{i}}\tau_{y_{j}}=(\sigma_{x_{i}}\tau_{x_{i}})(\sigma_{y_{j}}\tau_{y_{j}})=(-1)\times 1=-1.

Hence, {xi,yj}Wσ,τ\{x_{i},y_{j}\}\in W_{\sigma,\tau}. It follows that

Wσ,τ={{x,y}E:x=xi,y=yj for some i,j}.\displaystyle W_{\sigma,\tau}=\{\{x,y\}\in E:x=x_{i},y=y_{j}\text{ for some }i,j\}.

Therefore, we have

|Wσ,τ|(degG)min{|Dσ,τ|,|VDσ,τ|}.\displaystyle\lvert W_{\sigma,\tau}\rvert\leq(\deg G)\min\{\lvert D_{\sigma,\tau}\rvert,\lvert V\setminus D_{\sigma,\tau}\rvert\}.

Proposition 3.6.

For any graph GG, let σ\sigma and τ\tau be two spin configurations. Then, we have

|Wσ,τ|+|Dσ,τ|(degG+1)|V|2.\displaystyle\lvert W_{\sigma,\tau}\rvert+\lvert D_{\sigma,\tau}\rvert\leq\frac{(\deg G+1)\lvert V\rvert}{2}. (3.7)
Proof.

Using Lemma 3.5, if |Dσ,τ||V|/2\lvert D_{\sigma,\tau}\rvert\leq\lvert V\rvert/2 then

|Dσ,τ|+|Wσ,τ|(degG+1)|Dσ,τ|degG+12|V|,\displaystyle\lvert D_{\sigma,\tau}\rvert+\lvert W_{\sigma,\tau}\rvert\leq(\deg G+1)\lvert D_{\sigma,\tau}\rvert\leq\frac{\deg G+1}{2}\lvert V\rvert,

otherwise, if |Dσ,τ|>|V|/2\lvert D_{\sigma,\tau}\rvert>\lvert V\rvert/2, it follows that

|Dσ,τ|+|Wσ,τ|\displaystyle\lvert D_{\sigma,\tau}\rvert+\lvert W_{\sigma,\tau}\rvert (degG)|VDσ,τ|+|Dσ,τ|\displaystyle\leq(\deg G)\lvert V\setminus D_{\sigma,\tau}\rvert+\lvert D_{\sigma,\tau}\rvert
=(degG)|V|(degG1)|Dσ,τ|\displaystyle=(\deg G)\lvert V\rvert-(\deg G-1)\lvert D_{\sigma,\tau}\rvert
(degG)|V|degG12|V|\displaystyle\leq(\deg G)\lvert V\rvert-\frac{\deg G-1}{2}\lvert V\rvert
=degG+12|V|.\displaystyle=\frac{\deg G+1}{2}\lvert V\rvert.

Thus, we have the following theorem which is another answer for the question (2’) (see Theorem 3.4 for an alternative approach to the question (2’)).

Theorem 3.7.

Let {Jb}bE\{J_{b}\}_{b\in E} and {hx}xV\{h_{x}\}_{x\in V} be mutually independent standard Gaussian random variables. Then, we have

(H(σ~G)H(σG)εRH)1γ(kG;(δ|V|(degG+1)ε)2).\mathbb{P}\left(H(\tilde{\sigma}_{G})-H(\sigma_{G})\leq\varepsilon R_{H}\right)\geq 1-\gamma\left(k_{G};\left(\frac{\delta\lvert V\rvert(\deg G+1)}{\varepsilon}\right)^{2}\right). (3.8)
Proof.

Analogously as in the proof of Theorem 3.4, it follows from equation (3.6), RHvHR_{H}\geq\sqrt{v_{H}} and Proposition 3.6 that

(H(σ~G)H(σG)εRH)\displaystyle\mathbb{P}\left(H(\tilde{\sigma}_{G})-H(\sigma_{G})\leq\varepsilon R_{H}\right){} (δεvH2(|WσG,σ~G|+|DσG,σ~G|))\displaystyle\geq\mathbb{P}\left(\delta\leq\frac{\varepsilon\sqrt{v_{H}}}{2(\lvert W_{\sigma_{G},\tilde{\sigma}_{G}}\rvert+\lvert D_{\sigma_{G},\tilde{\sigma}_{G}}\rvert)}\right)
=(vH(2δ(|WσG,σ~G|+|DσG,σ~G|)ε)2)\displaystyle=\mathbb{P}\left(v_{H}\geq\left(\frac{2\delta(\lvert W_{\sigma_{G},\tilde{\sigma}_{G}}\rvert+\lvert D_{\sigma_{G},\tilde{\sigma}_{G}}\rvert)}{\varepsilon}\right)^{2}\right)
(vH(δ|V|(degG+1)ε)2)\displaystyle\geq\mathbb{P}\left(v_{H}\geq\left(\frac{\delta\lvert V\rvert(\deg G+1)}{\varepsilon}\right)^{2}\right)
=1γ(kG;(δ|V|(degG+1)ε)2),\displaystyle=1-\gamma\left(k_{G};\left(\frac{\delta\lvert V\rvert(\deg G+1)}{\varepsilon}\right)^{2}\right),

where we used the fact that vHv_{H} is distributed according to a chi-square distribution with kGk_{G} degrees of freedom. ∎

3.4 Comparison between approaches

In the rest of this section, we compare the methods presented in Sections 3.2 and 3.3 passing through several examples to which we apply Proposition 3.6.

The first example is the case where we consider complete graphs including the SK model. If we consider complete graphs, then Theorem 3.7 provides us with better results if compared to Theorem 3.4.

Example 3.8.

If GG is a complete graph (that is, all vertices are connected to each other) with NN vertices, then we have

degG+12|V|=N22.\displaystyle\frac{\deg G+1}{2}\lvert V\rvert=\frac{N^{2}}{2}.

On the other hand, the value of kGk_{G} will be given by

kG:=|E|+|V|=N(N1)2+N=N(N+1)2.\displaystyle k_{G}:=\lvert E\rvert+\lvert V\rvert=\frac{N(N-1)}{2}+N=\frac{N(N+1)}{2}.

Therefore,

degG+12|V|<kG.\displaystyle\frac{\deg G+1}{2}\lvert V\rvert<k_{G}.

Hence the uniform upper bound for |Wσ,τ|+|Dσ,τ|\lvert W_{\sigma,\tau}\rvert+\lvert D_{\sigma,\tau}\rvert we obtained in Proposition 3.6 is always better than kGk_{G}. Furthermore, we can calculate the explicit value of |Wσ,τ|+|Dσ,τ|\lvert W_{\sigma,\tau}\rvert+\lvert D_{\sigma,\tau}\rvert when GG is a complete graph. From the proof of Lemma 3.5, by assuming that GG is a complete graph, we can say that |Wσ,τ|=|Dσ,τ|(|V||Dσ,τ|)\lvert W_{\sigma,\tau}\rvert=\lvert D_{\sigma,\tau}\rvert(\lvert V\rvert-\lvert D_{\sigma,\tau}\rvert). Therefore,

|Wσ,τ|+|Dσ,τ|=|Dσ,τ|(N+1|Dσ,τ|)(N+1)24,\displaystyle\lvert W_{\sigma,\tau}\rvert+\lvert D_{\sigma,\tau}\rvert=\lvert D_{\sigma,\tau}\rvert(N+1-\lvert D_{\sigma,\tau}\rvert)\leq\frac{(N+1)^{2}}{4},

and the proof of Theorem 3.7 implies

(H(σ~G)H(σG)εRH)\displaystyle\mathbb{P}\left(H(\tilde{\sigma}_{G})-H(\sigma_{G})\leq\varepsilon R_{H}\right) (vH(2δ(|WσG,σ~G|+|DσG,σ~G|)ε)2)\displaystyle\geq\mathbb{P}\left(v_{H}\geq\left(\frac{2\delta(\lvert W_{\sigma_{G},\tilde{\sigma}_{G}}\rvert+\lvert D_{\sigma_{G},\tilde{\sigma}_{G}}\rvert)}{\varepsilon}\right)^{2}\right)
(vHδ2(N+1)44ε2)\displaystyle\geq{\mathbb{P}}\left(v_{H}\geq\frac{\delta^{2}(N+1)^{4}}{4\varepsilon^{2}}\right)
=1γ(kG;δ2(N+1)44ε2).\displaystyle=1-\gamma\left(k_{G};\frac{\delta^{2}(N+1)^{4}}{4\varepsilon^{2}}\right).

The following example considers King’s graphs and Theorem 3.7 works better than Theorem 3.4 as well as the above example.

Example 3.9.

Let GG be an N×MN\times M King’s graph. The N×MN\times M King’s graph can be visualized as an N×MN\times M chessboard where each of its squares corresponds to a vertex of the graph, and each edge represents a legal move of a king in a chess game. In that way, the inner vertices of the graph have 88 neighbors each, while the vertices in the corners have 33 neighbors each, and each of the remaining vertices on the sides of the graph has 55 neighbors. For an N×MN\times M King’s graph, we have

degG+12|V|=92MN,\displaystyle\frac{\deg G+1}{2}\lvert V\rvert=\frac{9}{2}MN,

since degG=8\deg G=8. Moreover, we have

kG=|E|+|V|=5MN3(M+N)+2.\displaystyle k_{G}=\lvert E\rvert+\lvert V\rvert=5MN-3(M+N)+2.

If MM and NN are sufficiently large, then we have

degG+12|V|<kG.\displaystyle\frac{\deg G+1}{2}\lvert V\rvert<k_{G}.

In the following example, differently from the previous ones, we can see that the estimate provided by Theorem 3.4 suits better than that of Theorem 3.7.

Example 3.10.

If GG is a star graph with degree k3k\geq 3, that is, GG consists of one vertex placed in the center and other kk vertices connected only with the center, then

degG+12|V|=(k+1)22.\displaystyle\frac{\deg G+1}{2}\lvert V\rvert=\frac{(k+1)^{2}}{2}.

Furthermore, we have

kG=|E|+|V|=2k+1.\displaystyle k_{G}=\lvert E\rvert+\lvert V\rvert=2k+1.

Therefore, we obtain

degG+12|V|>kG.\displaystyle\frac{\deg G+1}{2}\lvert V\rvert>k_{G}.
Refer to caption
(a) Complete graph.
Refer to caption
(b) N×NN\times N King’s graph.
Refer to caption
(c) Star graph.
Figure 1: Minimal number of digits to be considered in the binary expansions of the parameters so that with probability higher than 99%99\% the difference H(σ~G)H(σG)H(\tilde{\sigma}_{G})-H(\sigma_{G}) represents a value smaller than 1%1\% of RHR_{H}, as a function of the size of the graph.

According to the above examples, we conclude that it is not always possible to guarantee that the uniform upper bound of |Wσ,τ|+|Dσ,τ|\lvert W_{\sigma,\tau}\rvert+\lvert D_{\sigma,\tau}\rvert provided by Proposition 3.6 works better than kG=|E|+|V|k_{G}=\lvert E\rvert+\lvert V\rvert. Thus, we may have to consider such bounds separately when considering different graphs in order to obtain an optimal estimate for the probability that inequality H(σ~G)H(σG)<εRHH(\tilde{\sigma}_{G})-H(\sigma_{G})<\varepsilon R_{H} holds.

Let us consider again the problem of stability where we take into account only a finite number of terms in the binary expansions of the parameters (Jb)bE(J_{b})_{b\in E} and (hx)xV(h_{x})_{x\in V} as we illustrated in the beginning of Section 2. In Figure 1, corresponding to the sizes of different graphs, we show the minimum number of digits necessary to be considered in the binary expansions of such parameters such that with probability at least 99%99\% the difference H(σ~G)H(σG)H(\tilde{\sigma}_{G})-H(\sigma_{G}) represents a value smaller than 1%1\% of RHR_{H}. On each plot we compare the different methods developed in this paper, where the first method corresponds to the estimate from Theorem 3.4 and the second method corresponds to the estimate from Theorem 3.7. In Figure 1(a), we also included a third estimate from Example 3.8 which is sharper and gives us better results when compared to the other methods. As we expected, the second method provides us with better results when compared to the first one for complete graphs and for N×NN\times N King’s graphs when NN is sufficiently large. On the other hand, for star graphs the first method is more appropriate, moreover, a certain discrepancy of performance is easily observed.

4 Stability under a perturbed graph

In this section, we consider the stability of energy landscape when a given spin system defined on a graph is compressed into a smaller subsystem. Differently from the previous sections, we fix a given Hamiltonian and we assume a sufficient condition that guarantees the existence of a nontrivial subset of the entire vertex set outside of which we can randomly assign any spin configuration and the energy of the system is kept under control up to a certain error margin.

Let G=(V,E)G=(V,E) be a finite simple graph, and let HH be the Hamiltonian on GG given by

H(σ)={x,y}EJx,yσxσy\displaystyle H(\sigma)=-\sum_{\{x,y\}\in E}J_{x,y}\sigma_{x}\sigma_{y}

for every configuration σ{1,1}V\sigma\in\{-1,1\}^{V}, where {Jx,y}{x,y}E\{J_{x,y}\}_{\{x,y\}\in E} is a collection of mutually independent standard Gaussian random variables. What we would like to show is that we can compress the whole system into a nontrivial subsystem so that the energy landscape of such subsystem is close to the original one up to a given error margin. More precisely, our goal is to find a class of examples for which given a positive constant ε\varepsilon, there is a positive δ\delta such that the subsystem V0=V0(δ)V_{0}=V_{0}(\delta) of VV, defined from the relation

VV0{xV:|Jx,y|<δ holds for every y such that {x,y}E},\displaystyle V\setminus V_{0}\coloneqq\left\{x\in V:\text{$\lvert J_{x,y}\rvert<\delta$ holds for every $y$ such that $\{x,y\}\in E$}\right\},

is non-trivial, has size comparable to the size of VV, and satisfies

supσ,η{1,1}V|H(σ)H(σV0,ηVV0)|<εRH\sup_{\sigma,\eta\in\{-1,1\}^{V}}\left\lvert H(\sigma)-H(\sigma_{V_{0}},\eta_{V\setminus V_{0}})\right\rvert<\varepsilon R_{H} (4.1)

with high probability, see Figure 2.

Refer to caption
Figure 2: We want to approximate the energy of a configuration σ\sigma defined in the whole vertex set VV by the energy of a configuration that coincides with σ\sigma in V0V_{0} and whose spins ηi\eta_{i}’s in the set VV0V\setminus V_{0} are arbitrary.

4.1 One-dimensional discrete torus

Let us solve the problem stated above in the particular case where the graph GG is a one-dimensional discrete torus.

Theorem 4.1.

Let G=(V,E)G=(V,E) be a one-dimensional discrete torus with NN vertices, that is, V={1,2,,N}V=\{1,2,\dots,N\} and E={{1,2},{2,3},,{N1,N},{N,1}}E=\{\{1,2\},\{2,3\},\dots,\{N-1,N\},\{N,1\}\}. Given ε>0\varepsilon>0, let δ\delta be a positive number such that δ<ε/2π\delta<\varepsilon/\sqrt{2\pi}. Then, if AA is a subset of the event {0<|V\V0|<N}\{0<|V\backslash V_{0}|<N\}, it follows that

({supσ,τ{1,1}N|H(σ)H(σV0,τVV0)|<εRH}A)(A)12πN(2π2δε)2{\mathbb{P}}\left(\Bigg{\{}\sup_{\sigma,\tau\in\{-1,1\}^{N}}\left\lvert H(\sigma)-H(\sigma_{V_{0}},\tau_{V\setminus V_{0}})\right\rvert<\varepsilon R_{H}\Bigg{\}}\cap A\right)\geq{\mathbb{P}}(A)-\frac{1-\frac{2}{\pi}}{N\left(\sqrt{\frac{2}{\pi}}-\frac{2\delta}{\varepsilon}\right)^{2}} (4.2)

holds for each N3N\geq 3. In particular, given constants C>0C>0 and α[0,1)\alpha\in{[}0,1), we have

({supσ,τ{1,1}N|H(σ)H(σV0,τVV0)|<εRH}{CNα|V\V0|<N})\displaystyle{\mathbb{P}}\left(\Bigg{\{}\sup_{\sigma,\tau\in\{-1,1\}^{N}}\left\lvert H(\sigma)-H(\sigma_{V_{0}},\tau_{V\setminus V_{0}})\right\rvert<\varepsilon R_{H}\Bigg{\}}\cap\Big{\{}CN^{\alpha}\leq|V\backslash V_{0}|<N\Big{\}}\right)
(1CN1αθ2)211+1+2θ3θ2Nθ2θN\displaystyle\geq\left(1-\frac{C}{N^{1-\alpha}\theta^{2}}\right)^{2}\frac{1}{1+\frac{1+2\theta-3\theta^{2}}{N\theta^{2}}}-\theta^{N}- 12πN(2π2δε)2\displaystyle\frac{1-\frac{2}{\pi}}{N\left(\sqrt{\frac{2}{\pi}}-\frac{2\delta}{\varepsilon}\right)^{2}} (4.3)

for NN sufficiently large, where

θ=δδeξ2/22π𝑑ξ.\theta=\int_{-\delta}^{\delta}\frac{e^{-\xi^{2}/2}}{\sqrt{2\pi}}d\xi. (4.4)

Before we follow to the proof of the result above, let us clarify the theorem by providing the reader with some practical results. Let us consider the particular case where C=1C=1 and α(0,1)\alpha\in(0,1). Corresponding to different values of ε\varepsilon and δ\delta, we obtain lower bounds for the probability that the size of V\V0V\backslash V_{0} is at least NαN^{\alpha} and condition (4.1) holds, see the table below.

Examples
NN ε\varepsilon δ\delta α\alpha Minimum size of V\V0V\backslash V_{0} Right-hand side of (4.1)
10810^{8} 0.050.05 0.01980.0198 0.40.4 15841584 0.8770.877
10810^{8} 0.050.05 0.01980.0198 0.50.5 10410^{4} 0.3610.361
10810^{8} 0.10.1 0.03980.0398 0.50.5 10410^{4} 0.8100.810
101210^{12} 0.010.01 0.003980.00398 0.50.5 10610^{6} 0.8110.811
101210^{12} 0.050.05 0.01990.0199 0.50.5 10610^{6} 0.9920.992
101210^{12} 0.10.1 0.03990.0399 0.50.5 10610^{6} 0.9980.998
101210^{12} 0.050.05 0.01990.0199 0.60.6 1.58×107\approx 1.58\times 10^{7} 0.8790.879
101210^{12} 0.050.05 0.01990.0199 0.650.65 6.31×107\approx 6.31\times 10^{7} 0.5630.563
Table 1: Applications of Theorem 4.1.

Let us observe that for any pair σ,τ\sigma,\tau of spin configurations, we have

|H(σ)H(σV0,τV\V0)|\displaystyle|H(\sigma)-H(\sigma_{V_{0}},\tau_{V\backslash V_{0}})| =\displaystyle= |xV0yV\V0{x,y}EJx,yσx(σyτy)+{x,y}V\V0{x,y}EJx,y(σxσyτxτy)|\displaystyle\left|\sum_{x\in V_{0}}\sum_{\begin{subarray}{c}y\in V\backslash V_{0}\\ \{x,y\}\in E\end{subarray}}J_{x,y}\sigma_{x}(\sigma_{y}-\tau_{y})+\sum_{\begin{subarray}{c}\{x,y\}\subseteq V\backslash V_{0}\\ \{x,y\}\in E\end{subarray}}J_{x,y}(\sigma_{x}\sigma_{y}-\tau_{x}\tau_{y})\right|
=\displaystyle= |xV0yV\V0{x,y}EJx,yσxσy(1σyτy)+{x,y}V\V0{x,y}EJx,yσxσy(1σxτxσyτy)|\displaystyle\left|\sum_{x\in V_{0}}\sum_{\begin{subarray}{c}y\in V\backslash V_{0}\\ \{x,y\}\in E\end{subarray}}J_{x,y}\sigma_{x}\sigma_{y}(1-\sigma_{y}\tau_{y})+\sum_{\begin{subarray}{c}\{x,y\}\subseteq V\backslash V_{0}\\ \{x,y\}\in E\end{subarray}}J_{x,y}\sigma_{x}\sigma_{y}(1-\sigma_{x}\tau_{x}\sigma_{y}\tau_{y})\right|
=\displaystyle= |yV\V0xV{x,y}EJx,yσxσy[𝟙xV0(1σyτy)+𝟙xV\V0(1σxτxσyτy)/2]|\displaystyle\left|\sum_{y\in V\backslash V_{0}}\sum_{\begin{subarray}{c}x\in V\\ \{x,y\}\in E\end{subarray}}J_{x,y}\sigma_{x}\sigma_{y}\left[\mathbbm{1}_{x\in V_{0}}(1-\sigma_{y}\tau_{y})+\mathbbm{1}_{x\in V\backslash V_{0}}(1-\sigma_{x}\tau_{x}\sigma_{y}\tau_{y})/2\right]\right|
\displaystyle\leq 2yV\V0xV{x,y}E|Jx,y|2δyV\V0deg(y).\displaystyle 2\sum_{y\in V\backslash V_{0}}\sum_{\begin{subarray}{c}x\in V\\ \{x,y\}\in E\end{subarray}}|J_{x,y}|\leq 2\delta\sum_{y\in V\backslash V_{0}}\text{deg}(y).

In particular, if GG is the one-dimensional torus as in Theorem 4.1, it follows that

supσ,τ{1,1}N|H(σ)H(σV0,τVV0)|4δ|V\V0|.\sup_{\sigma,\tau\in\{-1,1\}^{N}}\left\lvert H(\sigma)-H(\sigma_{V_{0}},\tau_{V\setminus V_{0}})\right\rvert\leq 4\delta|V\backslash V_{0}|. (4.5)

Now, let us prepare two lemmas in order to prove Theorem 4.1.

Lemma 4.2.

For RH=supξ,η|H(ξ)H(η)|R_{H}=\sup_{\xi,\eta}\lvert H(\xi)-H(\eta)\rvert, we have

2x=1N|Jx,x+1|4minx=1,,N|Jx,x+1|RH2x=1N|Jx,x+1|,\displaystyle 2\sum_{x=1}^{N}\lvert J_{x,x+1}\rvert-4\min_{x=1,\dots,N}\lvert J_{x,x+1}\rvert\leq R_{H}\leq 2\sum_{x=1}^{N}\lvert J_{x,x+1}\rvert, (4.6)

hence, with probability 1,

RH22πNas N approaches infinity.\displaystyle R_{H}\sim 2\sqrt{\frac{2}{\pi}}N\quad\text{as $N$ approaches infinity}.
Proof.

Without loss of generality, we assume minx|Jx,x+1|=|JN,1|\min_{x}\lvert J_{x,x+1}\rvert=\lvert J_{N,1}\rvert. Let us fix σ1=1\sigma_{1}=1. Then, depending on the sign of J1,2J_{1,2}, we can determine σ2\sigma_{2} to minimize (or maximize) H(σ)H(\sigma). We continue this procedure up to σN\sigma_{N} and we have

minσ{1,1}NH(σ)\displaystyle\min_{\sigma\in\{-1,1\}^{N}}H(\sigma) x=1N|Jx,x+1|+2minx=1,,N|Jx,x+1|,\displaystyle\leq-\sum_{x=1}^{N}\lvert J_{x,x+1}\rvert\ +2\min_{x=1,\dots,N}\lvert J_{x,x+1}\rvert,
maxσ{1,1}NH(σ)\displaystyle\max_{\sigma\in\{-1,1\}^{N}}H(\sigma) x=1N|Jx,x+1|2minx=1,,N|Jx,x+1|\displaystyle\geq\sum_{x=1}^{N}\lvert J_{x,x+1}\rvert\ -2\min_{x=1,\dots,N}\lvert J_{x,x+1}\rvert

(the additional terms exist if frustration exist at σN\sigma_{N} and σ1\sigma_{1}). Hence the inequality of the lemma is proven.

To show the last statement, we divide all terms by NN and use the law of large numbers for the folded normal distribution. ∎

Lemma 4.3.

If we assume N3N\geq 3, then it follows that

𝔼[|VV0|]=Nθ2\displaystyle\mathbb{E}\left[\lvert V\setminus V_{0}\rvert\right]=N\theta^{2}

and

𝔼[|VV0|2]=Nθ2(1+2θ3θ2+Nθ2).\displaystyle\mathbb{E}\left[\lvert V\setminus V_{0}\rvert^{2}\right]=N\theta^{2}\left(1+2\theta-3\theta^{2}+N\theta^{2}\right).
Proof.

For each i=1,,Ni=1,\dots,N, let us define a random variable XiX_{i} by letting

Xi={1if |Ji1,i|<δ and |Ji,i+1|<δ,0otherwise.\displaystyle X_{i}=\begin{cases}1&\text{if $\lvert J_{i-1,i}\rvert<\delta$ and $\lvert J_{i,i+1}\rvert<\delta$,}\\ 0&\text{otherwise}.\end{cases}

Then, by the definition of V0V_{0}, the condition iVV0i\in V\setminus V_{0} is equivalent to Xi=1X_{i}=1. Therefore, the expected value of the size of VV0V\setminus V_{0} will be given by

𝔼[|VV0|]=𝔼[i=1NXi]=i=1N𝔼[Xi]=Nθ2.\displaystyle\mathbb{E}\left[\lvert V\setminus V_{0}\rvert\right]=\mathbb{E}\left[\sum_{i=1}^{N}X_{i}\right]=\sum_{i=1}^{N}\mathbb{E}\left[X_{i}\right]=N\theta^{2}.

Furthermore, we write

|VV0|2=i=1NXi2+2i=1NXiXi+1+i=1Nj{i1,i,i+1}XiXj.\displaystyle\lvert V\setminus V_{0}\rvert^{2}=\sum_{i=1}^{N}X_{i}^{2}+2\sum_{i=1}^{N}X_{i}X_{i+1}+\sum_{i=1}^{N}\sum_{\begin{subarray}{c}j\notin\{i-1,\,i,\,i+1\}\end{subarray}}X_{i}X_{j}.

Here, the random variables XiX_{i} and Xi+1X_{i+1} are not mutually independent but we have

XiXi+1={1if |Ji1,i|<δ|Ji,i+1|<δ and |Ji+1,i+2|<δ,0otherwise.\displaystyle X_{i}X_{i+1}=\begin{cases}1&\text{if $\lvert J_{i-1,i}\rvert<\delta$,\ $\lvert J_{i,i+1}\rvert<\delta$ and $\lvert J_{i+1,i+2}\rvert<\delta$},\\ 0&\text{otherwise}.\end{cases}

Thus, it follows that the identity

𝔼[|VV0|2]\displaystyle\mathbb{E}\left[\lvert V\setminus V_{0}\rvert^{2}\right] =Nθ2+2Nθ3+N(N3)θ4\displaystyle=N\theta^{2}+2N\theta^{3}+N(N-3)\theta^{4}
=Nθ2(1+2θ3θ2+Nθ2)\displaystyle=N\theta^{2}\left(1+2\theta-3\theta^{2}+N\theta^{2}\right)

holds, and we complete the proof. ∎

Proof of Theorem 4.1.

Let us start by splitting the probability in the left-hand side of equation (4.2) as

\displaystyle{\mathbb{P}} ({supσ,τ{1,1}N|H(σ)H(σV0,τVV0)|<εRH}A)\displaystyle\left(\Bigg{\{}\sup_{\sigma,\tau\in\{-1,1\}^{N}}\left\lvert H(\sigma)-H(\sigma_{V_{0}},\tau_{V\setminus V_{0}})\right\rvert<\varepsilon R_{H}\Bigg{\}}\cap A\right)
({supσ,τ{1,1}N|H(σ)H(σV0,τVV0)|<εRH}{|1Nx=1N|Jx,x+1|2π|<2π2δε}A)\displaystyle\geq{\mathbb{P}}\left(\Bigg{\{}\sup_{\sigma,\tau\in\{-1,1\}^{N}}\left\lvert H(\sigma)-H(\sigma_{V_{0}},\tau_{V\setminus V_{0}})\right\rvert<\varepsilon R_{H}\Bigg{\}}\cap\left\{\left|\frac{1}{N}\sum_{x=1}^{N}|J_{x,x+1}|-\sqrt{\frac{2}{\pi}}\right|<\sqrt{\frac{2}{\pi}}-\frac{2\delta}{\varepsilon}\right\}\cap A\right)
=(supσ,τ{1,1}N|H(σ)H(σV0,τVV0)|<εRH|B)(B),\displaystyle={\mathbb{P}}\left(\sup_{\sigma,\tau\in\{-1,1\}^{N}}\left\lvert H(\sigma)-H(\sigma_{V_{0}},\tau_{V\setminus V_{0}})\right\rvert<\varepsilon R_{H}\,\middle|\,B\right)\,{\mathbb{P}}\left(B\right),

where BB is the event given by

B={|1Nx=1N|Jx,x+1|2π|<2π2δε}A.B=\left\{\left|\frac{1}{N}\sum_{x=1}^{N}|J_{x,x+1}|-\sqrt{\frac{2}{\pi}}\right|<\sqrt{\frac{2}{\pi}}-\frac{2\delta}{\varepsilon}\right\}\cap A. (4.7)

From equation (4.5), Lemma 4.2, and the fact that, under condition BB (subset of AA), minx=1,,N|Jx,x+1|δ\min_{x=1,\dots,N}|J_{x,x+1}|\leq\delta, it follows that the conditional probability above satisfies

(supσ,τ{1,1}N|H(σ)H(σV0,τVV0)|<εRH|B)\displaystyle{\mathbb{P}}\left(\sup_{\sigma,\tau\in\{-1,1\}^{N}}\left\lvert H(\sigma)-H(\sigma_{V_{0}},\tau_{V\setminus V_{0}})\right\rvert<\varepsilon R_{H}\,\middle|\,B\right) (2δ|VV0|<ε(x=1N|Jx,x+1|2minx=1,,N|Jx,x+1|)|B)\displaystyle\geq{\mathbb{P}}\left(2\delta\lvert V\setminus V_{0}\rvert<\varepsilon\left(\sum_{x=1}^{N}\lvert J_{x,x+1}\rvert-2\min_{x=1,\dots,N}\lvert J_{x,x+1}\rvert\right)\,\middle|\,B\right)
(2δε|VV0|<(x=1N|Jx,x+1|2δ)|B)\displaystyle\geq{\mathbb{P}}\left(\frac{2\delta}{\varepsilon}\lvert V\setminus V_{0}\rvert<\left(\sum_{x=1}^{N}\lvert J_{x,x+1}\rvert-2\delta\right)\,\middle|\,B\right)
=(2δε|VV0|+εN<1Nx=1N|Jx,x+1||B)\displaystyle={\mathbb{P}}\left(\frac{2\delta}{\varepsilon}\frac{\lvert V\setminus V_{0}\rvert+\varepsilon}{N}<\frac{1}{N}\sum_{x=1}^{N}\lvert J_{x,x+1}\rvert\,\middle|\,B\right)
(2δε<1Nx=1N|Jx,x+1||B)=1,\displaystyle\geq{\mathbb{P}}\left(\frac{2\delta}{\varepsilon}<\frac{1}{N}\sum_{x=1}^{N}\lvert J_{x,x+1}\rvert\,\middle|\,B\right)=1,

then

(supσ,τ{1,1}N|H(σ)H(σV0,τVV0)|<εRH|B)=1.{\mathbb{P}}\left(\sup_{\sigma,\tau\in\{-1,1\}^{N}}\left\lvert H(\sigma)-H(\sigma_{V_{0}},\tau_{V\setminus V_{0}})\right\rvert<\varepsilon R_{H}\,\middle|\,B\right)=1. (4.8)

The rest of the proof consists of estimating the probability of the event BB. Let us write

(B)(A)+(|1Nx=1N|Jx,x+1|2π|<2π2δε)1.{\mathbb{P}}(B)\geq{\mathbb{P}}(A)+{\mathbb{P}}\left(\left|\frac{1}{N}\sum_{x=1}^{N}|J_{x,x+1}|-\sqrt{\frac{2}{\pi}}\right|<\sqrt{\frac{2}{\pi}}-\frac{2\delta}{\varepsilon}\right)-1. (4.9)

It follows from Chebyshev’s inequality that

(|1Nx=1N|Jx,x+1|2π|2π2δε)σFG2N(2π2δε)2,{\mathbb{P}}\left(\left|\frac{1}{N}\sum_{x=1}^{N}|J_{x,x+1}|-\sqrt{\frac{2}{\pi}}\right|\geq\sqrt{\frac{2}{\pi}}-\frac{2\delta}{\varepsilon}\right)\leq\frac{\sigma_{FG}^{2}}{N\left(\sqrt{\frac{2}{\pi}}-\frac{2\delta}{\varepsilon}\right)^{2}}, (4.10)

where σFG2\sigma_{FG}^{2} is the variance of the folded Gaussian random variable Y=|J1,2|Y=|J_{1,2}| which is equal to 12π1-\frac{2}{\pi}. By using equations (4.8), (4.9) and (4.10), equation (4.2) follows.

In particular, if AA is the event given by A={CNα|V\V0|<N}A=\{CN^{\alpha}\leq|V\backslash V_{0}|<N\}. Note that

(A)=(|V\V0|CNα)(|V\V0|=N),{\mathbb{P}}(A)={\mathbb{P}}(|V\backslash V_{0}|\geq CN^{\alpha})-{\mathbb{P}}(|V\backslash V_{0}|=N), (4.11)

where (|V\V0|=N)=θN{\mathbb{P}}(|V\backslash V_{0}|=N)=\theta^{N}. By the Paley-Zygmund inequality and Lemma 4.3, we have

(|VV0|CNα)\displaystyle{\mathbb{P}}\left(\lvert V\setminus V_{0}\rvert\geq CN^{\alpha}\right) (1CNα𝔼[|VV0|])2𝔼[|VV0|]2𝔼[|VV0|2]\displaystyle\geq{\left(1-\frac{CN^{\alpha}}{\mathbb{E}[\lvert V\setminus V_{0}\rvert]}\right)^{2}}\frac{\mathbb{E}[\lvert V\setminus V_{0}\rvert]^{2}}{\mathbb{E}[\lvert V\setminus V_{0}\rvert^{2}]}
=(1CNαNθ2)211+1+2θ3θ2Nθ2\displaystyle={\left(1-\frac{CN^{\alpha}}{N\theta^{2}}\right)^{2}}\frac{1}{1+\frac{1+2\theta-3\theta^{2}}{N\theta^{2}}}

for NN sufficiently large, therefore, equation (4.1) holds. ∎

4.2 Generalizations

The most natural step in further investigations is to extend the results obtained in Section 4.1 to the case where we include i.i.d. standard Gaussian external fields, and also extend such results to a larger class of examples such as to a dd-dimensional torus or even to finite graphs with bounded degree. Note that, by assuming the absence of external fields, in the same way as we obtained inequality (4.5), one can show that

|H(σ)H(σV0,τV\V0)|2δyV\V0deg(y)|H(\sigma)-H(\sigma_{V_{0}},\tau_{V\backslash V_{0}})|\leq 2\delta\sum_{y\in V\backslash V_{0}}\text{deg}(y) (4.12)

holds for any graph. So, analogously as in the one-dimensional torus case, it is expected that if we find a lower bound for RHR_{H}, as we did in Lemma 4.2, which is comparable to the right-hand side of equation (4.12), then we may derive an extension of our results for a larger class of graphs. Some numerical results suggest that, for an Ising spin system in a dd-dimensional torus with i.i.d. standard Gaussian spin-spin couplings and without external fields, RHR_{H} is still of order NN, but it still lacks a rigorous proof of that observation due to the difficulty of dealing with frustrated configurations in a higher dimensional torus.

The simulations presented in this section were performed by using a modified version of the stochastic cellular automata algorithm studied in [5, 6] to estimate the maximum and minimum value of the Hamiltonian HH in order to find an approximation of RHR_{H} corresponding to different values of NN. Note that, in such plots, each dot represents the value of RHR_{H} (resp. RH/NR_{H}/N) corresponding to a torus with NN vertices for a realization of the random values of spin-spin couplings (i.i.d. standard Gaussian random variables). In the one-dimensional case (see Figure 3), we see that the value of RH/NR_{H}/N approximates the value 22/π1.59572\sqrt{2/\pi}\approx 1.5957, as expected due to Lemma 4.2.

Refer to caption
(a)
Refer to caption
(b)
Figure 3: The dependence of RHR_{H} with respect to the size of the system NN in the one dimensional case and its asymptotic behavior as NN grows.

Now, for the two and three dimensional cases (see Figure 4), when we consider larger values of NN, the value of RH/NR_{H}/N seems to approximate the values 2.5642.564 and 3.3293.329, respectively. Note that such simulated values represent lower bounds for the real value of the limit RH/NR_{H}/N as NN approaches infinity, so the true limits are still unknown. Furthermore, we conjecture that such limit exists in any dimension and the random variable RH/NR_{H}/N converges almost surely due to the fact that, in higher dimension, its simulated values seem to fluctuate less around an asymptotic limit as compared to the one-dimensional case.

It is straightforward to show that, for the dd-dimensional torus, we have

RH2k=1diV|Ji,i+𝐞𝐤|,R_{H}\leq 2\sum_{k=1}^{d}\sum_{i\in V}|J_{i,i+\mathbf{e_{k}}}|,

where 𝐞𝐤\mathbf{e_{k}} stands for the kk-th canonical vector of the dd-dimensional Euclidean space, then

lim supNRHN2d2π.\limsup_{N}\frac{R_{H}}{N}\leq 2d\sqrt{\frac{2}{\pi}}. (4.13)

Moreover, it follows from the fact that RHbJb2R_{H}\geq\sqrt{\sum_{b}J_{b}^{2}} (see Lemma 3.1) and the Cauchy–Schwarz inequality that

1Nb|Jb|RH.\frac{1}{\sqrt{N}}\sum_{b}|J_{b}|\leq R_{H}. (4.14)

Therefore, we see that there is still room for improvement and the need of rigorous proofs about the existence and determination of the limit limNRH/N\lim_{N\to\infty}R_{H}/N, originating a mathematical problem which is interesting by itself.

Refer to caption
(a) Two dimensional.
Refer to caption
(b) Two dimensional.
Refer to caption
(c) Three dimensional.
Refer to caption
(d) Three dimensional
Figure 4: The dependence of RHR_{H} with respect to the size of the system NN in the two dimensional and three dimensional cases and their asymptotic behavior as NN grows.

Acknowledgment

This work was supported by JST CREST Grant Number JP22180021, Japan. We would like to thank Takashi Takemoto and Normann Mertig of Hitachi, Ltd., for providing us with a stimulating platform for the weekly meeting at Global Research Center for Food & Medical Innovation (FMI) of Hokkaido University. We would also like to thank Hiroshi Teramoto of Kansai University, as well as Masamitsu Aoki, Yoshinori Kamijima, Katsuhiro Kamakura, Suguru Ishibashi and Takuka Saito of Mathematics Department, for valuable comments and encouragement at the aforementioned meetings at FMI.

References

  • [1] A. Lucas, Ising formulations of many NP problems. Front. Phys.  12 (2014): https://doi.org/10.3389/fphy.2014.00005.
  • [2] P. Dai Pra, B. Scoppola, E. Scoppola, Sampling from a Gibbs measure with pair interaction by means of PCA, J. Stat. Phys.  149 (2012): 722-737.
  • [3] B. Hajek, Cooling schedules for optimal annealing. Math. Oper. Res. 13 (2), (1988): 311-329.
  • [4] M. C. Robini, Theoretically ground acceleration techniques for simulated annealing, Handbook of Optimization (2013).
  • [5] S. Handa, K. Kamakura, Y. Kamijima, A. Sakai, Finding optimal solutions by stochastic cellular automata, arXiv:1906.06645.
  • [6] B. H. Fukushima-Kimura, S. Handa, K. Kamakura, Y. Kamijima, A. Sakai, Mixing time and simulated annealing for the stochastic cellular automata, arXiv:2007.11287v2.
  • [7] T. Albash, V. M. Mayor, I. Hen, Analog Errors in Ising Machines, arXiv:1806.03744.