This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Delocalization for the 3–D discrete random Schrödinger operator at weak disorder

Westin King Robert C. Kirby  and  Constanze Liaw Department of Mathematics, Baylor University, One Bear Place #\#97328, Waco, TX 76798-7328, USA Westin¯\underline{\,\,\,\,}King@baylor.edu, Robert¯\underline{\,\,\,\,}Kirby@baylor.edu, Constanze¯\underline{\,\,\,\,}Liaw@baylor.edu
Abstract.

We apply a recently developed approach [18] to study the existence of extended states for the three dimensional discrete random Schrödinger operator at small disorder. The conclusion of delocalization at small disorder agrees with other numerical and experimental observations (see e.g. [17]). Further the work furnishes a verification of the numerical approach and its implementation.
Not being based on scaling theory, this method eliminates problems due to boundary conditions, common to previous numerical methods in the field. At the same time, as with any numerical experiment, one cannot exclude finite-size effects with complete certainty. Our work can be thought of as a new and quite different use of Lanczos’ algorithm; a posteriori tests to show that the orthogonality loss is very small.
We numerically track the “bulk distribution” (here: the distribution of where we most likely find an electron) of a wave packet initially located at the origin, after iterative application of the discrete random Schrödinger operator.

Key words and phrases:
Discrete random Schrödinger operator, Anderson localization, Extended states, Lanczos algorithm
2010 Mathematics Subject Classification:
47A16, 47B80, 81Q10
The work of Liaw and King was supported by DMS-1261687.

1. Introduction

Consider the discrete three dimensional Schrödinger operator, given by:

(1) f(x)=|i|=1(f(x+i)f(x)),\displaystyle-\bigtriangleup f(x)=-\sum\limits_{|i|=1}(f(x+i)-f(x)),

when ii is of the form (i1,i2,i3)T,ik(i_{1},i_{2},i_{3})^{T},i_{k}\in{\mathbb{Z}}, and consider an element δi(x)\delta_{i}(x) of l2(3)l^{2}({\mathbb{Z}}^{3}) given by

δi(x)={1x=i3,0else.\displaystyle\delta_{i}(x)=\left\{\begin{array}[]{ll}1&x=i\in{\mathbb{Z}}^{3},\\ 0&\text{else.}\end{array}\right.

Let the random variables ωi\omega_{i} be i.i.d. with uniform distribution in [c/2,c/2][-c/2,c/2], i.e. according to the probability distribution =c1Πiχ[c/2,c/2]dx\mathbb{P}=c^{-1}\Pi_{i}\chi_{{}_{\scriptstyle[-c/2,c/2]}}dx.

The 3–D random discrete Schrödinger operator, formally given by

Hω=+i3ωi<,δi>δi on l2(3),H_{\omega}=-\bigtriangleup+\sum_{i\in{\mathbb{Z}}^{3}}\omega_{i}<\,\cdot\,,\delta_{i}>\delta_{i}\text{ on }l^{2}({\mathbb{Z}}^{3}),

is the main object of study.

This operator has been studied extensively, see e.g. [16, 25] and the references therein. The first part of the operator -\bigtriangleup describes the movement of an electron inside a crystal with atoms located at all integer lattice points 3{\mathbb{Z}}^{3}. The perturbation i3ωi<,δi>δi\sum_{i\in{\mathbb{Z}}^{3}}\omega_{i}<\,\cdot\,,\delta_{i}>\delta_{i} can be interpreted as having the atoms randomly displaced around the lattice points. It is important to notice that the perturbation is almost surely non-compact, so that classical perturbation theory (e.g. Kato–Rosenblum Theorem, which states the invariance of the absolutely continuous spectrum under compact perturbations) cannot be applied almost surely. It is known that the absolutely continuous spectrum is deterministic, i.e. it occurs with probability one or zero, see e.g. [19]. Localization in the sense of exponentially decaying eigenfunctions was proved analytically for disorders cc above some threshold C0C_{0} (see e.g. [2], [9], and [25]). Currently, the smallest threshold in 3 dimensions is C0=100.6C_{0}=100.6 (see Table 1 in [21]).

Diffusion is expected but not proved for small disorder c>0c>0. We numerically determine a regime of disorders for which the three dimensional discrete random Schrödinger operator does not exhibit localization. Our calculations are based on the Lanczos algorithm [13] for determining orthogonal bases for Krylov spaces [27]. Although we are not the first to use this method (see e.g. [17, 23] and the references therein), our application of it is quite different. In particular, our method is not based on scaling theory (for further discussion see [18]). In [20], the Lanczos algorithm is employed to compute a set of eigenvalues and eigenvectors. However, we test for localization without computing eigenvalues or eigenvectors, but only compute the distance between δ111\delta_{111} and the orbit of δ000\delta_{000}. The orbit is the span of {Hωkδ000:k{0}}\left\{H_{\omega}^{k}\delta_{000}\mathchar 12346\relax k\in{\mathbb{N}}\cup\{0\}\right\}, which is exactly a Krylov subspace. At each step of the Lanczos iteration, we use the orthogonality of the generated vectors to update the distance of interest. In this way, we maintain the low memory cost of a three-term recurrence, bypassing the need to store any eigenvectors at all. In addition to this, we have performed some a posteriori tests of the Lanczos algorithm on smaller cases to measure the degree to which orthogonality may be lost.

Besides computational advantages, our approach also offers a different mathematical perspective. By utilizing eigenvectors, it is (tacitly) assumed that all spectral points are in fact eigenvalues, while our approach merely generates an orbit without attempting to rule out other kinds of spectral points.

While the contributions of this paper are numeric, the method (see [18]) provides an explicit analytic expression, which may yield a proof of the following numerically supported Main Result.

Main Result 1.1.

For disorder c3.5c\lesssim 3.5, numerical experiments indicate that the three dimensional discrete random Schrödinger operator does not exhibit Anderson localization with positive probability, in the sense that it has non-zero absolutely continuous spectrum with probability 1. (In particular, we do not have what is usually referred to as “strong dynamical localization” implying delocalization in most or even all of the other senses, see [12].)

The key analytical tool to our method is stated in Proposition 2.1 below. Section 3 is devoted to a description of the numerical experiment. The numerical testing criterion we applied is given by Numerical Criterion 3.1 below. Our numerical findings and the conclusions can be found in Section 4. In Subsections 4.1 and 4.2, we study the averaged data and find further numerical validation of our method. In Section 5 we verify the performance of the method in many examples. In Subsection 5.3, we present the distribution of energies after repeated application of the random operator of a wave packet initially located at the origin. We briefly remark on computing and memory requirements in Section 6.

2. Preliminaries

2.1. Singular and absolutely continuous parts of normal operators

Recall that an operator in a separable Hilbert space is called normal if TT=TTT^{*}T=TT^{*}. By the spectral theorem operator TT is unitarily equivalent to MzM_{z}, multiplication by the independent variable zz, in a direct sum of Hilbert spaces

=(z)𝑑μ(z)\mathcal{H}=\oplus\int\mathcal{H}(z)\,d\mu(z)

where μ\mu is a scalar positive measure on {\mathbb{C}}, called a scalar spectral measure of TT.

If TT is a unitary or self-adjoint operator, its spectral measure μ\mu is supported on the unit circle or on the real line, respectively. Via Radon decomposition, μ\mu can be decomposed into a singular and absolutely continuous parts μ=μs+μac\mu=\mu_{\scriptstyle\text{\rm s}}+\mu_{\scriptstyle\text{\rm ac}}. The singular component μs\mu_{\scriptstyle\text{\rm s}} can be further split into singular continuous and pure point parts. For unitary or self-adjoint TT we denote by TacT_{\scriptstyle\text{\rm ac}} the restriction of TT to its absolutely continuous part, i.e. TacT_{\scriptstyle\text{\rm ac}} is unitarily equivalent to Mt|(t)𝑑μac(t).M_{t}\big{|}_{\oplus\int\mathcal{H}(t)d\mu_{\scriptstyle\text{\rm ac}}(t)}. Similarly, define the singular, singular continuous and the pure point parts of TT, denoted by TsT_{\scriptstyle\text{\rm s}}, TscT_{\scriptstyle\text{\rm sc}} and TppT_{\scriptstyle\text{\rm pp}}, respectively.

2.2. Key tool

As mentioned above delocalization is deterministic. Therefore demonstrating that it does not occur with probability zero is sufficient to determine delocalization.

This following result makes our numerical experiment possible as it suffices to check the evolution of only one vector through repeated operations by the Anderson Hamiltonian and 3 dimensional random Schrödinger operator.

Fix the vectors δ000l2(3)\delta_{000}\in l^{2}({\mathbb{Z}}^{3}) and δ111l2(3)\delta_{111}\in l^{2}({\mathbb{Z}}^{3}), i.e. 3–tensors with zero entries, except for the (0,0,0)(0,0,0)-position and the (1,1,1)(1,1,1)-position, respectively, which equal 11.

Notice that

(2) Dω,cn:=dist(δ111,span{Hωkδ000:k=0,1,2,,n})\displaystyle D_{\omega,c}^{n}\mathrel{\mathop{\mathchar 12346\relax}}=\operatorname{dist}(\delta_{111},\text{span}\{H_{\omega}^{k}\delta_{000}\mathchar 12346\relax k=0,1,2,\ldots,n\})

describes the distance between the unit vector δ111\delta_{111} and the subspace obtained taking the closure of the span of the vectors δ000,Hωδ000,Hω2δ000,,Hωnδ000\delta_{000},H_{\omega}\delta_{000},H_{\omega}^{2}\delta_{000},\ldots,H_{\omega}^{n}\delta_{000}.

In numerical linear algebra, this space is called a Krylov subspace, and the Lanczos algorithm [13] provides a classical approach for finding an orthonormal basis. Our distance calculation (2) relies on the orthogonality of these vectors, iteratively updating the distance with each new Krylov vector.

Proposition 2.1.

Consider the discrete random Schrödinger operator given by equation (1). Let ωi\omega_{i}, i3i\in{\mathbb{Z}}^{3}, be i.i.d. random variables with uniform (Lebesgue) distribution on [c,c][-c,c], c>0c>0. To prove delocalization (i.e. the existence of absolutely continuous spectrum with positive probability), it suffices to find c>0c>0 for which the distance

(3) Dω,c:=limnDω,cn=0\displaystyle D_{\omega,c}\mathrel{\mathop{\mathchar 12346\relax}}=\lim_{n\to\infty}D_{\omega,c}^{n}=0

with non-zero probability. (Notice that the limit exists by the monotone convergence theorem.)

The proposition follows immediately from Theorems 1.1 and 1.2 of [14] and is stated in more generality in [18].

Remark 2.2.

The converse of Proposition 2.1 is not true. And we cannot draw any conclusions, if the distance between a fixed (unit) vector and the subspace generated by the orbit of another vector tends to zero. In particular, we cannot conclude that there must be localization. Even if we show (3) for many or ‘all’ vectors (instead of just δ111\delta_{111}), it could be possible that the absolutely continuous part has multiplicity one and that δ000\delta_{000} is cyclic, that is, l2(3)=closspan{Hωkδ000:k{0}}l^{2}({\mathbb{Z}}^{3})=\operatorname{clos}\operatorname{span}\{H_{\omega}^{k}\delta_{000}\mathchar 12346\relax k\in{\mathbb{N}}\cup\{0\}\}.

3. Method of numerical experiment

Consider the discrete Schrödinger operator given by (1) with random variable ω\omega distributed according to the hypotheses of Proposition 2.1.

By Proposition 2.1, we obtain delocalization if we can find c>0c>0 for which (3) happens with non-zero probability. Let us now explain precisely how we verify delocalization numerically, leading up to the Numerical Criterion 3.1 below.

In the numerical experiment, we initially fix cc and fix one computer-generated realization of the random variable ω\omega (with distribution in accordance to the hypotheses of Proposition 2.1). We then calculate the distances Dω,cnD_{\omega,c}^{n} for n{0,1,2,}n\in\{0,1,2,\ldots\}.

Assuming that we know Dω,cnD_{\omega,c}^{n} for n=0,,500n=0,\ldots,500, let us find a lower estimate for the limit

Dω,c=limnDω,cn.D_{\omega,c}=\lim_{n\to\infty}D_{\omega,c}^{n}.
Refer to caption
Figure 1. Typical trend for Dω,cnD^{n}_{\omega,c} for n=500n=500 iterations and c=0.3c=0.3.

Figure 1 displays a typical trend for the distance Dω,cnD_{\omega,c}^{n} as nn increases. Because the first n=119n=119 points were irregular and do not contribute to the above limit, they were omitted. Notice that the graph is decreasing, as is expected. Although it certainly appears that the limit does not go to 0, the graph could have logarithmic decay, approaching zero very slowly. To attain an estimate for Dω,cD_{\omega,c}, which excludes the case of such slow decay, we re-scaled the graph by nan^{-a}, 0.1a20.1\leq a\leq 2, so that the x-axis is inverted and the yy-intercept, yω,cy_{\omega,c}, of a line of best fit will estimate Dω,cD_{\omega,c}.

Refer to caption
Figure 2. Figure 1 re-scaled using a=1.65a=1.65. Notice the fine yy-scale and proximity to the yy-axis.

Figure 2 shows the re-scaled graph for n=120,121,122,,500n=120,121,122,\ldots,500. Subsection 3.1 describes the choice of aa and why, for small values of cc, Dω,cD_{\omega,c} does not decay to 0.

Since an approximating line is only an estimate, for further confidence in our results, we also calculated the minimum yy-intercept of all lines through two consecutive points and call it Lω,cL_{\omega,c} (see the steep line in Figure 3). This is essentially the “worst case,” and ought to underestimate Dω,cD_{\omega,c}, yielding the relationship

Lω,cyω,cDω,c.\displaystyle L_{\omega,c}\leq y_{\omega,c}\approx D_{\omega,c}\,.
Refer to caption
Figure 3. A closer look at a window in Figure 2. The steeper line is used to determine Lω,cL_{\omega,c}, and the line of best fit provides yω,cy_{\omega,c}. For this realization we have Lω,c=0.9585894L_{\omega,c}=0.9585894 and yω,c=0.9586354y_{\omega,c}=0.9586354.

We repeated this process for many values of cc and multiple, different, computer-generated instances of the random variable ω\omega. We took the minimum of yω,cy_{\omega,c} and Lω,cL_{\omega,c} across all instances of ω\omega, with the intent to demonstrate that Dω,cD_{\omega,c} is above 0 for many different ω\omega.

In order to give confidence to our calculations to account for random error occurring in the computer, we introduce the following restrictions even though Proposition 2.1 only requires that Lω,c>0L_{\omega,c}>0 .

Numerical Criterion 3.1.

For a fixed value of cc, we say that we have delocalization, if for at least 90% realizations we obtain Lω,c>.9>0L_{\omega,c}>.9>0 and yω,cLω,cy_{\omega,c}-L_{\omega,c} is of order 10310^{-3}. (Notice that we only need non-zero probability by Proposition 2.1, and Remark 2.2.)

3.1. Choice of the re-scaling parameter

For each fixed cc and ω\omega, the re-scaling exponent aa is chosen so that the re-scaled graph of the distance function (see Figure 2) satisfies the least square property; that is, the error with respect to square–norm when approximating the graph by a line is minimal. With this exponent we then find the corresponding linear approximation for the re-scaled distance function.

To find optimal aa, we used the mesh a=0.05:0.05:2a=0.05\mathchar 12346\relax 0.05\mathchar 12346\relax 2. Below is a table, see equation (8), for many values of cc, giving the percentage of usable trials (those for which an optimal a0.1a\geq 0.1 was found) for each value of cc. Trials are not usable if the re-scaling parameter a=0.05a=0.05 yields a concave graph. If this happens, we do not obtain any information (according to Remark 2.2). See Figure 4 below. Note that a small value (0.05\leq 0.05) of aa is “bad”, since the graph rescaled with a=0.05a=0.05 will be concave, and thus it is not expected for a line of best fit to underestimate the limit of the distance.

Refer to caption
Figure 4. A concave graph that yields no usable data (c=6c=6). The concave shape of the data implies that yω,cy_{\omega,c} is not necessarily a lower bound for Dω,cD_{\omega,c}.

A positive re-scaling factor implies that the graph in Figure 1 will not decay to zero. Indeed, using a re-scaling factor smaller than the optimal one will result in a convex graph for the distances Dω,cnD_{\omega,c}^{n}. And the yy-intercept of the line lies below the value expected for Dω,cD_{\omega,c}^{\infty}.

4. Conclusions

As mentioned in Section 3, for a fixed cc we chose many realizations ω\omega. For every value of cc, we took the minimum of the resulting quantities for Lω,cL_{\omega,c} and yω,cy_{\omega,c} (the yy-intercept of the approximating line and the minimum yy-intercept of the lines passing through any two consecutive points, respectively).

We present our observations for the Numerical Criterion 3.1 for n=500n=500. For fixed disorder, we will comment in Subsection 4.1 on the re-scaling parameters of averages over the distances Dω,cnD_{\omega,c}^{n}, n=0,1,2,,200n=0,1,2,\ldots,200 and n=0,1,2,,500n=0,1,2,\ldots,500.

The following tables (8) document the data obtained for n=200 by taking 15 realizations for each cc between 0.10.1 and 55, and 4 realizations for each c10c\geq 10 and c=0c=0. By \mathbb{P} we denote the probability of finding a re-scaling factor a[0.1,2]a\in[0.1,2].

(8) c00.10.20.30.40.50.6111111.93yω,c.95869.95866.95859.95842.9582.9580.9573Lω,c.95869.95865.95858.95738.9580.9563.9504\displaystyle\begin{array}[]{|c||c|c|c|c|c|c|c|}\hline\cr c&0&0.1&0.2&0.3&0.4&0.5&0.6\\ \hline\cr\mathbb{P}&1&1&1&1&1&1&.93\\ \hline\cr y_{\omega,c}&.95869&.95866&.95859&.95842&.9582&.9580&.9573\\ \hline\cr L_{\omega,c}&.95869&.95865&.95858&.95738&.9580&.9563&.9504\\ \hline\cr\end{array}
c0.70.80.911.522.5.931.93.81.871yω,c.9573.9576.9559.9554.9540.9404.9447Lω,c.9529.9574.9488.9487.9519.9063.9359\displaystyle\begin{array}[]{|c||c|c|c|c|c|c|c|}\hline\cr c&0.7&0.8&0.9&1&1.5&2&2.5\\ \hline\cr\mathbb{P}&.93&1&.93&.8&1&.87&1\\ \hline\cr y_{\omega,c}&.9573&.9576&.9559&.9554&.9540&.9404&.9447\\ \hline\cr L_{\omega,c}&.9529&.9574&.9488&.9487&.9519&.9063&.9359\\ \hline\cr\end{array}
c33.544.551015.8.67.73.83.61.75yω,c.9298.9105.8857.8843.8020.6458.0935Lω,c.8443.7364.5389.5699.1095.30812.4974\displaystyle\begin{array}[]{|c||c|c|c|c|c|c|c|}\hline\cr c&3&3.5&4&4.5&5&10&15\\ \hline\cr\mathbb{P}&.8&.67&.73&.83&.6&1&.75\\ \hline\cr y_{\omega,c}&.9298&.9105&.8857&.8843&.8020&.6458&.0935\\ \hline\cr L_{\omega,c}&.8443&.7364&.5389&.5699&.1095&.3081&-2.4974\\ \hline\cr\end{array}
c2025303540.25.75.25.75.25yω,c2.15071.94123.32713.33019.7171Lω,c14.160811.889515.057738.600740.4666\displaystyle\begin{array}[]{|c||c|c|c|c|c|}\hline\cr c&20&25&30&35&40\\ \hline\cr\mathbb{P}&.25&.75&.25&.75&.25\\ \hline\cr y_{\omega,c}&-2.1507&-1.9412&-3.3271&-3.3301&-9.7171\\ \hline\cr L_{\omega,c}&-14.1608&-11.8895&-15.0577&-38.6007&-40.4666\\ \hline\cr\end{array}

While for some c2.5c\geq 2.5, we have .9\mathbb{P}\geq.9 the difference between yω,cy_{\omega,c} and Lω,cL_{\omega,c} is relatively large, which means that the line from taking the least square approximation is likely not a good approximation for the distances.

We also repeated the experiment for n=500n=500 and the tables in equation (13) below documents the findings. In these trials, the first 119 entries were removed instead of the first 44, as in the n=200n=200 case. This larger crop makes the data more stable by giving better estimates for yω,cy_{\omega,c} and Lω,cL_{\omega,c} and by more consistently finding a usable rescaling factor aa. We ran 13 trials for c1c\leq 1 and 4 trials for all other values.

(13) c00.10.20.30.40.50.61111111yω,c.95869.95866.95856.95838.95813.95782.9575Lω,c.95869.95866.95855.95835.95809.95776.9574\displaystyle\begin{array}[]{|c||c|c|c|c|c|c|c|}\hline\cr c&0&0.1&0.2&0.3&0.4&0.5&0.6\\ \hline\cr\mathbb{P}&1&1&1&1&1&1&1\\ \hline\cr y_{\omega,c}&.95869&.95866&.95856&.95838&.95813&.95782&.9575\\ \hline\cr L_{\omega,c}&.95869&.95866&.95855&.95835&.95809&.95776&.9574\\ \hline\cr\end{array}
c0.70.80.911.522.51111111yω,c.9570.9565.9559.95523.9520.9518.9523Lω,c.9569.9564.9558.95515.9514.9504.9438\displaystyle\begin{array}[]{|c||c|c|c|c|c|c|c|}\hline\cr c&0.7&0.8&0.9&1&1.5&2&2.5\\ \hline\cr\mathbb{P}&1&1&1&1&1&1&1\\ \hline\cr y_{\omega,c}&.9570&.9565&.9559&.95523&.9520&.9518&.9523\\ \hline\cr L_{\omega,c}&.9569&.9564&.9558&.95515&.9514&.9504&.9438\\ \hline\cr\end{array}
c33.544.55101511111.5.75yω,c.9451.9556.9405.9271.9149.2600.5149Lω,c.9417.9492.9302.9041.80552.32179.8398\displaystyle\begin{array}[]{|c||c|c|c|c|c|c|c|}\hline\cr c&3&3.5&4&4.5&5&10&15\\ \hline\cr\mathbb{P}&1&1&1&1&1&.5&.75\\ \hline\cr y_{\omega,c}&.9451&.9556&.9405&.9271&.9149&.2600&-.5149\\ \hline\cr L_{\omega,c}&.9417&.9492&.9302&.9041&.8055&-2.3217&-9.8398\\ \hline\cr\end{array}
c2025303540.5.5.5.75.75yω,c2.85711.59572.51882.24073.1759Lω,c16.637924.070718.116818.782926.3506\displaystyle\begin{array}[]{|c||c|c|c|c|c|}\hline\cr c&20&25&30&35&40\\ \hline\cr\mathbb{P}&.5&.5&.5&.75&.75\\ \hline\cr y_{\omega,c}&-2.8571&-1.5957&-2.5188&-2.2407&-3.1759\\ \hline\cr L_{\omega,c}&-16.6379&-24.0707&-18.1168&-18.7829&-26.3506\\ \hline\cr\end{array}

A good rescaling factor aa was found for all 143 of the trials for c1c\leq 1 and all c3.5c\leq 3.5 satisfy Criterion 3.1, an improvement from the n=200n=200 case. Hence the final conclusion of this numerical experiment is precisely the Main Result 1.1. According to Remark 2.2 and Criterion 3.1, for c4c\geq 4, we do not have any conclusion.

4.1. Averages

In the tables in equation (18) below, for each fixed cc, we averaged the distances Dω,cnD_{\omega,c}^{n}, n=0,1,2,,200n=0,1,2,\ldots,200, of all our realizations. For those averaged distances, we determined the re-scaling parameters a~\tilde{a}, as well as y~c\tilde{y}_{c} and L~c\tilde{L}_{c} in analogy. The significance of our findings is that the re-scaling factors a~\tilde{a} are “roughly” decreasing and rather well-behaved for c1.5c\leq 1.5. For larger disorder, a~\tilde{a} becomes even less stable, and can’t even be found for large enough disorder.

(18) c00.10.20.30.40.50.6a~21.91.651.51.31.1.95y~c.95869.95869.95865.95861.95853.95846.95843L~c.95869.95868.95864.95861.95852.95846.95841\displaystyle\begin{array}[]{|c||c|c|c|c|c|c|c|}\hline\cr c&0&0.1&0.2&0.3&0.4&0.5&0.6\\ \hline\cr\tilde{a}&2&1.9&1.65&1.5&1.3&1.1&.95\\ \hline\cr\tilde{y}_{c}&.95869&.95869&.95865&.95861&.95853&.95846&.95843\\ \hline\cr\tilde{L}_{c}&.95869&.95868&.95864&.95861&.95852&.95846&.95841\\ \hline\cr\end{array}
c0.70.80.911.522.5a~1.31.8.91.6.85y~c.95818.9584.9579.95779.95714.9539.9544L~c.95816.9583.9578.95777.95708.9537.9541\displaystyle\begin{array}[]{|c||c|c|c|c|c|c|c|}\hline\cr c&0.7&0.8&0.9&1&1.5&2&2.5\\ \hline\cr\tilde{a}&1.3&1&.8&.9&1&.6&.85\\ \hline\cr\tilde{y}_{c}&.95818&.9584&.9579&.95779&.95714&.9539&.9544\\ \hline\cr\tilde{L}_{c}&.95816&.9583&.9578&.95777&.95708&.9537&.9541\\ \hline\cr\end{array}
c33.544.551015a~.65.3.55.65.5.5.3y~c.9485.9466.9345.9414.9217.8332.5312L~c.9478.9442.9332.9399.9137.7648.2063\displaystyle\begin{array}[]{|c||c|c|c|c|c|c|c|}\hline\cr c&3&3.5&4&4.5&5&10&15\\ \hline\cr\tilde{a}&.65&.3&.55&.65&.5&.5&.3\\ \hline\cr\tilde{y}_{c}&.9485&.9466&.9345&.9414&.9217&.8332&.5312\\ \hline\cr\tilde{L}_{c}&.9478&.9442&.9332&.9399&.9137&.7648&.2063\\ \hline\cr\end{array}
c2025303540a~.1.85.45N/AN/Ay~c.3300.2928.07223.09905.1751L~c1.8583.0051.406012.271212.7084\displaystyle\begin{array}[]{|c||c|c|c|c|c|}\hline\cr c&20&25&30&35&40\\ \hline\cr\tilde{a}&.1&.85&.45&N/A&N/A\\ \hline\cr\tilde{y}_{c}&-.3300&.2928&-.0722&-3.0990&-5.1751\\ \hline\cr\tilde{L}_{c}&-1.8583&-.0051&-.4060&-12.2712&-12.7084\\ \hline\cr\end{array}

In equation (23) below we document the analogous quantities for the n=500n=500 trials. Note that there is no rescaling factor for c=20c=20, while there is for that cc in the n=200n=200 trials. The data sets are not related to each other, aside from sharing the same disorder cc.

(23) c00.10.20.30.40.50.6a~21.751.351.21.111.05y~c.95869.95868.95864.95861.95855.95847.95823L~c.95869.95868.95864.95861.95854.95846.95822\displaystyle\begin{array}[]{|c||c|c|c|c|c|c|c|}\hline\cr c&0&0.1&0.2&0.3&0.4&0.5&0.6\\ \hline\cr\tilde{a}&2&1.75&1.35&1.2&1.1&1&1.05\\ \hline\cr\tilde{y}_{c}&.95869&.95868&.95864&.95861&.95855&.95847&.95823\\ \hline\cr\tilde{L}_{c}&.95869&.95868&.95864&.95861&.95854&.95846&.95822\\ \hline\cr\end{array}
c0.70.80.911.522.5a~1.051.451.15.65.7.6.9y~c.95806.95786.95805.95795.9558.9543.9561L~c.95805.95785.95803.95792.9555.9538.9556\displaystyle\begin{array}[]{|c||c|c|c|c|c|c|c|}\hline\cr c&0.7&0.8&0.9&1&1.5&2&2.5\\ \hline\cr\tilde{a}&1.05&1.45&1.15&.65&.7&.6&.9\\ \hline\cr\tilde{y}_{c}&.95806&.95786&.95805&.95795&.9558&.9543&.9561\\ \hline\cr\tilde{L}_{c}&.95805&.95785&.95803&.95792&.9555&.9538&.9556\\ \hline\cr\end{array}
c33.544.551015a~.55.6.65.35.15.25.3y~c.9506.9571.9479.9390.9244.7053.4991L~c.9497.9531.9451.9348.9140.4796.0327\displaystyle\begin{array}[]{|c||c|c|c|c|c|c|c|}\hline\cr c&3&3.5&4&4.5&5&10&15\\ \hline\cr\tilde{a}&.55&.6&.65&.35&.15&.25&.3\\ \hline\cr\tilde{y}_{c}&.9506&.9571&.9479&.9390&.9244&.7053&.4991\\ \hline\cr\tilde{L}_{c}&.9497&.9531&.9451&.9348&.9140&.4796&-.0327\\ \hline\cr\end{array}
c2025303540a~N/A.1.8N/AN/Ay~c.05691.96581.1264.09062.1984L~c.82546.53143.3376.18236.8235\displaystyle\begin{array}[]{|c||c|c|c|c|c|}\hline\cr c&20&25&30&35&40\\ \hline\cr\tilde{a}&N/A&.1&.8&N/A&N/A\\ \hline\cr\tilde{y}_{c}&.0569&-1.9658&-1.1264&.0906&-2.1984\\ \hline\cr\tilde{L}_{c}&-.8254&-6.5314&-3.3376&-.1823&-6.8235\\ \hline\cr\end{array}

4.2. Comparing n=200n=200 with n=500n=500.

The n=500n=500 data gave better results than the n=200n=200 data. The probability of finding a useable rescaling factor for n=500n=500 was higher than that of n=200n=200 for all but two values of cc. The average rescaling factor a~\tilde{a} was similar between the two data sets. Finally, yω,cLω,cy_{\omega,c}-L_{\omega,c} was smaller for the n=500n=500 data for small cc, suggesting that the approximation given by yω,cy_{\omega,c} is better.

5. Further validation of the method and the numerical experiments

Apart from the usual tests (the program is running stably, checking all subroutines, many verifications for small nn), we have conducted the following tests. Most important is the a posteriori test of orthogonality in the Lanczos algorithm in subsection 5.4.

5.1. Free discrete three dimensional Schrödinger operator

When we apply the free discrete Schrödinger operator H=H𝟎H=H_{\bf 0} to the vector δ000\delta_{000}, it immediately becomes clear that Hδ000H\delta_{000} as well as all vectors Hnδ000H^{n}\delta_{000}, n{0}n\in{\mathbb{N}}\cup\{0\}, are symmetric with respect to the origin. In dimension d=3d=3, it is not hard to see that the distance between δ111\delta_{111} and the orbit of δ000\delta_{000} under HH is at least 7220.9354\frac{\sqrt{7}}{{2\sqrt{2}}}\approx 0.9354. Indeed, we have

dist(δ111,closspan{Hnδ000:n{0}})>minxdist(ux,δ111)=722,\operatorname{dist}(\delta_{111},\operatorname{clos}\operatorname{span}\{H^{n}\delta_{000}\mathchar 12346\relax n\in{\mathbb{N}}\cup\{0\}\})>\min_{x}\operatorname{dist}(u_{x},\delta_{111})=\frac{\sqrt{7}}{{2\sqrt{2}}},

where

ux=xδ111+xδ111+xδ111+xδ111xδ111+xδ111+xδ111+xδ111,u_{x}=x\delta_{-1-1-1}+x\delta_{1-1-1}+x\delta_{-11-1}+x\delta_{-1-11}x\delta_{-1-11}+x\delta_{-11-1}+x\delta_{1-1-1}+x\delta_{111},

the eight vertices of the length 2 cube centered at (0,0,0)(0,0,0).

In the experiments for the free discrete two dimensional Schrödinger operator we obtained a yy-intercept of the approximating line approximately equals 0.95869360.9586936. The re-scaled graph of distances still had a convex shape, so the actual distance as nn\to\infty would be bigger. In fact, we have extracted our data an upper estimate of 0.9586939Dω,05000.9586939\approx D^{500}_{\omega,0}. Therefore, the distance must lie in the interval [0.9586936,0.9586939][0.9586936,0.9586939].

5.2. Orthogonalization Process

The c=0c=0 case shows a decrease in distance on only every other step. The symmetry caused by the absence of random perturbations means the 3-tensor after orthogonalization has alternating diamonds of zero and nonzero entries radiating from the origin, meaning the distance decreases every second application of the operator, when there is a nonzero entry in the (1,1,1)(1,1,1)-position.

5.3. Evolution under HωH_{\omega} of the bulk for small values of cc

We observe the bulk distribution which determines the distance from the origin where we are most likely to find an electron. Here, distance is measured by the taxicab method, so elements of the same distance form a diamond in the 3-D integer lattice. The bulk at this distance is the Euclidean norm of the elements constituting the diamond.

To be precise, we consider the elements of the vector m500m_{500} and define

(24) E(l,n)=|i|+|j|+|k|=l(mn)i,j,k2\displaystyle E(l,n)=\sqrt{\sum_{|i|+|j|+|k|=l}(m_{n})_{i,j,k}^{2}}

for the bulk E(l,n)E(l,n) of the vector mnm_{n} at taxicab distance ll from the origin. Here (mn)i,j,k(m_{n})_{i,j,k} refers to the (i,j,k)(i,j,k)-entry of the 22-tensor mnm_{n}. Slightly abusing notation, we normalize mnm_{n} and use the same notation for the normalized sequence of vectors.

Figure 5 is the result of averaging four sets of data for values of cc ranging from 0.10.1 to 11. As expected, the energy remains closer to the origin as disorder increases.

Refer to caption
Figure 5. Energy distribution of m500m_{500} for the disorders c=0.1:0.1:1c=0.1\mathchar 12346\relax 0.1\mathchar 12346\relax 1, averaged over the 11 realizations for each value of cc.

5.4. Lanczos and orthogonality

The Lanczos algorithm is known to lose orthogonality in many instances, which could cast doubt on our distance calculations. To test the accuracy for our problem, we stored the entire Krylov subspace generated on a smaller problem instance (N=150N=150) and stored these as columns of a matrix KK. The quantity Q=KTKIQ=\left\|K^{T}K-I\right\|_{\infty} should deviate with zero in proportion to the loss of orthogonality. In Tables (27) and (30), we measure the matrix \infty norm for realizations for several cases of cc near the critical value. We see that the Krylov vectors in these cases are in fact quite close to orthogonal especially for c3.5c\leq 3.5, although the orthogonality seems to decrease as cc grows.

(27) c0.00.51.01.52.02.53.03.5Q2.01e117.2e114.8e114.4e116.9e119.6e113.1e114.8e11\displaystyle\begin{array}[]{|c|c|c|c|c|c|c|c|c|c|c|c|c|c|c|c|c|c|}\hline\cr c&0.0&0.5&1.0&1.5&2.0&2.5&3.0&3.5\\ \hline\cr Q&2.01e{-11}&7.2e{-11}&4.8e{-11}&4.4e{-11}&6.9e{-11}&9.6e{-11}&3.1e{-11}&4.8e{-11}\\ \hline\cr\end{array}
(30) c4.04.55.05.56.06.57.07.58.0Q4.5e112.4e116.1e91.3e104.0e109.2e111.2e97.3e85.1e9\displaystyle\begin{array}[]{|c|c|c|c|c|c|c|c|c|c|c|c|c|c|c|c|c|c|}\hline\cr c&4.0&4.5&5.0&5.5&6.0&6.5&7.0&7.5&8.0\\ \hline\cr Q&4.5e{-11}&2.4e{-11}&6.1e{-9}&1.3e{-10}&4.0e{-10}&9.2e{-11}&1.2e{-9}&7.3e{-8}&5.1e{-9}\\ \hline\cr\end{array}

6. On computing and memory requirements

Using methodology similar to that in [18], all of the information contained in the 3-tensor is stored in one information vector. For this method, because of how the Hamiltonian acts, it is important for computing purposes that each point in the 3-tensor is stored in a position such that its neighbors along a coordinate axis are a consistent distance from that point in the vector. This methodology allows the vector to be half the size necessary for containing every point in a 3-tensor, but still approximately twice as large as is necessary. In order to explore localization in higher dimensions, a more efficient method is needed since a generalization of this code for dimension dd has time complexity 𝒪(nd)\mathcal{O}(n^{d}).

After prototyping our approach in MATLAB, we translated the code into FORTRAN90. This allowed us a smaller memory footprint and hence larger and more efficient runs. We then wrapped this routine into Python using the f2py package [22]. By doing so, we were able to run several cases concurrently on our workstation by using Python’s multiprocessing module.

Our simulations were run on a Dell Precision workstation with dual eight-core Intel Xeon E5-2680 processors running at 2.7GHz with 128GB of RAM. We used gfortran version 4.4.7 with flags -O3 -ftree-vectorizer-verbose=2 -msse2 -funroll-loops -ffast-math,
which, among other optimizations, enables instruction-level superscalar parallelism.

7. Further Projects

An immediate area for further exploration would be to consider various geometries, rather than simply the n-dimensional lattice. One geometry of interest is the Sierpinski gasket, starting at one corner and building the various triangles as nn increases. Preliminary results indicate that a program modeling the free random Schrödinger operator on this geometry should run with time complexity 𝒪(nln(3)ln(2)).\mathcal{O}\left(n^{\frac{\ln(3)}{\ln(2)}}\right).

References

  • [1] M. Aizenamn, G. M. Graf, Localization bounds for an electron. J. Phys. A: Math. Gen. 31 (1998), 6783–6806.
  • [2] M. Aizenman, S. Molchanov, Localization at large disorder and at extreme energies: An elementary derivation, Comm. Math. Phys. 157 (1993), no. 2, 245–278.
  • [3] P. W. Anderson, Absence of Diffusion in Certain Random Lattices, Phys. Rev., 109 (1958), 1492–1505.
  • [4] M. S. Birman, M. Z. Solomjak, Spectral theory of self-adjoint operators in Hilbert space, 1986.
  • [5] R. Carmona, J. Lacroix, Spectral theory of random Schrödinger operators, Birkhäuser, 1990.
  • [6] H. Cycon, R. Froese, W. Kirsh, B. Simon, Topics in the Theory of Schrödinger Operators, Springer Verlag, 1987.
  • [7] R. del Rio, S. Jitomirskaya, Y. Last, B. Simon, Operators with singular continuous spectrum. IV. Hausdorff dimensions, rank-one perturbations, and localization, J. Anal. Math. 69 (1996), 153–200. MR 1428099 (97m:47002)
  • [8] A. Figotin, L. Pastur, Spectral properties of disordered systems in the one-body approximation, Springer Verlag, 1991.
  • [9] J. Fröhlich, T. Spencer, Absence of Diffusion in the tight binding model for large disorder of low energy, Commun. Math. Phys. 88 (1983), 151–184.
  • [10] F. Germinet, A. Klein, J. H. Schenker, Dynamical delocalization in random Landau Hamiltonians, Ann. of Math. (2) 166 (2007), no. 1, 215–244. MR 2342695 (2008k:82060)
  • [11] F. Ghribi, P. D. Hislop, F. Klopp, Localization for Schrödinger operators with random vector potentials, 447 (2007), 123–138. MR 2423576 (2009d:82067)
  • [12] D. Hundertmark, A short introduction to Anderson localization, Analysis and stochastics of growth processes and interface models, 194–218, Oxford Univ. Press, Oxford, 2008. MR 2603225 (2011c:82038)
  • [13] C. Lanczos, An iteration method for the solution of the eigenvalue problem of linear differential and integral operators, United States Governm. Press Office 1950.
  • [14] by same author, Simplicity of singular spectrum in Anderson-type Hamiltonians, Duke Math. J. 133 (2006), no. 1, 185–204. MR 2219273 (2007g:47062)
  • [15] T. Kato, Perturbation theory for linear operators, Classics in Mathematics, Springer Verlag, Berlin, 1995, Reprint of the 1980 edition. MR 1335452 (96a:47025)
  • [16] W. Kirsh, An invitation to random Schrödinger operators, Panoramas et synthèses 25 (2008), 1–119.
  • [17] A. Lagendijk, B. van Tigglen, D.S. Wiersma, Fifty years of Anderson localization, Physics Today 82 Featured article (August 2009) no. 8, 24–29.
  • [18] C. Liaw, Approach to the Extended States Conjecture. Journal of Statistical Physics, 153 (2013) no. 6, 1022–1038.
  • [19] C. Liaw, Deterministic spectral properties of Anderson-type Hamiltonians. For preprint see arXiv:1009.1353.
  • [20] O. Schenk, M. Bollhöfer, R.A. Römer, On large-scale diagonalization techniques for the Anderson model of localization. SIAM review 50.1 (2008) 91–112.
  • [21] J. Schenker, How large is large? Estimating the critical disorder for the Anderson model. For preprint see arXiv-math:1305.6987v1.
  • [22] P. Peterson, F2PY: a tool for connecting Fortran and Python programs, International Journal of Computational Science and Engineering 4 (2009), no. 4, 296–305.
  • [23] J. Stein, U. Krey, Numerical Studies on the Anderson Localization Problem, Zeitschrift Für Physik B 34 (1979), 287–296.
  • [24] B. Simon, Cyclic vectors in the Anderson model, Rev. Math. Phys. 6 (1994), no. 5A, 1183–1185, Special issue dedicated to Elliott H. Lieb. MR 1301372 (95i:82058)
  • [25] by same author, Spectral analysis of rank-one perturbations and applications, Mathematical Quantum Theory I: Field Theory and Many-Body Theory (1994).
  • [26] B. Simon and T. Wolff, Singular continuous spectrum under rank-one perturbations and localization for random Hamiltonians, Comm. Pure Appl. Math., 39 (1986), no. 1, 75–90. MR 820340 (87k:47032).
  • [27] L. N. Trefethen and D. Bau, Numerical linear algebra, SIAM (1997).