We consider quasiperiodic operators on Zd with unbounded monotone sampling functions (“Maryland-type”), which are not required to be strictly monotone and are allowed to have flat segments. Under several geometric conditions on the frequencies, lengths of the segments, and their positions, we show that these operators enjoy Anderson localization at large disorder.

This paper can be considered as a direct continuation of the earlier publication.12 We consider quasiperiodic Schrödinger operators on 2(Zd),

(H(x)ψ)n=ε(Δψ)n+f(x+ωn)ψn,
(1.1)

where ω = (ω1, …, ωd) is the frequency vector and Δ is the discrete Laplacian,

(Δψ)n=mZd:|mn|1=1ψm.

We will consider the regime of large disorder, which, after rescaling, corresponds to small ɛ > 0. The function f, which generates the quasiperiodic potential, is a non-decreasing function, given by

f:(1/2,1/2)(,+),f(1/2±0)=,
(1.2)

and is extended into R\(Z+1/2) by 1-periodicity. As usual for quasiperiodic operators, the numbers {1, ω1, …, ωd} are assumed to be linearly independent over Q.

Potentials of form (1.2) will be called Maryland-type, after the classical Maryland model with f(x) = tan(πx) (we refer the reader to the following, certainly not exhaustive, list of related publications1,4–11). In Ref. 12, Sec. VI, we considered a case where f was not strictly monotone and was allowed to have a single flat piece of sufficiently small length h; in particular, |h| < minj |ωj|. In this situation, we can treat this flat piece as a single isolated resonance. In the present paper, we consider several more elaborate situations where the resonance is not isolated but has, in some sense, finite multiplicity.

The main result of this paper is Theorem 5.2, where we establish some sufficient conditions under which the operator H admits Anderson localization. We also list several particular examples and refinements in Sec. VI. The simplest new example (see Theorem 6.1 for a precise statement) is as follows. On 2(Z), consider operator (1.1) with the function f satisfying

f(x)E,x[a,a+L](1/2,1/2).

Additionally, assume that ω is Diophantine (see Sec. II B for the definition) and that [a − 2ω, a + L + 2ω] ⊂ (−1/2, 1/2). Suppose that f is Lipschitz monotone outside [a, a + L], with some regularity conditions similar to Ref. 12. Assume also that L is not a rational multiple of ω. Then, operator (1.1) has Anderson localization for 0 < ɛ < ɛ0(ω, L).

We would also like to mention the simplest possible class of operators that are not covered by our approach. Suppose that f is a constant on an interval [a, a + L] ⊂ (−1/2, 1/2), and suppose that the set

S{nZd:x+ωn[a,a+L]+Z}

has an unbounded connected component for some x (here, we define connectedness on the Zd graph with nearest neighbor edges). For example, this will happen if 0 < ω1 < ω2 < L. Our methods cannot cover such operators. In fact, it seems possible that such models do not demonstrate Anderson localization around the energy E=f[a,a+L].

Our approach, essentially, is based on assuming the opposite of the above: that is, all connected components of the set S are bounded and are sufficiently far away from each other. If we are able to surround each component by a layer of lattice points n such that f(x + ω · n) has good monotonicity properties, then, after a partial diagonalization, their monotonicity will “propagate,” in a weaker form, into the interior of the set S. Afterward, we can apply a modified version of the main result of Ref. 12 to finalize the diagonalization; see Proposition 2.1.

While the functions f under consideration will have flat pieces, a necessary assumption for all proofs below would be the existence of sufficiently many pieces with a good control of monotonicity. The corresponding regularity conditions and convergence results are summarized in this section.

It will be convenient to not exclude the case x1/2+Z+ωZd, where exactly one value of the potential in operator (1.1) becomes infinite: say, f(x + ω · n). In this case, the natural limiting object is the operator on 2(Zd\{n}) obtained from (1.1) by enforcing the Dirichlet condition ψ(n) = 0. The results of Ref. 12, which we will be using, extend “continuously” into these values of x (see Ref. 12, Remark 4.14), with a reasonable interpretation of infinities, if one adds an infinite eigenvalue with an eigenvector en. Here, {en:nZd} is the standard basis in 2(Zd). We will also use the notation {ej: j = 1, 2, …, d} for the standard basis in Rd. For a subset ARd, it will be convenient to use the notation 2(A) for 2(AZd). For example, 2([0, N]) = 2{0, 1, 2, …, N}.

Similarly to Ref. 12, we will always assume the following:

  • (f1)

    f:(1/2,1/2)R is continuous, non-decreasing, f(−1/2 + 0) = −, f(1/2 − 0) = +, and f is extended by 1-periodicity into R\(Z+1/2).

    Suppose that f satisfies (f1). Let Creg > 0 and x0 ∈ (−1/2, 1/2). We say that f is Creg-regular at x0, if the following holds:

  • (cr0)

    The pre-image f−1((f(x0) − 2, f(x0) + 2)) ∩ (−1/2, 1/2) is an open interval [denoted by (a, b)], and f(a,b) is a one-to-one map between (a, b) and (f(x0) − 2, f(x0) + 2).

  • (cr1)
    Let Dmin(x0)infx(a,b)f(x)1 (for points where f′ does not exist, consider the smallest of the derivative numbers). Then,
    Dmin(x0)f(x)CregDmin(x0),x(a,b).
    (2.1)
  • (cr2)
    Define (a1, b1) = f−1(f(x0) − 1, f(x0) + 1) ⊂ (a, b) and
    g(x)=1f(x)f(x0),x(b1,a1+1),
    extended by continuity to g(±1/2) = 0 (recall that we also assume f(x + 1) = f(x) so that the interval (b1, a1 + 1) is essentially (−1/2, 1/2)\(a1, b1) together with the point 1/2 = −1/2 mod 1). Then, under the same conventions on the existence of derivatives,
    |g(x)|CregDmin(x0),x(b1,a1+1).
    For convenience, we will require Dmin(x0) ≥ 1 as a necessary condition for Creg-regularity. This condition can always be achieved by rescaling.

The frequency vector ωRd is called Diophantine if there exist Cdio, τdio > 0 such that

nωdist(nω,Z)Cdio|n|τdio,nZd\{0}.
(2.2)

Without loss of generality, we will always assume that 0 < ω1 < … < ωd < 1/2. The set of Diophantine vectors with the above property will be denoted by DCCdio,τdio, implying that the dependence on d will be clear from the context. We will only use this definition with τdio > d + 1.

In Ref. 12, it was shown that if f is Creg-regular on (−1/2, 1/2) and ω is Diophantine, then the Rayleigh–Schrödinger perturbation series (see below) converges for sufficiently small ɛ > 0. However, one can also apply the construction from Ref. 12 in the case where Creg and Dmin themselves depend on ɛ, under some additional restrictions on the off-diagonal terms of H(x). In Ref. 12, Sec. VI, an example of such an operator was considered. In this section, we will describe a slightly more general class of operators for which the construction from the end of Ref. 12, Sec. VI can be applied, virtually, without any changes. The main results of the present paper will be obtained by reducing various operators to the class described in this section.

We will consider long range quasiperiodic operators with variable hopping terms. A quasiperiodic hopping matrix is, by definition, a matrix with elements of the following form:

Φmn(x)=φmn(x+ω(m+n)/2),m,nZd,
(2.3)

where φm:RC are Lipschitz 1-periodic functions, satisfying the self-adjointness condition

φm=φm̄.

Let also

φεmax{supkφk,εsupkφk}.

Define Range(Φ) to be the smallest number L ≥ 0 such that Φnm ≡ 0 for |mn| > L. We will only consider hopping matrices of finite range. Note that (2.3) can be reformulated as the following covariance property:

Φm+a,n+a(x)=Φmn(x+aω),m,n,aZd.

Fix some RN, and suppose that Φ1, Φ2, … is a family of quasiperiodic hopping matrices with Rangek) ≤ kR, defined by a family of functions φm1,φm2,. The class of operators we would like to consider will be of the following form:

H=V+εΦ1+ε2Φ2+,0ε<1,
(2.4)

where

(V(x)ψ)n=f(x+nω)ψn.

One can easily check that, assuming

φ=supjφj=supj,mφmj<+,0ε<1,

the part Φ = ɛΦ1 + ɛ2Φ2 +⋯ defines a bounded operator on 2(Zd).

The central object of Ref. 12 is the Rayleigh–Schrödinger perturbation series, which is a formal series of eigenvalues and eigenvectors

E=E0+εE1+ε2E2+,
(2.5)
ψ=ψ0+εψ1+ε2ψ2+,
(2.6)

where, in a small departure from the notation of Ref. 12, we assume that

λ0=f(x0+nω),ψ0=en,ψjψ0forj0.
(2.7)

Under the above assumptions, we consider the eigenvalue equation

H(x0)ψ=Eψ

as equality of coefficients of two power series in the variable ɛ,

(V+εΦ1+ε2Φ2+)(ψ0+εψ1+ε2ψ2+)=(E0+εE1+ε2E2+)(ψ0+εψ1+ε2ψ2+).

Assuming that f is strictly monotone, the above system of equations has a unique solution satisfying (2.7).

It will also be convenient to consider a graph Γ(x0) whose set of vertices is Zd, and there is an edge between m and n if Φmn(x0) ≠ 0. The length of that edge is the smallest j > 0 such that Φmnj(x0)0.

We can now describe the class of operators with convergent perturbation series, which will be used in this paper. Let I1, …, Ik ⊂ (−1/2, 1/2) be disjoint closed intervals and μ1, …, μk > 0. We will only consider μjN in applications, but the argument can be easily extended for μj ∈ [0, +). We will make the following assumptions.

  • (conv1)

    f satisfies (f1) from Sec. II A. Additionally, f is Creg-regular on (0, 1)\(I1 ∪…∪Ik) with Dmin ≥ 1.

  • (conv2)

    c1εμjf(x)c2 on Ij. Here, f is allowed to be non-differentiable at some values of x, but then, the inequality is required for all derivative numbers. As a consequence, f is strictly monotone on (−1/2, 1/2).

  • (conv3)

    Suppose that x0 + ω · nIj. Any edge of Γ(x0) that starts at n has a length of at least μj + 1.

The proof of the following result can be done along the same lines as the argument in Ref. 12, Proof of Theorem 6.2.

Proposition 2.1.
Under the above assumptions [(conv1)–(conv4)], there exists
ε0=ε0(Creg,{μj}j=1k,Cdio,τdio,φε,c1,c2)>0
such that, for 0 < ɛ < ɛ0, the perturbation series for operator(2.4)is convergent and the operatorH(x) satisfies Anderson localization for allxR.

Remark 2.2.

Here, we use an additional refinement of Ref. 12, by introducing ‖φɛ instead of considering ‖φ and ‖φ′‖ separately. The reason is that differentiating φ is “cheaper” than differentiating the denominators, and the latter has already been accounted for in (conv2).

One can also consider the perturbation series in finite volume. By “finite volume,” we mean the case when the operator is restricted to a finite box in Zd. The results of Ref. 12, as well as Proposition 2.1, do not extend directly into that case (in the language of Ref. 12, certain loops required for cancellation are forbidden). However, in all our applications, the volume will be fixed before choosing a small ɛ, in which case one can apply the regular perturbation theory of isolated eigenvalues.

Proposition 2.3.
LetH = V + ɛΦ be a self-adjoint operator on2([0, N]d),
(Hψ)n=Vnψn+εmΦnmψm,
with the diagonal partV(in other words, Φnn ≡ 0). Assume that all valuesVkare disjoint,
δ=maxm,k:mk|VmVk|>0.
Then, the coefficients of perturbation series (2.5), (2.6) satisfy
|Ej|C(N,d)jΦjδj,ψjC(N,d)jΦjδj.
As a consequence, both series converge for smallɛ > 0.

Corollary 2.4.
Suppose thatfsatisfies (f1) and {1, ω1, …, ωd} are rationally independent. For anyM > 0 andCreg > 0, there exists
ε0=ε0(M,Creg,ω)>0,δ=δ(M,Creg,ω)
such that the following is true. For anyx0R, any boxBinZdwith |B| ≤ M, anyɛwith 0 < ɛ < ɛ0, and anynB, such thatfisCreg-regular atx0 + ω · n, the operatorHrestricted to the boxBhas a unique simple eigenvalueE = E(x0), with
|E(x0)f(x0+ωn)|εδ1φ.
The corresponding eigenvectorψ,ψ2(B)=1, can be chosen to have
|ψ(x0)en|εδ1φ.
Moreover, bothEandψare Lipschitz continuous atx0and satisfy similar bounds,
|E(x0)f(x0+ωn)|εδ1φ,|ψ(x0)|εδ1φ.

Proof.

Since the components of ω are rationally independent and f is Creg-regular at x0 + ω · n, we have f(x0 + ω · n) separated from the other diagonal entries by a constant that only depends on ω, Creg, and the size of the box. Then, the argument follows from Proposition 2.3. The differentiability follows from differentiating the perturbation series in the variable x0, which, in turn, follows from (cr2).□

Remark 2.5.

Exact same bounds on the derivatives also hold for the infinite volume eigenvectors [in that case, δ = δ(Cdio, τdio, d)]. As a consequence, for most practical purposes involving at most one derivative, regular eigenvalues can be considered to be small perturbations of the original diagonal entries, in both finite and infinite volumes. In a finite volume, the constants depend on the size of the box but can be made independent of the location of the box.

The following is a simple consequence of rational independence.

Proposition 2.6.

Suppose thatfsatisfies (f1). For anyM > 0, there existsW = W(f, M, ω) such that, for any boxBwith |B| ≤ M, all eigenvalues ofHB(x), except for at most one, are bounded byW + ɛφin absolute value.

Proof.

The only possibility for the operator to have a very large eigenvalue is for x0 + ω · n to be close to Z+1/2. However, this implies that all the remaining values of x0 + ω · m, where mn is from the same box, are away from Z+1/2. Ultimately, the dependence on ω is only through Cdio and τdio.□

In this section, we will summarize some basic properties of the eigenvectors of discrete Schrödinger operators in bounded domains. Let CZd. Denote the Laplace operator on 2(C) (with the Dirichlet boundary conditions on Zd\C) by

(ΔCψ)(n)=mC:|mn|1=1ψ(m).

Definition 3.1.

Let A,BZd. We will define nAB to be directly reachable fromB using the following recurrent definition:

  1. Any point, nB is directly reachable in 0 steps.

  2. Let nAB, mA, and |mn| = 1, and the set
    {mZd:|nm|1,mm}
    is contained in AB and only consists of directly reachable points in at most k − 1 steps. Then, we declare m to be directly reachable in at most k steps.

The above definition can be understood as follows. The eigenvalue equation

m:|mn|1=1ψ(m)=(EV(n))ψ(n)
(3.1)

allows us to determine the value ψ(m) from the values

{ψ(m):|mn|1,mm},

where nAB is some point with |nm| = 1. A directly reachable point, therefore, is a point mAB such that ψ(m) can be determined from the values of ψ on B using a finite sequence of such operations.

Definition 3.2.

We will say that BZd satisfies direct unique continuation property (DUCP) for AZd if all points of A are directly reachable from B.

We will need the following very coarse quantitative version of the unique continuation. Better modifications to this lemma can be made in particular situations.

Lemma 3.3.
Assume thatBZdsatisfies DUCP forAZd, and any point ofAis directly reachable fromBin at mostNsteps, and |Vn| ≤ Wfor allnABfor someW > 0. Assume thatHABψ = andψ2(AB)1. Then, there existsmBsuch that
|ψ(m)||AB|1(2d+|E|+W)N.

Proof.

By contradiction, in each step, we can only increase the value of the eigenfunction by a factor of 2d + |E| + W.□

Corollary 3.4.
In Lemma 3.3, replace Δ byɛΔ. Then, the conclusion can be restated as
|ψ(m)||AB|1(2dε+(|E|+W))NεN.

Remark 3.5.

In Lemma 3.3, one can generalize the Laplace operator in the following way: the weight of any edge connecting two adjacent points of B can be replaced by any number between −1 and 1. The reason is that, in the process of calculating the new values of ψm, mB\A, one would not have to divide by these weights.

Remark 3.6.
For a subset AZd and kN, let
kA{nZd\A:dist(n,A)k}.
One can easily check that B2A satisfies DUCP for any bounded A. Now, suppose that A is a box. Then, in Corollary 3.4, one can take N = ⌈diam(B)⌉. Moreover, in Corollary 3.4, one can take W to be the second largest element of the set
{|Vn|:nA1A}.
In other words, for any mA, one can calculate ψ(m) from the values of ψ on B through a sequence of applications of (3.1) without using the value V(n) of the potential at any particular point n. Thus, the value of the potential at one point is allowed to be arbitrary large and yet excluded from consideration in the bound from 3.4.

This argument will be particularly useful in combination with Proposition 2.6 since it states that, while we may not be able to control the largest eigenvalue on a box of given size, we can control all remaining eigenvalues in terms of ω, f, and the size of the box.

We will also need some information about the eigenfunction decay, which follows from elementary perturbation theory.

Proposition 3.7.
LetAandHAbe as above, and letA = A1 ⊔…⊔Anbe a disjoint union. Assume that
|VmVn|η,mAj,nAk,jk.
Then, there exists an orthonormal basis {ψm: mA} of eigenfunctions ofHAsuch that ifmAj, then
|ψm(n)|C(A,{Aj})ηdist(n,Aj).
(3.2)

In other words, if the values of V are clustered, then the corresponding eigenfunctions decay exponentially away from the clusters. The proof immediately follows from the standard perturbation theory.13 The statement is only meaningful for η ≫ 1; otherwise, one can absorb the bound into C(A, {Aj}). The constant does not depend on V and on η, as long as (3.2) is satisfied. In particular, it does not depend on the energy range inside a cluster; additionally, any dependence on the number of lattice points in each cluster can also be absorbed into C.

Remark 3.8.

In Proposition 3.7, one can replace the Laplacian by a weighted Laplacian as long as the weights are bounded above by 1 in absolute value, with the same constants. One can also replace it by a long range operator if the distance function is modified accordingly.

The following statement, which also follows from perturbation theory for 2 × 2 matrices, is useful.

Proposition 3.9.
Under the assumptions of Proposition 3.7, replace ΔAby a weighted Laplacian, with all weights bounded by 1 in absolute value, and the weights of all edges connecting points from different clusters are bounded byɛ. Denote byHAthe operatorHAwith weights of all edges between different clusters replaced by zero. Then,
|λj(HA)λj(HA)|C(A,{Aj})ε2η1,
whereλj(H) denotes thejth eigenvalue ofHin the increasing order.

Proof.
This is a standard argument from perturbation theory. For each pair of points n, m from different clusters, let
P=en,en+em,em,
and consider the operator
H{m,n}PHAPranP.
In other words, restrict HA to 2{en, em}. Since |VmVn| ≥ η, for ɛ < η, we have two eigenvectors of H{m,n}, which are small perturbations of em, en, respectively. Denote these (orthonormal) eigenvectors by fm, fn, and consider the unitary matrix with columns fm, fn,
V=(fm,fn).
Let
V1V12A\{m,n}
and
HA1=V11HAV.
The above procedure is called a partial 2 × 2 diagonalization. One can check that, after such a transformation, the edge between m and n of length ɛ would disappear, and all other entries of the operator are changed at most by O(ɛ2η−1) [it is possible that the edge between m and n does not completely disappear, but its new length will be at most O(ɛ2η−1)].

Apply partial 2 × 2 diagonalization to each edge that connects a pair of vertices from different clusters. By doing this, one will reduce all weights of those edges to at most O(ɛ2η−1). Afterward, additional perturbation of the whole operator of size O(ɛ2η−1) would completely eliminate those edges. Finally, one can restore the entries inside of each cluster to their original form HA′ by another perturbation of the order O(ɛ2η−1). Note that the constants do depend on the size of the domain and geometry of clusters; however, they can be chosen uniformly in the magnitude of the potential (in fact, large values of the potential only make the situation easier) as long as the clusters are η-separated.□

We will also need the following property related to the differentiation of eigenvalues with respect to a parameter. The following is often referred to as Hellmann–Feynman variational formula; see Ref. 13, Remark II.2.2.

Proposition 3.10.
LettA(t) be a Lipschitz continuous family of self-adjoint matrices. Letλ(t) be a branch of an isolated eigenvalue ofA(t), and letψ(t) be a corresponding2-normalized eigenvector. Then,
λ(t)=ddtA(t)ψ(t),ψ(t)=A(t)ψ(t),ψ(t).
(3.3)

Proof.

The remaining term in the right-hand side of (3.3) is λ(t)ddtψ(t),ψ(t), which is identically zero due to normalization of ψ(t).□

In particular, if A(t) = A0 + f(t)⟨ej, ·⟩ej, then

λ(t)=f(t)|ψ(t),ej|2.

Proposition 3.11.
Under the assumptions of Proposition 3.7, assume that
H(t)=H0+f(t)ek,ek.
Suppose thatE = E(t) is an isolated eigenvalue ofH(t) that is separated from the rest of the spectrum byδ > 0. Assume thatmArandnAs. Denote byP(t) the spectral projection onto the eigenspace associated with the eigenvalueE(t). Then,
dPmn(t)dtC(A,{Aj})|f(t)|δ1ηdist(Ar,{k})dist({k},As).

Proof.
The spectral theorem implies that
|(H(t)E)1em,ek|Cδ1ηdist(Ar,{k}).
(3.4)
We also have
ddtP(t)mn=ddt12πi|λE|=δ/2(HAλ)1dλmn,
from which one can get the inequality by differentiating.□

Remark 3.12.

Our use of discrete unique continuation is relatively elementary. We refer the reader to Refs. 2 and 3 for more advanced applications of unique continuation for discrete Schrödinger operators.

Let f:(1/2,1/2)R satisfy (f1) from Sec. II. We will say that f is locally Lipschitz monotone atx0 ∈ (−1/2, 1/2) if

f(x)1,x:|f(x)f(x0)|1.

This condition is a weaker version of Creg-regularity at x0 with Dmin ≥ 1 and will be sufficient for the construction in the current section. Let also

ft(x)=f(x)+t{x1/2},
(4.1)

where {·} denotes the fractional part. Assume that f, ft are extended into R by 1-periodicity. It is easy to see that f1 is locally Lipschitz monotone at all x0 ∈ (−1/2, 1/2). Moreover, if ft is locally Lipschitz monotone at x0, then all ft, t′ ≥ t, are also locally Lipschitz monotone at x0.

In the following theorem, we consider N × N matrix-valued functions whose diagonal entries are locally Lipschitz monotone. Let {e1, …, eN} denote the standard basis in CN.

Theorem 4.1.
Let
d1,,dN(1/2,1/2),minij|didj|=δ>0.
Let also
At(x)=diag{ft(xd1),,ft(xdn)}+εΦ(x),A(x)=A0(x),
where Φ(·) is a continuous real symmetric 1-periodicN × Nmatrix-valued function onR. Assume that all eigenvalues ofAtare simple for allt ∈ [0, 1] and allxR. Then, there existcj = cj(N) > 0,j = 1, 2, such that, for
0εΦ<c1(N)δ,
we have the following:
  1. Suppose thatfis locally Lipschitz monotone atx0dj. Then,A(x) has a unique eigenvalueEj(x) and a unique2-normalized real eigenvectorψj(x) satisfying
    |Ej(x)f(xdj)|c2(N)εΦδ1,|ψj(x)ej|c2(N)εΦδ1.
    (4.2)
  2. There exists a continuous 1-periodic real orthogonal matrix functionU():RO(N,R)such thatU−1(x)A(x)U(x) is a diagonal matrix and the jth column ofU(x) coincides withψj(x) from (1) for allxwhere (1) is applicable.

Remark 4.2.

One needs to specify the exact meaning of “simple” in the case when one of the matrix entries becomes infinite. It is easy to see that, as x approaches dj, exactly one eigenvalue of At(x) approaches ±, and the remaining eigenvalues remain bounded. In fact, they approach the eigenvalues of At with the jth row and column removed and thus can be extended continuously through dj. Thus, we require all those finite eigenvalues to be simple. Equivalently, we require the distances between eigenvalues of At(x) to be bounded by some constant ϰ > 0 for all t ∈ [0, 1] and all xR\({d1,,dN}+Z).

Proof.

Part (1) follows from the standard perturbation theory of isolated eigenvalues; see also Proposition 2.3. Note that, once we fix a real 2-normalized branch, it must be close to ej or −ej, and requiring it to be close to ej completely determines this branch.

Since f1(·) is locally Lipschitz monotone on (−1/2, 1/2), there exist (unique real normalized) continuous branches of eigenvectors,
ψj():(1/2+dj,1/2+dj)Rd,ψj(x)ejc3(N)εΦδ1.
Since the spectra of At are simple, each ψj has a unique real 2-normalized extension in the parameter t into [0, 1]. Denote this extension by ψj(x, t) and the corresponding eigenvalue branch by Ej(x, t). For any x0 such that f is locally Lipschitz monotone at x0dj, any real eigenvector of At(x0) with eigenvalue Ej(x0, t) must be close to ej or −ej, uniformly in the parameter t. Since it is close to ej for t = 1, the same holds for all t.

A similar argument can be applied to x0 near the endpoints. Since one (and only one) eigenvalue of At(x) approaches infinity, the corresponding branch of eigenvector must approach ej or −ej as x → ±1/2 + dj, uniformly in t. Since for t = 1, it approaches +ej from both sides, the same holds for all t. In other words, it cannot suddenly change from a vector close to ej to a vector close to −ej due to continuity in t and the fact that it has to be close to ej or −ej for any particular t.□

Remark 4.3.

Under the assumptions of Theorem 4.1, one can take t → + in (4.1). In this case, ψj(x, t) will be well defined for all t ≥ 0 and uniformly approach ej as t. As a consequence, if the family {At(x)} satisfies the assumptions of Theorem 4.1, there is a unique family {Ut(x)} of real orthogonal matrices such that Ut(x) diagonalizes At(x) and Ut(x) → I as t → +. In this notation, we will have U(x) = U0(x). One can use this as an alternative definition of U(x).

Proposition 4.4.

FixL > 0,aR. LetJbe anN × Nmatrix satisfying the following properties:Jij = 0 for |ij| > 1 andJij=JjiR. Assume thatJi,i+1 ≥ 1 for all applicable values ofiand thatJii ∈ [a, a + L] for alli. Then, the spectrum ofJis simple. Moreover, ifEiEjare two eigenvalues ofJ, then |EiEj| ≥ c(N, L) > 0.

Proof.

The fact that the spectrum is simple is standard and follows from the fact that any solution to the eigenvalue equation is uniquely determined from its value at one point. The rest follows from compactness.□

Remark 4.5.

By rescaling, one can obtain a version of the proposition with |EiEj| ≥ c(N, ɛL)ɛ, provided that |Ji,i+1| > ɛ > 0.

Remark 4.6.

In Proposition 4.4, one can relax the requirement Jii ∈ [a, a + L] for the endpoint values J11 and JNN: one can remove this restriction for one of them or for both if one of them goes to + and the other to −.

Remark 4.7.

In higher dimensions, Proposition 4.4 will be often used in combination with Proposition 3.9: if a cluster from the latter proposition has a linear shape (or is somewhat one-dimensional), then Proposition 4.4 guarantees that its eigenvalues will be -separated (after restricting the operator to the cluster). The presence of coupling terms with other clusters will shift the eigenvalues by at most O(ɛ2) and will thus preserve the separation condition.

As a reminder, the standard basis of Rd (and Zd) is denoted by {e1, e2, …, ed}. The translation operator on functions is defined by

Tn:2(Zd)2(Zd),(Tnψ)(m)=ψ(m+n).

Since we will use the following specific translations more often, we also denote

TTe1,T*T1=Te1.

Consider operator (1.1) on 2(Zd), and assume that f satisfies (f1). Recall that, without loss of generality, we assumed that

0<ω1<ω2<<ωd<1/2.

Fix some x0R and Creg > 0. The construction below will depend on the choice of x0; however, any x0 would be equally applicable. A lattice point nZd will be called regular forH(x) if f is Creg-regular at x + ω · n. As discussed in Corollary 2.4 and Remark 2.5, for sufficiently small ɛ > 0, H(x) has an eigenvalue close to f(x + ω · n) and the corresponding eigenvector close to en at any regular point n. The same also holds for a restriction of H onto any box containing n. The smallness of ɛ depends on the size of the box. Additionally, “close” in this case also applies to derivatives in x. A lattice point that is not regular for H(x) will be, naturally, called singular forH(x). Let

Ssing(x){nZd:n is singular for H(x)}.

For the purposes of the construction, the point with infinite potential will be considered regular. Since the components of ω are rationally independent, there is at most one such point. We will always assume that singular points are contained in a finite range of energies:

(gen0) There is Ereg > 0 such that f is Creg-regular at any x with |f(x)| ≥ Ereg.

Let SZd be a finite subset possibly containing some singular points. Let also RS be another finite subset, with the following properties:

(gen1) S ∪ (S + e1) ⊂ R ∩ (R + e1).

(gen2) For every x ∈ [x0, x0 + ω1], all points of (R ∪ (R + e1))\(S ∪ (S + e1)) are regular for H(x).

Let also

RR\(R+e1),R+(R+e1)\R,R0R(R+e1),RR(R+e1).
(5.1)

We can represent R′ as a disjoint union R′ = RR+R0. Denote by HR(x) the following family of operators on 2(R′):

HR(x0)VR(x0)HR0R+(x0),
HR(x0+ω1)HRR0(x0+ω1)VR+(x0+ω1),

where VR±(x) denotes just the potential part of HR±(x), without the Laplacian, restricted onto the domains R±. For the values x ∈ [x0, x0 + ω1], we will linearly interpolate,

HR(x)(1s)VR(x0)HR0R+(x0)+sHRR0(x0+ω1)VR+(x0+ω1),
x=x0+sω1fors[0,1].

Assume the following condition:

(gen3) Matrix families HR(x), HR0R+(x), and HRR0(x) satisfy the assumptions of Theorem 4.1 for all x ∈ [x0, x0 + ω1].

Technically, in order to apply Theorem 4.1, HR(x) has to be defined (possibly with infinite entries) and be periodic for all xR. The exact form of the extension will not matter. For the sake of clarity, assume that for x ∈ [x0 + ω1, x0 + 1], the off-diagonal parts of HR(x) are linearly interpolated between those of HR(x0+ω1) and HR(x0) so that HR(x0+1)=HR(x0) and then extended 1-periodically to R. For the later construction [see the definition of the operator U0mb(x) below], we will only use HR(x) for [x0, x0 + ω1].

From the definitions, we have

HR0R+(x0)=T*(HRR0(x0+ω1))T,
(5.2)

where T: 2(R0R+) → 2(RR0) is the restriction of the translation map (which is a bijection between R0R+ and RR0). Denote by UR(x), x ∈ [x0, x0 + ω1], the result of applying Theorem 4.1 to the family HR(x). For x = x0 and x = x0 + ω, UR(x) splits into a direct sum,

UR(x0)=1RUR0R+(x0),
UR(x0+ω1)=URR0(x0+ω1)1R+.

The matrices UR0R+(x0) and URR0(x0+ω1) have real entries and, due to (5.2), diagonalize the same operator, up to the identification of RR0 with R0R+. Therefore, they consist of the same eigenvectors, up to the ordering and the choice of the signs. On the other hand, (gen3) guarantees that both approach the identity as t → +; see Remark 4.3. Therefore, they must be related in the same way as operators (5.2) they diagonalize,

UR0R+(x0)=T*URR0(x0+ω1)T.
(5.3)

Extend UR(x) into 2(Zd\R) by identity

U0mb(x)UR(x)1Zd\R.
(5.4)

Then, we have

T*U0mb(x0+ω1)T=U0mb(x0).
(5.5)

We start from U0mb(x) defined for x ∈ [x0, x0 + ω1]. One can easily check that there is a unique extension of this family into xR satisfying

U0mb(x+ω1)=T1U0mb(x)T,xR.
(5.6)

This above construction will be called the diagonalization of a moving block, which is reflected in the superscript mb. Note that the singular set also has the following covariance property:

Ssing(x0)=Ssing(x0ωn)+n.
(5.7)

As a consequence, if the above-mentioned set S from the moving block construction contains singular points for H(x0), they will naturally correspond to singular points in Se1 for H(x0 + ω1). In the above construction, as x increases, the set S “moves” along Zd with speed 1/ω1 and traces a part of Ssing, thus justifying the name of the construction.

It will also be convenient to use the following language: a normal operator on 2(Zd) is supported on R′ if 2(R′) is its invariant subspace and the operator acts as an identity on 2(Zd\R). The above construction implies that U0mb(x) is supported on R′ for x ∈ [x0, x0 + ω1] and is supported on R′ − me1 for x ∈ [x0 + 1, x0 + (m + 1)ω1], where mZ.

The operator family U0mb(x) takes care of only a part of Ssing(x) that corresponds to a finite block S that moves as x moves around R. We now need to cover the whole singular set by such blocks and multiply the corresponding operators. At the same time, the family U0mb(x) is not a quasiperiodic operator family: its entries are not even periodic in x. We will need to modify this family in order for it to be quasiperiodic. First, note that the set Ssing(x) is not only translationally covariant (5.7) but also translationally invariant,

Ssing(x+1)=Ssing(x).

Therefore, with each moving block, we actually have a “train” of blocks, with a translation by 1. Thus, it is natural to consider

U1mb(x)mZU0mb(x+m).
(5.8)

We will later require [see (gen4)] that the supports of U0mb(x+m) are all disjoint. As a consequence, the above product will be well defined.

Unlike U0mb(x), the family U1mb(x) is actually a periodic function in x and is quasiperiodic with respect to translations by ω1,

U1mb(x)=U1mb(x+1),U1mb(x+ω1)=T1U1mb(x)T,xR.
(5.9)

We now enforce the full covariance (quasiperiodicity) condition. Let ω′ = (ω2, …, ωd). For any nZd1\{0}, the family xT(0,n)V1(x + n′ · ω′)T−(0,n) describes a different moving block (in fact, a different train of moving blocks described above). Let

U2mb(x)nZd1T(0,n)U1(x+nω)T(0,n).
(5.10)

Each factor in (5.10) contains countably many copies of U0mb(x). If well defined, U2mb(x) is a fully quasiperiodic operator family,

U2mb(x)=U2mb(x+1),U2mb(x+ωn)=TnU2mb(x)Tn,xR.
(5.11)

If order for U1mb and U2mb is to be well defined, we will require the last condition:

(gen4) The supports of each copy of U0mb inside of operators (5.10) and (5.8) are disjoint.

The above construction will be referred to as a covariant family of moving blocks. The set S will be called the base block.

Remark 5.1.

Once we fix S and R′ for x ∈ [x0, x0 + ω1], the position of other copies of S and R′, appearing in different factors of (5.10) and (5.8), is completely determined by S, R, and ω.

One may need more than one such family to take care of all Ssing(x), in which case we will require all involved operators U2mb to commute with each other (in other words, all sets R′ under consideration will not overlap with each other). For simplicity, we will now consider the case with only one covariant moving block. Let

H2(x)U2mb(x)1H(x)U2mb(x).

We will now outline the general strategy of the proof of localization. Our goal is to show that, under some additional conditions on the sets R0 (which will impose some separation conditions on S), the operator family H2(x) will satisfy the assumptions of Proposition 2.1. For the convenience of referencing, we will denote the steps of the strategy by (s1)–(s5).

  • (s1)
    Quasiperiodicity of U2mb(x) (5.11) implies that H2(x) is also a quasiperiodic operator family, with diagonal entries defined by a new function f2,
    (H2(x))nn=f2(x+ωn).
    For sufficiently small ɛ (depending on the size of R′), the function f2 will be, say, Creg/2-regular at any x where the original f was Creg-regular; see Corollary 2.4 and Remark 2.5.
  • (s2)

    As a consequence, our attention should be toward the diagonal entries of H2 corresponding to singular points for H(x). Assume that dist((S(S+e1)),Zd\R0)1 so that we avoid the boundary effects. Then, for any m ∈ (S ∪ (S + e1)), f2(x + ω · m) is the exact eigenvalue of HR(x) that corresponds to the lattice point m. Here, we use the relation between eigenvalues of HR(x) and lattice points of R′ provided in Theorem 4.1.

  • (s3)

    Suppose that, for a single block, we have a subset BR0\(S ∪ (S + e1)) that satisfies DUCP for S ∪ (S + e1). For example, if R0 is a box whose boundary is not too close to S ∪ (S + e1), then we can surround S ∪ (S + e1) by a layer of thickness two inside R0; see Remark 3.6. For any eigenvalue of HR(x), corresponding to a lattice point mS ∪ (S + e1), apply Lemma 3.3 to the corresponding eigenfunction and conclude that it must be non-trivially supported on B. Since every point of B is regular, the corresponding diagonal entry of the operator is Lipschitz monotone. Proposition 3.10 guarantees that there is a non-trivial positive contribution to the derivative of each singular eigenvalue from some diagonal entries of B. The contribution from the remaining diagonal entries of HR(x) to the derivative is also non-negative.

  • (s4)

    The previous part almost guarantees (conv2), with one caveat: the interpolation process in the definition of HR(x) creates some x-dependent off-diagonal entries, which also contribute to the derivatives of singular eigenvalues not necessarily in a positive way. However, these entries are outside of R0. If we require S ∪ (S + e1) to be away from Zd\R0, we can make this contribution arbitrarily small, independent of the lower bound on the contribution from B described above. This will complete the verification of (conv2). The particular selection of the set B does not affect the definition of U2mb, but it only affects our choice of what to use as a lower bound on the derivatives. Thus, it can be chosen in an x-dependent way as long as the ultimate lower bounds are uniform in x.

  • (s5)

    In order to verify (conv3), note that, after diagonalization, singular points will only be coupled with points outside of R0. The strength of this coupling (i.e., in the language of Sec. II, the length of the corresponding edges) will be in terms of the decay of the corresponding eigenfunctions of HR, which, in turn, can be obtained using Proposition 3.7 and Proposition 3.11 [the latter is used to control the derivatives, as explained in (conv4)]. Ultimately, one can make it bounded by an arbitrarily large power of ɛ, by requiring dist((S(S+e1)),Zd\R0) to be large enough.

We will summarize the above construction in the following theorem.

Theorem 5.2.
Fixd,Creg,Cdio,τdio,Dmin ≥ 1, andcsep > 0. Let alsoωDCCdio,τdiosatisfy
0<ω1<ω2<<ωd<1/2.
LetSZdbe a finite subset. There existsr=r(S)Nand
ε0=ε0(d,Creg,Cdio,τdio,csep,S,Ereg,f)>0
such that, for 0 < ɛ < ɛ0and allxR, the operatorH(x) satisfies Anderson localization and is unitarily equivalent to an operator with convergent perturbation series if the following are satisfied for somex0R:
  1. fsatisfies (f1) and (gen0).

  2. There exists a boxRSsuch that the family of covariant moving blocks, constructed above, satisfies (gen1)(gen4).

  3. DefineR0,Ras in(5.1). We will requiredist(S(S+e1),Zd\R0)r.

  4. Ssing(x) is contained in the support ofU2mb(x), constructed above in(5.10). Due to (gen2), this means thatSsing(x) is contained in the union of the copies of the setsS ∪ (S + e1) that appear in the definition ofU2mb(x).

  5. The eigenvalues ofHS(S+e1)(x)are uniformlycsepɛ-separated for allx ∈ [x0, x0 + ω1]. Moreover, this property holds withfreplaced byftfrom(4.1), uniformly int ∈ [0, 1].

Remark 5.3.

There is additional dependence of ɛ0 on f, besides the dependence through Creg and Ereg. This dependence can be made explicit; we will specify it here in order to avoid overloading the statement of the main result. Consider the smallest box B such that dist(S(S+e1),Zd\B)4. Then, ɛ0 depends on f through the quantity W(f, M, ω) from Proposition 2.6. Note that the dependence on ω can be reduced to dependence on Cdio, τdio, and the dependence on M is accounted for in the dependence on S. See also Remark 3.6.

Proof.
We will show that the procedure described in the beginning of this section results in an operator unitarily equivalent to H(x) and satisfying the assumptions of Proposition 2.1. Fix r ≥ 4, and take some box R such that dist(S(S+e1),Zd\R)r, where R′ is defined in (5.1). We will specify additional conditions on r later. Fix some x0, and construct the covariant family of moving blocks, as described earlier in this section. For x ∈ [x0, x0 + ω1], let B(x) be the smallest box in Zd such that the following holds:
  1. B(x) ⊃ (S ∪ (S + e1)) and dist((S(S+e1)),Zd\B(x))1.

  2. For any box RS with dist(Zd\R,S(S+e1))4, the set

{nB(x):dist(n,Zd\B(x))1}
does not contain the point of R′ with largest absolute value of the potential.

The choice of B(x) can be made independently of R. Indeed, we can always choose two non-overlapping layers of thickness two around S and then pick one of them with the smaller maximal value of the potential. For example, let B1(x) be the smallest box such that dist((S(S+e1)),Zd\B1(x))2 and B2(x) be the smallest box such that dist(B1(x),Zd\B2(x))2. Then, for any x, we can either pick B = B1(x) or B = B2(x).

Construct the operator
H2(x)=U2mb(x)1H(x)U2mb(x)
as mentioned earlier in this section, using S and R′. The diagonal entries of H2(x) correspond to the lattice points and are ordered in the same way as the values of the potential. The entries at distance at least 2 from Zd\R are exact eigenvalues of the block HR(x) (for the respective lattice points). The corresponding eigenvectors have a cluster structure, in the sense of Proposition 3.7: one cluster would be (S ∪ (S + e1)) ∩ Ssing(x), and each of the remaining lattice points of R′ corresponds to a separate cluster, with separation bounded below by a quantity that, ultimately, depends only on r but not on ɛ. Proposition 3.7 implies that each regular eigenvector ψ decays as
ψ(m)Cεdist(m,n),
where n is its support point and each singular eigenvector decays as
ψ(m)Cεdist(m,S(S+e1)).
Here, C depends on r, but not on ɛ. Moreover, the derivatives of the eigenvectors in x satisfy similar decay with a loss of one ɛ; see Proposition 3.11.

As a consequence, each singular diagonal entry of H2(x) located on S ∪ (S + e1) has only outgoing edges of length bounded by r−2, with a derivative bound r−3. It would be sufficient to show that the assumption (conv2) is satisfied with μ independent of r. Note that we already have μ = 1 for each regular eigenvalue.

In order to establish the above assumption, apply Lemma 3.3 with
B={nB(x):distn,Zd\B(x)2},A=S(S+e1).
Proposition 3.7 implies that any 2(R′)-normalized eigenfunction ψ associated with a singular lattice point satisfies
ψ2(S(S+e1))Ssing(x)1O(ε)1/2
for sufficiently small ɛ. Lemma 3.3, in view of Remark 3.6 (see also Remark 5.3), implies that |ψ(b)|c1εc2 for some bB, where c2 only depends on S and c1 depends on Ereg. Since b is a regular point, Proposition 3.10 implies that the derivative of the singular eigenvalue, corresponding to ψ, has a contribution bounded below by C1ɛ2c, where C1 > 0 can be chosen to only be dependent on S, ω, and Ereg. The contributions from all other diagonal entries of H′(x) are also non-negative. Following (s4), we also need to account for the terms associated with the interpolated diagonal entries. These entries are located somewhere in R+R, which are outside of R0 and therefore away from S. By combining Propositions 3.7 and 3.10, we can see that their contribution to the derivatives of singular eigenvalues is bounded above by a power of ɛ that grows with r. Thus, as long as, say, r > 3c + 2, the assumptions are satisfied.□

One can naturally extend the result of Theorem 5.2 to the case where one requires more than one type of blocks to cover Ssing, assuming that they do not “interact” with each other.

Corollary 5.4.
Let
S={S1,,Sk}
be a collection of subsets ofZd. There existsr=r(S)Nsuch that the following is true. Suppose that the assumptions of Theorem 5.2 are satisfied for eachSj, with a modification of (4) thatSsingis contained in the union of supports of the corresponding operatorsU2mb,j, and those supports do not overlap. Then, the conclusion of Theorem 5.2 holds.

As the title suggests, in this section, we will discuss several more concrete examples of the construction from Theorem 5.2. In each particular case, one can make some optimizations in choosing the unique continuation sets and obtain weaker requirements on the separation between different blocks.

We will illustrate the above procedure, first, on the case where f has a single flat segment and is Creg-regular outside of a neighborhood of it. Assume the following.

(z1) f(x) = E for x ∈ [a, a + L] = [bL/2, b + L/2] ⊂ (−1/2, 1/2).

(z2) L is not an integer multiple of ω. Let L = + z, pN, z ∈ (0, 1). Let also β = min{z, 1 − z}.

(z3) f is Creg-regular outside of [aβ, a + L + β].

(z4) Let M = ⌈L/2ω⌉ + 2. Assume that the points

bMω,b(M1)ω,,b,b+ω,,b+(M+1)ω

are contained inside the interval (−1/2, 1/2).

Condition (z4) is the most important one and has to be present in some form. It manifests the separation condition (2) from Theorem 5.2. Let also

Eregmax{|f(bMω)|,|f(b+(M+1)ω)|}.

Theorem 6.1.
FixCreg,Cdio,τdio,Dmin = 1,β > 0, andMN. Let alsoωDCCdio,τdio(0,1/2). Letfsatisfy (f1) and (z1)–(z4) with the M,βfixed previously. There exists
ε0=ε0(Creg,Cdio,τdio,β,M,Ereg)>0
such that, for 0 < ɛ < ɛ0, operator(1.1)is unitarily equivalent to an operator with convergent perturbation series. As a consequence,H(x) satisfies Anderson localization for allxR.

Proof.

Let HM(x) be the restriction of H(x) into the interval [−M, …, M + 1], a (2M + 2) × (2M + 2)-matrix. The asymmetry is caused by the fact that we will consider H(x) for x ∈ [b, b + ω].

For x ∈ [b, b + ω], consider the matrix HM(x) and modify it as follows: the entries that couple −M with −M + 1 will be linearly interpolated from ɛ at x = b to 0 at x = b + ω and the entries coupling M and M + 1, respectively, in the opposite way: 0 at x = b and ɛ at x = 0. Denote the resulting operator HM′(x). Denote also by UM(x) the diagonalizing operator obtained in Theorem 4.1 for HM′(x). Proposition 4.4 combined with Proposition 3.9 guarantees that the family UM(x) satisfies the assumptions of Theorem 4.1 (note that we cannot apply 4.4 directly since, after rescaling, the range of energies will be of the size ɛ−1; however, one can use 3.9 to treat large eigenvalues corresponding to near the edges of the interval).

Denote by U0mb(x) the matrix UM(x) extended as the identity into 2(Z). At the moment, it has only been defined for x ∈ [b, b + ω]. However, the block structure of U0mb(x) implies (for a single value of x = b for which the equality makes sense) that
U0mb(x+ω)=T1U0mb(x)T.
(6.1)
As in Sec. V, we will use (6.1) as the definition of U0mb(x) for xR. We now perform the above procedure with each “copy” of the singular interval appearing in H(x). That is, define
U1mb(x)=mZU0mb(x+m).
Note that (z4) guarantees that the supports of the factors in the above product are disjoint. We do not need to introduce U2mb since there are no other lattice directions.
Now, let
H1(x)=U1mb(x)1H(x)U1mb(x).
At x = a, the diagonal matrix elements of H(a) at [−M + 3, M − 3] correspond to the values of f calculated on the flat segment. They are exactly equal to the eigenvalues of H′(a). In the notation of Sec. V, we have S = [−M + 3, M − 3].
As x goes from a to a + ω, the position of the flat segment on Z shifts by 1: the entry located at −M + 3 becomes regular, and the one located at M − 2 becomes singular, and the flat segment now covers the interval [−M + 2, M − 2]. The entries at −M, −M + 1, −M + 2, M − 1, M, M + 1 are always regular and are small perturbations of the entries of H(x) and therefore inherit their monotonicity properties. The remaining eigenvalues correspond to the set [−M + 3, M + 2], which will play the role of S ∪ (S + e1) from Sec. V,
S(S+e1)=[M+3,M2].
For each x ∈ [b, b + ω], consider the points
xMω,x(M1)ω,,x,x+ω,,x+(M+1)ω.
(6.2)
Assumption (z2) guarantees that both endpoints of [a, a + L] cannot be close to points (6.2) at the same time. More precisely, either a or a + L is at least β/2 away from points (6.2). As a consequence, one can pick two adjacent points p, p + 1 ∈ [−M, M + 1] such that x + and x + (p + 1)ω are two closest points to [a, a + L] on one side but still not too close (both are β/2-away). For example, if x = b, then the points p = −M + 1, p + 1 = −M + 2 would work. As x increases, we may have to switch to a pair of points on the other side. The set {p, p + 1} will play the role of B, the unique continuation set from Sec. V. As discussed previously, this set is x-dependent; however, all cases can be treated similarly, and we will assume that B = {−M + 1, −M + 2} for simplicity.

Let ψ be an eigenfunction corresponding to one of the lattice points on S ∪ (S + e1) = [−M + 3, M − 2], normalized in, say, 2([−M, M + 1]), with eigenvalue Ej(x), j ∈ [−M + 3, M − 2] (here, the eigenvalues are identified with lattice points through Theorem 4.1). By Proposition 3.7, ψ is supported on [−M + 3, M − 2] up to O(ɛ). Moreover, ψ(−M + 2) = O(ɛ) and ψ(M + 1) = O(ɛ2). The values f(x + ), mS, are all equal to E, except maybe for f(x + (M − 2)ω).

We will now discuss the unique continuation procedure from B = {−M + 1, −M + 2} into [−M + 3, M − 2]. Proposition 4.4 combined with Proposition 3.7 implies that
Ej+1(x)Ej(x)c1(ω,M)ε.
On the other hand, by considering the eigenvalues as perturbations of the diagonal elements, we also have
|Ej(x)E|3ε,j[M+3,M3],|EM2(x)f(x+(M2)ω)|3ε.
In both cases, the right-hand side is actually 2ɛ + c2(ω, M)ɛ2, and we assumed ɛ to be small enough for simplicity. Assume first that ψ is a singular eigenfunction corresponding to Ej, with j ∈ [−M + 3, M − 3]. By iterating the eigenvalue equation, it is easy to see that
|ψ(M+2)|c(M,ω,Ereg)ε.
(6.3)
Indeed, if |ψ(−M + 2)| is too small, then since |ψ(−M + 1)| = O(ɛ2), iterating the eigenvalue equation will contradict the 2-normalization condition. Apply Proposition 3.10 and conclude that
Ej(x)c(M,ω,Ereg)ε2,j[M+3,M3].
For the remaining eigenvalue EM−3, we need to consider two cases. If, say, |Ef(x + (M − 2)ω)| < 10ɛ, then the previous argument also applies to the eigenfunction corresponding to EM−3. If |Ef(x + (M − 2)ω)| ≥ 10ɛ, then the point M − 2 can be considered as an individual cluster, and the corresponding eigenfunction must be non-trivially supported at that point [one can check that |ψ(M2)|12ψ2]. Then, one can apply Proposition 3.10 directly at that point and obtain an even stronger conclusion,
EM2(x)c(M,ω,Ereg)>0.
In other words, such a point can be considered as a regular point for practical purposes.
The above argument implies (conv2) from Proposition 2.1 with μ = 2 for our operator. We will illustrate (conv3) on the example with M = 5. The result of applying Proposition 3.7 to the eigenvectors of HM(x) can be described as follows:
UM(x)1εε2ε3ε3ε3ε3ε3ε3ε9ε10ε11ε1εε2ε2ε2ε2ε2ε2ε8ε9ε10ε2ε1εεεεεεε7ε8ε9ε3ε2ε111111ε6ε7ε8ε4ε3ε2111111ε5ε6ε7ε5ε4ε3111111ε4ε5ε6ε6ε5ε4111111ε3ε4ε5ε7ε6ε5111111ε2ε3ε4ε8ε7ε6111111εε2ε3ε9ε8ε7εεεεεε1εε2ε10ε9ε8ε2ε2ε2ε2ε2ε2ε1εε11ε10ε9ε3ε3ε3ε3ε3ε3ε2ε1.
An entry ɛk means that the corresponding component of the eigenvector is bounded above by k, where c does not depend on ɛ. The above matrix illustrates the worst possible case for x ∈ [a, a + ω]: for example, at least one entry in edge of the central block is, actually, ɛ instead of 1 and some of the remaining entries in that column are actually better by one order of ɛ. Remark 3.11 implies that the derivatives in x of each eigenvector are at most ɛ worse than the actual components. The central 6 × 6 block corresponds to the interval [−2, 3], which covers the locations of singular diagonal entries corresponding to the flat piece.
Since UM(x) diagonalizes a block of H′(x) rather than H(x), the conjugation of H(x) by U0mb(x) will result in some extra off-diagonal entries. To calculate them, let us subtract H′(x) from H(x) and calculate the result of the conjugation of the remaining part,
1000UM(x)T0001O(1)ε00000ε0ε00000ε001×80000008×108×808×10000001×80ε00000ε0ε00000εO(1)1000UM(x)0001.
Each element of the matrix product is made out of products of three entries, one from each matrix. Even without multiplying three (14 × 14)-matrices, one can see the following: in order to produce a non-zero element that starts from an entry inside the [−2, 3] block, one needs to use a matrix entry from the right factor that connects the point entry from [−2, 3] to a point outside of the interval [−3, 4], thus gaining at least ɛ2. If we gained exactly ɛ2, then one will gain at least one ɛ from the central factor. Thus, any jump that starts inside of the [−2, 3] block will cost at least ɛ3, which is sufficient to satisfy (conv3) and apply Proposition 2.1 [in view of (conv4) and Remark 3.11].□

Remark 6.2.

One can check that, under the assumptions of Theorem 6.1, the derivative of the integrated density of states of the family H(x) has a spike of height ≍ɛ−2 and width ≍ɛ2 around energy E. One can produce larger spikes by combining multiple flat intervals at different energies; see Examples 5 and 6.

The approach from Subsection VI A can be extended to several higher-dimensional examples.

Example 2: Let d ≥ 2. Suppose that f and ω1 satisfy all assumptions from Subsection VI A (with ω = ω1). Additionally, assume the following:

(z5) For any x ∈ [a, a + L] and any nZd, nZe1, and |n|1 ≤ 6, we have f being Creg-regular at x + n · ω.

Theorem 6.3.
Suppose that the assumptions of Theorem 6.1 are satisfied withωreplaced byω1. Assume also thatωDCCdio,τdioand (z5). Then, there exists
ε0=ε0(d,Creg,Cdio,τdio,β,M,Ereg)>0
such that operator(1.1)satisfies Anderson localization onZd.

Proof.
We only outline the changes in the proof. Instead of the interval [−M, M + 1], consider the box
[3,3]d1×[M,M+1]
and construct UM(x) by applying Theorem 4.1 to the above box. One will need to interpolate the Laplacian terms that couple [−3, 3]d−1 × {−M} and [−3, 3]d−1 × {−M + 1} between ɛ and 0 and the terms coupling [−3, 3]d−1 × {M} and [−3, 3]d−1 × {M + 1} between 0 and ɛ as x goes from b to b + ω1. The eigenvalues on the “central” box {0}d−1 × [−M, M + 1] are -separated due to Proposition 4.4. Eigenvalues corresponding to the remaining entries are separated from each other by a constant independent of ɛ for ɛ ≪ 1 due to Proposition 3.7. Therefore, expansion of the box in the directions perpendicular to e1 will shift the eigenvalues of the block {0}d−1 × [−M, M + 1] by at most O(ɛ2) due to Proposition 3.9 and therefore will preserve O(ɛ)-separation property.
The rest follows from the Proof of Theorem 6.1. One needs to specify the unique continuation subset. For x = a, it will be
B={0}d1×{M+1,M+2}{nZd1:|n|1=1}×[M+1,M3].
In other words, we are allowed to make up to two steps to the left from the singular subset and one step in one other direction. The set B can be imagined as a “claw” surrounding the singular set S.

Proposition 3.7 implies that any singular eigenfunction restricted to {M+1}×{nZd1:|n|1=1} will be O(ɛ2). The rest of the set B is one step away from S, and therefore, any eigenfunction will have, ultimately, at least of mass somewhere on B. One may have to treat the right-most point of S separately, and the argument does not change from Theorem 6.1, in view of Proposition 4.4 and Proposition 3.9.

Once UM(x) is constructed for x ∈ [b, b + ω1], we repeat the steps in Sec. V in order to construct the operator U2mb. Condition (z3) will guarantee that the supports of different copies of U0mb(x) will not overlap.□

In this subsection, we outline several additional cases, which, ultimately, are covered by Theorem 5.2.

Example 3: One can consider multiple intervals of the type described in Theorem 6.1 as long as they are separated enough so that the supports of the operator U1 are disjoint. Similarly, one can combine multiple “non-interacting” cases of Theorem 6.3.

Example 4: More interestingly, if two intervals from Theorem 6.1 are not sufficiently disjoint, one can consider them as one singular set and apply the argument from the general theorem. One can also do it in the setting of Theorem 6.3, with a strengthened version of (z5).

Example 5: The following example can be called “a chain of intervals.” Let d = 1 and I1, …, IN ⊂ (−1/2, 1/2) be a collection of disjoint intervals,

f(x)=EjforxIj,Ej<Ekforj<k

and for j = 1, …, N − 1. Assume also that, for sufficiently large r = r(N) (r = 2N should be enough), we have

1/2<I1rω<IN+rω<1/2.

Then, the assumptions of Theorem 5.2 are satisfied, and we have Anderson localization for small ɛ.

Remark 6.4.

Suppose that, on top of all these assumptions, Ij+1 = Ij + ω, and we also assume for simplicity that N = 2k − 1. Then, it is easy to see that f2′(x) ≍ ɛ2k for xIk. As a result, the derivative of the integrated density of states has a spike of height ≍ɛ−2k and width ≍ɛ2k around Ek.

Example 6: The following example is a combination of Examples 2 and 5. It can be called “a tree of intervals.” Let I1, …, IN be a collection of disjoint intervals,

f(x)=EjforxIj,EjEkforjk.

Suppose that for some xI1, we have, say, x2x + ω2I2. Additional shifts by ω2, …, ωd may produce points, say, x3 = x2 + ω3I3, …. Assume that, for any xI1, we are guaranteed to escape all intervals after, say, q steps in the directions perpendicular to e1 and would only be able to come back to I1 after additional r steps. In other words, for xI1, f is regular at x + ω′ · n′ for all nZd1, q ≤ |n′| ≤ q + r. One can refine this in each particular case if the number of steps needed to escape the system of intervals depends on the direction. In all cases, we would require a “layer” of thickness r that consists of regular points and surrounds the collection of singular points described above. The thickness of the layer will depend on the number of steps required to escape the singular intervals. For r sufficiently large, the assumptions of Theorem 5.2 will be satisfied. In particular, to establish the separation condition (5) from Theorem 5.2, note that each interval produces a block with -separated eigenvalues due to Proposition 4.4. Different intervals are located at different energies and therefore do not affect each other (each lies in its own cluster). The coupling between these intervals satisfies the assumptions of Proposition 3.9 and therefore also does not affect the separation condition. After the conjugation, we will have, for the new diagonal function f2(x),

f2(x)εμj,xIj,

where μj depends on the number of steps in the directions perpendicular to e1 required to escape the interval (by more careful analysis, one can show that it is equal to twice the number of steps). As a consequence, the derivative of the integrated density of states of H will have spikes of order εμj around the energies close to Ej.

The research of L.P. was partially supported by the EPSRC, Grant Nos. EP/J016829/1 and EP/P024793/1. R.S. was partially supported by the NSF, Grant No. DMS–1814664. I.K. was partially supported by the NSF, Grant No. DMS–1846114.

At various stages of their careers, all authors were deeply inspired by the groundbreaking work of Jean Bourgain. This paper is dedicated to his memory.

Data sharing is not applicable to this article as no new data were created or analyzed in this study.

1.
J.
Béllissard
,
R.
Lima
, and
E.
Scoppola
, “
Localization in ν-dimensional incommensurate structures
,”
Commun. Math. Phys.
88
(
4
),
465
477
(
1983
).
2.
J.
Bourgain
and
A.
Klein
, “
Bounds on the density of states for Schrödinger operators
,”
Inventiones Math.
194
,
41
72
(
2013
).
3.
J.
Ding
and
C. K.
Smart
, “
Localization near the edge for the Anderson Bernoulli model on the two dimensional lattice
,”
Inventiones Math.
219
,
467
506
(
2020
).
4.
A. L.
Figotin
and
L. A.
Pastur
, “
An exactly solvable model of a multidimensional incommensurate structure
,”
Commun. Math. Phys.
95
(
4
),
401
425
(
1984
).
5.
D. R.
Grempel
,
S.
Fishman
, and
R. E.
Prange
, “
Localization in an incommensurate potential: An exactly solvable model
,”
Phys. Rev. Lett.
49
,
833
836
(
1982
).
6.
B.
Simon
, “
Almost periodic Schrödinger operators. IV. The Maryland model
,”
Ann. Phys.
159
(
1
),
157
183
(
1985
).
7.
S.
Jitomirskaya
and
W.
Liu
, “
Arithmetic spectral transitions for the Maryland model
,”
Commun. Pure Appl. Math.
70
(
6
),
1025
1051
(
2017
).
8.
S.
Jitomirskaya
and
I.
Kachkovskiy
, “
All couplings localization for quasiperiodic operators with Lipschitz monotone potentials
,”
J. Eur. Math. Soc.
21
(
3
),
777
795
(
2019
).
9.
S.
Jitomirskaya
and
F.
Yang
, “
Pure point spectrum for the Maryland model: A constructive proof
,”
Ergodic Theory Dyn. Syst.
41
(
1
),
283
294
(
2021
).
10.
S.
Jitomirskaya
and
F.
Yang
, “
Singular continuous spectrum for singular potentials
,”
Commun. Math. Phys.
351
(
3
),
1127
1135
(
2017
).
11.
I.
Kachkovskiy
, “
Localization for quasiperiodic operators with unbounded monotone potentials
,”
J. Funct. Anal.
277
(
10
),
3467
3490
(
2019
).
12.
I.
Kachkovskiy
,
L.
Parnovski
, and
R.
Shterenberg
, “
Convergence of perturbation series for unbounded monotone quasiperiodic operators
,” arXiv:2006.00346.
13.
T.
Kato
,
Perturbation Theory for Linear Operators
(
Springer
,
1995
).