Buckets:

|
download
raw
255 kB

Title: The Power of Preconditioning in Overparameterized Low-Rank Matrix Sensing

URL Source: https://arxiv.org/html/2302.01186

Markdown Content: Back to arXiv

This is experimental HTML to improve accessibility. We invite you to report rendering errors. Use Alt+Y to toggle on accessible reporting links and Alt+Shift+Y to toggle off. Learn more about this project and help improve conversions.

Why HTML? Report Issue Back to Abstract Download PDF Abstract 1Introduction 2Problem formulation 3Main results 4Analysis 5Numerical experiments 6Discussions

HTML conversions sometimes display errors due to content that did not convert correctly from the source. This paper uses the following packages that are not yet supported by the HTML conversion tool. Feedback on these issues are not necessary; they are known and are being worked on.

failed: arydshln.sty

Authors: achieve the best HTML results from your LaTeX submissions by following these best practices.

License: CC BY 4.0 arXiv:2302.01186v4 [cs.LG] 30 Dec 2025 The Power of Preconditioning in Overparameterized Low-Rank Matrix Sensing Xingyu Xu CMU Carnegie Mellon University; Email: {xingyuxu,yandis}@andrew.cmu.edu. Yandi Shen1 CMU Yuejie Chi Yale Yale University; Email: yuejie.chi@yale.edu. Cong Ma UChicago University of Chicago; Email: congm@uchicago.edu. (February 2023; Revised December 2025) Abstract

We propose 𝖲𝖼𝖺𝗅𝖾𝖽𝖦𝖣 ​ ( πœ† ) , a preconditioned gradient descent method to tackle the low-rank matrix sensing problem when the true rank is unknown, and when the matrix is possibly ill-conditioned. Using overparameterized factor representations, 𝖲𝖼𝖺𝗅𝖾𝖽𝖦𝖣 ​ ( πœ† ) starts from a small random initialization, and proceeds by gradient descent with a specific form of damped preconditioning to combat bad curvatures induced by overparameterization and ill-conditioning. 𝖲𝖼𝖺𝗅𝖾𝖽𝖦𝖣 ​ ( πœ† ) is remarkably robust to ill-conditioning compared to vanilla gradient descent ( 𝖦𝖣 ) even with overparameterization. Specifically, we show that, under the restricted isometry property (RIP) of the sensing operator, 𝖲𝖼𝖺𝗅𝖾𝖽𝖦𝖣 ​ ( πœ† ) converges to the true low-rank matrix at a constant linear rate after a small number of iterations that scales only logarithmically with respect to the condition number and the problem dimension. This significantly improves over the convergence rate of vanilla 𝖦𝖣 which suffers from a polynomial dependency on the condition number. Furthermore, we show that in the presence of measurement noise, 𝖲𝖼𝖺𝗅𝖾𝖽𝖦𝖣 ​ ( πœ† ) converges to the minimax optimal error up to a multiplicative factor of the condition number at the same rate as in the noiseless setting, which is the first nearly minimax-optimal overparameterized gradient method for low-rank matrix sensing scaling with the true rank rather than the (possibly much larger) overparameterized rank. Our results also extend to the setting when the matrix is only approximately low-rank under the Gaussian design. Our work provides evidence on the power of preconditioning in accelerating the convergence without hurting generalization in overparameterized learning.

Keywords: low-rank matrix sensing, overparameterization, preconditioned gradient descent method, random initialization, ill-conditioning

Contents 1Introduction 2Problem formulation 3Main results 4Analysis 5Numerical experiments 6Discussions 1Introduction

Low-rank matrix recovery plays an essential role in modern machine learning and signal processing. To fix ideas, let us consider estimating a rank- π‘Ÿ ⋆ positive semidefinite matrix 𝑀 ⋆ ∈ ℝ 𝑛 Γ— 𝑛 based on a few linear measurements 𝑦 ≔ π’œ ​ ( 𝑀 ⋆ ) , where π’œ : ℝ 𝑛 Γ— 𝑛 β†’ ℝ π‘š models the measurement process. Significant research efforts have been devoted to tackling low-rank matrix recovery in a statistically and computationally efficient manner in recent years. Perhaps the most well-known method is convex relaxation (candes2011tight; recht2010guaranteed; davenport2016overview), which seeks the matrix with lowest nuclear norm to fit the observed measurements:

min 𝑀 βͺ° 0 β€– 𝑀 β€– βˆ— s.t. 𝑦

π’œ ​ ( 𝑀 ) .

While statistically optimal, convex relaxation is prohibitive in terms of both computation and memory as it directly operates in the ambient matrix domain, i.e., ℝ 𝑛 Γ— 𝑛 . To address this challenge, nonconvex approaches based on low-rank factorization have been proposed (burer2005LRSDP):

min 𝑋 ∈ ℝ 𝑛 Γ— π‘Ÿ 1 4 ​ β€– π’œ ​ ( 𝑋 ​ 𝑋 ⊀ ) βˆ’ 𝑦 β€– 2 2 ,

(1)

where π‘Ÿ is a user-specified rank parameter. Despite nonconvexity, when the rank is correctly specified, i.e., when π‘Ÿ

π‘Ÿ ⋆ , the problem (1) admits computationally efficient solvers (chi2019nonconvex), e.g., gradient descent (GD) with spectral initialization or with small random initialization. However, three main challenges remain when applying the factorization-based nonconvex approach (1) in practice.

β€’

Unknown rank. First, the true rank π‘Ÿ ⋆ is often unknown, which makes it infeasible to set π‘Ÿ

π‘Ÿ ⋆ . One necessarily needs to consider an overparameterized setting in which π‘Ÿ is set conservatively, i.e., one sets π‘Ÿ β‰₯ π‘Ÿ ⋆ or even π‘Ÿ

𝑛 .

β€’

Poor conditioning. Second, the ground truth matrix 𝑀 ⋆ may be ill-conditioned, which is commonly encountered in practice. Existing approaches such as gradient descent are still computationally expensive in such settings as the number of iterations necessary for convergence increases with the condition number.

β€’

Robustness to noise and approximate low-rankness. Last but not least, it is desirable that the performance is robust when the measurement 𝑦 is contaminated by noise and when 𝑀 ⋆ is approximately low-rank.

In light of these two challenges, the main goal of this work is to address the following question:

Can one develop an efficient and robust method for solving ill-conditioned matrix recovery in the overparameterized setting?

parameterization reference algorithm init. iteration complexity

π‘Ÿ

π‘Ÿ ⋆ stoger2021small GD random πœ… 8 + πœ… 6 ​ log ⁑ ( πœ… ​ 𝑛 / πœ€ )

zhang2021preconditioned 𝖯𝗋𝖾𝖼𝖦𝖣 spectral log ⁑ ( 1 / πœ€ )

Theorem 2 ScaledGD( πœ† ) random log ⁑ πœ… β‹… log ⁑ ( πœ… ​ 𝑛 ) + log ⁑ ( 1 / πœ€ )

π‘Ÿ

π‘Ÿ ⋆ tong2021accelerating ScaledGD spectral log ⁑ ( 1 / πœ€ )

stoger2021small GD random πœ… 8 ​ log ⁑ ( πœ… ​ 𝑛 ) + πœ… 2 ​ log ⁑ ( 1 / πœ€ )

Theorem 3 ScaledGD( πœ† ) random log ⁑ πœ… β‹… log ⁑ ( πœ… ​ 𝑛 ) + log ⁑ ( 1 / πœ€ ) Table 1:Comparison of iteration complexity with existing algorithms for low-rank matrix sensing under Gaussian designs. Here, 𝑛 is the matrix dimension, π‘Ÿ ⋆ is the true rank, π‘Ÿ is the overparameterized rank, and πœ… is the condition number of the problem instance (see Section 2 for a formal problem formulation). It is important to note that in the overparameterized setting ( π‘Ÿ

π‘Ÿ ⋆ ), the sample complexity of zhang2021preconditioned scales polynomially with the overparameterized rank π‘Ÿ , while that of stoger2021small and ours only scale polynomially with the true rank π‘Ÿ ⋆ . 1.1Our contributions: a preview

The main contribution of the current paper is to answer the question affirmatively by developing a preconditioned gradient descent method (ScaledGD( πœ† )) that converges to the (possibly ill-conditioned) low-rank matrix in a fast and global manner, even with overparamterized rank π‘Ÿ β‰₯ π‘Ÿ ⋆ .

Theorem 1 (Informal).

Under overparameterization π‘Ÿ β‰₯ π‘Ÿ ⋆ and mild statistical assumptions, ScaledGD( πœ† )β€”starting from a sufficiently small random initialization with a sample complexity depending polynomially with the true rank π‘Ÿ ⋆ β€”achieves a relative πœ€ -accuracy, i.e., β€– 𝑋 𝑇 ​ 𝑋 𝑇 ⊀ βˆ’ 𝑀 ⋆ β€– π–₯ ≀ πœ€ ​ β€– 𝑀 ⋆ β€– , with no more than an order of

log ⁑ πœ… β‹… log ⁑ ( πœ… ​ 𝑛 ) + log ⁑ ( 1 / πœ€ )

iterations, where πœ… is the condition number of the problem. Moreover, in the presence of per-entry Gaussian measurement noise 𝒩 ​ ( 0 , 𝜎 2 ) , ScaledGD( πœ† ) converges to the nearly minimax-optimal error

β€– 𝑋 𝑇 ​ 𝑋 𝑇 ⊀ βˆ’ 𝑀 ⋆ β€– π–₯ ≲ πœ… 4 ​ 𝜎 ​ 𝑛 ​ π‘Ÿ ⋆

with the same rate as above.

The above theorem suggests that from a small random initialization, ScaledGD( πœ† ) converges at a constant linear rateβ€”independent of the condition numberβ€”after a small logarithmic number of iterations. Overall, the iteration complexity is nearly independent of the condition number and the problem dimension, making it extremely suitable for solving large-scale and ill-conditioned problems. To the best of our knowledge, ScaledGD( πœ† ) is the first provably minimax-optimal overparameterized gradient method for low-rank matrix sensing, where both the sample complexity and the error bound depend on the true rank π‘Ÿ ⋆ . In contrast, prior error bounds for nonconvex gradient methods zhang2024fast; zhuo2021computational scale with the overparameterized rank π‘Ÿ , which can be significantly larger. Our results also extend to the setting when the matrix 𝑀 ⋆ is only approximately low-rank under the Gaussian design, which is new. See Table 1 for a summary of comparisons with prior art in the noiseless setting.

Our algorithm ScaledGD( πœ† ) is closely related to scaled gradient descent (ScaledGD) (tong2021accelerating), a recently proposed preconditioned gradient descent method that achieves a πœ… -independent convergence rate under spectral initialization and exact parameterization. We modify the preconditioner design by introducing a fixed damping term, which prevents the preconditioner itself from being ill-conditioned due to overparameterization; the modified preconditioner preserves the low computational overhead when the overparameterization is moderate. In the exact parameterization setting, our result extends ScaledGD beyond local convergence by characterizing the number of iterations it takes to enter the local basin of attraction from a small random initialization.

Moreover, our results shed light on the power of preconditioning in accelerating the optimization process over vanilla GD while still guaranteeing generalization in overparameterized learning models (amari2020does). Remarkably, despite the existence of an infinite number of global minima in the landscape of (1) that do not generalize, i.e., not corresponding to the ground truth, starting from a small random initialization, GD (li2018algorithmic; stoger2021small) is known to converge to a generalizable solution without explicit regularization. However, GD takes 𝑂 ​ ( πœ… 8 + πœ… 6 ​ log ⁑ ( πœ… ​ 𝑛 / πœ€ ) ) iterations to reach πœ€ -accuracy, which is unacceptable even for moderate condition numbers. On the other hand, while common wisdom suggests that preconditioning accelerates convergence, it is yet unclear if it still converges to a generalizable global minimum. Our work answers this question in the affirmative for overparameterized low-rank matrix sensing, where ScaledGD( πœ† ) significantly accelerates the convergence against the poor condition numberβ€”both in the initial phase and in the local phaseβ€”without hurting generalization, which is corroborated in Figure 1.

Figure 1:Comparison between ScaledGD( πœ† ) and GD. The learning rate of GD has been fine-tuned to achieve fastest convergence for each πœ… , while that of ScaledGD( πœ† ) is fixed to 0.3 . The initialization scale 𝛼 in each case has been fine-tuned so that the final accuracy is 10 βˆ’ 9 . The details of the experiment are deferred to Section 5. 1.2Related work

Significant efforts have been devoted to understanding nonconvex optimization for low-rank matrix estimation in recent years, see chi2019nonconvex and chen2018harnessing for recent overviews. By reparameterizing the low-rank matrix into a product of factor matrices, also known as the Burer-Monteiro factorization (burer2005LRSDP), the focus point has been examining if the factor matrices can be recoveredβ€”up to invertible transformationsβ€”faithfully using simple iterative algorithms in a provably efficient manner. However, the majority of prior efforts suffer from the limitations that they assume an exact parameterization where the rank of the ground truth is given or estimated somewhat reliably, and rely on a carefully constructed initialization (e.g., using the spectral method (chen2021spectral)) in order to guarantee global convergence in a polynomial time. The analyses adopted in the exact parameterization case fail to generalize when overparameterization presents, and drastically new approaches are called for.

Overparameterization in low-rank matrix sensing.

li2018algorithmic made a theoretical breakthrough that showed that gradient descent converges globally to any prescribed accuracy even in the presence of full overparameterization ( π‘Ÿ

𝑛 ), with a small random initialization, where their analyses were subsequently adapted and extended in stoger2021small and zhuo2021computational. ding2021rank investigated robust low-rank matrix recovery with overparameterization from a spectral initialization, and ma2022global examined the same problem from a small random initialization with noisy measurements. zhang2021preconditioned; zhang2022preconditioned developed a preconditioned gradient descent method for overparameterized low-rank matrix sensing, where an adaptive damping parameter is introduced in ScaledGD. A variant with global convergence guarantee is studied in zhang2022preconditioned, which requires adding perturbation at the initial stage to first converge to a second-order stationary point before switching to a fast local convergence. Last but not least, a number of other notable works that study overparameterized low-rank models include, but are not limited to, soltanolkotabi2018theoretical; geyer2020low; oymak2019overparameterized; zhang2021sharp; zhang2022improved.

Global convergence from random initialization without overparameterization.

Despite nonconvexity, it has been established recently that several structured learning models admit global convergence via simple iterative methods even when initialized randomly even without overparameterization. For example, chen2019gradient showed that phase retrieval converges globally from a random initialization using a near-minimal number of samples through a delicate leave-one-out analysis. In addition, the efficiency of randomly initialized GD is established for complete dictionary learning (gilboa2019efficient; bai2018subgradient), multi-channel sparse blind deconvolution (qu2019nonconvex; shi2021manifold), asymmetric low-rank matrix factorization (ye2021global), and rank-one matrix completion (kim2022rank). Moving beyond GD, lee2022randomly showed that randomly initialized alternating least-squares converges globally for rank-one matrix sensing, whereas chandrasekher2022alternating developed sharp recovery guarantees of alternating minimization for generalized rank-one matrix sensing with sample-splitting and random initialization.

Algorithmic or implicit regularization.

Our work is related to the phenomenon of algorithmic or implicit regularization (gunasekar2017implicit), where the trajectory of simple iterative algorithms follows a path that maintains desirable properties without explicit regularization. Along this line, ma2017implicit; chen2019nonconvex; li2021nonconvex highlighted the implicit regularization of GD for several statistical estimation tasks, ma2021beyond showed that GD automatically balances the factor matrices in asymmetric low-rank matrix sensing, where jiang2022algorithmic analyzed the algorithmic regularization in overparameterized asymmetric matrix factorization in a model-free setting.

2Problem formulation

Section 2.1 introduces the problem of low-rank matrix sensing, and Section 2.2 provides background on the proposed ScaledGD( πœ† ) algorithm developed for the possibly overparameterized case.

2.1Model and assumptions

Suppose that the ground truth 𝑀 ⋆ ∈ ℝ 𝑛 Γ— 𝑛 is a positive-semidefinite (PSD) matrix of rank π‘Ÿ ⋆ β‰ͺ 𝑛 , whose (compact) eigendecomposition is given by

𝑀 ⋆

π‘ˆ ⋆ ​ Ξ£ ⋆ 2 ​ π‘ˆ ⋆ ⊀ .

Here, the columns of π‘ˆ ⋆ ∈ ℝ 𝑛 Γ— π‘Ÿ ⋆ specify the set of eigenvectors, and Ξ£ ⋆ ∈ ℝ π‘Ÿ ⋆ Γ— π‘Ÿ ⋆ is a diagonal matrix where the diagonal entries are ordered in a non-increasing fashion. Setting 𝑋 ⋆ ≔ π‘ˆ ⋆ ​ Ξ£ ⋆ ∈ ℝ 𝑛 Γ— π‘Ÿ ⋆ , we can rewrite 𝑀 ⋆ as

𝑀 ⋆

𝑋 ⋆ ​ 𝑋 ⋆ ⊀ .

(2)

We call 𝑋 ⋆ the ground truth low-rank factor matrix, whose condition number πœ… is defined as

πœ… ≔ 𝜎 max ​ ( 𝑋 ⋆ ) 𝜎 min ​ ( 𝑋 ⋆ ) .

(3)

Here we recall that 𝜎 max ​ ( 𝑋 ⋆ ) and 𝜎 min ​ ( 𝑋 ⋆ ) are the largest and the smallest singular values of 𝑋 ⋆ , respectively.

Instead of having access to 𝑀 ⋆ directly, we wish to recover 𝑀 ⋆ from a set of random linear measurements π’œ ​ ( 𝑀 ⋆ ) , where π’œ : Sym 2 ⁑ ( ℝ 𝑛 ) β†’ ℝ π‘š is a linear map from the space of 𝑛 Γ— 𝑛 symmetric matrices to ℝ π‘š , namely

𝑦

π’œ ​ ( 𝑀 ⋆ ) ,

(4)

or equivalently,

𝑦 𝑖

⟨ 𝐴 𝑖 , 𝑀 ⋆ ⟩ , 1 ≀ 𝑖 ≀ π‘š .

We are interested in recovering 𝑀 ⋆ based on the measurements 𝑦 and the sensing operator π’œ in a provably efficient manner, even when the true rank π‘Ÿ ⋆ is unknown.

2.2ScaledGD( πœ† ) for overparameterized low-rank matrix sensing

Inspired by the factorized representation (2), we aim to recover the low-rank matrix 𝑀 ⋆ by solving the following optimization problem (burer2005LRSDP):

min 𝑋 ∈ ℝ 𝑛 Γ— π‘Ÿ 𝑓 ​ ( 𝑋 ) ≔ 1 4 ​ β€– π’œ ​ ( 𝑋 ​ 𝑋 ⊀ ) βˆ’ 𝑦 β€– 2 2 ,

(5)

where π‘Ÿ is a predetermined rank parameter, possibly different from π‘Ÿ ⋆ . It is evident that for any rotation matrix 𝑂 ∈ π’ͺ π‘Ÿ , it holds that 𝑓 ​ ( 𝑋 )

𝑓 ​ ( 𝑋 ​ 𝑂 ) , leading to an infinite number of global minima of the loss function  𝑓 .

A prelude: exact parameterization.

When π‘Ÿ is set to be the true rank π‘Ÿ ⋆ of 𝑀 ⋆ , tong2021accelerating set forth a provable algorithmic approach called scaled gradient descent (ScaledGD)β€”gradient descent with a specific form of preconditioningβ€”that adopts the following update rule

𝖲𝖼𝖺𝗅𝖾𝖽𝖦𝖣 : 𝑋 𝑑 + 1

𝑋 𝑑 βˆ’ πœ‚ ​ βˆ‡ 𝑓 ​ ( 𝑋 𝑑 ) ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 ) βˆ’ 1

(6)

= 𝑋 𝑑 βˆ’ πœ‚ ​ π’œ βˆ— ​ π’œ ​ ( 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ ) ​ 𝑋 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 ) βˆ’ 1 .

Here, 𝑋 𝑑 is the 𝑑 -th iterate, βˆ‡ 𝑓 ​ ( 𝑋 𝑑 ) is the gradient of 𝑓 at 𝑋

𝑋 𝑑 , and πœ‚ > 0 is the learning rate. Moreover, π’œ βˆ— : ℝ π‘š ↦ Sym 2 ⁑ ( ℝ 𝑛 ) is the adjoint operator of π’œ , that is π’œ βˆ— ​ ( 𝑦 )

βˆ‘ 𝑖

1 π‘š 𝑦 𝑖 ​ 𝐴 𝑖 for 𝑦 ∈ ℝ π‘š .

At the expense of light computational overhead, ScaledGD is remarkably robust to ill-conditioning compared with vanilla gradient descent (GD). It is established in tong2021accelerating that ScaledGD, when starting from spectral initialization, converges linearly at a constant rateβ€”independent of the condition number πœ… of 𝑋 ⋆ (cf. (3)); in contrast, the iteration complexity of GD (tu2015low; zheng2015convergent) scales on the order of πœ… 2 from the same initialization, therefore GD becomes exceedingly slow when the problem instance is even moderately ill-conditioned, a scenario that is quite commonly encountered in practice.

ScaledGD( πœ† ): overparametrization under unknown rank.

In this paper, we are interested in the so-called overparameterization regime, where π‘Ÿ ⋆ ≀ π‘Ÿ ≀ 𝑛 . From an operational perspective, the true rank π‘Ÿ ⋆ is related to model order, e.g., the number of sources or targets in a scene of interest, which is often unavailable and makes it necessary to consider the misspecified setting. Unfortunately, in the presence of overparameterization, the original ScaledGD algorithm is no longer appropriate, as the preconditioner ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 ) βˆ’ 1 might become numerically unstable to calculate. Therefore, we propose a new variant of ScaledGD by adjusting the preconditioner as

𝖲𝖼𝖺𝗅𝖾𝖽𝖦𝖣 ( πœ† ) : 𝑋 𝑑 + 1

𝑋 𝑑 βˆ’ πœ‚ ​ βˆ‡ 𝑓 ​ ( 𝑋 𝑑 ) ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 ,

(7)

= 𝑋 𝑑 βˆ’ πœ‚ ​ π’œ βˆ— ​ π’œ ​ ( 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ ) ​ 𝑋 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 ,

where πœ† > 0 is a fixed damping parameter. The new algorithm is dubbed as ScaledGD( πœ† ), and it recovers the original ScaledGD when πœ†

0 . Similar to ScaledGD, a key property of ScaledGD( πœ† ) is that the iterates { 𝑋 𝑑 } are equivariant with respect to the parameterization of the factor matrix. Specifically, taking a rotationally equivalent factor 𝑋 𝑑 ​ 𝑂 with an arbitrary 𝑂 ∈ π’ͺ π‘Ÿ , and feeding it into the update rule (7), the next iterate

𝑋 𝑑 ​ 𝑂 βˆ’ πœ‚ ​ π’œ βˆ— ​ π’œ ​ ( 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ ) ​ 𝑋 𝑑 ​ 𝑂 ​ ( 𝑂 ⊀ ​ 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 ​ 𝑂 + πœ† ​ 𝐼 ) βˆ’ 1

𝑋 𝑑 + 1 ​ 𝑂

is rotated simultaneously by the same rotation matrix 𝑂 . In other words, the recovered matrix sequence 𝑀 𝑑

𝑋 𝑑 ​ 𝑋 𝑑 ⊀ is invariant with respect to the parameterization of the factor matrix.

Remark 1.

We note that a related variant of ScaledGD, called 𝖯𝗋𝖾𝖼𝖦𝖣 , has been proposed recently in zhang2021preconditioned; zhang2022preconditioned for the overparameterized setting, which follows the update rule

𝖯𝗋𝖾𝖼𝖦𝖣 : 𝑋 𝑑 + 1

𝑋 𝑑 βˆ’ πœ‚ π’œ βˆ— π’œ ( 𝑋 𝑑 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ ) 𝑋 𝑑 ( 𝑋 𝑑 ⊀ 𝑋 𝑑 + πœ† 𝑑 𝐼 ) βˆ’ 1 ,

(8)

where the damping parameters πœ† 𝑑

𝑓 ​ ( 𝑋 𝑑 ) are selected in an iteration-varying manner. In contrast, ScaledGD( πœ† ) assumes a fixed damping parameter πœ† throughout the iterations. We defer more detailed comparisons with 𝖯𝗋𝖾𝖼𝖦𝖣 in Section 3.

3Main results

Before formally presenting our theorems, let us introduce several key assumptions that will be in effect throughout this paper.

Restricted Isometry Property.

A key property of the operator π’œ ​ ( β‹… ) is the celebrated Restricted Isometry Property (RIP) (recht2010guaranteed), which says that the operator π’œ ​ ( β‹… ) approximately preserves the distances between low-rank matrices. The formal definition is given as follows.

Definition 1 (Restricted Isometry Property).

The linear map π’œ ​ ( β‹… ) is said to obey rank- π‘Ÿ RIP with a constant 𝛿 π‘Ÿ ∈ [ 0 , 1 ) , if for all matrices 𝑀 ∈ Sym 2 ⁑ ( ℝ 𝑛 ) of rank at most π‘Ÿ , it holds that

( 1 βˆ’ 𝛿 π‘Ÿ ) ​ β€– 𝑀 β€– π–₯ 2 ≀ β€– π’œ ​ ( 𝑀 ) β€– 2 2 ≀ ( 1 + 𝛿 π‘Ÿ ) ​ β€– 𝑀 β€– π–₯ 2 .

(9)

The Restricted Isometry Constant (RIC) is defined to be the smallest positive 𝛿 π‘Ÿ such that (9) holds.

The RIP is a standard assumption in low-rank matrix sensing, which has been verified to hold with high probability for a wide variety of measurement operators. The following lemma establishes the RIP for the Gaussian design.

Lemma 1.

(stoger2024non, Lemma 1) If the sensing operator π’œ ​ ( β‹… ) follows the Gaussian design, i.e., the entries of { 𝐴 𝑖 } 𝑖

1 π‘š are independent up to symmetry with diagonal elements sampled from 𝒩 ​ ( 0 , 1 / π‘š ) and off-diagonal elements from 𝒩 ​ ( 0 , 1 / 2 ​ π‘š ) , then with high probability, π’œ ​ ( β‹… ) satisfies rank- π‘Ÿ RIP with constant 𝛿 π‘Ÿ , as long as π‘š β‰₯ 𝐢 ​ 𝑛 ​ π‘Ÿ / 𝛿 π‘Ÿ 2 for some sufficiently large universal constant 𝐢

0 .

We make the following assumption about the operator π’œ ​ ( β‹… ) .

Assumption 1.

The operator π’œ ​ ( β‹… ) satisfies the rank- ( π‘Ÿ ⋆ + 1 ) RIP with 𝛿 π‘Ÿ ⋆ + 1 ≕ 𝛿 . Furthermore, there exist a sufficiently small constant 𝑐 𝛿

0 and a sufficiently large constant 𝐢 𝛿

0 such that

𝛿 ≀ 𝑐 𝛿 ​ π‘Ÿ ⋆ βˆ’ 1 / 2 ​ πœ… βˆ’ 𝐢 𝛿 .

(10) Small random initialization.

Similar to li2018algorithmic; stoger2021small, we set the initialization 𝑋 0 to be a small random matrix, i.e.,

𝑋 0

𝛼 ​ 𝐺 ,

(11)

where 𝐺 ∈ ℝ 𝑛 Γ— π‘Ÿ is some matrix considered to be normalized and 𝛼

0 controls the magnitude of the initialization. To simplify exposition, we take 𝐺 to be a standard random Gaussian matrix, that is, 𝐺 is a random matrix with i.i.d. entries distributed as 𝒩 ​ ( 0 , 1 / 𝑛 ) .

Choice of parameters.

Last but not least, the parameters of ScaledGD( πœ† ) are selected according to the following assumption.

Assumption 2.

There exist some universal constants 𝑐 πœ‚ , 𝑐 πœ† , 𝐢 𝛼

0 such that ( πœ‚ , πœ† , 𝛼 ) in ScaledGD( πœ† ) satisfy the following conditions:

( 𝗅𝖾𝖺𝗋𝗇𝗂𝗇𝗀 ​ 𝗋𝖺𝗍𝖾 )

πœ‚ ≀ 𝑐 πœ‚ ,

(12a)

( 𝖽𝖺𝗆𝗉𝗂𝗇𝗀 ​ 𝗉𝖺𝗋𝖺𝗆𝖾𝗍𝖾𝗋 )

1 100 ​ 𝑐 πœ† ​ πœ… βˆ’ 4 ​ 𝜎 min 2 ​ ( 𝑋 ⋆ ) ≀ πœ† ≀ 𝑐 πœ† ​ 𝜎 min 2 ​ ( 𝑋 ⋆ ) ,

(12b)

( π—‚π—‡π—‚π—π—‚π–Ίπ—…π—‚π—“π–Ίπ—π—‚π—ˆπ—‡ ​ π—Œπ—‚π—“π–Ύ )

log ⁑ β€– 𝑋 ⋆ β€– 𝛼 β‰₯ 𝐢 𝛼 max ⁑ ( πœ‚ , πœ… βˆ’ 2 ) ​ log ⁑ ( 2 ​ πœ… ) β‹… log ⁑ ( 2 ​ πœ… ​ 𝑛 ) .

(12c)

We are now in place to present the main theorems.

3.1The overparameterization setting

We begin with our main theorem, which characterizes the performance of ScaledGD( πœ† ) with overparameterization.

Theorem 2.

Suppose Assumptions 1 and 2 hold. With high probability (with respect to the realization of the random initialization 𝐺 ), there exists a universal constant 𝐢 min

0 such that for some 𝑇 ≀ 𝑇 min ≔ 𝐢 min πœ‚ ​ log ⁑ β€– 𝑋 ⋆ β€– 𝛼 , we have

β€– 𝑋 𝑇 ​ 𝑋 𝑇 ⊀ βˆ’ 𝑀 ⋆ β€– π–₯ ≀ 𝛼 1 / 3 ​ β€– 𝑋 ⋆ β€– 5 / 3 .

In particular, for any prescribed accuracy target πœ€ ∈ ( 0 , 1 ) , by choosing a sufficiently small 𝛼 fulfilling both (12c) and 𝛼 ≀ πœ€ 3 ​ β€– 𝑋 ⋆ β€– , we have β€– 𝑋 𝑇 ​ 𝑋 𝑇 ⊀ βˆ’ 𝑀 ⋆ β€– π–₯ ≀ πœ€ ​ β€– 𝑀 ⋆ β€– .

A few remarks are in order.

Iteration complexity.

Theorem 2 shows that by choosing an appropriate 𝛼 , ScaledGD( πœ† ) finds an πœ€ -accurate solution, i.e., β€– 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ β€– π–₯ ≀ πœ€ ​ β€– 𝑀 ⋆ β€– , in no more than an order of

log ⁑ πœ… β‹… log ⁑ ( πœ… ​ 𝑛 ) + log ⁑ ( 1 / πœ€ )

iterations. Roughly speaking, this asserts that ScaledGD( πœ† ) converges at a constant linear rate after an initial phase of approximately 𝑂 ​ ( log ⁑ πœ… β‹… log ⁑ ( πœ… ​ 𝑛 ) ) iterations. Most notably, the iteration complexity is nearly independent of the condition number πœ… , with a small overhead only through the poly-logarithmic additive term 𝑂 ​ ( log ⁑ πœ… β‹… log ⁑ ( πœ… ​ 𝑛 ) ) . In contrast, GD requires 𝑂 ​ ( πœ… 8 + πœ… 6 ​ log ⁑ ( πœ… ​ 𝑛 / πœ€ ) ) iterations to converge from a small random initialization to πœ€ -accuracy; see stoger2021small; li2018algorithmic. Thus, the convergence of GD is much slower than ScaledGD( πœ† ) even for mildly ill-conditioned matrices.

Sample complexity.

The sample complexity of ScaledGD( πœ† ) hinges upon the Assumption 1. When the sensing operator π’œ ​ ( β‹… ) follows the Gaussian design, this assumption is fulfilled as long as π‘š ≳ 𝑛 ​ π‘Ÿ ⋆ 2 β‹… π—‰π—ˆπ—…π—’ ​ ( πœ… ) . Notably, our sample complexity depends only on the true rank π‘Ÿ ⋆ , but not on the overparameterized rank π‘Ÿ β€” a crucial feature in order to provide meaningful guarantees when the overparameterized rank π‘Ÿ is close to the full dimension 𝑛 . The dependency on πœ… in the sample complexity, on the other end, is believed to be an artifact of the proof, as empirically shown in some related settings (see e.g., Figure 4 of chen2019noisy). Rigorously proving this, however, remains an open problem in nonconvex low-rank estimation (chi2019nonconvex).

Comparison with zhang2021preconditioned; zhang2022preconditioned.

As mentioned earlier, our proposed algorithm ScaledGD( πœ† ) is similar to 𝖯𝗋𝖾𝖼𝖦𝖣 proposed in zhang2021preconditioned that adopts an iteration-varying damping parameter in ScaledGD tong2021accelerating, with several important distinctions. In terms of theoretical guarantees, zhang2021preconditioned only provides the local convergence for 𝖯𝗋𝖾𝖼𝖦𝖣 assuming an initialization close to the ground truth; in contrast, we provide global convergence guarantees where a small random initialization is used. More critically, the sample complexity of 𝖯𝗋𝖾𝖼𝖦𝖣 zhang2021preconditioned depends on the overparameterized rank π‘Ÿ , while ours only depends on the true rank π‘Ÿ ⋆ . While zhang2022preconditioned also studied variants of 𝖯𝗋𝖾𝖼𝖦𝖣 with global convergence guarantees, they require additional operations such as gradient perturbations and switching between different algorithmic stages, which are harder to implement in practice. Furthermore, their convergence rate is much more pessimistic than ours. Our theory suggests that additional perturbation is unnecessary to ensure the global convergence of ScaledGD( πœ† ), as ScaledGD( πœ† ) automatically adapts to different curvatures of the optimization landscape throughout the entire trajectory.

3.2The exact parameterization setting

We now single out the exact parametrization case, i.e., when π‘Ÿ

π‘Ÿ ⋆ . In this case, our theory suggests that ScaledGD( πœ† ) converges to the ground truth even from a random initialization with a fixed scale 𝛼

0 .

Theorem 3.

Assume that π‘Ÿ

π‘Ÿ ⋆ . Suppose Assumptions 1 and 2 hold. With high probability (with respect to the realization of the random initialization 𝐺 ), there exist some universal constants 𝐢 min > 0 and 𝑐 > 0 such that for some 𝑇 ≀ 𝑇 min

𝐢 min πœ‚ ​ log ⁑ ( β€– 𝑋 ⋆ β€– / 𝛼 ) , we have for any 𝑑 β‰₯ 𝑇

β€– 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ β€– π–₯ ≀ ( 1 βˆ’ 𝑐 ​ πœ‚ ) 𝑑 βˆ’ 𝑇 ​ β€– 𝑀 ⋆ β€– .

Theorem 3 shows that with some fixed initialization scale 𝛼 , ScaledGD( πœ† ) takes at most an order of

log ⁑ πœ… β‹… log ⁑ ( πœ… ​ 𝑛 ) + log ⁑ ( 1 / πœ€ )

iterations to converge to πœ€ -accuracy for any πœ€

0 in the exact parameterization case. Compared with ScaledGD (tong2021accelerating) which takes 𝑂 ​ ( log ⁑ ( 1 / πœ€ ) ) iterations to converge from a spectral initialization, we only pay a logarithmic order 𝑂 ​ ( log ⁑ πœ… β‹… log ⁑ ( πœ… ​ 𝑛 ) ) of additional iterations to converge from a random initialization. In addition, once the algorithms enter the local regime, both ScaledGD( πœ† ) and ScaledGD behave similarly and converge at a fast constant linear rate, suggesting the effect of damping is locally negligible. Furthermore, compared with GD (stoger2021small) which requires 𝑂 ​ ( πœ… 8 ​ log ⁑ ( πœ… ​ 𝑛 ) + πœ… 2 ​ log ⁑ ( 1 / πœ€ ) ) iterations to achieve πœ€ -accuracy, our theory again highlights the benefit of ScaledGD( πœ† ) in boosting the global convergence even for mildly ill-conditioned matrices.

3.3The noisy setting

We next consider the case where the measurements are contaminated by noise πœ‰

( πœ‰ 𝑖 ) 𝑖

1 π‘š , that is

𝑦

π’œ ​ ( 𝑀 ⋆ ) + πœ‰ , or more concretely 𝑦 𝑖

⟨ 𝐴 𝑖 , 𝑀 ⋆ ⟩ + πœ‰ 𝑖 , 1 ≀ 𝑖 ≀ π‘š .

(13)

Instantiating (7) with the noisy measurements, the update rule of ScaledGD( πœ† ) can be written as

𝑋 𝑑 + 1

𝑋 𝑑 βˆ’ πœ‚ ​ ( π’œ βˆ— ​ π’œ ​ ( 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ ) βˆ’ π’œ βˆ— ​ ( 𝑦 ) ) ​ 𝑋 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 .

(14)

For simplicity, we make the following mild assumption on the noise.

Assumption 3.

We assume that πœ‰ 𝑖 ’s are independent with π’œ ​ ( β‹… ) , and are i.i.d. Gaussian, i.e.,

πœ‰ 𝑖 ∼ i.i.d. 𝒩 ​ ( 0 , 𝜎 2 ) , 1 ≀ 𝑖 ≀ π‘š .

Our theory demonstrates that ScaledGD( πœ† ) achieves the minimax-optimal error in this noisy setting as long as the noise is not too large.

Theorem 4.

Assume that 𝜎 ​ 𝑛 ≀ 𝑐 𝜎 ​ πœ… βˆ’ 𝐢 𝜎 ​ β€– 𝑀 ⋆ β€– for some sufficiently small universal constant 𝑐 𝜎

0 and some sufficiently large universal constant 𝐢 𝜎

0 . Then the following holds with high probability (with respect to the realization of the random initialization 𝐺 and the noise πœ‰ ). Suppose Assumptions 1, 2 and 3 hold. Given a prescribed accuracy target πœ€ ∈ ( 0 , 1 ) , suppose further that 𝛼 ≀ πœ€ 3 ​ β€– 𝑋 ⋆ β€– . There exist universal constants 𝐢 min

0 , 𝐢 4

0 , such that for some 𝑇 ≀ 𝑇 min ≔ 𝐢 min πœ‚ ​ log ⁑ β€– 𝑋 ⋆ β€– 𝛼 , we have

β€– 𝑋 𝑇 ​ 𝑋 𝑇 ⊀ βˆ’ 𝑀 ⋆ β€–

≀ max ⁑ ( πœ€ ​ β€– 𝑀 ⋆ β€– , 𝐢 4 ​ πœ… 4 ​ 𝜎 ​ 𝑛 ) ,

β€– 𝑋 𝑇 ​ 𝑋 𝑇 ⊀ βˆ’ 𝑀 ⋆ β€– π–₯

≀ max ⁑ ( πœ€ ​ β€– 𝑀 ⋆ β€– , 𝐢 4 ​ πœ… 4 ​ 𝜎 ​ 𝑛 ​ π‘Ÿ ⋆ ) .

A few remarks are in order.

Minimax optimality.

Theorem 4 suggests that as long as the noise level is not too large, by setting the optimization error πœ€ sufficiently small, i.e., πœ€ ​ β€– 𝑀 ⋆ β€– ≍ πœ… 4 ​ 𝜎 ​ 𝑛 , ScaledGD( πœ† ) finds a solution that satisfies

β€– 𝑋 𝑇 ​ 𝑋 𝑇 ⊀ βˆ’ 𝑀 ⋆ β€– ≲ πœ… 4 ​ 𝜎 ​ 𝑛 , β€– 𝑋 𝑇 ​ 𝑋 𝑇 ⊀ βˆ’ 𝑀 ⋆ β€– π–₯ ≲ πœ… 4 ​ 𝜎 ​ 𝑛 ​ π‘Ÿ ⋆

(15)

in no more than log ⁑ πœ… β‹… log ⁑ ( πœ… ​ 𝑛 ) + log ⁑ ( β€– 𝑀 ⋆ β€– πœ… 4 ​ 𝜎 ​ 𝑛 ) iterations, the number of which again only depends logarithmically on the problem parameters. When πœ… is upper bounded by a constant, our result is minimax optimal, in the sense that the final error matches the minimax lower bound in the classical work of candes2011tight, which we recall here for completeness: for any estimator 𝑀 ^ ​ ( 𝑦 ) based on the measurement 𝑦 defined in (13), for any π‘Ÿ ⋆ ≀ 𝑛 , there always exists some 𝑀 ⋆ ∈ ℝ 𝑛 Γ— 𝑛 of rank π‘Ÿ ⋆ such that

β€– 𝑀 ^ ​ ( 𝑦 ) βˆ’ 𝑀 ⋆ β€– ≳ 𝜎 ​ 𝑛 , β€– 𝑀 ^ ​ ( 𝑦 ) βˆ’ 𝑀 ⋆ β€– π–₯ ≳ 𝜎 ​ 𝑛 ​ π‘Ÿ ⋆

with probability at least 0.99 (with respect to the realization of the noise πœ‰ ). To the best of our knowledge, Theorem 4 is the first result to establish the minimax optimality (up to multiplicative factors of πœ… ) of overparameterized gradient methods in the context of low-rank matrix sensing. We remark that similar sub-optimality with respect to πœ… is also observed in chen2019noisy.

Consistency.

It is often desirable that the estimator is (asymptotically) consistent, i.e., the estimation error converges to 0 as the number of samples π‘š β†’ ∞ . To see that Theorem 4 indicates ScaledGD( πœ† ) indeed produces a consistent estimator, let us consider again the Gaussian design. In this case, ⟨ 𝐴 𝑖 , 𝑀 ⋆ ⟩ is on the order of β€– 𝑀 ⋆ β€– / π‘š , thus the signal-to-noise ratio can be measured by 𝖲𝖭𝖱 ≔ ( β€– 𝑀 ⋆ β€– / π‘š ) 2 / 𝜎 2

β€– 𝑀 ⋆ β€– 2 / ( π‘š ​ 𝜎 2 ) . With this notation, Theorem 4 asserts that the final error is 𝑂 ​ ( 𝖲𝖭𝖱 βˆ’ 1 / 2 ​ 𝑛 π‘š ​ β€– 𝑀 ⋆ β€– ) in operator norm and 𝑂 ​ ( 𝖲𝖭𝖱 βˆ’ 1 / 2 ​ 𝑛 ​ π‘Ÿ ⋆ π‘š ​ β€– 𝑀 ⋆ β€– ) in Frobenius norm, both of which converge to 0 at a rate of 𝑛 ​ π‘Ÿ ⋆ π‘š as π‘š β†’ ∞ when 𝖲𝖭𝖱 is fixed.

3.4The approximately low-rank setting

Last but not least, we examine a more general model of 𝑀 ⋆ , which does not need to be exactly low-rank, but only approximately low-rank. Instead of recovering 𝑀 ⋆ exactly, one seeks to find a low-rank approximation to 𝑀 ⋆ from its linear measurements.

To set up, let 𝑀 ⋆ ∈ ℝ 𝑛 Γ— 𝑛 be a general PSD ground truth matrix, where its spectral decomposition is given by 𝑀 ⋆

βˆ‘ 𝑖

1 𝑛 𝜎 𝑖 ​ 𝑒 𝑖 ​ 𝑒 𝑖 ⊀ , with

𝜎 1 β‰₯ 𝜎 2 β‰₯ β‹― β‰₯ 𝜎 𝑛 .

For any given π‘Ÿ ≀ 𝑛 , let 𝑀 π‘Ÿ be the best rank- π‘Ÿ approximation of 𝑀 ⋆ and 𝑀 π‘Ÿ β€² be the residual, i.e.,

𝑀 ⋆

βˆ‘ 𝑖

1 π‘Ÿ 𝜎 𝑖 ​ 𝑒 𝑖 ​ 𝑒 𝑖 ⊀ ⏟ ≕ 𝑀 π‘Ÿ + βˆ‘ 𝑖

π‘Ÿ + 1 𝑛 𝜎 𝑖 ​ 𝑒 𝑖 ​ 𝑒 𝑖 ⊀ ⏟ ≕ 𝑀 π‘Ÿ β€² .

(16)

If 𝑀 ^ π‘Ÿ is a rank- π‘Ÿ approximation to 𝑀 ⋆ , the approximation error can be measured by β€– 𝑀 ^ π‘Ÿ βˆ’ 𝑀 ⋆ β€– π–₯ . It is well-known that the best rank- π‘Ÿ approximation in this sense is exactly 𝑀 π‘Ÿ , and the optimal error is thus β€– 𝑀 π‘Ÿ β€² β€– π–₯ . By picking a larger π‘Ÿ , one has a smaller approximation error β€– 𝑀 π‘Ÿ β€² β€– π–₯ , but a higher memory footprint for the low-rank approximation 𝑀 π‘Ÿ whose condition number also grows with π‘Ÿ .

For simplicity, we consider the Gaussian design (cf. Lemma 1) in this subsection, which is less general than the RIP. The following theorem demonstrates that, as long as the sample size satisfies π‘š ≳ 𝑛 ​ π‘Ÿ ⋆ 2 β‹… π—‰π—ˆπ—…π—’ ​ ( πœ… ) , ScaledGD( πœ† ) automatically adapts to the available sample size and produces a near-optimal rank- π‘Ÿ ⋆ approximation to 𝑀 ⋆ in spite of overparameterization.

Theorem 5.

Assume that 𝑀 ⋆ is given in (16) and the sensing operator π’œ follows the Gaussian design with π‘š β‰₯ 𝐢 ​ 𝑛 ​ π‘Ÿ ⋆ 2 ​ πœ… 𝐢 , where πœ…

𝜎 1 / 𝜎 π‘Ÿ ⋆ is the condition number of 𝑀 π‘Ÿ ⋆ . In addition, assume β€– 𝑀 π‘Ÿ ⋆ β€² β€– ≀ 𝑐 𝜎 ​ πœ… βˆ’ 𝐢 𝜎 ​ β€– 𝑀 ⋆ β€– and β€– 𝑀 π‘Ÿ ⋆ β€² β€– π–₯ ≀ 𝑐 𝜎 ​ πœ… βˆ’ 𝐢 𝜎 ​ π‘š 𝑛 ​ β€– 𝑀 ⋆ β€– . Then the following holds with high probability (with respect to the realization of the random initialization 𝐺 and the sensing operator π’œ ). Suppose Assumption 2 holds for 𝑀 π‘Ÿ ⋆

𝑋 ⋆ ​ 𝑋 ⋆ ⊀ . Given a prescribed accuracy target πœ€ ∈ ( 0 , 1 ) , suppose further that 𝛼 ≀ πœ€ 3 ​ β€– 𝑋 ⋆ β€– . there exist universal constants 𝐢 min

0 , 𝐢 5

0 , such that for some 𝑇 ≀ 𝑇 min ≔ 𝐢 min πœ‚ ​ log ⁑ β€– 𝑋 ⋆ β€– 𝛼 , we have

β€– 𝑋 𝑇 ​ 𝑋 𝑇 ⊀ βˆ’ 𝑀 ⋆ β€– π–₯

≀ max ⁑ ( πœ€ ​ β€– 𝑀 ⋆ β€– , 𝐢 5 ​ πœ… 4 ​ β€– 𝑀 π‘Ÿ ⋆ β€² β€– π–₯ ) .

Here, 𝐢

0 , 𝐢 𝜎

0 are some sufficiently large universal constants, and 𝑐 𝜎

0 is some sufficiently small universal constant.

Remark 2.

Theorem 5 also holds in the matrix factorization setting, i.e., when π’œ is the identity operator.

Theorem 5 suggests that as long as 𝑀 ⋆ is well approximated by a low-rank matrix, by setting the optimization error πœ€ sufficiently small, i.e., πœ€ ​ β€– 𝑀 ⋆ β€– ≍ πœ… 4 ​ β€– 𝑀 π‘Ÿ ⋆ β€² β€– π–₯ , ScaledGD( πœ† ) finds a solution that satisfies

β€– 𝑋 𝑇 ​ 𝑋 𝑇 ⊀ βˆ’ 𝑀 ⋆ β€– π–₯ ≲ πœ… 4 ​ β€– 𝑀 π‘Ÿ ⋆ β€² β€– π–₯

(17)

in no more than log ⁑ πœ… β‹… log ⁑ ( πœ… ​ 𝑛 ) + log ⁑ ( β€– 𝑀 ⋆ β€– πœ… 4 ​ β€– 𝑀 π‘Ÿ ⋆ β€² β€– π–₯ ) iterations, which again only depend on the problem parameters logarithmically. This suggests that if the residual 𝑀 π‘Ÿ ⋆ β€² is small, ScaledGD( πœ† ) produces an approximate solution to the best rank- π‘Ÿ ⋆ approximation problem with near-optimal error, up to a multiplicative factor depending only on πœ… , without knowing the rank π‘Ÿ ⋆ a priori. To our best knowledge, this is the first near-optimal theoretical guarantee for approximate low-rank matrix sensing using gradient-based methods.

4Analysis

In this section, we present the main steps for proving Theorem 2 and Theorem 3. The proofs of Theorem 4 and Theorem 5 will follow the same ideas with minor modification. The detailed proofs are collected in the appendix. All of our statements will be conditioned on the following high probability event regarding the initialization matrix 𝐺 :

β„°

{ β€– 𝐺 β€– ≀ 𝐢 𝐺 } ∩ { 𝜎 min ​ ( π‘ˆ ^ ⊀ ​ 𝐺 ) β‰₯ ( 2 ​ 𝑛 ) βˆ’ 𝐢 𝐺 } ,

(18)

where π‘ˆ ^ ∈ ℝ 𝑛 Γ— π‘Ÿ ⋆ is an orthonormal basis of the eigenspace associated with the π‘Ÿ ⋆ largest eigenvalues of π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) , and 𝐢 𝐺

0 is some sufficiently large universal constant. It is a standard result in random matrix theory that β„° happens with high probability, as verified by the following lemma.

Lemma 2.

With respect to the randomness in 𝐺 , the event β„° happens with probability at least 1 βˆ’ ( 𝑐 ​ 𝑛 ) βˆ’ 𝐢 𝐺 ​ ( π‘Ÿ βˆ’ π‘Ÿ ⋆ + 1 ) / 2 βˆ’ 2 ​ exp ⁑ ( βˆ’ 𝑐 ​ 𝑛 ) , where 𝑐

0 is some universal constant.

Proof.

See Appendix A.1. ∎

4.1Preliminaries: decomposition of the iterates

Before embarking on the main proof, we present a useful decomposition (cf. (19)) of the iterate 𝑋 𝑑 into a signal term, a misalignment error term, and an overparametrization error term. Choose some matrix π‘ˆ ⋆ , βŸ‚ ∈ ℝ 𝑛 Γ— ( 𝑛 βˆ’ π‘Ÿ ⋆ ) such that [ π‘ˆ ⋆ , π‘ˆ ⋆ , βŸ‚ ] is orthonormal. Then we can define

𝑆 𝑑 ≔ π‘ˆ ⋆ ⊀ ​ 𝑋 𝑑 ∈ ℝ π‘Ÿ ⋆ Γ— π‘Ÿ , and 𝑁 𝑑 ≔ π‘ˆ ⋆ , βŸ‚ ⊀ ​ 𝑋 𝑑 ∈ ℝ ( 𝑛 βˆ’ π‘Ÿ ⋆ ) Γ— π‘Ÿ .

Let the SVD of 𝑆 𝑑 be

𝑆 𝑑

π‘ˆ 𝑑 ​ Ξ£ 𝑑 ​ 𝑉 𝑑 ⊀ ,

where π‘ˆ 𝑑 ∈ ℝ π‘Ÿ ⋆ Γ— π‘Ÿ ⋆ , Ξ£ 𝑑 ∈ ℝ π‘Ÿ ⋆ Γ— π‘Ÿ ⋆ , and 𝑉 𝑑 ∈ ℝ π‘Ÿ Γ— π‘Ÿ ⋆ . Similar to π‘ˆ ⋆ , βŸ‚ , we define the orthogonal complement of 𝑉 𝑑 as 𝑉 𝑑 , βŸ‚ ∈ ℝ π‘Ÿ Γ— ( π‘Ÿ βˆ’ π‘Ÿ ⋆ ) . When π‘Ÿ

π‘Ÿ ⋆ we simply set 𝑉 𝑑 , βŸ‚

0 .

We are now ready to present the main decomposition of 𝑋 𝑑 , which we use repeatedly in later analysis. This decomposition is inspired by stoger2021small. A similar decomposition also appeared in ma2022global.

Proposition 1.

The following decomposition holds:

𝑋 𝑑

π‘ˆ ⋆ ​ 𝑆 ~ 𝑑 ​ 𝑉 𝑑 ⊀ ⏟ π—Œπ—‚π—€π—‡π–Ίπ—… + π‘ˆ ⋆ , βŸ‚ ​ 𝑁 ~ 𝑑 ​ 𝑉 𝑑 ⊀ ⏟ π—†π—‚π—Œπ–Ίπ—…π—‚π—€π—‡π—†π–Ύπ—‡π— + π‘ˆ ⋆ , βŸ‚ ​ 𝑂 ~ 𝑑 ​ 𝑉 𝑑 , βŸ‚ ⊀ ⏟ π—ˆπ—π–Ύπ—‹π—‰π–Ίπ—‹π–Ίπ—†π–Ύπ—π—‹π—‚π—“π–Ίπ—π—‚π—ˆπ—‡ ,

(19)

where

𝑆 ~ 𝑑 ≔ 𝑆 𝑑 ​ 𝑉 𝑑 ∈ ℝ π‘Ÿ ⋆ Γ— π‘Ÿ ⋆ , 𝑁 ~ 𝑑 ≔ 𝑁 𝑑 ​ 𝑉 𝑑 ∈ ℝ ( 𝑛 βˆ’ π‘Ÿ ⋆ ) Γ— π‘Ÿ ⋆ , and 𝑂 ~ 𝑑 ≔ 𝑁 𝑑 ​ 𝑉 𝑑 , βŸ‚ ∈ ℝ ( 𝑛 βˆ’ π‘Ÿ ⋆ ) Γ— ( π‘Ÿ βˆ’ π‘Ÿ ⋆ ) .

(20) Proof.

See Appendix A.2. ∎

Several remarks on the decomposition are in order.

β€’

First, since 𝑉 𝑑 , βŸ‚ spans the obsolete subspace arising from overparameterization, 𝑂 ~ 𝑑 naturally represents the error incurred by overparameterization; in particular, in the well-specified case (i.e., π‘Ÿ

π‘Ÿ ⋆ ), one has zero overparameterization error, i.e., 𝑂 ~ 𝑑

0 .

β€’

Second, apart from the rotation matrix 𝑉 𝑑 , 𝑆 ~ 𝑑 documents the projection of the iterates 𝑋 𝑑 onto the signal space π‘ˆ ⋆ . Similarly, 𝑁 ~ 𝑑 characterizes the misalignment of the iterates with the signal subspace π‘ˆ ⋆ . It is easy to observe that in order for 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ β‰ˆ 𝑀 ⋆ , one must have 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ β‰ˆ Ξ£ ⋆ 2 , and 𝑁 ~ 𝑑 β‰ˆ 0 .

β€’

Last but not least, the extra rotation induced by 𝑉 𝑑 is extremely useful in making the signal/misalignment terms rationally invariant. To see this, suppose that we rotate the current iterate by 𝑋 𝑑 ↦ 𝑋 𝑑 ​ 𝑄 with some rotational matrix 𝑄 ∈ π’ͺ π‘Ÿ , then 𝑆 𝑑 ↦ 𝑆 𝑑 ​ 𝑄 but 𝑆 ~ 𝑑 remains unchanged, and similarly for 𝑁 ~ 𝑑 .

4.2Proof roadmap

Our analysis breaks into a few phases that characterize the dynamics of the key terms in the above decomposition, which we provide a roadmap to facilitate understanding. Denote

𝐢 max ≔ { 4 ​ 𝐢 min ,

π‘Ÿ

π‘Ÿ ⋆ ,

∞ ,
π‘Ÿ

π‘Ÿ ⋆ , and 𝑇 max ≔ 𝐢 max πœ‚ ​ log ⁑ ( β€– 𝑋 ⋆ β€– / 𝛼 ) ,

where 𝑇 max represents the largest index of the iterates that we maintain error control. The analysis boils down to the following phases, indicated by time points 𝑑 1 , 𝑑 2 , 𝑑 3 , 𝑑 4 that satisfy

𝑑 1 ≀ 𝑇 min / 16 , 𝑑 1 ≀ 𝑑 2 ≀ 𝑑 1 + 𝑇 min / 16 , 𝑑 2 ≀ 𝑑 3 ≀ 𝑑 2 + 𝑇 min / 16 , 𝑑 3 ≀ 𝑑 4 ≀ 𝑑 3 + 𝑇 min / 16 .

β€’

Phase I: approximate power iterations. In the initial phase, ScaledGD( πœ† ) behaves similarly to GD, which is shown in stoger2021small to approximate the power method in the first few iterations up to 𝑑 1 . After this phase, namely for 𝑑 ∈ [ 𝑑 1 , 𝑇 max ] , although the signal strength is still quite small, it begins to be aligned with the ground truth with the overparameterization error kept relatively small.

β€’

Phase II: exponential amplification of the signal. In this phase, ScaledGD( πœ† ) behaves somewhat as a mixture of GD and ScaledGD with a proper choice of the damping parameter πœ† ≍ 𝜎 min 2 ​ ( 𝑋 ⋆ ) , which ensures the signal strength first grows exponentially fast to reach a constant level no later than 𝑑 2 , and then reaches the desired level no later than 𝑑 3 , i.e., 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ β‰ˆ Ξ£ ⋆ 2 .

β€’

Phase III: local linear convergence. At the last phase, ScaledGD( πœ† ) behaves similarly to ScaledGD, which converges linearly at a rate independent of the condition number. Specifically, for 𝑑 ∈ [ 𝑑 3 , 𝑇 max ] , the reconstruction error β€– 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ β€– π–₯ converges at a linear rate up to some small overparameterization error, until reaching the desired accuracy for any 𝑑 ∈ [ 𝑑 4 , 𝑇 max ] .

4.3Phase I: approximate power iterations

It has been observed in stoger2021small that when initialized at a small scaled random matrix, the first few iterations of GD mimic the power iterations on the matrix π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) . When it comes to ScaledGD( πœ† ), since the initialization size 𝛼 is chosen to be much smaller than the damping parameter πœ† , the preconditioner ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 behaves like ( πœ† ​ 𝐼 ) βˆ’ 1 in the beginning. This renders ScaledGD( πœ† ) akin to gradient descent in the initial phase. As a result, we also expect the first few iterations of ScaledGD( πœ† ) to be similar to the power iterations, i.e.,

𝑋 𝑑 β‰ˆ ( 𝐼 + πœ‚ πœ† ​ π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) ) 𝑑 ​ 𝑋 0 , when  ​ 𝑑 ​  is small .

Such proximity between ScaledGD( πœ† ) and power iterations can indeed be justified in the beginning period, which allows us to deduce the following nice properties after the initial iterates of ScaledGD( πœ† ).

Lemma 3.

Under the same setting as Theorem 2, there exists an iteration number 𝑑 1 : 𝑑 1 ≀ 𝑇 min / 16 such that

𝜎 min ​ ( 𝑆 ~ 𝑑 1 ) β‰₯ 𝛼 2 / β€– 𝑋 ⋆ β€– ,

(21)

and that, for any 𝑑 ∈ [ 𝑑 1 , 𝑇 max ] , 𝑆 ~ 𝑑 is invertible and one has

β€– 𝑂 ~ 𝑑 β€–

≀ ( 𝐢 3 . 𝑏 ​ πœ… ​ 𝑛 ) βˆ’ 𝐢 3 . 𝑏 ​ β€– 𝑋 ⋆ β€– ​ 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 ) ,

(22a)

β€– 𝑂 ~ 𝑑 β€–

≀ ( 1 + πœ‚ 12 ​ 𝐢 max ​ πœ… ) 𝑑 βˆ’ 𝑑 1 ​ 𝛼 5 / 6 ​ β€– 𝑋 ⋆ β€– 1 / 6 ,

(22b)

β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€–

≀ 𝑐 3 ​ πœ… βˆ’ 𝐢 𝛿 / 2 ​ β€– 𝑋 ⋆ β€– ,

(22c)

β€– 𝑆 ~ 𝑑 β€–

≀ 𝐢 3 . π‘Ž ​ πœ… 3 ​ β€– 𝑋 ⋆ β€– ,

(22d)

where 𝐢 3 . π‘Ž , 𝐢 3 . 𝑏 , 𝑐 3 are some positive constants satisfying 𝐢 3 . π‘Ž ≲ 𝑐 πœ† βˆ’ 1 / 2 , 𝑐 3 ≲ 𝑐 𝛿 / 𝑐 πœ† , and 𝐢 3 . 𝑏 can be made arbitrarily large by increasing 𝐢 𝛼 .

Proof.

See Appendix C. ∎

Remark 3.

Let us record two immediate consequences of (22), which sometimes are more convenient for later analysis. From (22a), we may deduce

β€– 𝑂 ~ 𝑑 β€–

≀ ( 𝐢 3 . 𝑏 ​ πœ… ​ 𝑛 ) βˆ’ 𝐢 3 . 𝑏 ​ β€– 𝑋 ⋆ β€– ​ 𝜎 min ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝜎 min ​ ( 𝑆 ~ 𝑑 )

≀ πœ… ​ ( 𝐢 3 . 𝑏 ​ πœ… ​ 𝑛 ) βˆ’ 𝐢 3 . 𝑏 ​ 𝜎 min ​ ( 𝑆 ~ 𝑑 )

≀ ( 𝐢 3 . 𝑏 β€² ​ πœ… ​ 𝑛 ) βˆ’ 𝐢 3 . 𝑏 β€² ​ 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ,

(23)

where 𝐢 3 . 𝑏 β€²

𝐢 3 . 𝑏 / 2 , provided 𝐢 3 . 𝑏

4 . It is clear that 𝐢 3 . 𝑏 β€² can also be made arbitrarily large by enlarging 𝐢 𝛼 . Similarly, from (22b), we may deduce

β€– 𝑂 ~ 𝑑 β€– ≀ ( 1 + πœ‚ 12 ​ 𝐢 max ​ πœ… ) 𝑑 βˆ’ 𝑑 1 ​ 𝛼 5 / 6 ​ β€– 𝑋 ⋆ β€– 1 / 6
≀ ( 1 + πœ‚ 12 ​ 𝐢 max ​ πœ… ) 𝐢 max πœ‚ ​ log ⁑ ( β€– 𝑋 ⋆ β€– / 𝛼 ) ​ 𝛼 5 / 6 ​ β€– 𝑋 ⋆ β€– 1 / 6

≀ ( β€– 𝑋 ⋆ β€– / 𝛼 ) 1 / 12 ​ 𝛼 5 / 6 ​ β€– 𝑋 ⋆ β€– 1 / 6

𝛼 3 / 4 ​ β€– 𝑋 ⋆ β€– 1 / 4 .

(24)

Lemma 3 ensures the iterates of ScaledGD( πœ† ) maintain several desired properties after iteration 𝑑 1 , as summarized in (22). In particular, for any 𝑑 ∈ [ 𝑑 1 , 𝑇 max ] : (i) the overparameterization error β€– 𝑂 ~ 𝑑 β€– remains small relatively to the signal strength measured in terms of the scaled minimum singular value 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 ) , and remains bounded with respect to the size of the initialization 𝛼 (cf. (22a) and (22b) and their consequences (3) and (24)); (ii) the scaled misalignment-to-signal ratio remains bounded, suggesting the iterates remain aligned with the ground truth signal subspace π‘ˆ ⋆ (cf. (22c)); (iii) the size of the signal component 𝑆 ~ 𝑑 remains bounded (cf. (22d)). These properties play an important role in the follow-up analysis.

Remark 4.

It is worth noting that, the scaled minimum singular value 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 ) plays a key role in our analysis, which is in sharp contrast to the use of the vanilla minimum singular value 𝜎 min ​ ( 𝑆 ~ 𝑑 ) in the analysis of gradient descent (stoger2021small). This new measure of signal strength is inspired by the scaled distance for ScaledGD introduced in tong2021accelerating; tong2022scaling, which carefully takes the preconditioner design into consideration. Similarly, the metrics β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– in (22c) and β€– Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 β€– (to be seen momentarily) are also scaled for similar considerations to unveil the fast convergence (almost) independent of the condition number.

4.4Phase II: exponential amplification of the signal

By the end of Phase I, the signal strength is still quite small (cf. (21)), which is far from the desired level. Fortunately, the properties established in Lemma 3 allow us to establish an exponential amplification of the signal term 𝑆 ~ 𝑑 thereafter, which can be further divided into two stages.

1.

In the first stage, the signal is boosted to a constant level, i.e., 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βͺ° 1 10 ​ Ξ£ ⋆ 2 ;

2.

In the second stage, the signal grows further to the desired level, i.e., 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ β‰ˆ Ξ£ ⋆ 2 .

We start with the first stage, which again uses 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 ) as a measure of signal strength in the following lemma.

Lemma 4.

For any 𝑑 such that (22) holds, we have

𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 + 1 ) β‰₯ ( 1 βˆ’ 2 ​ πœ‚ ) ​ 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 ) .

Moreover, if 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 ) ≀ 1 / 3 , then

𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 + 1 ) β‰₯ ( 1 + 1 8 ​ πœ‚ ) ​ 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 ) .

Proof.

See Appendix D.1. ∎

The second half of Lemma 4 uncovers the exponential growth of the signal strength 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 ) until a constant level after several iterations, which resembles the exponential growth of the signal strength in GD (stoger2021small). This is formally established in the following corollary.

Corollary 1.

There exists an iteration number 𝑑 2 : 𝑑 1 ≀ 𝑑 2 ≀ 𝑑 1 + 𝑇 min / 16 such that for all 𝑑 ∈ [ 𝑑 2 , 𝑇 max ] , we have

𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βͺ° 1 10 ​ Ξ£ ⋆ 2 .

(25) Proof.

See Appendix D.2. ∎

We next aim to show that 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ β‰ˆ Ξ£ ⋆ 2 after the signal strength is above the constant level. To this end, the behavior of ScaledGD( πœ† ) becomes closer to that of ScaledGD, and it turns out to be easier to work with β€– Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 β€– as a measure of the scaled recovery error of the signal component. We establish the approximate exponential shrinkage of this measure in the following lemma.

Lemma 5.

For all 𝑑 ∈ [ 𝑑 2 , 𝑇 max ] with 𝑑 2 given in Corollary 1, one has

β€– Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 β€– ≀ ( 1 βˆ’ πœ‚ ) ​ β€– Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 β€– + 1 100 ​ πœ‚ .

(26) Proof.

See Appendix D.3. ∎

With the help of Lemma 5, it is straightforward to establish the desired approximate recovery guarantee of the signal component, i.e., 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ β‰ˆ Ξ£ ⋆ 2 .

Corollary 2.

There exists an iteration number 𝑑 3 : 𝑑 2 ≀ 𝑑 3 ≀ 𝑑 2 + 𝑇 min / 16 such that for any 𝑑 ∈ [ 𝑑 3 , 𝑇 max ] , one has

9 10 ​ Ξ£ ⋆ 2 βͺ― 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βͺ― 11 10 ​ Ξ£ ⋆ 2 .

(27) Proof.

See Appendix D.4. ∎

4.5Phase III: local convergence

Corollary 2 tells us that after iteration 𝑑 3 , we enter a local region in which 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ is close to the ground truth Ξ£ ⋆ 2 . In this local region, the behavior of ScaledGD( πœ† ) becomes closer to that of ScaledGD analyzed in tong2021accelerating. We turn attention to the reconstruction error β€– 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ β€– π–₯ that measures the generalization performance, and show it converges at a linear rate independent of the condition number up to some small overparameterization error.

Lemma 6.

There exists some universal constant 𝑐 6

0 such that for any 𝑑 : 𝑑 3 ≀ 𝑑 ≀ 𝑇 max , we have

βˆ₯ 𝑋 𝑑 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ βˆ₯ π–₯ ≀ ( 1 βˆ’ 𝑐 6 πœ‚ ) 𝑑 βˆ’ 𝑑 3 π‘Ÿ ⋆ βˆ₯ 𝑀 ⋆ βˆ₯ + 8 𝑐 6 βˆ’ 1 βˆ₯ 𝑀 ⋆ βˆ₯ max 𝑑 3 ≀ 𝜏 ≀ 𝑑 ( β€– 𝑂 ~ 𝜏 β€– β€– 𝑋 ⋆ β€– ) 1 / 2 .

(28)

In particular, there exists an iteration number 𝑑 4 : 𝑑 3 ≀ 𝑑 4 ≀ 𝑑 3 + 𝑇 min / 16 such that for any 𝑑 ∈ [ 𝑑 4 , 𝑇 max ] , we have

β€– 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ β€– π–₯ ≀ 𝛼 1 / 3 ​ β€– 𝑋 ⋆ β€– 5 / 3 ≀ πœ€ ​ β€– 𝑀 ⋆ β€– .

(29)

Here, πœ€ and 𝛼 are as stated in Theorem 2.

Proof.

See Appendix E. ∎

4.6Proofs of main theorems

Now we are ready to collect the results in the preceding sections to prove our main results, i.e., Theorem 2 and Theorem 3. The proofs of Theorem 4 and Theorem 5 follows from similar ideas but with additional technicality, thus are postponed to Appendix F.

We start with proving Theorem 2. By Lemma 3, Corollary 1, Corollary 2 and Lemma 6, the final 𝑑 4 given by Lemma 6 is no more than 4 Γ— 𝑇 min / 16 ≀ 𝑇 min / 2 , thus (29) holds for all 𝑑 ∈ [ 𝑇 min / 2 , 𝑇 max ] , in particular, for some 𝑇 ≀ 𝑇 min , as claimed.

Now we consider Theorem 3. In case that π‘Ÿ

π‘Ÿ ⋆ , it follows from definition that 𝑂 ~ 𝑑

0 vanishes for all 𝑑 . It follows from Lemma 6, in particular from (28), that

β€– 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ β€– π–₯ ≀ ( 1 βˆ’ 𝑐 6 ​ πœ‚ ) 𝑑 βˆ’ 𝑑 3 ​ π‘Ÿ ⋆ ​ β€– 𝑀 ⋆ β€– ,

for any 𝑑 β‰₯ 𝑑 3 (recall that 𝑇 max

∞ by definition when π‘Ÿ

π‘Ÿ ⋆ ). Note that ( 1 βˆ’ 𝑐 6 ​ πœ‚ ) 𝑑 ​ π‘Ÿ ⋆ ≀ ( 1 βˆ’ 𝑐 6 ​ πœ‚ ) 𝑑 βˆ’ 𝑇 + 𝑑 3 if 𝑇 βˆ’ 𝑑 3 β‰₯ 4 ​ log ⁑ ( π‘Ÿ ⋆ ) / ( 𝑐 6 ​ πœ‚ ) given that πœ‚ ≀ 𝑐 πœ‚ is sufficiently small. Thus for any 𝑑 β‰₯ 𝑇 we have

β€– 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ β€– π–₯ ≀ ( 1 βˆ’ 𝑐 6 ​ πœ‚ ) 𝑑 βˆ’ 𝑇 ​ β€– 𝑀 ⋆ β€– .

It is clear that one may choose such 𝑇 which also satisfies 𝑇 ≀ 𝑑 3 + 8 / ( 𝑐 6 ​ πœ‚ ) ≀ 𝑑 3 + 𝑇 min / 16 . We have already shown in the proof of Theorem 2 that 𝑑 3 ≀ 4 Γ— 𝑇 min / 16 ≀ 𝑇 min / 4 , thus 𝑇 ≀ 𝑇 min as desired.

Early stopping.

In the overparameterized setting, our theory guarantees the reconstruction error to be small until some iteration 𝑇 max . This is consistent with the phenomenon known as early stopping in prior works of learning with overparameterized models (stoger2021small; li2018algorithmic). Given the form of (22b), one may wonder if the early stopping needs to be precisely controlled, if β€– 𝑂 ~ 𝑑 β€– could grow excessively. Fortunately, this is not the case, as the following proposition – proved in Appendix E – demonstrates.

Proposition 2.

Under the same setting as Theorem 2, we have

β€– 𝑂 ~ 𝑑 β€– ≀ 𝛼 7 / 10 ​ β€– 𝑋 ⋆ β€– 3 / 10 , βˆ€ 𝑑 ≀ ( β€– 𝑋 ⋆ β€– 𝛼 ) 3 / 10 .

As we pick a very small 𝛼 , this means one does not need to do early stopping for all practical purposes.

5Numerical experiments

In this section, we conduct numerical experiments to demonstrate the efficacy of ScaledGD( πœ† ) for solving overparameterized low-rank matrix sensing. We set the ground truth matrix 𝑋 ⋆

π‘ˆ ⋆ ​ Ξ£ ⋆ ∈ ℝ 𝑛 Γ— π‘Ÿ ⋆ where π‘ˆ ⋆ ∈ ℝ 𝑛 Γ— π‘Ÿ ⋆ is a random orthogonal matrix and Ξ£ ⋆ ∈ ℝ π‘Ÿ ⋆ Γ— π‘Ÿ ⋆ is a diagonal matrix whose condition number is set to be πœ… . We set 𝑛

150 and π‘Ÿ ⋆

3 , and use random Gaussian measurements with π‘š

10 ​ 𝑛 ​ π‘Ÿ ⋆ . The overparameterization rank π‘Ÿ is set to be 5 unless otherwise specified.

Throughout our experiments, to choose πœ† , we estimate 𝜎 min ​ ( 𝑋 ⋆ ) using a simple rule of thumb. Let 𝜎 ^ 1 β‰₯ 𝜎 ^ 2 β‰₯ β‹― β‰₯ 𝜎 ^ 𝑛 be the singular values of π’œ βˆ— ​ ( 𝑦 ) . Let 𝑖 0 be the smallest number such that

βˆ‘ 𝑖 ≀ 𝑖 0 𝜎 ^ 𝑖 β‰₯ 0.95 ​ βˆ‘ 𝑖 ≀ 𝑛 𝜎 ^ 𝑖 .

Then we estimate 𝜎 ^ min 2 ​ ( 𝑋 ⋆ )

𝜎 ^ 𝑖 0 . This heuristic also applies to noisy or approximately low-rank matrices, thanks to our Theorem 4 and Theorem 5. In practice, the 0.95 threshold can be tuned towards a desired accuracy level.

Comparison with overparameterized GD.

We run ScaledGD( πœ† ) and GD with random initialization and compare their convergence speeds under different condition numbers πœ… of the ground truth 𝑋 ⋆ ; the result is depicted in Figure 1. Even for a moderate range of πœ… , GD slows down significantly while the convergence speed of ScaledGD( πœ† ) remains almost the same with a almost negligible initial phase, which is consistent with our theory. The advantage of ScaledGD( πœ† ) enlarges as πœ… increase, and is already more than 10x times faster than GD when πœ…

7 .

Effect of initialization size.

We study the effect of the initialization scale 𝛼 on the reconstruction accuracy of ScaledGD( πœ† ).

We fix the learning rate πœ‚ to be a constant and vary the initialization scale. We run ScaledGD( πœ† ) until it converges.1 The resulting reconstruction errors and their corresponding initialization scales are plotted in Figure 2. It can be inferred that the reconstruction error increases with respect to 𝛼 , which is consistent with our theory.

Figure 2:Relative reconstruction error versus initialization scale 𝛼 . The slope of the dashed line is approximately 1 . Comparison with zhang2021preconditioned.

We compare ScaledGD( πœ† ) with the algorithm 𝖯𝗋𝖾𝖼𝖦𝖣 proposed in zhang2021preconditioned, which also has a πœ… -independent convergence rate assuming a sufficiently good initialization using spectral initialization. However, 𝖯𝗋𝖾𝖼𝖦𝖣 requires RIP of rank π‘Ÿ , thus demanding 𝑂 ​ ( 𝑛 ​ π‘Ÿ 2 ) many samples instead of 𝑂 ​ ( 𝑛 ​ π‘Ÿ ⋆ 2 ) as in GD and ScaledGD( πœ† ). This can be troublesome for larger π‘Ÿ . To demonstrate this point, we run ScaledGD( πœ† ) and 𝖯𝗋𝖾𝖼𝖦𝖣 with different overparameterization rank π‘Ÿ while fixing all other parameters. The results are shown in Figure 3. It can be seen that the convergence rate of 𝖯𝗋𝖾𝖼𝖦𝖣 and ScaledGD( πœ† ) are almost the same when the rank is exactly specified ( π‘Ÿ

π‘Ÿ ⋆

3 ), though ScaledGD( πœ† ) requires a few more iterations for the initial phases2. When π‘Ÿ goes higher, ScaledGD( πœ† ) is almost unaffected, while 𝖯𝗋𝖾𝖼𝖦𝖣 suffers from a significant drop in the convergence rate and even breaks down with a moderate overparameterization π‘Ÿ

20 .

Figure 3:Relative reconstruction error versus the number of iterates with different overparameterization rank π‘Ÿ for ScaledGD( πœ† ) and 𝖯𝗋𝖾𝖼𝖦𝖣 . Noisy setting.

Though our theoretical results here are formulated in the noiseless setting, empirical evidence indicates our algorithm ScaledGD( πœ† ) also works in the noisy setting. Modifying the equation (4) for noiseless measurements, we assume the noisy measurements 𝑦 𝑖

⟨ 𝐴 𝑖 , 𝑀 ⟩ + πœ‰ 𝑖 where πœ‰ 𝑖 ∼ 𝒩 ​ ( 0 , 𝜎 2 ) are i.i.d. Gaussian noises. The minimax lower bound for the reconstruction error β€– 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ β€– π–₯ is denoted by β„° π—Œπ—π–Ίπ—

𝜎 ​ 𝑛 ​ π‘Ÿ ⋆ (candes2011tight). We compare the reconstruction error of ScaledGD( πœ† ) with β„° π—Œπ—π–Ίπ— under different noise levels 𝜎 . The results are shown in Figure 4. It can be seen that the final error of ScaledGD( πœ† ) matches the minimax optimal error β„° π—Œπ—π–Ίπ— within a small multiplicative factor for all noise levels.

Figure 4:The relative reconstruction error of ScaledGD( πœ† ) versus the number of iterates for ScaledGD( πœ† ) in the noisy setting, where it is observed that the final error of ScaledGD( πœ† ) approaches the minimax error. 6Discussions

This paper demonstrates that an appropriately preconditioned gradient descent method, called ScaledGD( πœ† ), guarantees an accelerated convergence to the ground truth low-rank matrix in overparameterized low-rank matrix sensing, when initialized from a sufficiently small random initialization. Furthermore, in the case of exact parameterization, our analysis guarantees the fast global convergence of ScaledGD( πœ† ) from a small random initialization. Collectively, this complements and represents a major step forward from prior analyses of ScaledGD (tong2021accelerating) by allowing overparametrization and small random initialization for noisy and approximately low-rank settings. This works opens up a few exciting future directions that are worth further exploring.

β€’

Asymmetric case. Our current analysis is confined to the recovery of low-rank positive semidefinite matrices, with only one factor matrix to be recovered. It remains to generalize this analysis to the recovery of general low-rank matrices with overparameterization.

β€’

Robust setting. Many applications encounter corrupted measurements that call for robust recovery algorithms that optimize nonsmooth functions such as the least absolute deviation loss. One such example is the scaled subgradient method (tong2021low), which is the nonsmooth counterpart of ScaledGD robust to ill-conditioning, and it’ll be interesting to study its performance under overparameterization.

β€’

Other overparameterized learning models. Our work provides evidence on the power of preconditioning in accelerating the convergence without hurting generalization in overparameterized low-rank matrix sensing, which is one kind of overparameterized learning models. It will be greatly desirable to extend the insights developed herein to other overparameterized learning models, for example low-rank matrix optimization (boumal2016smooth), tensors (tong2022scaling; dong2022fast), and neural networks (wang2021deep).

We believe the analysis framework put forth in this paper can be extended to analyze these general issues, by leveraging similar error decompositions and tailoring the treatment to the corresponding measurement or data models, see an overview ma2024provably and some recent works giampouras2024guarantees; diaz2025preconditioned along this line after the initial version of this paper.

Acknowledgements

The work of X. Xu and Y. Chi is supported in part by Office of Naval Research under N00014-19-1-2404, by Air Force Office of Scientific Research under award number FA9550-25-1-0060, and by National Science Foundation under CCF-1901199, DMS-2134080 and ECCS-2126634. The work of C. Ma is partially supported National Science Foundation via grant DMS-2311127 and DMS CAREER Award 2443867. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the United States Air Force.

Appendix APreliminaries

This section collects several preliminary results that are useful in later proofs. In general, for a matrix 𝐴 , we will denote by π‘ˆ 𝐴 the first factor in its compact SVD 𝐴

π‘ˆ 𝐴 ​ Ξ£ 𝐴 ​ 𝑉 𝐴 ⊀ , unless otherwise specified.

A.1Proof of Lemma 2

It is a standard result in random matrix theory (vershynin2010nonasym; rudelson2009smallest) that an 𝑀 Γ— 𝑁 ( 𝑀 β‰₯ 𝑁 ) random matrix 𝐺 0 with i.i.d. standard Gaussian entries satisfies

β„™ ​ ( β€– 𝐺 0 β€– ≀ 4 ​ ( 𝑀 + 𝑁 ) )

β‰₯ 1 βˆ’ exp ⁑ ( βˆ’ 𝑀 / 𝐢 ) ,

(30a)

β„™ ​ ( 𝜎 min ​ ( 𝐺 0 ) β‰₯ πœ€ ​ ( 𝑀 βˆ’ 𝑁 βˆ’ 1 ) )

β‰₯ 1 βˆ’ ( 𝐢 ​ πœ€ ) 𝑀 βˆ’ 𝑁 + 1 βˆ’ exp ⁑ ( βˆ’ 𝑀 / 𝐢 ) ,

(30b)

for some universal constant 𝐢

0 and for any πœ€

0 . Applying (30a) to the random matrix 𝑛 ​ 𝐺 which is an 𝑛 Γ— π‘Ÿ random matrix with i.i.d. standard Gaussian entries, we have

β€– 𝐺 β€– ≀ 4 ​ ( 𝑛 + π‘Ÿ ) / 𝑛 ≀ 8

with probability at least 1 βˆ’ exp ⁑ ( βˆ’ 𝑛 / 𝐢 ) .

Turning to the bound on 𝜎 min βˆ’ 1 ​ ( π‘ˆ ^ ⊀ ​ 𝐺 ) , observe that 𝑛 ​ π‘ˆ ^ ⊀ ​ 𝐺 is a π‘Ÿ ⋆ Γ— π‘Ÿ random matrix with i.i.d. standard Gaussian entries, thus applying (30b) to 𝑛 ​ π‘ˆ ^ ⊀ ​ 𝐺 with πœ€

( 2 ​ 𝑛 ) βˆ’ 𝐢 𝐺 + 1 yields

𝜎 min βˆ’ 1 ​ ( π‘ˆ ^ ⊀ ​ 𝐺 ) ≀ ( 2 ​ 𝑛 ) 𝐢 𝐺 βˆ’ 1 ​ ( π‘Ÿ βˆ’ π‘Ÿ ⋆ βˆ’ 1 ) βˆ’ 1 ≀ ( 2 ​ 𝑛 ) 𝐢 𝐺 βˆ’ 1 ​ ( 2 ​ π‘Ÿ ) ≀ ( 2 ​ 𝑛 ) 𝐢 𝐺

with probability at least 1 βˆ’ ( 2 ​ 𝑛 / 𝐢 ) βˆ’ ( 𝐢 𝐺 βˆ’ 1 ) ​ ( π‘Ÿ βˆ’ π‘Ÿ ⋆ + 1 ) βˆ’ exp ⁑ ( βˆ’ 𝑛 / 𝐢 ) . Here, the second inequality follows from

1 π‘Ÿ βˆ’ π‘Ÿ ⋆ βˆ’ 1 ≀ 1 π‘Ÿ βˆ’ π‘Ÿ βˆ’ 1

π‘Ÿ + π‘Ÿ βˆ’ 1 < 2 ​ π‘Ÿ .

Combining the above two bounds directly implies the desired probability bound if we choose 𝑐

1 / 𝐢 and choose a large 𝐢 𝐺 such that 𝐢 𝐺 β‰₯ 8 and 𝐢 𝐺 βˆ’ 1 β‰₯ 𝐢 𝐺 / 2 .

A.2Proof of Proposition 1

Using the definitions of 𝑆 𝑑 and 𝑁 𝑑 , we have

𝑋 𝑑

( π‘ˆ ⋆ ​ π‘ˆ ⋆ ⊀ + π‘ˆ ⋆ , βŸ‚ ​ π‘ˆ ⋆ , βŸ‚ ⊀ ) ​ 𝑋 𝑑

π‘ˆ ⋆ ​ 𝑆 𝑑 + π‘ˆ ⋆ , βŸ‚ ​ 𝑁 𝑑

π‘ˆ ⋆ ​ 𝑆 ~ 𝑑 ​ 𝑉 𝑑 ⊀ + π‘ˆ ⋆ , βŸ‚ ​ 𝑁 𝑑 ​ ( 𝑉 𝑑 ​ 𝑉 𝑑 ⊀ + 𝑉 𝑑 , βŸ‚ ​ 𝑉 𝑑 , βŸ‚ ⊀ )

π‘ˆ ⋆ ​ 𝑆 ~ 𝑑 ​ 𝑉 𝑑 ⊀ + π‘ˆ ⋆ , βŸ‚ ​ 𝑁 ~ 𝑑 ​ 𝑉 𝑑 ⊀ + π‘ˆ ⋆ , βŸ‚ ​ 𝑂 ~ 𝑑 ​ 𝑉 𝑑 , βŸ‚ ⊀ ,

where in the second line, we used the relation 𝑆 ~ 𝑑

𝑆 𝑑 ​ 𝑉 𝑑

π‘ˆ 𝑑 ​ Ξ£ 𝑑 ​ 𝑉 𝑑 ⊀ ​ 𝑉 𝑑

π‘ˆ 𝑑 ​ Ξ£ 𝑑 and thus

𝑆 𝑑

𝑆 ~ 𝑑 ​ 𝑉 𝑑 ⊀ .

(31) A.3Consequences of RIP

The first result is a standard consequence of RIP, see, for example stoger2021small.

Lemma 7.

Suppose that the linear map π’œ : Sym 2 ⁑ ( ℝ 𝑛 ) β†’ ℝ π‘š satisfies Assumption 1. Then we have

β€– ( ℐ βˆ’ π’œ βˆ— ​ π’œ ) ​ ( 𝑍 ) β€– ≀ 𝛿 ​ β€– 𝑍 β€– π–₯

for any 𝑍 ∈ Sym 2 ⁑ ( ℝ 𝑛 ) with rank at most π‘Ÿ ⋆ . Consequently, with πœ† ^ 1 β‰₯ β‹― β‰₯ πœ† ^ 𝑛 denoting the eigenvalues of π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) , it holds that

| πœ† ^ 𝑖 βˆ’ 𝜎 𝑖 2 ​ ( 𝑋 ⋆ ) | ≀ 𝛿 ​ π‘Ÿ ⋆ ​ β€– 𝑋 ⋆ β€– 2 .

We need another straightforward consequence of RIP, given by the following lemma.

Lemma 8.

Under the same setting as Lemma 7, we have

β€– ( ℐ βˆ’ π’œ βˆ— ​ π’œ ) ​ ( 𝑍 ) β€– ≀ 2 ​ 𝛿 ​ ( π‘Ÿ ∨ π‘Ÿ ⋆ ) / π‘Ÿ ⋆ ​ β€– 𝑍 β€– π–₯ ≀ 2 ​ ( π‘Ÿ ∨ π‘Ÿ ⋆ ) ​ 𝛿 π‘Ÿ ⋆ ​ β€– 𝑍 β€–

for any 𝑍 ∈ Sym 2 ⁑ ( ℝ 𝑛 ) with rank at most π‘Ÿ .

Proof.

Without loss of generality we may assume π‘Ÿ β‰₯ π‘Ÿ ⋆ , thus π‘Ÿ ∨ π‘Ÿ ⋆

π‘Ÿ . We claim that it is possible to decompose 𝑍

βˆ‘ 𝑖 ≀ ⌈ π‘Ÿ / π‘Ÿ ⋆ βŒ‰ 𝑍 𝑖 where 𝑍 𝑖 ∈ Sym 2 ⁑ ( ℝ 𝑛 ) , rank ⁑ ( 𝑍 𝑖 ) ≀ π‘Ÿ ⋆ and 𝑍 𝑖 ​ 𝑍 𝑗

0 if 𝑖 β‰  𝑗 . To see why this is the case, notice the spectral decomposition of 𝑍 gives π‘Ÿ rank-one components that are mutually orthogonal, thus we may divide them into ⌈ π‘Ÿ / π‘Ÿ ⋆ βŒ‰ subgroups indexed by 𝑖

1 , … , ⌈ π‘Ÿ / π‘Ÿ ⋆ βŒ‰ , such that each subgroup contains at most π‘Ÿ ⋆ components. Let 𝑍 𝑖 be the sum of the components in the subgroup 𝑖 , it is easy to check that 𝑍 𝑖 has the desired property.

The property of the decomposition yields

β€– 𝑍 β€– π–₯ 2

tr ⁑ ( 𝑍 2 )

βˆ‘ 𝑖 , 𝑗 ≀ ⌈ π‘Ÿ / π‘Ÿ ⋆ βŒ‰ tr ⁑ ( 𝑍 𝑖 ​ 𝑍 𝑗 )

βˆ‘ 𝑖 ≀ ⌈ π‘Ÿ / π‘Ÿ ⋆ βŒ‰ β€– 𝑍 𝑖 β€– π–₯ 2 .

(32)

But for each 𝑍 𝑖 , Lemma 7 implies

β€– ( ℐ βˆ’ π’œ βˆ— ​ π’œ ) ​ ( 𝑍 𝑖 ) β€– ≀ 𝛿 ​ β€– 𝑍 𝑖 β€– π–₯ .

Summing up for 𝑖 ≀ ⌈ π‘Ÿ / π‘Ÿ ⋆ βŒ‰ yields

β€– ( ℐ βˆ’ π’œ βˆ— ​ π’œ ) ​ ( 𝑍 ) β€– ≀ βˆ‘ 𝑖 ≀ ⌈ π‘Ÿ / π‘Ÿ ⋆ βŒ‰ β€– ( ℐ βˆ’ π’œ βˆ— ​ π’œ ) ​ ( 𝑍 𝑖 ) β€– ≀ 𝛿 ​ βˆ‘ 𝑖 ≀ ⌈ π‘Ÿ / π‘Ÿ ⋆ βŒ‰ β€– 𝑍 𝑖 β€– π–₯ ≀ 𝛿 ​ ⌈ π‘Ÿ / π‘Ÿ ⋆ βŒ‰ ​ β€– 𝑍 β€– π–₯ ,

where the last inequality follows from (32) and from Cauchy-Schwarz inequality.

The first inequality in Lemma 8 follows from the above inequality by noting that ⌈ π‘Ÿ / π‘Ÿ ⋆ βŒ‰ ≀ 2 ​ π‘Ÿ / π‘Ÿ ⋆ given π‘Ÿ β‰₯ π‘Ÿ ⋆ which was assumed in the beginning of the proof. The second inequality in Lemma 8 follows from β€– 𝑍 β€– π–₯ ≀ π‘Ÿ ​ β€– 𝑍 β€– . ∎

A.4Matrix perturbation results

The next few results are all on matrix perturbations. We first present a perturbation result on matrix inverse.

Lemma 9.

Assume that 𝐴 , 𝐡 are square matrices of the same dimension, and that 𝐴 is invertible. If β€– 𝐴 βˆ’ 1 ​ 𝐡 β€– ≀ 1 / 2 , then

( 𝐴 + 𝐡 ) βˆ’ 1

𝐴 βˆ’ 1 + 𝐴 βˆ’ 1 ​ 𝐡 ​ 𝑄 ​ 𝐴 βˆ’ 1 , for some  ​ β€– 𝑄 β€– ≀ 2 .

Similarly, if β€– 𝐡 ​ 𝐴 βˆ’ 1 β€– ≀ 1 / 2 , then we have

( 𝐴 + 𝐡 ) βˆ’ 1

𝐴 βˆ’ 1 + 𝐴 βˆ’ 1 ​ 𝑄 ​ 𝐡 ​ 𝐴 βˆ’ 1 , for some  ​ β€– 𝑄 β€– ≀ 2 .

In particular, if β€– 𝐡 β€– ≀ 𝜎 min ​ ( 𝐴 ) / 2 , then both of the above equations hold.

Proof.

The claims follow from the identity

( 𝐴 + 𝐡 ) βˆ’ 1

𝐴 βˆ’ 1 βˆ’ 𝐴 βˆ’ 1 ​ 𝐡 ​ ( 𝐼 + 𝐴 βˆ’ 1 ​ 𝐡 ) βˆ’ 1 ​ 𝐴 βˆ’ 1

𝐴 βˆ’ 1 βˆ’ 𝐴 βˆ’ 1 ​ ( 𝐼 + 𝐡 ​ 𝐴 βˆ’ 1 ) βˆ’ 1 ​ 𝐡 ​ 𝐴 βˆ’ 1 .

For the first claim when β€– 𝐴 βˆ’ 1 ​ 𝐡 β€– ≀ 1 / 2 , we set 𝑄 := βˆ’ ( 𝐼 + 𝐴 βˆ’ 1 ​ 𝐡 ) βˆ’ 1 , which satisfies β€– 𝑄 β€–

β€– ( 𝐼 + 𝐴 βˆ’ 1 ​ 𝐡 ) βˆ’ 1 β€– ≀ 1 1 βˆ’ β€– 𝐴 βˆ’ 1 ​ 𝐡 β€– ≀ 2 . The second claim follows similarly. Finally, we note that when β€– 𝐡 β€– ≀ 𝜎 min ​ ( 𝐴 ) / 2 , it holds

β€– 𝐴 βˆ’ 1 ​ 𝐡 β€– ≀ 1 𝜎 min ​ ( 𝐴 ) ​ β€– 𝐡 β€– ≀ 1 2 and β€– 𝐡 ​ 𝐴 βˆ’ 1 β€– ≀ β€– 𝐡 β€– ​ 1 𝜎 min ​ ( 𝐴 ) ≀ 1 2 ,

thus completing the proof. ∎

Next, we focus on the minimum singular value of certain matrix of form 𝐼 + 𝐴 ​ 𝐡 .

Lemma 10.

If 𝐴 , 𝐡 are positive definite matrices of the same size, we have

𝜎 min ​ ( 𝐼 + 𝐴 ​ 𝐡 ) β‰₯ πœ… βˆ’ 1 / 2 ​ ( 𝐴 ) , where  ​ πœ… ​ ( 𝐴 ) ≔ β€– 𝐴 β€– 𝜎 min ​ ( 𝐴 ) .

Proof.

Writing 𝐼 + 𝐴 ​ 𝐡

𝐴 1 / 2 ​ ( 𝐼 + 𝐴 1 / 2 ​ 𝐡 ​ 𝐴 1 / 2 ) ​ 𝐴 βˆ’ 1 / 2 , we obtain

𝜎 min ​ ( 𝐼 + 𝐴 ​ 𝐡 ) β‰₯ 𝜎 min ​ ( 𝐴 1 / 2 ) ​ 𝜎 min ​ ( 𝐴 βˆ’ 1 / 2 ) ​ 𝜎 min ​ ( 𝐼 + 𝐴 1 / 2 ​ 𝐡 ​ 𝐴 1 / 2 ) .

The proof is completed by noting that 𝜎 min ​ ( 𝐴 1 / 2 )

𝜎 min 1 / 2 ​ ( 𝐴 ) , 𝜎 min ​ ( 𝐴 βˆ’ 1 / 2 )

β€– 𝐴 β€– βˆ’ 1 / 2 , and that 𝜎 min ​ ( 𝐼 + 𝐴 1 / 2 ​ 𝐡 ​ 𝐴 1 / 2 ) β‰₯ 1 since 𝐴 1 / 2 ​ 𝐡 ​ 𝐴 1 / 2 is positive semidefinite. ∎

The last result still concerns the minimum singular value of a matrix of interest.

Lemma 11.

There exists a universal constant 𝑐 11

0 such that if Ξ› is a positive definite matrix obeying β€– Ξ› β€– ≀ 𝑐 11 and 𝜎 min ​ ( π‘Œ ) ≀ 1 / 3 , then for any πœ‚ ≀ 𝑐 11 we have

𝜎 min ​ ( ( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( π‘Œ ​ π‘Œ ⊀ + Ξ› ) βˆ’ 1 ) ​ π‘Œ ) β‰₯ ( 1 + πœ‚ 6 ) ​ 𝜎 min ​ ( π‘Œ ) .

(33) Proof.

Denote 𝑍

π‘Œ ​ π‘Œ ⊀ and let π‘ˆ ​ Ξ£ ​ π‘ˆ ⊀

𝑍 + Ξ› be the spectral decomposition of 𝑍 + Ξ› . By a coordinate transform one may assume 𝑍 + Ξ›

Ξ£ . It suffices to show

πœ† min ​ ( ( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ Ξ£ βˆ’ 1 ) ​ 𝑍 ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ Ξ£ βˆ’ 1 ) ) β‰₯ ( 1 + 1 6 ​ πœ‚ ) 2 ​ πœ† min ​ ( 𝑍 ) .

(34)

For simplicity we denote 𝜁

πœ† min ​ ( 𝑍 ) , which is by assumption smaller than 1 / 9 . Fix 𝐾

1 / 4 so that 𝐾 β‰₯ 2 ​ 𝜁 + 4 ​ 𝑐 11 by choosing 𝑐 11 to be small enough. By permuting coordinates we may further assume that the diagonal matrix Ξ£ is of the following form:

Ξ£

[ Ξ£ ≀ 𝐾

Ξ£

𝐾 ] ,

(35)

where Ξ£ ≀ 𝐾 , Ξ£

𝐾 are diagonal matrices such that πœ† max ​ ( Ξ£ ≀ 𝐾 ) ≀ 𝐾 and πœ† min ​ ( Ξ£

𝐾 )

𝐾 . It suffices to consider the case where Σ

𝐾 is not degenerate, because otherwise πœ† max ​ ( Ξ£ ) ≀ 𝐾 ≀ 1 / 2 , and the desired (34) follows as

πœ† min ​ ( ( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ Ξ£ βˆ’ 1 ) ​ 𝑍 ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ Ξ£ βˆ’ 1 ) ) β‰₯ ( 1 βˆ’ πœ‚ + πœ‚ ​ πœ† max βˆ’ 1 ​ ( Ξ£ ) ) 2 ​ πœ† min ​ ( 𝑍 ) β‰₯ ( 1 + πœ‚ ) 2 ​ πœ† min ​ ( 𝑍 ) .

For the rest of the proof, we assume the block corresponding to Ξ£

𝐾 is not degenerate.

Divide 𝑍 into blocks of the same shape as (35):

𝑍

[ 𝑍 0

𝐴

𝐴 ⊀

𝑍 1 ] .

(36)

The purpose of such division is to facilitate computation of minimum eigenvalues by Schur’s complement lemma. For preparation, we make a few simple observations. Since 𝑍

Ξ£ βˆ’ Ξ› , we see that 𝐴 being an off-diagonal submatrix of 𝑍 satisfies β€– 𝐴 β€– ≀ β€– Ξ› β€– ≀ 𝑐 11 , and similarly β€– 𝑍 0 βˆ’ Ξ£ ≀ 𝐾 β€– ≀ 𝑐 11 , β€– 𝑍 1 βˆ’ Ξ£

𝐾 β€– ≀ 𝑐 11 . In particular, we have

πœ† min ​ ( 𝑍 1 ) β‰₯ πœ† min ​ ( Ξ£

𝐾 ) βˆ’ 𝑐 11

𝐾 βˆ’ 𝑐 11 β‰₯ 2 ​ 𝜁 + 3 ​ 𝑐 11

𝜁 ,

(37)

which implies 𝑍 1 βˆ’ 𝜁 ​ 𝐼 is positive definite and invertible. Thus by Schur’s complement lemma, 𝑍 βͺ° 𝜁 ​ 𝐼 is equivalent to

𝑍 0 βˆ’ 𝜁 ​ 𝐼 βˆ’ 𝐴 ​ ( 𝑍 1 βˆ’ 𝜁 ​ 𝐼 ) βˆ’ 1 ​ 𝐴 ⊀ βͺ° 0 ,

(38)

which provides an analytic characterization for the minimum eigenvalue 𝜁 of 𝑍 .

The rest of the proof follows from the following steps: we will first show again by Schur’s complement lemma that (34) admits a similar analytic characterization. More precisely, denoting 𝜁 β€²

( 1 + πœ‚ 6 ) 2 ​ 𝜁 , Ξ£ 0

( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ Ξ£ ≀ 𝐾 βˆ’ 1 and Ξ£ 1

( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ Ξ£

𝐾 βˆ’ 1 , then (34) is equivalent to

𝑍 0 βˆ’ 𝜁 β€² ​ Ξ£ 0 βˆ’ 2 βˆ’ 𝐴 ​ ( 𝑍 1 βˆ’ 𝜁 β€² ​ Ξ£ 1 βˆ’ 2 ) βˆ’ 1 ​ 𝐴 ⊀ βͺ° 0 .

(39)

After proving they are equivalent, we will prove that (39) holds as long as the following sufficient condition holds

𝑍 0 βˆ’ ( 1 + 3 ​ πœ‚ ) βˆ’ 2 ​ 𝜁 β€² ​ 𝐼 βˆ’ 𝐴 ​ ( 𝑍 1 βˆ’ 𝜁 ​ 𝐼 ) βˆ’ 1 ​ 𝐴 ⊀ βˆ’ 10 ​ πœ‚ ​ 𝜁 ​ 𝐴 ​ ( 𝑍 1 βˆ’ 𝜁 ​ 𝐼 ) βˆ’ 2 ​ 𝐴 ⊀ βͺ° 0 .

(40)

In the last step, we establish the above sufficient condition to complete the proof.

Step 1: equivalence between (34) and (39).

First notice that

( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ Ξ£ βˆ’ 1 ) ​ 𝑍 ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ Ξ£ βˆ’ 1 )

[ Ξ£ 0 ​ 𝑍 0 ​ Ξ£ 0

Ξ£ 0 ​ 𝐴 ​ Ξ£ 1

Ξ£ 1 ​ 𝐴 ⊀ ​ Ξ£ 0

Ξ£ 1 ​ 𝑍 1 ​ Ξ£ 1 ] .

(41)

In order to invoke Schur’s complement lemma, we need to verify Ξ£ 1 ​ 𝑍 1 ​ Ξ£ 1 βˆ’ 𝜁 β€² ​ 𝐼 ≻ 0 . Observe that by definition we have

Ξ£ 0 βͺ° ( 1 + ( 𝐾 βˆ’ 1 βˆ’ 1 ) ​ πœ‚ ) ​ 𝐼

( 1 + 3 ​ πœ‚ ) ​ 𝐼 , Ξ£ 1 βͺ° ( 1 βˆ’ πœ‚ ) ​ 𝐼 .

(42)

Hence

Ξ£ 1 ​ 𝑍 1 ​ Ξ£ 1 βˆ’ 𝜁 β€² ​ 𝐼 βͺ° ( 1 βˆ’ πœ‚ ) 2 ​ 𝑍 1 βˆ’ ( 1 + 1 6 ​ πœ‚ ) 2 ​ 𝜁 ​ 𝐼 ≻ 2 ​ ( 1 βˆ’ πœ‚ ) 2 ​ 𝜁 ​ 𝐼 βˆ’ ( 1 + 1 6 ​ πœ‚ ) 2 ​ 𝜁 ​ 𝐼 ≻ 0 ,

where in the second inequality we used 𝑍 1 βˆ’ 2 ​ 𝜁 ​ 𝐼 ≻ 0 proved in (37), and in the last inequality we used πœ‚ ≀ 𝑐 πœ‚ with 𝑐 πœ‚ sufficiently small. This completes the verification that Ξ£ 1 ​ 𝑍 1 ​ Ξ£ 1 βˆ’ 𝜁 β€² ​ 𝐼 ≻ 0 . Now, invoking Schur’s complement lemma yields that (34) is equivalent to

Ξ£ 0 ​ 𝑍 0 ​ Ξ£ 0 βˆ’ 𝜁 β€² ​ 𝐼 βˆ’ Ξ£ 0 ​ 𝐴 ​ Ξ£ 1 ​ ( Ξ£ 1 ​ 𝑍 1 ​ Ξ£ 1 βˆ’ 𝜁 β€² ​ 𝐼 ) βˆ’ 1 ​ Ξ£ 1 ​ 𝐴 ⊀ ​ Ξ£ 0 βͺ° 0 ,

which simplifies easily to (39), as claimed.

Step 2: establishing (40) as a sufficient condition for (39).

By (42), it follows that

( 𝑍 1 βˆ’ 𝜁 β€² ​ Ξ£ 1 βˆ’ 2 ) βˆ’ 1
βͺ― ( 𝑍 1 βˆ’ ( 1 βˆ’ πœ‚ ) βˆ’ 2 ​ 𝜁 β€² ​ 𝐼 ) βˆ’ 1

( 𝑍 1 βˆ’ 𝜁 ​ 𝐼 βˆ’ ( ( 1 βˆ’ πœ‚ ) βˆ’ 2 ​ 𝜁 β€² βˆ’ 𝜁 ) ​ 𝐼 ) βˆ’ 1 ,

(43)

where we used the well-known fact that 𝐴 βͺ― 𝐡 implies 𝐡 βˆ’ 1 βͺ― 𝐴 βˆ’ 1 for positive definite matrices 𝐴 and 𝐡 (cf. (bhatia1997matrix, Proposition V.1.6)). We aim to apply Lemma 9 to control the above term, by treating ( ( 1 βˆ’ πœ‚ ) βˆ’ 2 ​ 𝜁 β€² βˆ’ 𝜁 ) ​ 𝐼 as a perturbation term. For this purpose we need to verify

| ( 1 βˆ’ πœ‚ ) βˆ’ 2 ​ 𝜁 β€² βˆ’ 𝜁 | ≀ 1 2 ​ πœ† min ​ ( 𝑍 1 βˆ’ 𝜁 ​ 𝐼 ) .

(44)

Given πœ‚ ≀ 𝑐 πœ‚ with sufficiently small 𝑐 πœ‚ , we have ( 1 βˆ’ πœ‚ ) βˆ’ 2 ≀ 1 + 3 ​ πœ‚ , ( 1 + 1 6 ​ πœ‚ ) 2 ≀ 1 + πœ‚ , and ( 1 + 3 ​ πœ‚ ) ​ ( 1 + πœ‚ ) ≀ 1 + 5 ​ πœ‚ , thus

0 ≀ ( 1 βˆ’ πœ‚ ) βˆ’ 2 ​ ( 1 + 1 6 ​ πœ‚ ) 2 ​ 𝜁 βˆ’ 𝜁

( 1 βˆ’ πœ‚ ) βˆ’ 2 ​ 𝜁 β€² βˆ’ 𝜁 ≀ ( 1 + 3 ​ πœ‚ ) ​ ( 1 + πœ‚ ) ​ 𝜁 βˆ’ 𝜁 ≀ 5 ​ πœ‚ ​ 𝜁 < 𝜁 / 2 ,

where the last inequality follows from 𝑐 πœ‚ ≀ 1 / 10 . On the other hand, invoking (37), we obtain

1 2 ​ 𝜁 ≀ 1 2 ​ ( πœ† min ​ ( 𝑍 1 ) βˆ’ 𝜁 )

1 2 ​ πœ† min ​ ( 𝑍 1 βˆ’ 𝜁 ​ 𝐼 ) ,

which verifies (44). Thus we may apply Lemma 9 to show

β€– ( 𝑍 1 βˆ’ 𝜁 ​ 𝐼 ) ​ ( ( 𝑍 1 βˆ’ 𝜁 ​ 𝐼 ) βˆ’ 1 βˆ’ ( 𝑍 1 βˆ’ 𝜁 ​ 𝐼 βˆ’ ( ( 1 βˆ’ πœ‚ ) βˆ’ 2 ​ 𝜁 β€² βˆ’ 𝜁 ) ​ 𝐼 ) βˆ’ 1 ) ​ ( 𝑍 1 βˆ’ 𝜁 ​ 𝐼 ) β€– ≀ 2 ​ | ( 1 βˆ’ πœ‚ ) βˆ’ 2 ​ 𝜁 β€² βˆ’ 𝜁 | ≀ 10 ​ πœ‚ ​ 𝜁 ,

therefore

( 𝑍 1 βˆ’ 𝜁 ​ 𝐼 βˆ’ ( ( 1 βˆ’ πœ‚ ) βˆ’ 2 ​ 𝜁 β€² βˆ’ 𝜁 ) ​ 𝐼 ) βˆ’ 1 βͺ― ( 𝑍 1 βˆ’ 𝜁 ​ 𝐼 ) βˆ’ 1 + 10 ​ πœ‚ ​ 𝜁 ​ ( 𝑍 1 βˆ’ 𝜁 ​ 𝐼 ) βˆ’ 2 .

Together with (43), this implies

( 𝑍 1 βˆ’ 𝜁 β€² ​ Ξ£ 1 βˆ’ 2 ) βˆ’ 1 βͺ― ( 𝑍 1 βˆ’ 𝜁 ​ 𝐼 ) βˆ’ 1 + 10 ​ πœ‚ ​ 𝜁 ​ ( 𝑍 1 βˆ’ 𝜁 ​ 𝐼 ) βˆ’ 2 .

(45)

Combining (42) and (45), we see that a sufficient condition for (39) to hold is (40).

Step 3: establishing (40).

It is clear that (40) is implied by

𝜁 ​ 𝐼 βˆ’ ( 1 + 3 ​ πœ‚ ) βˆ’ 2 ​ 𝜁 β€² ​ 𝐼 βˆ’ 10 ​ πœ‚ ​ 𝜁 ​ 𝐴 ​ ( 𝑍 1 βˆ’ 𝜁 ​ 𝐼 ) βˆ’ 2 ​ 𝐴 ⊀ βͺ° 0 ,

(46)

by leveraging the relation 𝑍 0 βͺ° 𝜁 ​ 𝐼 + 𝐴 ​ ( 𝑍 1 βˆ’ 𝜁 ​ 𝐼 ) βˆ’ 1 ​ 𝐴 ⊀ from (38).

Hence, it boils down to prove (46). Recalling β€– 𝐴 β€– ≀ 𝑐 11 , and from (37), we know πœ† min ​ ( 𝑍 1 βˆ’ 𝜁 ​ 𝐼 ) β‰₯ 𝐾 βˆ’ 𝑐 11 βˆ’ 𝜁 β‰₯ 𝜁 + 3 ​ 𝑐 11 . Thus

β€– 𝐴 ​ ( 𝑍 1 βˆ’ 𝜁 ​ 𝐼 ) βˆ’ 2 ​ 𝐴 ⊀ β€– ≀ β€– 𝐴 β€– 2 ​ β€– ( 𝑍 1 βˆ’ 𝜁 ​ 𝐼 ) βˆ’ 2 β€– ≀ 𝑐 11 2 / ( 𝜁 + 3 ​ 𝑐 11 ) 2 ≀ 1 / 9 .

Therefore, to prove (46) it suffices to show

𝜁 βˆ’ ( 1 + 3 ​ πœ‚ ) βˆ’ 2 ​ 𝜁 β€² β‰₯ 10 9 ​ πœ‚ ​ 𝜁 .

(47)

It is easy to verify that the above inequality holds for our choice 𝜁 β€²

( 1 + 1 6 ​ πœ‚ ) 2 ​ 𝜁 . In fact, given πœ‚ ≀ 𝑐 πœ‚ for sufficiently small 𝑐 πœ‚ , we have ( 1 + 3 ​ πœ‚ ) βˆ’ 2 ≀ 1 βˆ’ 4 ​ πœ‚ , ( 1 + 1 6 ​ πœ‚ ) 2 ≀ 1 + πœ‚ . These together yield

𝜁 βˆ’ ( 1 + 3 ​ πœ‚ ) βˆ’ 2 ​ ( 1 + 1 6 ​ πœ‚ ) 2 ​ 𝜁 β‰₯ 𝜁 βˆ’ ( 1 βˆ’ 4 ​ πœ‚ ) ​ ( 1 + πœ‚ ) ​ 𝜁

3 ​ πœ‚ ​ 𝜁 + 4 ​ πœ‚ 2 ​ 𝜁 β‰₯ 3 ​ πœ‚ ​ 𝜁 β‰₯ 10 9 ​ πœ‚ ​ 𝜁 ,

establishing (47) as desired. ∎

Appendix BDecompositions of key terms

In this section, we first present a useful bound of a key error quantity

Ξ” 𝑑 ≔ ( ℐ βˆ’ π’œ βˆ— ​ π’œ ) ​ ( 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ ) ,

(48)

where 𝑋 𝑑 is the iterate of ScaledGD( πœ† ) given in (7).

Lemma 12.

Suppose π’œ ​ ( β‹… ) satisfies Assumption 1. For any 𝑑 β‰₯ 0 such that (22) holds, we have

β€– Ξ” 𝑑 β€– ≀ 8 ​ 𝛿 ​ ( β€– 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 β€– π–₯ + β€– 𝑆 ~ 𝑑 β€– ​ β€– 𝑁 ~ 𝑑 β€– π–₯ + 𝑛 ​ β€– 𝑂 ~ 𝑑 β€– 2 ) .

(49)

In particular, there exists some constant 𝑐 12 ≲ 𝑐 𝛿 / 𝑐 πœ† such that

β€– Ξ” 𝑑 β€– ≀ 16 ​ ( 𝐢 3 . π‘Ž + 1 ) 2 ​ 𝑐 𝛿 ​ πœ… βˆ’ 2 ​ 𝐢 𝛿 / 3 ​ β€– 𝑋 ⋆ β€– 2 ≀ 𝑐 12 ​ πœ… βˆ’ 2 ​ 𝐢 𝛿 / 3 ​ β€– 𝑋 ⋆ β€– 2 .

(50) Proof.

The decomposition (19) in Proposition 1 yields

𝑋 𝑑 ​ 𝑋 𝑑 ⊀

π‘ˆ ⋆ ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ ​ π‘ˆ ⋆ ⊀ + π‘ˆ ⋆ ​ 𝑆 ~ 𝑑 ​ 𝑁 ~ 𝑑 ⊀ ​ π‘ˆ ⋆ , βŸ‚ ⊀ + π‘ˆ ⋆ , βŸ‚ ​ 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ ​ π‘ˆ ⋆ ⊀ + π‘ˆ ⋆ , βŸ‚ ​ 𝑁 ~ 𝑑 ​ 𝑁 ~ 𝑑 ⊀ ​ π‘ˆ ⋆ , βŸ‚ ⊀ + π‘ˆ ⋆ , βŸ‚ ​ 𝑂 ~ 𝑑 ​ 𝑂 ~ 𝑑 ⊀ ​ π‘ˆ ⋆ , βŸ‚ ⊀ .

Since 𝑀 ⋆

π‘ˆ ⋆ ​ Ξ£ ⋆ 2 ​ π‘ˆ ⋆ ⊀ , we have

𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆

π‘ˆ ⋆ ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ π‘ˆ ⋆ ⊀ ⏟

⁣ : 𝑇 1 + π‘ˆ ⋆ ​ 𝑆 ~ 𝑑 ​ 𝑁 ~ 𝑑 ⊀ ​ π‘ˆ ⋆ , βŸ‚ ⊀ + π‘ˆ ⋆ , βŸ‚ ​ 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ ​ π‘ˆ ⋆ ⊀ ⏟

⁣ : 𝑇 2 + π‘ˆ ⋆ , βŸ‚ ​ 𝑁 ~ 𝑑 ​ 𝑁 ~ 𝑑 ⊀ ​ π‘ˆ ⋆ , βŸ‚ ⊀ ⏟

⁣ : 𝑇 3 + π‘ˆ ⋆ , βŸ‚ ​ 𝑂 ~ 𝑑 ​ 𝑂 ~ 𝑑 ⊀ ​ π‘ˆ ⋆ , βŸ‚ ⊀ ⏟

⁣ : 𝑇 4 .

(51)

Note that π‘ˆ ⋆ ∈ ℝ 𝑛 Γ— π‘Ÿ ⋆ is of rank π‘Ÿ ⋆ , thus 𝑇 1 has rank at most π‘Ÿ ⋆ and 𝑇 2 has rank at most 2 ​ π‘Ÿ ⋆ . Similarly, since 𝑁 ~ 𝑑

𝑁 𝑑 ​ 𝑉 𝑑 while 𝑉 𝑑 ∈ ℝ π‘Ÿ Γ— π‘Ÿ ⋆ is of rank π‘Ÿ ⋆ , 𝑇 3 has rank at most π‘Ÿ ⋆ . It is also trivial that 𝑇 4 as an 𝑛 Γ— 𝑛 matrix has rank at most 𝑛 . Invoking Lemma 8, we obtain

β€– ( ℐ βˆ’ π’œ βˆ— ​ π’œ ) ​ ( 𝑇 1 ) β€–

≀ 2 ​ 𝛿 ​ β€– π‘ˆ ⋆ ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ π‘ˆ ⋆ ⊀ β€– π–₯ ≀ 2 ​ 𝛿 ​ β€– 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 β€– π–₯ ,

β€– ( ℐ βˆ’ π’œ βˆ— ​ π’œ ) ​ ( 𝑇 2 ) β€–

≀ 2 ​ 3 ​ 𝛿 ​ β€– π‘ˆ ⋆ ​ 𝑆 ~ 𝑑 ​ 𝑁 ~ 𝑑 ⊀ ​ π‘ˆ ⋆ , βŸ‚ ⊀ + π‘ˆ ⋆ , βŸ‚ ​ 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ ​ π‘ˆ ⋆ ⊀ β€– π–₯ ≀ 4 ​ 2 ​ 𝛿 ​ β€– 𝑆 ~ 𝑑 β€– ​ β€– 𝑁 ~ 𝑑 β€– π–₯ ,

β€– ( ℐ βˆ’ π’œ βˆ— ​ π’œ ) ​ ( 𝑇 3 ) β€–

≀ 2 ​ 𝛿 ​ β€– π‘ˆ ⋆ , βŸ‚ ​ 𝑁 ~ 𝑑 ​ 𝑁 ~ 𝑑 ⊀ ​ π‘ˆ ⋆ , βŸ‚ ⊀ β€– π–₯ ≀ 2 ​ 𝛿 ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– ​ β€– 𝑆 ~ 𝑑 β€– ​ β€– Ξ£ ⋆ βˆ’ 1 β€– ​ β€– 𝑁 ~ 𝑑 β€– π–₯ ≀ 𝛿 ​ β€– 𝑆 ~ 𝑑 β€– ​ β€– 𝑁 ~ 𝑑 β€– π–₯ ,

β€– ( ℐ βˆ’ π’œ βˆ— ​ π’œ ) ​ ( 𝑇 4 ) β€–

≀ 2 ​ 𝛿 ​ 𝑛 ​ β€– π‘ˆ ⋆ , βŸ‚ ​ 𝑂 ~ 𝑑 ​ 𝑂 ~ 𝑑 ⊀ ​ π‘ˆ ⋆ , βŸ‚ ⊀ β€– ≀ 2 ​ 𝛿 ​ 𝑛 ​ β€– 𝑂 ~ 𝑑 β€– 2 ,

where the third line follows from β€– Ξ£ ⋆ βˆ’ 1 β€–

πœ… ​ β€– 𝑋 ⋆ β€– βˆ’ 1 and from (22c) in view that 𝐢 𝛿 is sufficiently large and 𝑐 3 is sufficiently small. The conclusion (49) follows from summing up the above inequalities.

For the remaining part of the lemma, note that the following inequalities that bound the individual terms of (49) can be inferred from (22): namely,

β€– 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ β€– π–₯ ≀ 2 ​ π‘Ÿ ⋆ ​ β€– 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ β€– ≀ 2 ​ π‘Ÿ ⋆ ​ ( 𝐢 3 . π‘Ž 2 ​ πœ… 2 + 1 ) ​ β€– 𝑋 ⋆ β€– 2

by (22d), and

β€– 𝑆 ~ 𝑑 β€– ​ β€– 𝑁 ~ 𝑑 β€– π–₯
≀ π‘Ÿ ⋆ ​ β€– 𝑆 ~ 𝑑 β€– ​ β€– 𝑁 ~ 𝑑 β€–

≀ π‘Ÿ ⋆ ​ ( 𝐢 3 . π‘Ž ​ πœ… 3 ​ β€– 𝑋 ⋆ β€– ) β‹… β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– β‹… β€– 𝑆 ~ 𝑑 β€– β‹… β€– Ξ£ ⋆ βˆ’ 1 β€–

≀ π‘Ÿ ⋆ ​ ( 𝐢 3 . π‘Ž ​ πœ… 3 ​ β€– 𝑋 ⋆ β€– ) β‹… ( 𝑐 3 ​ πœ… βˆ’ 𝐢 𝛿 / 2 ​ β€– 𝑋 ⋆ β€– ) β‹… ( 𝐢 3 . π‘Ž ​ πœ… 3 ​ β€– 𝑋 ⋆ β€– ) β‹… 𝜎 min βˆ’ 1 ​ ( Ξ£ ⋆ )

π‘Ÿ ⋆ ​ 𝑐 3 ​ 𝐢 3 . π‘Ž 2 ​ πœ… 6 ​ β€– 𝑋 ⋆ β€– 2 ​ πœ… βˆ’ 𝐢 𝛿 / 2

≀ π‘Ÿ ⋆ ​ 𝐢 3 . π‘Ž 2 ​ β€– 𝑋 ⋆ β€– 2 ,

where the first inequality uses the fact that 𝑁 ~ 𝑑

𝑁 𝑑 ​ 𝑉 𝑑 contains a rank- π‘Ÿ ⋆ factor 𝑉 𝑑 , hence has rank at most π‘Ÿ ⋆ ; the second line follows from (22d), the third line follows from (22c) and (22d), and the last line follows from choosing 𝑐 𝛿 sufficiently small such that 𝑐 3 ≀ 1 (which is possible since 𝑐 3 ≲ 𝑐 𝛿 / 𝑐 πœ† ) and from choosing 𝐢 𝛿 such that πœ… 6 ​ πœ… βˆ’ 𝐢 𝛿 / 2 ≀ 1 . Finally, from (22b) and its corollary (24), we have

2 ​ 𝑛 ​ β€– 𝑂 ~ 𝑑 β€– 2 ≀ 2 ​ 𝑛 ​ 𝛼 3 / 2 ​ β€– 𝑋 ⋆ β€– 1 / 2 ≀ β€– 𝑋 ⋆ β€– 2 ,

since from (12c) it is easy to show that 𝛼 ≀ ( 2 ​ 𝑛 ) βˆ’ 2 / 3 ​ β€– 𝑋 ⋆ β€– .

Combining these inequalities and (49) yields

β€– Ξ” 𝑑 β€– ≀ 8 ​ 𝛿 ​ π‘Ÿ ⋆ ​ ( 2 ​ 𝐢 3 . π‘Ž 2 ​ πœ… 2 + 1 + 𝐢 3 . π‘Ž 2 + 1 ) ​ β€– 𝑋 ⋆ β€– 2 ≀ 16 ​ 𝛿 ​ π‘Ÿ ⋆ ​ πœ… 2 ​ ( 𝐢 3 . π‘Ž 2 + 1 ) ​ β€– 𝑋 ⋆ β€– 2 .

(52)

Recalling that by (10) we have 𝛿 ​ π‘Ÿ ⋆ ​ πœ… 2 ≀ 𝑐 𝛿 ​ πœ… βˆ’ 𝐢 𝛿 + 2 ≀ 𝑐 𝛿 ​ πœ… βˆ’ 2 ​ 𝐢 𝛿 / 3 as long as 𝐢 𝛿 β‰₯ 6 , we obtain the desired conclusion. We may choose 𝑐 12

32 ​ ( 𝐢 3 . π‘Ž + 1 ) 2 ​ 𝑐 𝛿 , and the bound 𝑐 12 ≲ 𝑐 𝛿 / 𝑐 πœ† follows from 𝐢 3 . π‘Ž ≲ 𝑐 πœ† βˆ’ 1 / 2 . ∎

We next present several useful decompositions of the signal term 𝑆 𝑑 + 1 and the noise term  𝑁 𝑑 + 1 , which are extremely useful in later developments.

Lemma 13.

For any 𝑑 such that 𝑆 ~ 𝑑 is invertible and (22) holds, we have

𝑆 𝑑 + 1

( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 + 𝐸 𝑑 π‘Ž ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ) ​ 𝑆 ~ 𝑑 ​ 𝑉 𝑑 ⊀ + πœ‚ ​ 𝐸 𝑑 𝑏 ,

(53a)

𝑁 𝑑 + 1

𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 + πœ‚ ​ 𝐸 𝑑 𝑐 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑉 𝑑 ⊀

  • πœ‚ ​ 𝐸 𝑑 𝑒 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀
  • πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑉 𝑑 ⊀
  • 𝑂 ~ 𝑑 ​ 𝑉 𝑑 , βŸ‚ ⊀
  • πœ‚ ​ 𝐸 𝑑 𝑑 ,

(53b)

where the error terms satisfy

β€– | 𝐸 𝑑 π‘Ž | β€–

≀ 2 ​ 𝑐 3 ​ πœ… βˆ’ 4 ​ β€– 𝑋 ⋆ β€– β‹… β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | ​ β€– + 2 β€– ​ | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | β€– ,

(54a)

β€– | 𝐸 𝑑 𝑏 | β€–

≀ ( β€– 𝑂 ~ 𝑑 β€– 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ) 3 / 4 ​ 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ≀ 1 20 ​ πœ… βˆ’ 10 ​ 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ,

(54b)

β€– | 𝐸 𝑑 𝑐 | β€–

≀ πœ… βˆ’ 6 ​ β€– 𝑋 ⋆ β€– β‹… β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€– ,

(54c)

β€– | 𝐸 𝑑 𝑑 | β€–

≀ ( β€– 𝑂 ~ 𝑑 β€– 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ) 3 / 4 ​ 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ,

(54d)

β€– | 𝐸 𝑑 𝑒 | β€–

≀ 2 ​ β€– | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | β€– + 𝑐 12 ​ πœ… βˆ’ 6 ​ β€– 𝑋 ⋆ β€– β‹… β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€– .

(54e)

Moreover, we have

β€– 𝐸 𝑑 𝑏 β€– ≀ 1 24 ​ 𝐢 max ​ πœ… ​ β€– 𝑂 ~ 𝑑 β€– ,

(54f)

β€– 𝐸 𝑑 𝑑 β€– ≀ 1 24 ​ 𝐢 max ​ πœ… ​ β€– 𝑂 ~ 𝑑 β€– .

(54g)

Here, | | | β‹… | | | can either be the Frobenius norm or the spectral norm.

To proceed, we would need the approximate update equation of the rotated signal term 𝑆 ~ 𝑑 + 1 , and the rotated misalignment term 𝑁 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 βˆ’ 1 later in the proof. Since directly analyzing the evolution of these two terms seems challenging, we resort to two surrogate matrices 𝑆 𝑑 + 1 ​ 𝑉 𝑑 + 𝑆 𝑑 + 1 ​ 𝑉 𝑑 , βŸ‚ ​ 𝑄 , and ( 𝑁 𝑑 + 1 ​ 𝑉 𝑑 + 𝑁 𝑑 + 1 ​ 𝑉 𝑑 , βŸ‚ ​ 𝑄 ) ​ ( 𝑆 𝑑 + 1 ​ 𝑉 𝑑 + 𝑆 𝑑 + 1 ​ 𝑉 𝑑 , βŸ‚ ​ 𝑄 ) βˆ’ 1 , as documented in the following two lemmas.

Lemma 14.

For any 𝑑 such that 𝑆 ~ 𝑑 is invertible and (22) holds, and any matrix 𝑄 ∈ ℝ ( π‘Ÿ βˆ’ π‘Ÿ ⋆ ) Γ— π‘Ÿ ⋆ with β€– 𝑄 β€– ≀ 2 , we have

𝑆 𝑑 + 1 ​ 𝑉 𝑑 + 𝑆 𝑑 + 1 ​ 𝑉 𝑑 , βŸ‚ ​ 𝑄

( 𝐼 + πœ‚ ​ 𝐸 𝑑 14 ) ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ) ​ 𝑆 ~ 𝑑 ,

(55)

where 𝐸 𝑑 14 ∈ ℝ π‘Ÿ ⋆ Γ— π‘Ÿ ⋆ is a matrix (depending on 𝑄 ) satisfying

β€– 𝐸 𝑑 14 β€– ≀ 1 200 ​ ( 𝐢 3 . π‘Ž + 1 ) 4 ​ πœ… 6 .

Here, 𝐢 3 . π‘Ž

0 is given in Lemma 3.

Lemma 15.

For any 𝑑 such that 𝑆 ~ 𝑑 is invertible and (22) holds, and any matrix 𝑄 ∈ ℝ ( π‘Ÿ βˆ’ π‘Ÿ ⋆ ) Γ— π‘Ÿ ⋆ with β€– 𝑄 β€– ≀ 2 , we have

( 𝑁 𝑑 + 1 ​ 𝑉 𝑑 + 𝑁 𝑑 + 1 ​ 𝑉 𝑑 , βŸ‚ ​ 𝑄 ) ​ ( 𝑆 𝑑 + 1 ​ 𝑉 𝑑 + 𝑆 𝑑 + 1 ​ 𝑉 𝑑 , βŸ‚ ​ 𝑄 ) βˆ’ 1

𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ ( 1 + πœ‚ ​ 𝐸 𝑑 15 . π‘Ž ) ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 + πœ‚ ​ Ξ£ ⋆ 2 ) βˆ’ 1 ​ ( 1 + πœ‚ ​ 𝐸 𝑑 14 ) βˆ’ 1 + πœ‚ ​ 𝐸 𝑑 15 . 𝑏

where 𝐸 𝑑 15 . π‘Ž , 𝐸 𝑑 15 . 𝑏 are matrices (depending on 𝑄 ) satisfying

β€– 𝐸 𝑑 15 . π‘Ž β€–

≀ 1 200 ​ ( 𝐢 3 . π‘Ž + 1 ) 4 ​ πœ… 6 ,

(56a)

β€– | 𝐸 𝑑 15 . 𝑏 | β€–

≀ 400 ​ 𝑐 πœ† βˆ’ 1 ​ πœ… 6 ​ β€– 𝑋 ⋆ β€– βˆ’ 2 ​ β€– | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | ​ β€– + 1 64 ​ ( 𝐢 3 . π‘Ž + 1 ) 2 ​ πœ… 5 ​ β€– 𝑋 ⋆ β€– β€– ​ | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€–

  • 1 64 ​ ( β€– 𝑂 ~ 𝑑 β€– 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ) 2 / 3 .

(56b)

Here, | | | β‹… | | | can either be the Frobenius norm or the spectral norm, and 𝐢 3 . π‘Ž

0 is given in Lemma 3.

B.1Proof of Lemma 13

We split the proof into three steps: (1) provide several useful approximation results regarding the matrix inverses utilizing the facts that β€– 𝑂 ~ 𝑑 β€– and β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– are small (as shown by Lemma 3); (2) proving the claims (53a), (54a), (54b), and (54f) associated with the signal term 𝑆 𝑑 + 1 ; (3) proving the claims (53b), (54c), (54d), (54e), and (54g) associated with the noise term 𝑁 𝑑 + 1 . Note that our approximation results in step (1) include choices of some matrices { 𝑄 𝑖 } with small spectral norms, whose choices may be different from lemma to lemma for simplicity of presentation;

B.1.1Step 1: preliminaries

We know from (22) that the overparametrization error 𝑂 ~ 𝑑 is negligible compared to the signals 𝑆 ~ 𝑑 and 𝜎 min ​ ( 𝑋 ⋆ ) . This combined with the decomposition (19) reveals a desired approximation ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 β‰ˆ ( 𝑉 𝑑 ​ ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 ) ​ 𝑉 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 . This approximation is formalized in the lemma below.

Lemma 16.

If πœ† β‰₯ 4 ​ ( β€– 𝑂 ~ 𝑑 β€– 2 ∨ 2 ​ β€– 𝑁 ~ 𝑑 β€– ​ β€– 𝑂 ~ 𝑑 β€– ) for some 𝑑 , then

( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1

( 𝑉 𝑑 ​ ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 ) ​ 𝑉 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1

+ ( 𝑉 𝑑 ​ ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 ) ​ 𝑉 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝐸 𝑑 16 . π‘Ž ​ ( 𝑉 𝑑 ​ ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 ) ​ 𝑉 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1

( 𝑉 𝑑 ​ ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 ) ​ 𝑉 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ​ ( 𝐼 + 𝐸 𝑑 16 . 𝑏 )

(57)

where the error terms 𝐸 𝑑 16 . π‘Ž , 𝐸 𝑑 16 . 𝑏 can be expressed as

𝐸 𝑑 16 . π‘Ž

( 𝑉 𝑑 , βŸ‚ ​ 𝑂 ~ 𝑑 ⊀ ​ 𝑂 ~ 𝑑 ​ 𝑉 𝑑 , βŸ‚ ⊀ + 𝑉 𝑑 ​ 𝑁 ~ 𝑑 ⊀ ​ 𝑂 ~ 𝑑 ​ 𝑉 𝑑 , βŸ‚ ⊀ + 𝑉 𝑑 , βŸ‚ ​ 𝑂 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 ​ 𝑉 𝑑 ⊀ ) ​ 𝑄 1 ,

(58a)

𝐸 𝑑 16 . 𝑏

πœ† βˆ’ 1 ​ 𝐸 𝑑 16 . π‘Ž ​ 𝑄 2 ,

(58b)

for some matrices 𝑄 1 , 𝑄 2 such that max ⁑ { β€– 𝑄 1 β€– , β€– 𝑄 2 β€– } ≀ 2 .

Proof.

Expanding 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 according to (19), we have

𝑋 𝑑 ⊀ ​ 𝑋 𝑑

𝑉 𝑑 ​ ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 ) ​ 𝑉 𝑑 ⊀ + 𝑉 𝑑 , βŸ‚ ​ 𝑂 ~ 𝑑 ⊀ ​ 𝑂 ~ 𝑑 ​ 𝑉 𝑑 , βŸ‚ ⊀ + 𝑉 𝑑 ​ 𝑁 ~ 𝑑 ⊀ ​ 𝑂 ~ 𝑑 ​ 𝑉 𝑑 , βŸ‚ ⊀ + 𝑉 𝑑 , βŸ‚ ​ 𝑂 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 ​ 𝑉 𝑑 ⊀ .

The conclusion readily follows from Lemma 9 by setting therein 𝐴

𝑉 𝑑 ​ ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 ) ​ 𝑉 𝑑 ⊀ + πœ† ​ 𝐼 and 𝐡

𝑉 𝑑 , βŸ‚ ​ 𝑂 ~ 𝑑 ⊀ ​ 𝑂 ~ 𝑑 ​ 𝑉 𝑑 , βŸ‚ ⊀ + 𝑉 𝑑 ​ 𝑁 ~ 𝑑 ⊀ ​ 𝑂 ~ 𝑑 ​ 𝑉 𝑑 , βŸ‚ ⊀ + 𝑉 𝑑 , βŸ‚ ​ 𝑂 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 ​ 𝑉 𝑑 ⊀ , where the condition β€– 𝐴 βˆ’ 1 ​ 𝐡 β€– ≀ 1 / 2 is satisfied since

β€– 𝐴 βˆ’ 1 ​ 𝐡 β€– ≀ 𝜎 min ​ ( 𝐴 ) βˆ’ 1 ​ β€– 𝐡 β€– ≀ πœ† βˆ’ 1 β‹… ( β€– 𝑂 ~ 𝑑 β€– 2 + 2 ​ β€– 𝑂 ~ 𝑑 β€– ​ β€– 𝑁 ~ 𝑑 β€– ) ≀ 1 / 2 .

∎

Moreover, the dominating term on the right hand side of (57) can be equivalently written as

( 𝑉 𝑑 ​ ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 ) ​ 𝑉 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1

( 𝑉 𝑑 ​ ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 + πœ† ​ 𝐼 ) ​ 𝑉 𝑑 ⊀ + πœ† ​ 𝑉 𝑑 , βŸ‚ ​ 𝑉 𝑑 , βŸ‚ ⊀ ) βˆ’ 1

𝑉 𝑑 ​ ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑉 𝑑 ⊀ + πœ† βˆ’ 1 ​ 𝑉 𝑑 , βŸ‚ ​ 𝑉 𝑑 , βŸ‚ ⊀ .

(59)

When the misalignment error β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– is small, we expect ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 β‰ˆ ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 , which is formalized in the following lemma that establishes ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + 𝑆 ~ 𝑑 ​ 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 + πœ† ​ 𝐼 ) βˆ’ 1 β‰ˆ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 , due to the following approximation

( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1

𝑆 ~ 𝑑 βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + 𝑆 ~ 𝑑 ​ 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑

β‰ˆ 𝑆 ~ 𝑑 βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑

( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 .

Lemma 17.

If β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– ≀ 𝜎 min ​ ( 𝑋 ⋆ ) / 16 for some 𝑑 , then

( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + 𝑆 ~ 𝑑 ​ 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 + πœ† ​ 𝐼 ) βˆ’ 1

( 𝐼 + 𝐸 𝑑 17 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ,

(60)

where the error term 𝐸 𝑑 17 is a matrix defined as

𝐸 𝑑 17

πœ… 2 ​ β€– 𝑋 ⋆ β€– βˆ’ 2 ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– ​ 𝑄 1 ​ ( 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ ) ​ 𝑄 2 ,

(61)

where 𝑄 1 , 𝑄 2 are matrices of appropriate dimensions satisfying β€– 𝑄 1 β€– ≀ 1 , β€– 𝑄 2 β€– ≀ 2 . In particular, we have

β€– | 𝐸 𝑑 17 | β€– ≀ 2 ​ πœ… 2 ​ β€– 𝑋 ⋆ β€– βˆ’ 2 ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– β‹… β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€– ,

(62)

where | | | β‹… | | | can be either the operator norm or the Frobenius norm.

Proof.

In order to apply Lemma 9, setting 𝐴

𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 and 𝐡

𝑆 ~ 𝑑 ​ 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 , it is straightforward to verify that

β€– 𝐴 βˆ’ 1 ​ 𝐡 β€–

β€– ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 β€– ≀ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 β€– 2 ≀ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– 2 ​ β€– Ξ£ ⋆ βˆ’ 1 β€– 2 ≀ ( 1 / 16 ) 2 ,

where we use the obvious fact that β€– ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ β€– ≀ 1 . Applying Lemma 9, we obtain

( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + 𝑆 ~ 𝑑 ​ 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 + πœ† ​ 𝐼 ) βˆ’ 1 βˆ’ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1

( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ 𝑄 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1

( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ ​ Ξ£ ⋆ βˆ’ 1 ​ ( 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ ) ⊀ ​ ( 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ ) ​ Ξ£ ⋆ βˆ’ 1 ​ 𝑄 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1

for some matrix 𝑄 with β€– 𝑄 β€– ≀ 2 . Since one may further write

( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + 𝑆 ~ 𝑑 ​ 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 + πœ† ​ 𝐼 ) βˆ’ 1 βˆ’ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1

β€– Ξ£ ⋆ βˆ’ 1 β€– 2 ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ ​ Ξ£ ⋆ βˆ’ 1 β€– Ξ£ ⋆ βˆ’ 1 β€– ​ ( 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ ) ⊀ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– ​ ( 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ ) ​ Ξ£ ⋆ βˆ’ 1 β€– Ξ£ ⋆ βˆ’ 1 β€– ​ 𝑄 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ,

the conclusion follows by setting 𝐸 𝑑 17 as in (61) with

𝑄 1

( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ ​ Ξ£ ⋆ βˆ’ 1 β€– Ξ£ ⋆ βˆ’ 1 β€– ​ ( 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ ) ⊀ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– , 𝑄 2

Ξ£ ⋆ βˆ’ 1 β€– Ξ£ ⋆ βˆ’ 1 β€– ​ 𝑄 .

The last inequality (62) is then a direct consequence of (61). ∎

B.1.2Step 2: a key recursion

Recall the definition Ξ” 𝑑 in (48), we can rewrite the update equation (7) as

𝑋 𝑑 + 1

𝑋 𝑑 βˆ’ πœ‚ ​ ( 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ ) ​ 𝑋 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 + πœ‚ ​ Ξ” 𝑑 ​ 𝑋 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 .

(63)

Multiplying both sides of (63) by π‘ˆ ⋆ ⊀ on the left, we obtain

𝑆 𝑑 + 1

𝑆 𝑑 βˆ’ πœ‚ ​ 𝑆 𝑑 ​ 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 + πœ‚ ​ Ξ£ ⋆ 2 ​ 𝑆 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 + πœ‚ ​ π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 ​ 𝑋 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1

( 1 βˆ’ πœ‚ ) ​ 𝑆 𝑑 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 + π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 ​ π‘ˆ ⋆ ) ​ 𝑆 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 + πœ‚ ​ π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 ​ π‘ˆ ⋆ , βŸ‚ ​ 𝑁 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 .

(64)

Similarly, multiplying both sides of (63) by π‘ˆ ⋆ , βŸ‚ ⊀ , we obtain

𝑁 𝑑 + 1

𝑁 𝑑 ​ ( 𝐼 βˆ’ πœ‚ ​ 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 ) + πœ‚ ​ π‘ˆ ⋆ , βŸ‚ ⊀ ​ Ξ” 𝑑 ​ 𝑋 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1

( 1 βˆ’ πœ‚ ) ​ 𝑁 𝑑 + πœ‚ ​ πœ† ​ 𝑁 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 + πœ‚ ​ π‘ˆ ⋆ , βŸ‚ ⊀ ​ Ξ” 𝑑 ​ π‘ˆ ⋆ ​ 𝑆 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 + πœ‚ ​ π‘ˆ ⋆ , βŸ‚ ⊀ ​ Ξ” 𝑑 ​ π‘ˆ ⋆ , βŸ‚ ​ 𝑁 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 .

(65)

These expressions motivate the need to study the terms 𝑆 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 and 𝑁 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 , which we formalize in the following lemma.

Lemma 18.

Under the same setting as Lemma 13, we have

𝑆 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1

( 𝐼 + 𝐸 𝑑 17 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑉 𝑑 ⊀ + 𝐸 𝑑 18 . π‘Ž ,

(66a)

𝑁 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1

𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ ( 𝐼 + 𝐸 𝑑 17 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑉 𝑑 ⊀ + πœ† βˆ’ 1 ​ 𝑂 ~ 𝑑 ​ 𝑉 𝑑 , βŸ‚ ⊀ + 𝐸 𝑑 18 . 𝑏 ,

(66b)

where 𝐸 𝑑 17 is given in (61), and the error terms 𝐸 𝑑 18 . π‘Ž , 𝐸 𝑑 18 . 𝑏 can be expressed as

𝐸 𝑑 18 . π‘Ž

πœ… ​ πœ† βˆ’ 1 ​ β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– 𝑂 ~ 𝑑 β€– ​ 𝑄 1 ​ ( 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ ) ⊀ ​ 𝑄 2 ,

(67a)

𝐸 𝑑 18 . 𝑏

( 𝑁 ~ 𝑑 ​ ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑉 𝑑 ⊀ + πœ† βˆ’ 1 ​ 𝑂 ~ 𝑑 ​ 𝑉 𝑑 , βŸ‚ ⊀ ) ​ 𝐸 𝑑 16 . 𝑏

πœ† βˆ’ 1 ​ ( β€– 𝑁 ~ 𝑑 β€– ​ 𝑄 3 + β€– 𝑂 ~ 𝑑 β€– ​ 𝑄 4 ) ​ 𝐸 𝑑 16 . 𝑏 .

(67b)

for some matrices { 𝑄 𝑖 } 1 ≀ 𝑖 ≀ 4 with spectral norm bounded by 2 , and 𝐸 𝑑 16 . 𝑏 defined in (58b).

Proof.

To begin, combining Lemma 16 and the discussion thereafter (cf. (57)–(B.1.1)) and the fact that 𝑆 ~ 𝑑

𝑆 𝑑 ​ 𝑉 𝑑 , we have for some matrix 𝑄 with β€– 𝑄 β€– ≀ 2 that

𝑆 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1

𝑆 ~ 𝑑 ​ ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑉 𝑑 ⊀ ​ ( 𝐼 + 𝐸 𝑑 16 . 𝑏 )

𝑆 ~ 𝑑 ​ ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑉 𝑑 ⊀ + 𝑆 ~ 𝑑 ​ ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 ​ πœ† βˆ’ 1 ​ 𝑁 ~ 𝑑 ⊀ ​ 𝑂 ~ 𝑑 ​ 𝑄

( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + 𝑆 ~ 𝑑 ​ 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑉 𝑑 ⊀

  • 𝑆 ~ 𝑑 ​ ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑
  • 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑
  • πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 ⊀ ​ ( 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ) ⊀ ​ ( 𝑂 ~ 𝑑 / πœ† ) ​ 𝑄 .

(68)

Note that the condition of Lemma 16 can be verified as follows: since

β€– 𝑂 ~ 𝑑 β€–

≀ 𝐢 3 . 𝑏 βˆ’ 𝐢 3 . 𝑏 ​ πœ… βˆ’ 3 β‹… β€– 𝑋 ⋆ β€– β‹… 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ) β‹… β€– 𝑆 ~ 𝑑 β€– ≀ 𝐢 3 . 𝑏 βˆ’ 𝐢 3 . 𝑏 ​ 𝐢 3 . π‘Ž ​ 𝜎 min ​ ( 𝑋 ⋆ ) ,

β€– 𝑁 ~ 𝑑 β€–

≀ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– β‹… β€– Ξ£ ⋆ βˆ’ 1 β€– β‹… β€– 𝑆 ~ 𝑑 β€– ≀ 𝑐 3 ​ πœ… βˆ’ 𝐢 𝛿 / 2 ​ β€– 𝑋 ⋆ β€– β‹… 𝐢 3 . π‘Ž ​ πœ… 3 ​ β€– 𝑋 ⋆ β€– 𝜎 min ​ ( 𝑋 ⋆ ) ≀ 𝑐 3 ​ 𝐢 3 . π‘Ž ​ 𝜎 min ​ ( 𝑋 ⋆ )

provided 𝐢 𝛿 β‰₯ 6 , the bounds 𝑐 3 ≲ 𝑐 𝛿 / 𝑐 πœ† and 𝐢 3 . π‘Ž ≲ 𝑐 πœ† βˆ’ 1 / 2 imply that when we choose 𝐢 𝛼 to be large enough (depending on 𝑐 πœ† , 𝑐 𝛿 ),

2 ​ β€– 𝑁 ~ 𝑑 β€– ​ β€– 𝑂 ~ 𝑑 β€– ∨ β€– 𝑂 ~ 𝑑 β€– 2 ≀ πœ† / 4 ,

as desired.

Now the first term in (68) can be handled by invoking Lemma 17, since its condition is verified by β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– ≀ 𝑐 3 ​ πœ… βˆ’ ( 𝐢 𝛿 / 2 βˆ’ 1 ) ​ 𝜎 min ​ ( 𝑋 ⋆ ) ≀ 𝜎 min ​ ( 𝑋 ⋆ ) / 16 provided 𝐢 𝛿 β‰₯ 2 and 𝑐 3 ≀ 1 / 16 by choosing 𝑐 𝛿 sufficiently small (depending on 𝑐 πœ† ). Namely,

( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + 𝑆 ~ 𝑑 ​ 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑉 𝑑 ⊀

( 𝐼 + 𝐸 𝑑 17 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑉 𝑑 ⊀ .

For the second term, by noting that

β€– 𝑆 ~ 𝑑 ​ ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 ⊀ β€– ≀ β€– 𝑆 ~ 𝑑 ​ ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 ⊀ β€– ≀ 1 ,

it can be expressed as

πœ† βˆ’ 1 ​ β€– 𝑂 ~ 𝑑 β€– ​ 𝑆 ~ 𝑑 ​ ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 ⊀ ​ ( 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ) ⊀ ​ ( 𝑂 ~ 𝑑 / β€– 𝑂 ~ 𝑑 β€– ) ​ 𝑄

πœ… ​ πœ† βˆ’ 1 ​ β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– 𝑂 ~ 𝑑 β€– ​ 𝑄 1 ​ ( 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ ) ⊀ ​ 𝑄 2

for 𝑄 1

𝑆 ~ 𝑑 ​ ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 ⊀ β‹… πœ… βˆ’ 1 ​ β€– 𝑋 ⋆ β€– ​ Ξ£ ⋆ βˆ’ 1 with β€– 𝑄 1 β€– ≀ 1 and 𝑄 2

( 𝑂 ~ 𝑑 / β€– 𝑂 ~ 𝑑 β€– ) ​ 𝑄 which satisfies β€– 𝑄 2 β€– ≀ β€– 𝑄 β€– ≀ 2 . Applying the above two bounds to (68) yields (66a).

Similarly, moving to (66b), it follows that

𝑁 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1

( 𝑁 ~ 𝑑 ​ ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑉 𝑑 ⊀ + πœ† βˆ’ 1 ​ 𝑂 ~ 𝑑 ​ 𝑉 𝑑 , βŸ‚ ⊀ ) ​ ( 𝐼 + 𝐸 𝑑 16 . 𝑏 )

𝑁 ~ 𝑑 ​ ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑉 𝑑 ⊀ + πœ† βˆ’ 1 ​ 𝑂 ~ 𝑑 ​ 𝑉 𝑑 , βŸ‚ ⊀ + 𝐸 𝑑 18 . 𝑏 ,

(69)

where we have

𝐸 𝑑 18 . 𝑏

( 𝑁 ~ 𝑑 ​ ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑉 𝑑 ⊀ + πœ† βˆ’ 1 ​ 𝑂 ~ 𝑑 ​ 𝑉 𝑑 , βŸ‚ ⊀ ) ​ 𝐸 𝑑 16 . 𝑏

πœ† βˆ’ 1 ​ ( β€– 𝑁 ~ 𝑑 β€– ​ 𝑄 3 + β€– 𝑂 ~ 𝑑 β€– ​ 𝑄 4 ) ​ 𝐸 𝑑 16 . 𝑏

for some matrices 𝑄 3 , 𝑄 4 with β€– 𝑄 3 β€– , β€– 𝑄 4 β€– ≀ 1 . In the last line we used β€– ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 β€– ≀ πœ† βˆ’ 1 . For the first term of (69), we use Lemma 17 and obtain

𝑁 ~ 𝑑 ​ ( 𝑆 ~ 𝑑 ⊀ ​ 𝑆 ~ 𝑑 + 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑉 𝑑 ⊀

𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + 𝑆 ~ 𝑑 ​ 𝑁 ~ 𝑑 ⊀ ​ 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑉 𝑑 ⊀

𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ ( 𝐼 + 𝐸 𝑑 17 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑉 𝑑 ⊀ .

This yields the representation in (66b). ∎

B.1.3Step 3: proofs associated with 𝑆 𝑑 + 1 .

With the help of Lemma 18, we are ready to prove (53a) and the associated norm bounds (54a), (54b), and (54f). To begin with, we plug (66a), (66b) into (64) and use 𝑆 𝑑

𝑆 ~ 𝑑 ​ 𝑉 𝑑 ⊀ to obtain

𝑆 𝑑 + 1

( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 + 𝐸 𝑑 π‘Ž ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ) ​ 𝑆 ~ 𝑑 ​ 𝑉 𝑑 ⊀ + πœ‚ ​ 𝐸 𝑑 𝑏 ,

where the error terms 𝐸 𝑑 π‘Ž and 𝐸 𝑑 𝑏 are

𝐸 𝑑 π‘Ž

≔ π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 ​ π‘ˆ ⋆ + ( Ξ£ ⋆ 2 + π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 ​ π‘ˆ ⋆ + πœ† ​ 𝐼 ) ​ 𝐸 𝑑 17 + π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 ​ π‘ˆ ⋆ , βŸ‚ ​ 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ ( 𝐼 + 𝐸 𝑑 17 ) ,

𝐸 𝑑 𝑏

≔ ( Ξ£ ⋆ 2 + π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 ​ π‘ˆ ⋆ + πœ† ​ 𝐼 ) ​ 𝐸 𝑑 18 . π‘Ž + π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 ​ π‘ˆ ⋆ , βŸ‚ ​ ( πœ† βˆ’ 1 ​ 𝑂 ~ 𝑑 ​ 𝑉 𝑑 , βŸ‚ ⊀ + 𝐸 𝑑 18 . 𝑏 ) .

This establishes the identity (53a). To control β€– | 𝐸 𝑑 π‘Ž | β€– , we observe that

β€– | 𝐸 𝑑 π‘Ž | β€–

≀ β€– | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | β€– + β€– Ξ£ ⋆ 2 + π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 ​ π‘ˆ ⋆ + πœ† ​ 𝐼 β€– β‹… β€– | 𝐸 𝑑 17 | β€– + β€– | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | β€– β‹… β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– β‹… β€– Ξ£ ⋆ βˆ’ 1 β€– β‹… ( 1 + β€– 𝐸 𝑑 17 β€– )

≀ ( 1 + 𝑐 12 ​ πœ… βˆ’ 2 ​ 𝐢 𝛿 / 3 + 𝑐 πœ† ) ​ β€– 𝑋 ⋆ β€– 2 β‹… β€– | 𝐸 𝑑 17 | β€– + β€– | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | β€– + 𝑐 3 ​ πœ… βˆ’ 𝐢 𝛿 / 2 ​ β€– 𝑋 ⋆ β€– β‹… 𝜎 min βˆ’ 1 ​ ( 𝑋 ⋆ ) β‹… ( 1 + β€– 𝐸 𝑑 17 β€– ) β‹… β€– | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | β€–

≀ 2 ​ β€– 𝑋 ⋆ β€– 2 β‹… β€– | 𝐸 𝑑 17 | β€– + ( 1 + 𝑐 3 ​ ( 1 + β€– 𝐸 𝑑 17 β€– ) ) ​ β€– | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | β€– ,

where the second line follows from Lemma 12 and Equations (12b), (22c); the last line holds since 𝑐 12 , 𝑐 πœ† are sufficiently small and 𝐢 𝛿 is sufficiently large. Now we invoke the bound (62) in Lemma 17 to see

β€– | 𝐸 𝑑 17 | β€– ≀ 2 ​ πœ… 2 ​ β€– 𝑋 ⋆ β€– βˆ’ 2 ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– ​ β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€–

≀ 2 ​ 𝑐 3 ​ πœ… 2 ​ πœ… βˆ’ 𝐢 𝛿 / 2 ​ β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€–

≀ 2 ​ 𝑐 3 ​ πœ… βˆ’ 6 ​ β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€– ,

where the last line follows again by choosing sufficiently large 𝐢 𝛿 . Furthermore, since β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– ≀ 𝑐 3 ​ πœ… βˆ’ 𝐢 𝛿 / 2 ​ β€– 𝑋 ⋆ β€– for small enough 𝑐 3 , we obtain β€– 𝐸 𝑑 17 β€– ≀ 1 . Combining these inequalities yields the claimed bound

β€– | 𝐸 𝑑 π‘Ž | β€– ≀ 2 ​ 𝑐 3 ​ πœ… βˆ’ 4 ​ β€– 𝑋 ⋆ β€– β‹… β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | ​ β€– + 2 β€– ​ | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | β€– .

The bound of β€– | 𝐸 𝑑 𝑏 | β€– and β€– 𝐸 𝑑 𝑏 β€– can be proved in a similar way, utilizing the bound for β€– 𝑂 ~ 𝑑 β€– in (24). In fact, a computation similar to the above shows

β€– | 𝐸 𝑑 𝑏 | β€–

≀ 2 ​ β€– 𝑋 ⋆ β€– 2 β‹… β€– | 𝐸 18 . π‘Ž | β€– + πœ† βˆ’ 1 ​ β€– Ξ” 𝑑 β€– β‹… β€– | 𝑂 ~ 𝑑 | β€– + β€– Ξ” 𝑑 β€– β‹… β€– | 𝐸 18 . 𝑏 | β€–

≀ 2 ​ πœ… ​ πœ† βˆ’ 1 β‹… β€– 𝑋 ⋆ β€– β‹… β€– 𝑂 ~ 𝑑 β€– β‹… β€– 𝑄 1 β€– β‹… β€– 𝑄 2 β€– β‹… β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€– + 100 ​ 𝑐 πœ† βˆ’ 1 ​ 𝜎 min βˆ’ 1 ​ ( 𝑀 ⋆ ) ​ 𝑐 12 ​ πœ… βˆ’ 2 ​ 𝐢 𝛿 / 3 ​ β€– 𝑋 ⋆ β€– 2 β‹… β€– | 𝑂 ~ 𝑑 | β€–

  • 8 ​ πœ† βˆ’ 2 ​ 𝑐 12 ​ πœ… βˆ’ 2 ​ 𝐢 𝛿 / 3 ​ ( β€– 𝑁 ~ 𝑑 β€–
  • β€– 𝑂 ~ 𝑑 β€– ) ​ β€– 𝑁 ~ 𝑑 β€– β‹… β€– | 𝑂 ~ 𝑑 | β€–

≀ 800 ​ πœ… 7 ​ 𝑐 πœ† βˆ’ 1 ​ β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– 𝑂 ~ 𝑑 β€– β‹… β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | ​ β€– + 1 48 ​ ( 𝐢 max + 1 ) ​ πœ… β€– ​ | 𝑂 ~ 𝑑 | β€– .

Here, 𝐢 max is the constant given by Lemma 3. Similarly, we have

β€– 𝐸 𝑑 𝑏 β€– ≀ 800 ​ πœ… 7 ​ 𝑐 πœ† βˆ’ 1 ​ β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– 𝑂 ~ 𝑑 β€– β‹… β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– + 1 48 ​ ( 𝐢 max + 1 ) ​ πœ… ​ β€– 𝑂 ~ 𝑑 β€– .

The bound (54f) now follows directly from the bound of β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– in Lemma 3, provided 𝑐 𝛿 is sufficiently small and 𝐢 𝛿 is sufficiently large. To prove (54b), we note that

β€– | 𝐴 | β€– ≀ 𝑛 ​ β€– 𝐴 β€–

(70)

for any unitarily invariant norm | | | β‹… | | | and real matrix 𝐴 ∈ ℝ 𝑝 Γ— π‘ž with 𝑝 ∨ π‘ž ≀ 𝑛 (which can be easily verified when | | | β‹… | | |

βˆ₯ β‹… βˆ₯ or βˆ₯ β‹… βˆ₯ π–₯ ). Thus

β€– | 𝐸 𝑑 𝑏 | β€– ≀ ( 800 ​ πœ… 7 ​ 𝑐 πœ† βˆ’ 1 ​ 𝑐 3 ​ πœ… βˆ’ 𝐢 𝛿 / 2 + 1 24 ​ ( 𝐢 max + 1 ) ​ πœ… ) ​ 𝑛 ​ β€– 𝑂 ~ 𝑑 β€– ≀ ( β€– 𝑂 ~ 𝑑 β€– 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ) 3 / 4 ​ 𝜎 min ​ ( 𝑆 ~ 𝑑 )

(71)

where the last inequality follows from the control of β€– 𝑂 ~ 𝑑 β€– given by (3) provided 𝑐 3 is sufficiently small and 𝐢 3 . 𝑏 therein is sufficiently large. This establishes the first inequality in (54b), and the second inequality therein follows directly from (3).

B.1.4Step 4: proofs associated with 𝑁 ~ 𝑑 + 1 .

Now we move on to prove the identity (53b), and the norm controls (54c), (54d), (54e), and (54g) associated with the misalignment term 𝑁 ~ 𝑑 + 1 . Plugging (66a), (66b) into (65) and using the decomposition 𝑁 𝑑

𝑁 ~ 𝑑 ​ 𝑉 𝑑 ⊀ + 𝑂 ~ 𝑑 ​ 𝑉 𝑑 , βŸ‚ ⊀ , we have

𝑁 𝑑 + 1

𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 + πœ‚ ​ 𝐸 𝑑 𝑐 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑉 𝑑 ⊀

  • πœ‚ ​ 𝐸 𝑑 𝑒 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀
  • πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑉 𝑑 ⊀
  • 𝑂 ~ 𝑑 ​ 𝑉 𝑑 , βŸ‚ ⊀
  • πœ‚ ​ 𝐸 𝑑 𝑑 ,

where the error terms are defined to be

𝐸 𝑑 𝑐

≔ πœ† ​ 𝐸 𝑑 17 ,

𝐸 𝑑 𝑑

≔ ( πœ† ​ 𝐼 + π‘ˆ ⋆ , βŸ‚ ⊀ ​ Ξ” 𝑑 ​ π‘ˆ ⋆ , βŸ‚ ) ​ 𝐸 𝑑 18 . 𝑏 + πœ† βˆ’ 1 ​ π‘ˆ ⋆ , βŸ‚ ⊀ ​ Ξ” 𝑑 ​ π‘ˆ ⋆ , βŸ‚ ​ 𝑂 ~ 𝑑 ​ 𝑉 𝑑 , βŸ‚ ⊀ + π‘ˆ ⋆ , βŸ‚ ⊀ ​ Ξ” 𝑑 ​ π‘ˆ ⋆ ​ 𝐸 𝑑 18 . π‘Ž ,

𝐸 𝑑 𝑒

≔ π‘ˆ ⋆ , βŸ‚ ⊀ ​ Ξ” 𝑑 ​ π‘ˆ ⋆ ​ ( 𝐼 + 𝐸 𝑑 17 ) + π‘ˆ ⋆ , βŸ‚ ⊀ ​ Ξ” 𝑑 ​ π‘ˆ ⋆ , βŸ‚ ​ 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ ( 𝐼 + 𝐸 𝑑 17 ) .

This establishes the decomposition (53b). The remaining norm controls follow from the expressions above and similar computation as we have done for 𝑆 𝑑 + 1 . For the sake of brevity, we omit the details.

B.2Proof of Lemma 14

Use the identity (53a) in Lemma 13 and the fact that 𝑉 𝑑 and 𝑉 𝑑 , βŸ‚ have orthogonal columns to obtain

𝑆 𝑑 + 1 ​ 𝑉 𝑑 + 𝑆 𝑑 + 1 ​ 𝑉 𝑑 , βŸ‚ ​ 𝑄

( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 + 𝐸 𝑑 π‘Ž ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ) ​ 𝑆 ~ 𝑑 + πœ‚ ​ 𝐸 𝑑 𝑏 ​ ( 𝑉 𝑑 + 𝑉 𝑑 , βŸ‚ ​ 𝑄 )

( 𝐼 + πœ‚ ​ 𝐸 𝑑 14 ) ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ) ​ 𝑆 ~ 𝑑

( 𝐼 + πœ‚ ​ 𝐸 𝑑 14 ) ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 + πœ‚ ​ Ξ£ ⋆ 2 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 ,

(72)

where 𝐸 𝑑 14 is defined to be

𝐸 𝑑 14
≔ ( 𝐸 𝑑 π‘Ž ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 + 𝐸 𝑑 𝑏 ​ ( 𝑉 𝑑 + 𝑉 𝑑 , βŸ‚ ​ 𝑄 ) ​ 𝑆 ~ 𝑑 βˆ’ 1 ) ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ) βˆ’ 1

𝐸 𝑑 π‘Ž ​ ( ( 1 βˆ’ πœ‚ ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) ) βˆ’ 1

+ 𝐸 𝑑 𝑏 ​ ( 𝑉 𝑑 + 𝑉 𝑑 , βŸ‚ ​ 𝑄 ) ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ) βˆ’ 1

: 𝑇 1 + 𝑇 2 ,

where the invertibility of 𝑆 ~ 𝑑 follows from Lemma 3, and the invertibility of ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 follows from (113).

Since ( 1 βˆ’ πœ‚ ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βͺ° πœ† ​ 𝐼 and πœ† β‰₯ 1 100 ​ 𝑐 πœ† ​ 𝜎 min ​ ( 𝑀 ⋆ ) by (12b), we have

β€– 𝑇 1 β€– ≀ πœ† βˆ’ 1 ​ β€– 𝐸 𝑑 π‘Ž β€– ≀ 100 ​ 𝑐 πœ† βˆ’ 1 ​ 𝜎 min βˆ’ 1 ​ ( 𝑀 ⋆ ) ​ β€– 𝐸 𝑑 π‘Ž β€– .

In view of the bound (54a) on β€– 𝐸 𝑑 π‘Ž β€– in Lemma 13, we further have

β€– 𝑇 1 β€–

≀ 100 ​ 𝑐 πœ† βˆ’ 1 ​ 𝜎 min βˆ’ 2 ​ ( 𝑋 ⋆ ) ​ ( πœ… βˆ’ 4 ​ β€– 𝑋 ⋆ β€– β‹… β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– + β€– Ξ” 𝑑 β€– )

≀ 100 ​ 𝑐 πœ† βˆ’ 1 ​ πœ… 2 ​ β€– 𝑋 ⋆ β€– βˆ’ 2 ​ ( πœ… βˆ’ 4 ​ 𝑐 3 ​ πœ… βˆ’ 𝐢 𝛿 / 2 + 𝑐 12 ​ πœ… βˆ’ 2 ​ 𝐢 𝛿 / 3 ) ​ β€– 𝑋 ⋆ β€– 2

≀ 1 400 ​ ( 𝐢 3 . π‘Ž + 1 ) 4 ​ πœ… 5 ,

where the second inequality follows from (22c) in Lemma 3 and Lemma 12, and the last inequality holds as long as 𝑐 3 and 𝑐 12 are sufficiently small and 𝐢 𝛿 is sufficiently large (by first fixing 𝑐 πœ† and then choosing 𝑐 𝛿 to be sufficiently small).

The term 𝑇 2 can be controlled in a similar way. Since β€– 𝐴 ​ 𝐡 β€– ≀ β€– 𝐴 β€– β‹… β€– 𝐡 β€– , one has

β€– 𝑇 2 β€–

≀ β€– 𝐸 𝑑 𝑏 β€– β‹… ( β€– 𝑉 𝑑 β€– + β€– 𝑉 𝑑 , βŸ‚ β€– ​ β€– 𝑄 β€– ) β‹… β€– 𝑆 ~ 𝑑 βˆ’ 1 β€– β‹… 𝜎 min βˆ’ 1 ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 )

≀ ( i ) 3 ​ β€– 𝐸 𝑑 𝑏 β€– β‹… 𝜎 min βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ) β‹… πœ… 1 βˆ’ πœ‚ ≀ ( ii ) 6 ​ πœ… ​ ( β€– 𝑂 ~ 𝑑 β€– 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ) 3 / 4 ≀ ( iii ) 1 400 ​ ( 𝐢 3 . π‘Ž + 1 ) 4 ​ πœ… 5 .

Here, (i) follows from the bound (113) and the facts that β€– 𝑉 𝑑 β€– ∨ β€– 𝑉 𝑑 , βŸ‚ β€– ≀ 1 , β€– 𝑄 β€– ≀ 2 ; (ii) arises from the control (54b) on β€– 𝐸 𝑑 𝑏 β€– in Lemma 13 as well as the condition πœ‚ ≀ 𝑐 πœ‚ ≀ 1 / 2 ; and (iii) follows from the implication (3) of Lemma 3.

The proof is completed by summing up the bounds on β€– 𝑇 1 β€– and β€– 𝑇 2 β€– .

B.3Proof of Lemma 15

Similar to the proof of Lemma 14, we can use the identity (53b) in Lemma 13 and the fact that 𝑉 𝑑 and 𝑉 𝑑 , βŸ‚ have orthogonal columns to obtain

𝑁 𝑑 + 1 ​ 𝑉 𝑑 + 𝑁 𝑑 + 1 ​ 𝑉 𝑑 , βŸ‚ ​ 𝑄

𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 + πœ‚ ​ 𝐸 𝑑 𝑐 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 + πœ‚ ​ 𝐸 𝑑 15 . 𝑐

𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ ( 𝐼 + πœ‚ ​ 𝐸 𝑑 15 . π‘Ž ) ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 + πœ‚ ​ 𝐸 𝑑 15 . 𝑐 ,

(73)

where the error terms are defined to be

𝐸 𝑑 15 . 𝑐

≔ 𝐸 𝑑 𝑒 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 + πœ‚ βˆ’ 1 ​ 𝑂 ~ 𝑑 ​ 𝑄 + 𝐸 𝑑 𝑑 ​ ( 𝑉 𝑑 + 𝑉 𝑑 , βŸ‚ ​ 𝑄 ) ,

(74)

𝐸 𝑑 15 . π‘Ž

≔ 𝐸 𝑑 𝑐 ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 .

(75)

Combine (73) and (72) to arrive at

( 𝑁 𝑑 + 1 ​ 𝑉 𝑑 + 𝑁 𝑑 + 1 ​ 𝑉 𝑑 , βŸ‚ ​ 𝑄 ) ​ ( 𝑆 𝑑 + 1 ​ 𝑉 𝑑 + 𝑆 𝑑 + 1 ​ 𝑉 𝑑 , βŸ‚ ​ 𝑄 ) βˆ’ 1

𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ ( 𝐼 + πœ‚ ​ 𝐸 𝑑 15 . π‘Ž ) ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 + πœ‚ ​ Ξ£ ⋆ 2 ) βˆ’ 1 ​ ( 𝐼 + πœ‚ ​ 𝐸 𝑑 14 ) βˆ’ 1 + πœ‚ ​ 𝐸 𝑑 15 . 𝑏 ,

(76)

where, using

( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 + πœ‚ ​ Ξ£ ⋆ 2 ) βˆ’ 1

( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ) βˆ’ 1 ,

we have

𝐸 𝑑 15 . 𝑏
≔ 𝐸 𝑑 15 . 𝑐 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 + πœ‚ ​ Ξ£ ⋆ 2 ) βˆ’ 1 ​ ( 𝐼 + πœ‚ ​ 𝐸 𝑑 14 ) βˆ’ 1

𝐸 𝑑 𝑒 ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 + πœ‚ ​ Ξ£ ⋆ 2 ) βˆ’ 1 ​ ( 𝐼 + πœ‚ ​ 𝐸 𝑑 14 ) βˆ’ 1

  • πœ‚ βˆ’ 1 ​ 𝑂 ~ 𝑑 ​ 𝑄 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝐼

  • πœ‚ ​ ( Ξ£ ⋆ 2

  • πœ† ​ 𝐼 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀

  • πœ† ​ 𝐼 ) βˆ’ 1 ) βˆ’ 1 ​ ( 𝐼

  • πœ‚ ​ 𝐸 𝑑 14 ) βˆ’ 1

  • 𝐸 𝑑 𝑑 ​ ( 𝑉 𝑑

  • 𝑉 𝑑 , βŸ‚ ​ 𝑄 ) ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝐼

  • πœ‚ ​ ( Ξ£ ⋆ 2

  • πœ† ​ 𝐼 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀

  • πœ† ​ 𝐼 ) βˆ’ 1 ) βˆ’ 1 ​ ( 𝐼

  • πœ‚ ​ 𝐸 𝑑 14 ) βˆ’ 1

≕ 𝑇 1 + 𝑇 2 + 𝑇 3 .

It remains to bound β€– 𝐸 15 . π‘Ž β€– and β€– | 𝐸 15 . 𝑏 | β€– . By (54c), we have

β€– 𝐸 15 . π‘Ž β€– ≀ πœ† βˆ’ 1 ​ β€– 𝐸 𝑑 𝑐 β€–

≀ 100 ​ 𝑐 πœ† βˆ’ 1 ​ 𝜎 min βˆ’ 2 ​ ( 𝑋 ⋆ ) β‹… πœ… βˆ’ 4 ​ β€– 𝑋 ⋆ β€– ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€–

≀ 100 ​ 𝑐 πœ† βˆ’ 1 ​ 𝑐 3 ​ πœ… βˆ’ 2 ​ πœ… βˆ’ 𝐢 𝛿 / 2

≀ 1 200 ​ ( 𝐢 3 . π‘Ž + 1 ) 4 ​ πœ… 5 ,

where the penultimate inequality follows from (22c) and the last inequality holds with the proviso that 𝑐 3 is sufficiently small and 𝐢 𝛿 is sufficiently large.

Now we move to bound β€– | 𝐸 15 . 𝑏 | β€– . To this end, the relation β€– ( 𝐼 + πœ‚ ​ 𝐸 𝑑 14 ) βˆ’ 1 β€– ≀ 2 is quite helpful. This follows from Lemma 14 in which we have established that β€– 𝐸 𝑑 14 β€– ≀ 1 / 2 . As a result of this relation, we obtain

β€– | 𝑇 1 | β€–

≀ 2 ​ πœ† βˆ’ 1 ​ β€– | 𝐸 𝑑 𝑒 | β€– ,

β€– | 𝑇 2 | β€–

≀ 2 ​ β€– | 𝑂 ~ 𝑑 | β€– β‹… β€– 𝑄 β€– β‹… β€– 𝑆 ~ 𝑑 βˆ’ 1 β€– β‹… β€– ( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ) βˆ’ 1 β€– ,

β€– | 𝑇 3 | β€–

≀ 2 ​ β€– | 𝐸 𝑑 𝑑 | β€– β‹… ( 1 + β€– 𝑄 β€– ) β‹… β€– 𝑆 ~ 𝑑 βˆ’ 1 β€– β‹… β€– ( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ) βˆ’ 1 β€– .

Similar to the control of 𝑇 1 in the proof of Lemma 14, we can take the condition πœ† β‰₯ 1 100 ​ πœ… βˆ’ 4 ​ 𝑐 πœ† ​ 𝜎 min 2 ​ ( 𝑋 ⋆ ) and the bound (54e) collectively to see that

β€– | 𝑇 1 | β€– ≀ 400 ​ 𝑐 πœ† βˆ’ 1 ​ πœ… 6 ​ β€– 𝑋 ⋆ β€– βˆ’ 2 ​ β€– | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | ​ β€– + 1 64 ​ ( 𝐢 3 . π‘Ž + 1 ) 2 ​ πœ… 4 ​ β€– 𝑋 ⋆ β€– β€– ​ | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€– .

Regarding the terms 𝑇 2 and 𝑇 3 , we see from (113) that

β€– ( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ) βˆ’ 1 β€– ≀ πœ… 1 βˆ’ πœ‚ ≀ 2 ​ πœ… ,

as long πœ‚ is sufficiently small. Recalling the assumption β€– 𝑄 β€– ≀ 2 , this allows us to obtain

β€– | 𝑇 2 | β€–

≀ 8 ​ πœ‚ βˆ’ 1 ​ πœ… ​ β€– | 𝑂 ~ 𝑑 | β€– 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ≀ 8 ​ πœ‚ βˆ’ 1 ​ πœ… ​ 𝑛 ​ β€– 𝑂 ~ 𝑑 β€– 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ,

β€– | 𝑇 3 | β€–

≀ 12 ​ πœ… ​ β€– | 𝐸 𝑑 𝑑 | β€– / 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ,

where the first inequality again uses the elementary fact β€– | 𝑂 ~ 𝑑 | β€– ≀ 𝑛 ​ β€– 𝑂 ~ 𝑑 β€– in (70).

The desired bounds then follow from plugging in the bounds (54d) and (24).

Appendix CProofs for Phase I

The goal of this section is to prove Lemma 3 in an inductive manner. We achieve this goal in two steps. In Section C.1, we find an iteration number 𝑑 1 ≀ 𝑇 min / 16 such that the claim (22) is true at 𝑑 1 . This establishes the base case. Then in Section C.2, we prove the induction step, namely if the claim (22) holds for some iteration 𝑑 β‰₯ 𝑑 1 , we aim to show that (22) continues to hold for the iteration 𝑑 + 1 . These two steps taken collectively finishes the proof of Lemma 3.

C.1Establishing the base case: Finding a valid 𝑑 1

The following lemma ensures the existence of such an iteration number 𝑑 1 .

Lemma 19.

Under the same setting as Theorem 2, we have for some 𝑑 1 ≀ 𝑇 min / 16 such that (21) holds and that (22) hold with 𝑑

𝑑 1 .

The rest of this subsection is devoted to the proof of this lemma.

Define an auxiliary sequence

𝑋 ^ 𝑑 ≔ ( 𝐼 + πœ‚ πœ† ​ π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) ) 𝑑 ​ 𝑋 0 ,

(77)

which can be viewed as power iterations on the matrix π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) from the initialization 𝑋 0 .

In what follows, we first establish that the true iterates { 𝑋 𝑑 } stay close to the auxiliary iterates { 𝑋 ^ 𝑑 } as long as the initialization size 𝛼 is small; see Lemma 20. This proximity then allows us to invoke the result in stoger2021small (see Lemma 21) to establish Lemma 19. For the rest of the appendices, we work on the following event given in (18):

β„°

{ β€– 𝐺 β€– ≀ 𝐢 𝐺 } ∩ { 𝜎 min βˆ’ 1 ​ ( π‘ˆ ^ ⊀ ​ 𝐺 ) ≀ ( 2 ​ 𝑛 ) 𝐢 𝐺 } .

Step 1: controlling distance between 𝑋 𝑑 and 𝑋 ^ 𝑑 .

The following lemma guarantees the closeness between the two iterates { 𝑋 𝑑 } and { 𝑋 ^ 𝑑 } , with the proof deferred to Appendix C.1.1. Recall that 𝐢 𝐺 is the constant defined in the event β„° in (18), and 𝑐 πœ† is the constant given in Theorem 2.

Lemma 20.

Suppose that πœ† β‰₯ 1 100 ​ 𝑐 πœ† ​ 𝜎 min 2 ​ ( 𝑋 ⋆ ) . For any πœƒ ∈ ( 0 , 1 ) , there exists a large enough constant 𝐾

𝐾 ​ ( πœƒ , 𝑐 πœ† , 𝐢 𝐺 )

0 such that the following holds: As long as 𝛼 obeys

log ⁑ β€– 𝑋 ⋆ β€– 𝛼 β‰₯ 𝐾 max ⁑ ( πœ‚ , πœ… βˆ’ 2 ) ​ log ⁑ ( 2 ​ πœ… ​ 𝑛 ) β‹… ( 1 + log ⁑ ( 1 + πœ‚ πœ† ​ β€– π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) β€– ) ) ,

(78)

one has for all 𝑑 ≀ 1 πœƒ ​ πœ‚ ​ log ⁑ ( πœ… ​ 𝑛 ) :

βˆ₯ 𝑋 𝑑 βˆ’ 𝑋 ^ 𝑑 βˆ₯

≀ 𝑑 ​ ( 1 + πœ‚ πœ† ​ βˆ₯ π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) βˆ₯ ) 𝑑 ​ 𝛼 2 β€– 𝑋 ⋆ β€– .

(79)

Moreover, βˆ₯ 𝑋 𝑑 βˆ₯ ≀ βˆ₯ 𝑋 ⋆ βˆ₯ for all such 𝑑 .

Step 2: borrowing a lemma from stoger2021small.

Compared to the original sequence 𝑋 𝑑 , the behavior of the power iterates 𝑋 ^ 𝑑 is much easier to analyze. Now that we have sufficient control over β€– 𝑋 𝑑 βˆ’ 𝑋 ^ 𝑑 β€– , it is possible to show that 𝑋 𝑑 has the desired properties in Lemma 19 by first establishing the corresponding property of 𝑋 ^ 𝑑 and then invoking a standard matrix perturbation argument. Fortunately, such a strategy has been implemented by stoger2021small and wrapped into the following helper lemma.

Denote

𝑠 𝑗 ≔ 𝜎 𝑗 ​ ( 𝐼 + πœ‚ πœ† ​ π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) )

1 + πœ‚ πœ† ​ 𝜎 𝑗 ​ ( π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) ) , 𝑗

1 , 2 , … , 𝑛

and recall that π‘ˆ ^ (resp. π‘ˆ 𝑋 ~ 𝑑 ) is an orthonormal basis of the eigenspace associated with the π‘Ÿ ⋆ largest eigenvalues of π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) (resp. 𝑋 ~ 𝑑 ).

Lemma 21.

There exists some small universal 𝑐 21

0 such that the following hold. Assume that for some 𝛾 ≀ 𝑐 21 ,

βˆ₯ ( ℐ βˆ’ π’œ βˆ— ​ π’œ ) ​ ( 𝑀 ⋆ ) βˆ₯ ≀ 𝛾 ​ 𝜎 min 2 ​ ( 𝑋 ⋆ ) ,

(80)

and furthermore,

πœ™ ≔ 𝛼 ​ β€– 𝐺 β€– ​ 𝑠 π‘Ÿ ⋆ + 1 𝑑 + β€– 𝑋 𝑑 βˆ’ 𝑋 ^ 𝑑 β€– 𝛼 ​ 𝜎 min ​ ( π‘ˆ ^ ⊀ ​ 𝐺 ) ​ 𝑠 π‘Ÿ ⋆ 𝑑 ≀ 𝑐 21 ​ πœ… βˆ’ 2 .

(81)

Then there exists some universal 𝐢 21

0 such that the following hold:

𝜎 min ​ ( 𝑆 ~ 𝑑 )

β‰₯ 𝛼 4 ​ 𝜎 min ​ ( π‘ˆ ^ ⊀ ​ 𝐺 ) ​ 𝑠 π‘Ÿ ⋆ 𝑑 ,

(82a)

β€– 𝑂 ~ 𝑑 β€–

≀ 𝐢 21 ​ πœ™ ​ 𝛼 ​ 𝜎 min ​ ( π‘ˆ ^ ⊀ ​ 𝐺 ) ​ 𝑠 π‘Ÿ ⋆ 𝑑 ,

(82b)

β€– π‘ˆ ⋆ , βŸ‚ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 β€–

≀ 𝐢 21 ​ ( 𝛾 + πœ™ ) ,

(82c)

where 𝑋 ~ 𝑑 ≔ 𝑋 𝑑 ​ 𝑉 𝑑 ∈ ℝ 𝑛 Γ— π‘Ÿ ⋆ . Proof of Lemma 21.

This follows from the claims of stoger2021small by noting that βˆ₯ 𝑂 ~ 𝑑 βˆ₯

βˆ₯ π‘ˆ ⋆ , βŸ‚ ⊀ ​ 𝑋 𝑑 ​ 𝑉 𝑑 , βŸ‚ βˆ₯ ≀ βˆ₯ 𝑋 𝑑 ​ 𝑉 𝑑 , βŸ‚ βˆ₯ for (82b).3 ∎

Step 3: completing the proof.

Now, with the help of Lemma 21, we are ready to prove Lemma 19. We start with verifying the two assumptions in Lemma 21.

Verifying assumption (80).

By the RIP in (9), Lemma 8, and the condition of 𝛿 in (10), we have

βˆ₯ ( ℐ βˆ’ π’œ βˆ— π’œ ) ( 𝑀 ⋆ ) βˆ₯ ≀ π‘Ÿ ⋆ 𝛿 βˆ₯ 𝑀 ⋆ βˆ₯ ≀ 𝑐 𝛿 πœ… βˆ’ ( 𝐢 𝛿 βˆ’ 2 ) 𝜎 min 2 ( 𝑋 ⋆ )

: 𝛾 𝜎 min 2 ( 𝑋 ⋆ ) .

(83)

Here 𝛾

𝑐 𝛿 ​ πœ… βˆ’ ( 𝐢 𝛿 βˆ’ 2 ) ≀ 𝑐 21 , as 𝑐 𝛿 is assumed to be sufficiently small.

Verifying assumption (81).

By Weyl’s inequality and (83), we have

| 𝑠 𝑗 βˆ’ 1 βˆ’ πœ‚ πœ† ​ 𝜎 𝑗 ​ ( 𝑀 ⋆ ) | ≀ πœ‚ πœ† ​ βˆ₯ ( ℐ βˆ’ π’œ βˆ— ​ π’œ ) ​ ( 𝑀 ⋆ ) βˆ₯ ≀ πœ‚ πœ† ​ 𝑐 𝛿 ​ πœ… βˆ’ ( 𝐢 𝛿 βˆ’ 2 ) ​ 𝜎 min 2 ​ ( 𝑋 ⋆ ) ≀ 100 ​ 𝑐 𝛿 𝑐 πœ† ​ πœ‚ ,

where the last inequality follows from the condition πœ† β‰₯ 1 100 ​ πœ… βˆ’ 4 ​ 𝑐 πœ† ​ 𝜎 min 2 ​ ( 𝑋 ⋆ ) . Furthermore, using the condition πœ† ≀ 𝑐 πœ† ​ 𝜎 min 2 ​ ( 𝑋 ⋆ ) assumed in (12b), the above bound implies that, for some 𝐢

𝐢 ​ ( 𝑐 πœ† , 𝑐 𝛿 )

0 ,

𝑠 1

≀ 1 + πœ‚ πœ† ​ β€– 𝑀 ⋆ β€– + 100 ​ 𝑐 𝛿 𝑐 πœ† ​ πœ‚ ≀ 1 + 𝐢 ​ πœ‚ ​ πœ… 6 ,

(84a)

𝑠 π‘Ÿ ⋆

β‰₯ 1 + πœ‚ πœ† ​ 𝜎 min 2 ​ ( 𝑋 ⋆ ) βˆ’ 100 ​ 𝑐 𝛿 𝑐 πœ† ​ πœ‚ β‰₯ 1 + πœ‚ 2 ​ 𝑐 πœ† ,

(84b)

𝑠 π‘Ÿ ⋆

≀ 1 + πœ‚ πœ† ​ 𝜎 min 2 ​ ( 𝑋 ⋆ ) + 100 ​ 𝑐 𝛿 𝑐 πœ† ​ πœ‚ ≀ 1 + 2 ​ πœ‚ πœ† / 𝜎 min 2 ​ ( 𝑋 ⋆ ) ,

(84c)

𝑠 π‘Ÿ ⋆ + 1

≀ 1 + 100 ​ 𝑐 𝛿 𝑐 πœ† ​ πœ‚ ≀ 1 + πœ‚ 4 ​ 𝑐 πœ† ,

(84d)

where we use the fact that 𝜎 π‘Ÿ ⋆ + 1 ​ ( 𝑀 ⋆ )

0 , and 𝑐 𝛿 ≀ 1 / 400 . Consequently we have 𝑠 π‘Ÿ ⋆ / 𝑠 π‘Ÿ ⋆ + 1 β‰₯ 1 + 𝑐 β€² ​ πœ‚ for some 𝑐 β€²

𝑐 β€² ​ ( 𝑐 πœ† ) > 0 , assuming 𝑐 πœ‚ ≀ 𝑐 πœ† . Thus for any large constant 𝐿 > 0 , there is some constant 𝑐 β€²β€²

𝑐 β€²β€² ​ ( 𝑐 β€² ) > 0 such that, setting 𝐿 β€²

𝑐 β€²β€² ​ 𝐿 ​ log ⁑ ( 𝐿 ) we have

( 𝑠 π‘Ÿ ⋆ / 𝑠 π‘Ÿ ⋆ + 1 ) 𝑑 β‰₯ ( 𝐿 ​ πœ… ​ 𝑛 ) 𝐿 , βˆ€ 𝑑 β‰₯ 𝐿 β€² πœ‚ ​ log ⁑ ( πœ… ​ 𝑛 ) .

On the event β„° given in (18), we can choose 𝐿 large enough so that 𝐿 β‰₯ 2 ​ 𝐢 𝐺 , hence β€– 𝐺 β€– ≀ 𝐿 and 𝜎 min βˆ’ 1 ​ ( π‘ˆ ^ ⊀ ​ 𝐺 ) ≀ ( 2 ​ 𝑛 ) 𝐿 / 2 . Summarizing these inequalities, we see for 𝑑 β‰₯ 𝐿 β€² πœ‚ ​ log ⁑ ( πœ… ​ 𝑛 ) ,

𝛼 ​ β€– 𝐺 β€– ​ 𝑠 π‘Ÿ ⋆ + 1 𝑑 𝛼 ​ 𝜎 min ​ ( π‘ˆ ^ ⊀ ​ 𝐺 ) ​ 𝑠 π‘Ÿ ⋆ 𝑑

≀ 𝐿 ​ 𝜎 min βˆ’ 1 ​ ( π‘ˆ ^ ⊀ ​ 𝐺 ) ​ ( 𝑠 π‘Ÿ ⋆ + 1 / 𝑠 π‘Ÿ ⋆ ) 𝑑

≀ 𝐿 ​ ( 2 ​ 𝑛 ) 𝐿 / 2 ​ ( 𝐿 ​ πœ… ​ 𝑛 ) βˆ’ 𝐿 ≀ ( 𝐿 ​ πœ… ​ 𝑛 ) βˆ’ 𝐿 / 2 .

(85)

Furthermore, invoking Lemma 20 with πœƒ

1 / ( 2 ​ 𝐿 β€² ) (note that (78) is implied by the assumption (12c), where 𝐢 𝛼 is assumed sufficiently large, considering πœ† β‰₯ 1 100 ​ 𝑐 πœ† ​ 𝜎 min 2 ​ ( 𝑋 ⋆ ) and β€– π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) β€– ≀ β€– 𝑀 ⋆ β€– + 𝛾 ​ 𝜎 min 2 ​ ( 𝑋 ⋆ ) ≀ 2 ​ β€– 𝑋 ⋆ β€– 2 by (83)), we obtain for any 𝑑 ≀ 1 πœƒ ​ πœ‚ ​ log ⁑ ( πœ… ​ 𝑛 )

2 ​ 𝐿 β€² πœ‚ ​ log ⁑ ( πœ… ​ 𝑛 ) that β€– 𝑋 𝑑 βˆ’ 𝑋 ^ 𝑑 β€– ≀ 𝑑 ​ 𝑠 1 𝑑 ​ 𝛼 2 / βˆ₯ 𝑋 ⋆ βˆ₯ . This implies

β€– 𝑋 𝑑 βˆ’ 𝑋 ^ 𝑑 β€– 𝛼 ​ 𝜎 min ​ ( π‘ˆ ^ ⊀ ​ 𝐺 ) ​ 𝑠 π‘Ÿ ⋆ 𝑑

≀ ( 𝑠 1 / 𝑠 π‘Ÿ ⋆ ) 𝑑 ​ 𝜎 min βˆ’ 1 ​ ( π‘ˆ ^ ⊀ ​ 𝐺 ) ​ 𝛼 / β€– 𝑋 ⋆ β€–

≀ 𝑠 1 𝑑 ​ 𝜎 min βˆ’ 1 ​ ( π‘ˆ ^ ⊀ ​ 𝐺 ) ​ 𝛼 / β€– 𝑋 ⋆ β€–

≀ exp ⁑ ( 𝑑 ​ log ⁑ ( 𝑠 1 ) + 𝐿 ​ log ⁑ ( 𝐿 ​ πœ… ​ 𝑛 ) ) ​ 𝛼 / β€– 𝑋 ⋆ β€– ≀ ( 𝐿 ​ πœ… ​ 𝑛 ) βˆ’ 𝐿 / 2

(86)

where the second inequality follows from (84b), the penultimate inequality follows from our choice of 𝐿 which ensured 𝜎 min βˆ’ 1 ​ ( π‘ˆ ^ ⊀ ​ 𝐺 ) ≀ ( 2 ​ 𝑛 ) 𝐿 / 2 , and the last inequality follows from (84a), our choice 𝑑 ≀ 2 ​ 𝐿 β€² πœ‚ ​ log ⁑ ( πœ… ​ 𝑛 ) and our assumption (12c) on 𝛼 which implies 𝛼 / β€– 𝑋 ⋆ β€– ≀ ( 2 ​ πœ… ​ 𝑛 ) βˆ’ 𝐢 𝛼 , given that 𝐢 𝛼 is sufficiently large, e.g. 𝐢 𝛼 β‰₯ 𝐢 ​ ( 𝐿 , 𝑐 πœ† , 𝑐 πœ‚ ) . It may also be inferred from the above arguments that 𝐿 can be made arbitrarily large by increasing 𝐢 𝛼 .

Combining the above arguments, we conclude that for any 𝑑 ∈ [ ( 𝐿 β€² / πœ‚ ) ​ log ⁑ ( πœ… ​ 𝑛 ) , ( 2 ​ 𝐿 β€² / πœ‚ ) ​ log ⁑ ( πœ… ​ 𝑛 ) ] , both of (85), (86) hold, hence the condition in (81) can be verified by

πœ™

𝛼 ​ β€– 𝐺 β€– ​ 𝑠 π‘Ÿ ⋆ + 1 𝑑 + β€– 𝑋 𝑑 βˆ’ 𝑋 ^ 𝑑 β€– 𝛼 ​ 𝜎 min ​ ( π‘ˆ ^ ⊀ ​ 𝐺 ) ​ 𝑠 π‘Ÿ ⋆ 𝑑

≀ 2 ​ ( 𝐿 ​ πœ… ​ 𝑛 ) βˆ’ 𝐿 / 2

(87)

≀ 𝑐 21 ​ πœ… βˆ’ 2 ,

by choosing 𝐿 sufficiently large.

This completes the verification of both assumptions of Lemma 21. Upon noting that the upper threshold of 𝑑 satisfies ( 2 ​ 𝐿 β€² / πœ‚ ) ​ log ⁑ ( πœ… ​ 𝑛 ) ≀ 𝑇 min / 16 , we will now invoke the conclusions of Lemma 21 to prove Lemma 19 for some 𝑑 ∈ [ ( 𝐿 β€² / πœ‚ ) ​ log ⁑ ( πœ… ​ 𝑛 ) , 𝑇 min / 16 ] .

Proof of bound (21).

This can be inferred from (82a) in the following way. Recalling that 𝜎 min ​ ( π‘ˆ ^ ⊀ ​ 𝐺 ) β‰₯ ( 2 ​ 𝑛 ) βˆ’ 𝐢 𝐺 on the event β„° , and 𝑠 π‘Ÿ ⋆ β‰₯ 1 by (84b), we obtain from (82a) that

𝜎 min ​ ( 𝑆 ~ 𝑑 1 ) β‰₯ 1 4 ​ 𝛼 ​ ( 2 ​ 𝑛 ) βˆ’ 𝐢 𝐺 β‰₯ 𝛼 2 / βˆ₯ 𝑋 ⋆ βˆ₯ ,

given the condition (12c) which guarantees

𝛼 β€– 𝑋 ⋆ β€– ≀ ( 2 ​ 𝑛 ) βˆ’ 𝐢 𝛼 / πœ‚ ≀ 1 4 ​ ( 2 ​ 𝑛 ) βˆ’ 𝐢 𝐺 ,

as long as πœ‚ ≀ 𝑐 πœ‚ ≀ 1 and 𝐢 𝛼 β‰₯ 𝐢 𝐺 + 2 . The proof is complete.

Proof of bound (22a).

We combine (82a), (82b), and (87) to obtain

βˆ₯ 𝑂 ~ 𝑑 1 βˆ₯ 𝜎 min ​ ( 𝑆 ~ 𝑑 1 ) ≀ 4 ​ 𝐢 21 ​ πœ™ ≀ 4 ​ 𝐢 21 ​ ( 𝐿 ​ πœ… ​ 𝑛 ) βˆ’ 𝐿 / 2 ≀ ( 𝐿 ​ πœ… ​ 𝑛 / 2 ) βˆ’ 𝐿 / 2 ,

where the last inequality follows from taking 𝐿 sufficiently large. We further note that (12b) implies

𝜎 min ​ ( 𝑆 ~ 𝑑 1 ) ≀ β€– Ξ£ ⋆ 2 + πœ† ​ 𝐼 β€– 1 / 2 ​ 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 1 )

≀ ( 𝑐 πœ† + 1 ) 1 / 2 ​ β€– 𝑋 ⋆ β€– ​ 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 1 )

≀ 2 ​ β€– 𝑋 ⋆ β€– ​ 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 1 ) ,

assuming 𝑐 πœ† ≀ 1 , hence

βˆ₯ 𝑂 ~ 𝑑 1 βˆ₯ 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 1 ) ≀ 2 ​ β€– 𝑋 ⋆ β€– ​ ( 𝐿 ​ πœ… ​ 𝑛 / 2 ) βˆ’ 𝐿 / 2 ≀ ( 𝐢 3 . 𝑏 ​ πœ… ​ 𝑛 ) βˆ’ 𝐢 3 . 𝑏 ​ β€– 𝑋 ⋆ β€– ,

as desired, with 𝐢 3 . 𝑏

𝐿 / 4 as long as 𝐿 is sufficiently large. It is also clear that 𝐢 3 . 𝑏 can be made arbitrarily large by enlarging 𝐢 𝛼 as 𝐿 can be.

Proof of bound (22b).

We apply (82b) to yield

βˆ₯ 𝑂 ~ 𝑑 1 βˆ₯ ≀ 𝐢 21 ​ πœ™ ​ 𝛼 ​ 𝜎 min ​ ( π‘ˆ ^ ⊀ ​ 𝐺 ) ​ 𝑠 π‘Ÿ ⋆ 𝑑 1 ≀ 𝐢 𝐺 ​ 𝐢 21 ​ ( 𝐿 ​ πœ… ​ 𝑛 ) βˆ’ 𝐿 / 2 ​ ( 1 + 2 ​ πœ‚ 𝑐 πœ† ) 𝑑 1 ​ 𝛼 ≀ 𝛼 5 / 6 ​ βˆ₯ 𝑋 ⋆ βˆ₯ 1 / 6 ,

where the second inequality follows from 𝜎 min ​ ( π‘ˆ ^ ⊀ ​ 𝐺 ) ≀ β€– 𝐺 β€– ≀ 𝐢 𝐺 by assumption and from (84c); the last inequality follows from 𝑑 1 ≀ ( 2 ​ 𝐿 β€² / πœ‚ ) ​ log ⁑ ( πœ… ​ 𝑛 ) and from the condition (12c) on 𝛼 , provided that 𝐢 𝛼 is sufficiently large.

Proof of bound (22c).

We apply (82c) to yield that

βˆ₯ π‘ˆ ⋆ , βŸ‚ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 + 1 βˆ₯ ≀ 𝐢 21 ​ ( 𝛾 + πœ™ ) ≀ 𝑐 𝛿 𝑐 πœ† ​ πœ… βˆ’ 2 ​ 𝐢 𝛿 / 3 ,

using the bounds of 𝛾 and πœ™ in (83) and (87), provided that 𝑐 πœ† ≀ 1 2 ​ min ⁑ ( 1 , 𝐢 21 βˆ’ 1 ) and 𝐿 β‰₯ 2 ​ ( 𝐢 𝛿 + 1 ) . To further bound β€– 𝑁 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 βˆ’ 1 ​ Ξ£ ⋆ β€– we need the following lemma.

Lemma 22.

Assume 𝑆 ~ 𝑑 is invertible, and at least one of the following is true: (i) β€– π‘ˆ ⋆ , βŸ‚ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 β€– ≀ 1 / 4 ; (ii) β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– ≀ πœ… βˆ’ 1 ​ β€– 𝑋 ⋆ β€– / 4 . Then

πœ… βˆ’ 1 ​ β€– 𝑋 ⋆ β€– ​ β€– π‘ˆ ⋆ , βŸ‚ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 β€– ≀ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– ≀ 2 ​ β€– 𝑋 ⋆ β€– ​ β€– π‘ˆ ⋆ , βŸ‚ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 β€– .

The proof is postponed to Section C.1.2. Returning to the proof of bound (22c), the above lemma yields

β€– 𝑁 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 βˆ’ 1 ​ Ξ£ ⋆ β€– ≀ 2 ​ 𝑐 𝛿 𝑐 πœ† ​ β€– 𝑋 ⋆ β€– ​ πœ… βˆ’ 2 ​ 𝐢 𝛿 / 3 ≀ 𝑐 3 ​ β€– 𝑋 ⋆ β€– ​ πœ… βˆ’ 2 ​ 𝐢 𝛿 / 3 ,

for some 𝑐 3 ≲ 𝑐 𝛿 / 𝑐 πœ† , as desired.

Proof of bound (22d).

We have

βˆ₯ 𝑆 ~ 𝑑 1 βˆ₯

βˆ₯ π‘ˆ ⋆ ⊀ ​ 𝑋 𝑑 1 ​ 𝑉 𝑑 1 βˆ₯ ≀ βˆ₯ 𝑋 𝑑 1 βˆ₯ ≀ βˆ₯ 𝑋 ⋆ βˆ₯ ,

where the last step follows from Lemma 20.

C.1.1Proof of Lemma 20

We prove the claim (79) by induction and also show that β€– 𝑋 𝑑 β€– ≀ β€– 𝑋 ⋆ β€– follows from (79). For the base case 𝑑

0 , it holds by definition. Assume that (79) holds for some 𝑑 ≀ 1 πœƒ ​ πœ‚ ​ log ⁑ ( πœ… ​ 𝑛 ) βˆ’ 1 . We aim to prove that (i) βˆ₯ 𝑋 𝑑 βˆ₯ ≀ βˆ₯ 𝑋 ⋆ βˆ₯ and that (ii) the inequality (79) continues to hold for 𝑑 + 1 .

Proof of βˆ₯ 𝑋 𝑑 βˆ₯ ≀ βˆ₯ 𝑋 ⋆ βˆ₯ .

By the induction hypothesis we know

βˆ₯ 𝑋 𝑑 βˆ’ 𝑋 ^ 𝑑 βˆ₯ ≀ 𝑑 ​ ( 1 + πœ‚ πœ† ​ βˆ₯ π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) βˆ₯ ) 𝑑 ​ 𝛼 2 β€– 𝑋 ⋆ β€– .

In view of the constraint (78) on 𝛼 and the restriction 𝑑 ≀ 1 πœƒ ​ πœ‚ ​ log ⁑ ( πœ… ​ 𝑛 ) , we have

𝑑 ​ 𝛼 βˆ₯ 𝑋 ⋆ βˆ₯ ≀ 1 πœƒ ​ πœ‚ ​ log ⁑ ( πœ… ​ 𝑛 ) β‹… πœ‚ 𝐾 ​ 1 log ⁑ ( πœ… ​ 𝑛 )

1 𝐾 ​ πœƒ ≀ 1

as long as 𝐾

𝐾 ​ ( πœƒ , 𝑐 πœ† , 𝐢 𝐺 ) is sufficiently large. This further implies

β€– 𝑋 𝑑 βˆ’ 𝑋 ^ 𝑑 β€– ≀ ( 𝑑 ​ 𝛼 β€– 𝑋 ⋆ β€– ) ​ ( 1 + πœ‚ πœ† ​ β€– π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) β€– ) 𝑑 ​ 𝛼 ≀ ( 1 + πœ‚ πœ† ​ β€– π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) β€– ) 𝑑 ​ 𝛼 .

On the other hand, since βˆ₯ 𝑋 0 βˆ₯ ≀ 𝐢 𝐺 ​ 𝛼 under the event β„° (cf. (18)), in view of (77), we have

β€– 𝑋 ^ 𝑑 β€– ≀ ( 1 + πœ‚ πœ† ​ β€– π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) β€– ) 𝑑 ​ β€– 𝑋 0 β€– ≀ 𝐢 𝐺 ​ ( 1 + πœ‚ πœ† ​ β€– π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) β€– ) 𝑑 ​ 𝛼 .

Thus for a large enough 𝐾

𝐾 ​ ( πœƒ , 𝑐 πœ† , 𝐢 𝐺 ) , we have

β€– 𝑋 𝑑 β€– ≀ β€– 𝑋 𝑑 βˆ’ 𝑋 ^ 𝑑 β€– + β€– 𝑋 ^ 𝑑 β€– ≀ ( 1 + πœ‚ πœ† ​ β€– π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) β€– ) 𝑑 ​ ( 𝐢 𝐺 + 1 ) ​ 𝛼 ≀ 𝑐 πœ† / 200 β‹… πœ… βˆ’ 4 ​ β€– 𝑋 ⋆ β€– ,

(88)

where the last inequality follows from the condition on 𝑑 and the choice of 𝛼 in (78):

log ⁑ βˆ₯ 𝑋 ⋆ βˆ₯ 𝛼 β‰₯ log ⁑ 200 ​ ( 𝐢 𝐺 + 1 ) ​ πœ… 4 𝑐 πœ† + 𝑑 ​ log ⁑ ( 1 + πœ‚ πœ† ​ β€– π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) β€– ) .

The inequality (88) clearly implies βˆ₯ 𝑋 𝑑 βˆ₯ ≀ βˆ₯ 𝑋 ⋆ βˆ₯ .

Proof of (79) at the induction step.

The proof builds on a key recursive relation on βˆ₯ 𝑋 𝑑 + 1 βˆ’ 𝑋 ^ 𝑑 + 1 βˆ₯ , from which the induction follows readily from our assumption.

Step 1: building a recursive relation on βˆ₯ 𝑋 𝑑 + 1 βˆ’ 𝑋 ^ 𝑑 + 1 βˆ₯ .

By definition (77), we have 𝑋 ^ 𝑑 + 1

( 𝐼 + πœ‚ πœ† ​ π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) ) ​ 𝑋 ^ 𝑑 , which implies the following decomposition:

𝑋 𝑑 + 1 βˆ’ 𝑋 ^ 𝑑 + 1

[ 𝑋 𝑑 + 1 βˆ’ ( 𝐼 + πœ‚ πœ† ​ π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) ) ​ 𝑋 𝑑 ] ⏟ ≕ 𝑇 1 + ( 𝐼 + πœ‚ πœ† ​ π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) ) ​ ( 𝑋 𝑑 βˆ’ 𝑋 ^ 𝑑 ) ⏟ ≕ 𝑇 2 .

(89)

We shall control each term separately.

β€’

The second term 𝑇 2 can be trivially bounded as

β€– 𝑇 2 β€–

βˆ₯ ( 𝐼 + πœ‚ πœ† ​ π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) ) ​ ( 𝑋 𝑑 βˆ’ 𝑋 ^ 𝑑 ) βˆ₯ ≀ ( 1 + πœ‚ πœ† ​ βˆ₯ π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) βˆ₯ ) ​ βˆ₯ 𝑋 𝑑 βˆ’ 𝑋 ^ 𝑑 βˆ₯ .

(90) β€’

Turning to the first term 𝑇 1 , by the update rule (7) of 𝑋 𝑑 + 1 and the triangle inequality, we further have

β€– 𝑇 1 β€–

β€– 𝑋 𝑑 + 1 βˆ’ ( 𝐼 + πœ‚ πœ† ​ π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) ) ​ 𝑋 𝑑 β€–

≀ β€– πœ‚ ​ π’œ βˆ— ​ π’œ ​ ( 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ ) ​ 𝑋 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 β€–

  • β€– πœ‚ ​ π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) ​ 𝑋 𝑑 ​ ( ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑
  • πœ† ​ 𝐼 ) βˆ’ 1 βˆ’ πœ† βˆ’ 1 ​ 𝐼 ) β€– .

(91)

Since β€– ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 β€– ≀ πœ† βˆ’ 1 , it follows that the first term in (β€’ β€£ C.1.1) can be bounded by

β€– πœ‚ ​ π’œ βˆ— ​ π’œ ​ ( 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ ) ​ 𝑋 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 β€– ≀ πœ‚ πœ† ​ β€– π’œ βˆ— ​ π’œ ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 ) β€– ​ β€– 𝑋 𝑑 β€– .

In addition, since 𝑐 πœ† / 200 β‹… πœ… βˆ’ 4 ​ β€– 𝑋 ⋆ β€–

𝑐 πœ† ​ 𝜎 min 2 ​ ( 𝑋 ⋆ ) / 200 ≀ πœ† / 2 by the condition πœ† β‰₯ 1 100 ​ πœ… βˆ’ 4 ​ 𝑐 πœ† ​ 𝜎 min 2 ​ ( 𝑋 ⋆ ) , we have by (88) that βˆ₯ 𝑋 𝑑 βˆ₯ ≀ πœ† / 2 . Therefore, invoking Lemma 9 implies that

( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 βˆ’ πœ† βˆ’ 1 ​ 𝐼

πœ† βˆ’ 2 ​ 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 ​ 𝑄 , for some  ​ 𝑄 ​ with  ​ βˆ₯ 𝑄 βˆ₯ ≀ 2 .

As a result, the second term in (β€’ β€£ C.1.1) can be bounded by

β€– πœ‚ ​ π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) ​ 𝑋 𝑑 ​ ( ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 βˆ’ πœ† βˆ’ 1 ​ 𝐼 ) β€– ≀ 2 ​ πœ‚ πœ† 2 ​ β€– π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) β€– ​ β€– 𝑋 𝑑 β€– 3 .

Combining the above two inequalities leads to

β€– 𝑇 1 β€– ≀ πœ‚ πœ† ​ ( β€– π’œ βˆ— ​ π’œ ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 ) β€– + 2 πœ† ​ β€– π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) β€– ​ β€– 𝑋 𝑑 β€– 2 ) ​ β€– 𝑋 𝑑 β€– .

In view of Lemma 8, we know βˆ₯ π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) βˆ₯ ≲ π‘Ÿ ⋆ ​ βˆ₯ 𝑀 ⋆ βˆ₯ and βˆ₯ π’œ βˆ— ​ π’œ ​ ( 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ ) βˆ₯ ≲ π‘Ÿ ​ βˆ₯ 𝑋 𝑑 βˆ₯ 2 . Plugging these relations into the previous bound leads to

β€– 𝑇 1 β€– ≲ πœ‚ ​ π‘Ÿ πœ† ​ ( 1 + βˆ₯ 𝑀 ⋆ βˆ₯ πœ† ) ​ βˆ₯ 𝑋 𝑑 βˆ₯ 3 ≲ πœ‚ ​ πœ… 6 ​ π‘Ÿ β€– 𝑀 ⋆ β€– ​ πœ… 6 ​ β€– 𝑋 𝑑 β€– 3 ,

(92)

where the last inequality follows from πœ† ≳ πœ… βˆ’ 4 ​ 𝜎 min 2 ​ ( 𝑋 ⋆ )

πœ… βˆ’ 6 ​ β€– 𝑀 ⋆ β€– (cf. (12b)).

Putting the bounds on 𝑇 1 and 𝑇 2 together leads to

βˆ₯ 𝑋 𝑑 + 1 βˆ’ 𝑋 ^ 𝑑 + 1 βˆ₯ ≀ ( 1 + πœ‚ πœ† ​ β€– π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) β€– ) ​ βˆ₯ 𝑋 𝑑 βˆ’ 𝑋 ^ 𝑑 βˆ₯ + 𝐢 ​ πœ‚ ​ πœ… 1 ​ 2 ​ π‘Ÿ β€– 𝑀 ⋆ β€– ​ β€– 𝑋 𝑑 β€– 3

(93)

for some universal constant 𝐢

𝐢 ​ ( 𝑐 πœ† )

0 .

Step 2: finishing the induction.

By the bound of β€– 𝑋 𝑑 β€– in (88), it suffices to prove

𝑑 ​ ( 1 + πœ‚ πœ† ​ βˆ₯ π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) βˆ₯ ) 𝑑 + 1 ​ 𝛼 2 β€– 𝑋 ⋆ β€– + 𝐢 ​ ( 𝐢 𝐺 + 1 ) 3 ​ πœ‚ ​ πœ… 12 ​ π‘Ÿ β€– 𝑋 ⋆ β€– 2 ​ ( 1 + πœ‚ πœ† ​ βˆ₯ π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) βˆ₯ ) 3 ​ 𝑑 ​ 𝛼 3

≀ ( 𝑑 + 1 ) ​ ( 1 + πœ‚ πœ† ​ βˆ₯ π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) βˆ₯ ) 𝑑 + 1 ​ 𝛼 2 β€– 𝑋 ⋆ β€– .

This is equivalent to

𝐢 ​ ( 𝐢 𝐺 + 1 ) 3 ​ πœ‚ ​ πœ… 12 ​ π‘Ÿ ​ ( 1 + πœ‚ πœ† ​ β€– π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) β€– ) 2 ​ 𝑑 βˆ’ 1 ≀ β€– 𝑋 ⋆ β€– 𝛼 ,

which again follows readily from our assumption 𝑑 ≀ 1 πœƒ ​ πœ‚ ​ log ⁑ ( πœ… ​ 𝑛 ) and the assumption (78) on 𝛼 which implies

log ⁑ ( β€– 𝑋 ⋆ β€– 𝛼 )

β‰₯ ( 2 ​ 𝑑 βˆ’ 1 ) ​ log ⁑ ( 1 + πœ‚ πœ† ​ β€– π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) β€– ) + 12 ​ log ⁑ πœ… + log ⁑ 𝑛 + 𝐾

β‰₯ ( 2 ​ 𝑑 βˆ’ 1 ) ​ log ⁑ ( 1 + πœ‚ πœ† ​ β€– π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) β€– ) + 12 ​ log ⁑ ( 𝑛 ​ πœ… ​ π‘Ÿ ) + log ⁑ ( 𝐢 ​ ( 𝐢 𝐺 + 1 ) 3 )

provided 𝐾

𝐾 ​ ( πœƒ , 𝑐 πœ† , 𝐢 𝐺 ) is sufficiently large. The proof is complete.

C.1.2Proof of Lemma 22

We begin with the following observation:

𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1

π‘ˆ ⋆ , βŸ‚ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 ​ Ξ£ 𝑋 ~ 𝑑 ​ 𝑉 𝑋 ~ 𝑑 ⊀ ​ 𝑉 𝑋 ~ 𝑑 ​ Ξ£ 𝑋 ~ 𝑑 βˆ’ 1 ​ ( π‘ˆ ⋆ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 ) βˆ’ 1

π‘ˆ ⋆ , βŸ‚ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 ​ ( π‘ˆ ⋆ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 ) βˆ’ 1

(94)

where we use: (i) 𝑁 ~ 𝑑

π‘ˆ ⋆ , βŸ‚ ⊀ ​ ( π‘ˆ 𝑋 ~ 𝑑 ​ Ξ£ 𝑋 ~ 𝑑 ​ 𝑉 𝑋 ~ 𝑑 ⊀ ) and 𝑆 ~ 𝑑

π‘ˆ ⋆ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 ​ Ξ£ 𝑋 ~ 𝑑 ​ 𝑉 𝑋 ~ 𝑑 ⊀ ; (ii) 𝑋 ~ 𝑑 is invertible since 𝑆 ~ 𝑑 is invertible, and hence 𝑉 𝑋 ~ 𝑑 has rank π‘Ÿ ⋆ and Ξ£ 𝑋 ~ 𝑑 , π‘ˆ ⋆ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 are also invertible.

We will show that the above quantity is small if (and only if) π‘ˆ ⋆ , βŸ‚ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 is small.

Turning to the proof, we first show that (ii) implies (i), thus it suffices to prove the lemma under the condition (i). In fact, in virtue of (94) we have

β€– π‘ˆ ⋆ , βŸ‚ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 β€– ≀ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 β€– ​ β€– π‘ˆ ⋆ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 β€– ≀ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 β€– ≀ 𝜎 min ​ ( 𝑋 ⋆ ) βˆ’ 1 ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– ,

where we used β€– π‘ˆ ⋆ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 β€– ≀ β€– π‘ˆ ⋆ β€– ​ β€– π‘ˆ 𝑋 ~ 𝑑 β€– ≀ 1 . Consequently, β€– π‘ˆ ⋆ , βŸ‚ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 β€– ≀ 1 / 4 if β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– ≀ πœ… βˆ’ 1 ​ β€– 𝑋 ⋆ β€– / 4 , as claimed.

We proceed to show that the conclusion holds assuming condition (i). The first inequality has already been established above. For the second inequality, using (94) again, it suffices to prove β€– ( π‘ˆ ⋆ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 ) βˆ’ 1 β€– ≀ 2 , which is in turn equivalent to 𝜎 min ​ ( π‘ˆ ⋆ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 ) β‰₯ 1 / 2 . Now note that π‘ˆ 𝑋 ~ 𝑑

π‘ˆ ⋆ ​ π‘ˆ ⋆ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 + π‘ˆ ⋆ , βŸ‚ ​ π‘ˆ ⋆ , βŸ‚ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 , thus

𝜎 min ​ ( π‘ˆ ⋆ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 )

𝜎 π‘Ÿ ⋆ ​ ( π‘ˆ ⋆ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 )

β‰₯ 𝜎 π‘Ÿ ⋆ ​ ( π‘ˆ ⋆ ​ π‘ˆ ⋆ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 )

β‰₯ 𝜎 π‘Ÿ ⋆ ​ ( π‘ˆ 𝑋 ~ 𝑑 ) βˆ’ β€– π‘ˆ ⋆ , βŸ‚ ​ π‘ˆ ⋆ , βŸ‚ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 β€–

β‰₯ 1 βˆ’ β€– π‘ˆ ⋆ , βŸ‚ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 β€– β‰₯ 3 / 4 .

In the last line, we used 𝜎 π‘Ÿ ⋆ ​ ( π‘ˆ 𝑋 ~ 𝑑 )

1 , which follows from π‘ˆ 𝑋 ~ 𝑑 being a 𝑛 Γ— π‘Ÿ ⋆ orthonormal matrix, and the assumption (i). This completes the proof.

C.2Establishing the induction step

The claimed invertibility of 𝑆 ~ 𝑑 follows from induction and from Lemma 4. In fact, by (21) we know 𝑆 ~ 𝑑 1 is invertible, and by Lemma 4 we know that if 𝑆 ~ 𝑑 is invertible, 𝑆 ~ 𝑑 + 1 would also be invertible since 𝑆 ~ 𝑑 (resp. 𝑆 ~ 𝑑 + 1 ) has the same invertibility as ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 (resp. ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 + 1 ). For the rest of the proof we focus on establishing (22) by induction.

For the induction step we need to understand the one-step behaviors of β€– 𝑂 ~ 𝑑 β€– , β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– , and β€– 𝑆 ~ 𝑑 β€– , which are supplied by the following lemmas.

Lemma 23.

For any 𝑑 such that (22) holds,

β€– 𝑂 ~ 𝑑 + 1 β€– ≀ ( 1 + 1 12 ​ 𝐢 max ​ πœ… ​ πœ‚ ) ​ β€– 𝑂 ~ 𝑑 β€– .

(95) Lemma 24.

For any 𝑑 such that (22) holds, setting 𝑍 𝑑

Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) ​ Ξ£ ⋆ βˆ’ 1 , there exists some universal constant 𝐢 24

0 such that

β€– | 𝑁 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 βˆ’ 1 ​ Ξ£ ⋆ | β€– ≀ ( 1 βˆ’ πœ‚ 3 ​ ( β€– 𝑍 𝑑 β€– + πœ‚ ) ) ​ β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | ​ β€– + πœ‚ ​ 𝐢 24 ​ πœ… 6 𝑐 πœ† ​ β€– 𝑋 ⋆ β€– β€– ​ | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | β€– + πœ‚ ​ ( β€– 𝑂 ~ 𝑑 β€– 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ) 1 / 2 ​ β€– 𝑋 ⋆ β€– .

(96)

In particular, if 𝑐 3

100 ​ 𝐢 24 ​ ( 𝐢 3 . π‘Ž + 1 ) 4 ​ 𝑐 𝛿 / 𝑐 πœ† , then β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– ≀ 𝑐 3 ​ πœ… βˆ’ 𝐢 𝛿 / 2 ​ β€– 𝑋 ⋆ β€– implies β€– 𝑁 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 βˆ’ 1 ​ Ξ£ ⋆ β€– ≀ 𝑐 3 ​ πœ… βˆ’ 𝐢 𝛿 / 2 ​ β€– 𝑋 ⋆ β€– .

Lemma 25.

For any 𝑑 such that (22) holds,

β€– 𝑆 ~ 𝑑 + 1 β€– ≀ ( 1 βˆ’ πœ‚ 2 ) ​ β€– 𝑆 ~ 𝑑 β€– + 100 ​ 𝑐 πœ† βˆ’ 1 / 2 ​ πœ‚ ​ πœ… 3 ​ β€– 𝑋 ⋆ β€– .

(97)

In particular, if 𝐢 3 . π‘Ž

200 ​ 𝑐 πœ† βˆ’ 1 / 2 , then β€– 𝑆 ~ 𝑑 β€– ≀ 𝐢 3 . π‘Ž ​ πœ… 3 ​ β€– 𝑋 ⋆ β€– implies β€– 𝑆 ~ 𝑑 + 1 β€– ≀ 𝐢 3 . π‘Ž ​ πœ… 3 ​ β€– 𝑋 ⋆ β€– .

We now return to the induction step. Recall that we need to show (22a)–(22d) hold for 𝑑 + 1 . It is obvious that (22b)–(22d) hold for 𝑑 + 1 by the induction hypothesis and the above lemmas. It remains to prove (22a). To this end we distinguish two cases: 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 ) ≀ 1 / 3 and 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 )

1 / 3 . In the former case, (22a) for 𝑑 + 1 follows from Lemma 23 and Lemma 4 (to be proved in Appendix D.1), which imply (provided 𝐢 max β‰₯ 2 )

β€– 𝑂 ~ 𝑑 + 1 β€– 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 + 1 ) ≀ ( 1 + πœ‚ 4 ​ 𝐢 max ​ πœ… ) ( 1 + πœ‚ / 8 ) ​ β€– 𝑂 ~ 𝑑 β€– 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 ) ≀ β€– 𝑂 ~ 𝑑 β€– 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 ) ,

as desired. In the latter case where 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 )

1 / 3 , one may apply the first part of Lemma 4 to deduce that 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 + 1 ) β‰₯ 1 / 10 (given that πœ‚ ≀ 𝑐 πœ‚ for some sufficiently small constant 𝑐 πœ‚ ). This combined with (22b) for 𝑑 + 1 (already proved) yields desired inequality (22a) for 𝑑 + 1 , given our assumption (12c) on the smallness of 𝛼 . This completes the proof.

C.2.1Proof of Lemma 23

If π‘Ÿ

π‘Ÿ ⋆ , then we have β€– 𝑂 ~ 𝑑 β€–

0 for all 𝑑 β‰₯ 0 . The conclusion follows trivially. Therefore, we only consider the case when π‘Ÿ

π‘Ÿ ⋆ . By definition, we have

𝑂 ~ 𝑑 + 1

𝑁 𝑑 + 1 ​ 𝑉 𝑑 + 1 , βŸ‚

𝑁 𝑑 + 1 ​ 𝑉 𝑑 ​ 𝑉 𝑑 ⊀ ​ 𝑉 𝑑 + 1 , βŸ‚ + 𝑁 𝑑 + 1 ​ 𝑉 𝑑 , βŸ‚ ​ 𝑉 𝑑 , βŸ‚ ⊀ ​ 𝑉 𝑑 + 1 , βŸ‚

βˆ’ 𝑁 𝑑 + 1 ​ 𝑉 𝑑 ​ ( 𝑆 𝑑 + 1 ​ 𝑉 𝑑 ) βˆ’ 1 ​ 𝑆 𝑑 + 1 ​ 𝑉 𝑑 , βŸ‚ ​ 𝑉 𝑑 , βŸ‚ ⊀ ​ 𝑉 𝑑 + 1 , βŸ‚ + 𝑁 𝑑 + 1 ​ 𝑉 𝑑 , βŸ‚ ​ 𝑉 𝑑 , βŸ‚ ⊀ ​ 𝑉 𝑑 + 1 , βŸ‚ ,

where the last inequality uses the fact that 𝑉 𝑑 ⊀ ​ 𝑉 𝑑 + 1 , βŸ‚

βˆ’ ( 𝑆 𝑑 + 1 ​ 𝑉 𝑑 ) βˆ’ 1 ​ 𝑆 𝑑 + 1 ​ 𝑉 𝑑 , βŸ‚ ​ 𝑉 𝑑 , βŸ‚ ⊀ ​ 𝑉 𝑑 + 1 , βŸ‚ . To see this, note that

𝑆 𝑑 + 1 ​ 𝑉 𝑑 + 1 , βŸ‚

0 ⟹ 𝑆 𝑑 + 1 ​ 𝑉 𝑑 ​ 𝑉 𝑑 ⊀ ​ 𝑉 𝑑 + 1 , βŸ‚

βˆ’ 𝑆 𝑑 + 1 ​ 𝑉 𝑑 , βŸ‚ ​ 𝑉 𝑑 , βŸ‚ ⊀ ​ 𝑉 𝑑 + 1 , βŸ‚ .

Left-multiplying both sides by ( 𝑆 𝑑 + 1 ​ 𝑉 𝑑 ) βˆ’ 1 yields the desired identity. Note that the invertibility of 𝑆 𝑑 + 1 ​ 𝑉 𝑑 follows from the invertibility of 𝑆 ~ 𝑑 by inserting 𝑄

0 in Lemma 14.

By Lemma 13, we immediately obtain that 𝑆 𝑑 + 1 ​ 𝑉 𝑑 , βŸ‚

πœ‚ ​ 𝐸 𝑑 𝑏 ​ 𝑉 𝑑 , βŸ‚ , and 𝑁 𝑑 + 1 ​ 𝑉 𝑑 , βŸ‚

𝑂 ~ 𝑑 + πœ‚ ​ 𝐸 𝑑 𝑑 ​ 𝑉 𝑑 , βŸ‚ , where β€– 𝐸 𝑑 𝑏 β€– ∨ β€– 𝐸 𝑑 𝑑 β€– ≀ 1 24 ​ 𝐢 max ​ πœ… ​ β€– 𝑂 ~ 𝑑 β€– . Assume for now that

β€– 𝑁 𝑑 + 1 ​ 𝑉 𝑑 ​ ( 𝑆 𝑑 + 1 ​ 𝑉 𝑑 ) βˆ’ 1 β€– ≀ 1 .

(98)

In addition, notice that β€– 𝑉 𝑑 , βŸ‚ ⊀ ​ 𝑉 𝑑 + 1 , βŸ‚ β€– ≀ 1 since both factors are orthonormal matrices, we have

β€– 𝑂 ~ 𝑑 + 1 β€–

≀ β€– 𝑂 ~ 𝑑 β€– + πœ‚ ​ β€– 𝑁 𝑑 + 1 ​ 𝑉 𝑑 ​ ( 𝑆 𝑑 + 1 ​ 𝑉 𝑑 ) βˆ’ 1 β€– ​ β€– 𝐸 𝑑 𝑏 β€– + πœ‚ ​ β€– 𝐸 𝑑 𝑑 β€–

≀ ( 1 + 1 12 ​ 𝐢 max ​ πœ… ​ πœ‚ ) ​ β€– 𝑂 ~ 𝑑 β€– ,

as desired. It remains to prove (98).

Proof of bound (98).

This can be done by plugging 𝑄

0 into Lemma 15 and bounding the resulting expression. This (in fact, a much stronger inequality) will be done in detail in the proof of Lemma 24, to be presented soon in Section C.2.2. In fact, the resulting expression is the same as (103) there (albeit with different values of 𝐸 𝑑 14 . π‘Ž , 𝐸 𝑑 15 . π‘Ž , 𝐸 𝑑 15 . 𝑏 , which do not affect the proof). Following the same strategy to control (103) there, we may show that β€– 𝑁 𝑑 + 1 ​ 𝑉 𝑑 ​ ( 𝑆 𝑑 + 1 ​ 𝑉 𝑑 ) βˆ’ 1 ​ Ξ£ ⋆ β€– enjoys the same bound (108) as β€– 𝑁 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 βˆ’ 1 ​ Ξ£ ⋆ β€– , the right hand side of which is less than πœ… βˆ’ 1 ​ β€– 𝑋 ⋆ β€–

β€– Ξ£ ⋆ βˆ’ 1 β€– βˆ’ 1 given (22c) and (22d). Thus β€– 𝑁 𝑑 + 1 ​ 𝑉 𝑑 ​ ( 𝑆 𝑑 + 1 ​ 𝑉 𝑑 ) βˆ’ 1 β€– ≀ β€– 𝑁 𝑑 + 1 ​ 𝑉 𝑑 ​ ( 𝑆 𝑑 + 1 ​ 𝑉 𝑑 ) βˆ’ 1 ​ Ξ£ ⋆ β€– ​ β€– Ξ£ ⋆ βˆ’ 1 β€– ≀ 1 as claimed.

C.2.2Proof of Lemma 24

Denoting 𝑋 ~ 𝑑 ≔ 𝑋 𝑑 ​ 𝑉 𝑑 , we have 𝑁 ~ 𝑑

π‘ˆ ⋆ , βŸ‚ ⊀ ​ 𝑋 ~ 𝑑 and 𝑆 ~ 𝑑

π‘ˆ ⋆ ⊀ ​ 𝑋 ~ 𝑑 . Suppose for the moment that

β€– ( 𝑉 𝑑 ⊀ ​ 𝑉 𝑑 + 1 ) βˆ’ 1 β€– ≀ 2 ,

(99)

whose proof is deferred to the end of this section. We can write the update equation of 𝑋 ~ 𝑑 as

𝑋 ~ 𝑑 + 1

𝑋 𝑑 + 1 ​ 𝑉 𝑑 + 1

𝑋 𝑑 + 1 ​ 𝑉 𝑑 ​ 𝑉 𝑑 ⊀ ​ 𝑉 𝑑 + 1 + 𝑋 𝑑 + 1 ​ 𝑉 𝑑 , βŸ‚ ​ 𝑉 𝑑 , βŸ‚ ⊀ ​ 𝑉 𝑑 + 1

( 𝑋 𝑑 + 1 ​ 𝑉 𝑑 + 𝑋 𝑑 + 1 ​ 𝑉 𝑑 , βŸ‚ ​ 𝑉 𝑑 , βŸ‚ ⊀ ​ 𝑉 𝑑 + 1 ​ ( 𝑉 𝑑 ⊀ ​ 𝑉 𝑑 + 1 ) βˆ’ 1 ) ​ 𝑉 𝑑 ⊀ ​ 𝑉 𝑑 + 1 .

(100)

Left-multiplying both sides of (100) with π‘ˆ ⋆ , βŸ‚ (or π‘ˆ ⋆ ), we obtain

𝑁 ~ 𝑑 + 1

( 𝑁 𝑑 + 1 ​ 𝑉 𝑑 + 𝑁 𝑑 + 1 ​ 𝑉 𝑑 , βŸ‚ ​ 𝑄 ) ​ 𝑉 𝑑 ⊀ ​ 𝑉 𝑑 + 1 ,

(101a)

𝑆 ~ 𝑑 + 1

( 𝑆 𝑑 + 1 ​ 𝑉 𝑑 + 𝑆 𝑑 + 1 ​ 𝑉 𝑑 , βŸ‚ ​ 𝑄 ) ​ 𝑉 𝑑 ⊀ ​ 𝑉 𝑑 + 1 ,

(101b)

where we define 𝑄 ≔ 𝑉 𝑑 , βŸ‚ ⊀ ​ 𝑉 𝑑 + 1 ​ ( 𝑉 𝑑 ⊀ ​ 𝑉 𝑑 + 1 ) βˆ’ 1 . Consequently, we arrive at

𝑁 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 βˆ’ 1

( 𝑁 𝑑 + 1 ​ 𝑉 𝑑 + 𝑁 𝑑 + 1 ​ 𝑉 𝑑 , βŸ‚ ​ 𝑄 ) ​ ( 𝑆 𝑑 + 1 ​ 𝑉 𝑑 + 𝑆 𝑑 + 1 ​ 𝑉 𝑑 , βŸ‚ ​ 𝑄 ) βˆ’ 1 .

(102)

Since β€– 𝑄 β€– ≀ 2 (which is an immediate implication of (99)), we can invoke Lemma 15 to obtain

𝑁 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 βˆ’ 1 ​ Ξ£ ⋆

𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ ( 𝐼 + πœ‚ ​ 𝐸 𝑑 15 . π‘Ž ) ​ 𝐴 𝑑 ​ ( 𝐴 𝑑 + πœ‚ ​ Ξ£ ⋆ 2 ) βˆ’ 1 ​ ( 𝐼 + πœ‚ ​ 𝐸 𝑑 14 ) βˆ’ 1 ​ Ξ£ ⋆ + πœ‚ ​ 𝐸 𝑑 15 . 𝑏 ​ Ξ£ ⋆

𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ ​ ( 𝐼 + πœ‚ ​ Ξ£ ⋆ βˆ’ 1 ​ 𝐸 𝑑 15 . π‘Ž ​ Ξ£ ⋆ ) ​ 𝐻 𝑑 ​ ( 𝐻 𝑑 + πœ‚ ​ 𝐼 ) βˆ’ 1 ​ ( 𝐼 + πœ‚ ​ Ξ£ ⋆ βˆ’ 1 ​ 𝐸 𝑑 14 ​ Ξ£ ⋆ ) βˆ’ 1 + πœ‚ ​ 𝐸 𝑑 15 . 𝑏 ​ Ξ£ ⋆ ,

(103)

where for simplicity of notation, we denote

𝐴 𝑑 ≔ ( 1 βˆ’ πœ‚ ) ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 , and 𝐻 𝑑 ≔ Ξ£ ⋆ βˆ’ 1 ​ 𝐴 𝑑 ​ Ξ£ ⋆ βˆ’ 1 .

In addition, we have

β€– 𝐸 𝑑 14 β€– + β€– 𝐸 𝑑 15 . π‘Ž β€–

≀ 1 64 ​ πœ… 5 ,

β€– | 𝐸 𝑑 15 . 𝑏 | β€–

≀ 800 ​ 𝑐 πœ† βˆ’ 1 ​ πœ… 2 ​ β€– 𝑋 ⋆ β€– βˆ’ 2 ​ β€– | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | ​ β€– + 1 64 ​ ( 𝐢 3 . π‘Ž + 1 ) 2 ​ πœ… 5 ​ β€– 𝑋 ⋆ β€– β€– ​ | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€– + 1 64 ​ ( β€– 𝑂 ~ 𝑑 β€– 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ) 2 / 3 .

Moreover, it is clear that πœ‚ ≀ 𝑐 πœ‚ ≀ 1 ≀ πœ… 4 since πœ… β‰₯ 1 , and that β€– 𝐻 𝑑 β€– ≀ πœ… 2 ​ ( 1 + β€– 𝑆 ~ 𝑑 β€– 2 / β€– 𝑋 ⋆ β€– 2 ) ≀ ( 𝐢 3 . π‘Ž + 1 ) 2 ​ πœ… 4 . Hence we have

β€– 𝐻 𝑑 β€– + πœ‚ ≀ 2 ​ ( 𝐢 3 . π‘Ž + 1 ) 2 ​ πœ… 4

which implies

β€– 𝐸 𝑑 14 β€– + β€– 𝐸 𝑑 15 . π‘Ž β€– ≀ 1 24 ​ πœ… ​ 1 β€– 𝐻 𝑑 β€– + πœ‚ .

(104)

Similarly we may also show

β€– | 𝐸 𝑑 15 . 𝑏 | β€– ≀ 800 ​ 𝑐 πœ† βˆ’ 1 ​ πœ… 2 ​ β€– 𝑋 ⋆ β€– βˆ’ 2 ​ β€– | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | ​ β€– + 1 12 ​ ( β€– 𝐻 𝑑 β€– + πœ‚ ) ​ β€– 𝑋 ⋆ β€– β€– ​ | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€– + 1 2 ​ ( β€– 𝑂 ~ 𝑑 β€– 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ) 2 / 3 .

(105)

Since 𝐻 𝑑 is obviously positive definite, we have

β€– 𝐻 𝑑 ​ ( 𝐻 𝑑 + πœ‚ ​ 𝐼 ) βˆ’ 1 β€– ≀ 1 βˆ’ πœ‚ β€– 𝐻 𝑑 β€– + πœ‚ .

(106)

Thus

β€– | 𝑁 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 βˆ’ 1 ​ Ξ£ ⋆ | β€– ≀

( 1 βˆ’ πœ‚ β€– 𝐻 𝑑 β€– + πœ‚ ) ​ ( 1 βˆ’ πœ‚ ​ πœ… ​ β€– 𝐸 𝑑 14 β€– ) βˆ’ 1 ​ ( 1 + πœ‚ ​ πœ… ​ β€– 𝐸 𝑑 15 . π‘Ž β€– ) ​ β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | ​ β€– + πœ‚ β€– ​ | 𝐸 𝑑 15 . 𝑏 | β€– ​ β€– 𝑋 ⋆ β€– .

≀

( 1 βˆ’ πœ‚ β€– 𝐻 𝑑 β€– + πœ‚ ) ​ ( 1 + 1 12 ​ πœ‚ β€– 𝐻 𝑑 β€– + πœ‚ ) 2 ​ β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€–

  • πœ‚ ​ 800 ​ πœ… 2 𝑐 πœ† ​ β€– 𝑋 ⋆ β€– ​ β€– | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | ​ β€–
  • 1 12 ​ πœ‚ β€– 𝐻 𝑑 β€–
  • πœ‚ β€– ​ | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€–
  • 1 2 ​ πœ‚ ​ ( β€– 𝑂 ~ 𝑑 β€– 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ) 2 / 3 ​ β€– 𝑋 ⋆ β€–

≀

( 1 βˆ’ 5 6 ​ πœ‚ β€– 𝐻 𝑑 β€– + πœ‚ ) ​ β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | ​ β€– + 1 12 ​ πœ‚ β€– 𝐻 𝑑 β€– + πœ‚ β€– ​ | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€–

  • πœ‚ ​ 800 ​ πœ… 2 𝑐 πœ† ​ β€– 𝑋 ⋆ β€– ​ β€– | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | β€–
  • 1 2 ​ πœ‚ ​ ( β€– 𝑂 ~ 𝑑 β€– 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ) 2 / 3 ​ β€– 𝑋 ⋆ β€–

≀

( 1 βˆ’ 3 4 ​ πœ‚ β€– 𝐻 𝑑 β€– + πœ‚ ) ​ β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | ​ β€– + πœ‚ ​ 800 ​ πœ… 2 𝑐 πœ† ​ β€– 𝑋 ⋆ β€– β€– ​ | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | β€– + 1 2 ​ πœ‚ ​ ( β€– 𝑂 ~ 𝑑 β€– 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ) 2 / 3 ​ β€– 𝑋 ⋆ β€–

≀

( 1 βˆ’ 3 4 ​ πœ‚ β€– 𝑍 𝑑 β€– + πœ‚ ) ​ β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | ​ β€– + πœ‚ ​ 800 ​ πœ… 2 𝑐 πœ† ​ β€– 𝑋 ⋆ β€– β€– ​ | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | β€– + 1 2 ​ πœ‚ ​ ( β€– 𝑂 ~ 𝑑 β€– 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ) 2 / 3 ​ β€– 𝑋 ⋆ β€– ,

(107)

where in the second inequality we used ( 1 βˆ’ π‘₯ ) βˆ’ 1 ≀ 1 + π‘₯ for π‘₯ < 1 , in the penultimate inequality we used the elementary fact ( 1 βˆ’ π‘₯ ) ​ ( 1 + 1 16 ​ π‘₯ ) 2 ≀ 1 βˆ’ 5 6 ​ π‘₯ for π‘₯ ∈ [ 0 , 1 ] , and in the last inequality we used the obvious fact

β€– 𝐻 𝑑 β€–

β€– Ξ£ ⋆ βˆ’ 1 ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) ​ Ξ£ ⋆ βˆ’ 1 β€– ≀ β€– Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) ​ Ξ£ ⋆ βˆ’ 1 β€–

β€– 𝑍 𝑑 β€– .

The desired inequality (96) follows from the above inequality by setting 𝐢 24

800 .

For the remaining claim, we need to apply the conclusion of the first part with | | | β‹… | | |

βˆ₯ β‹… βˆ₯ . Then we note the following bounds:

(i)

β€– 𝑍 𝑑 β€– ≀ β€– Ξ£ ⋆ βˆ’ 1 β€– 2 ​ ( β€– 𝑆 ~ 𝑑 β€– 2 + πœ† ) ≀ ( 𝐢 3 . π‘Ž + 1 ) 2 ​ πœ… 4 by (22d) and (12b) (since we may choose 𝑐 πœ† ≀ 1 );

(ii)

πœ‚ ≀ 𝑐 πœ‚ ≀ ( 𝐢 3 . π‘Ž + 1 ) 2 ​ πœ… 4 ;

(iii)

β€– π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 β€– ≀ β€– Ξ” 𝑑 β€– ≀ 16 ​ ( 𝐢 3 . π‘Ž + 1 ) 2 ​ 𝑐 𝛿 ​ πœ… βˆ’ 2 ​ 𝐢 𝛿 / 3 ​ β€– 𝑋 ⋆ β€– 2 by Lemma 12;

(iv)

( β€– 𝑂 ~ 𝑑 β€– / 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ) 1 / 2 ≀ 𝑐 𝛿 ​ πœ… βˆ’ 2 ​ 𝐢 𝛿 / 3 by (22a), if we choose 𝐢 𝛼 β‰₯ 3 ​ 𝑐 𝛿 βˆ’ 1 + 3 ​ 𝐢 𝛿 + 3 .

These together imply

β€– 𝑁 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 βˆ’ 1 ​ Ξ£ ⋆ β€– ≀ ( 1 βˆ’ πœ‚ 6 ​ ( 𝐢 3 . π‘Ž + 1 ) 2 ​ πœ… 4 ) ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– + πœ‚ ​ 16 ​ 𝐢 24 ​ πœ… 2 𝑐 πœ† ​ ( 𝐢 3 . π‘Ž + 1 ) 2 ​ 𝑐 𝛿 ​ πœ… βˆ’ 2 ​ 𝐢 𝛿 / 3 ​ β€– 𝑋 ⋆ β€– + πœ‚ ​ 𝑐 𝛿 ​ πœ… βˆ’ 2 ​ 𝐢 𝛿 / 3 ​ β€– 𝑋 ⋆ β€– .

(108)

The conclusion follows easily by plugging in β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– ≀ 𝑐 3 ​ πœ… βˆ’ 𝐢 𝛿 / 2 ​ β€– 𝑋 ⋆ β€– and using πœ… 6 ​ πœ… βˆ’ 2 ​ 𝐢 𝛿 / 3 ≀ πœ… βˆ’ 𝐢 𝛿 / 2 when 𝐢 𝛿 is sufficiently large.

Proof of bound (99).

First, we observe that it is equivalent to show that 𝜎 min ​ ( 𝑉 𝑑 ⊀ ​ 𝑉 𝑑 + 1 ) β‰₯ 1 / 2 . But from 𝑉 𝑑 + 1 ​ 𝑉 𝑑 + 1 ⊀ + 𝑉 𝑑 + 1 , βŸ‚ ​ 𝑉 𝑑 + 1 , βŸ‚ ⊀

𝐼 we have

𝜎 min ​ ( 𝑉 𝑑 ⊀ ​ 𝑉 𝑑 + 1 )

𝜎 π‘Ÿ ⋆ ​ ( 𝑉 𝑑 ⊀ ​ 𝑉 𝑑 + 1 ) β‰₯ 𝜎 π‘Ÿ ⋆ ​ ( 𝑉 𝑑 ⊀ ​ 𝑉 𝑑 + 1 ​ 𝑉 𝑑 + 1 ⊀ )

𝜎 π‘Ÿ ⋆ ​ ( 𝑉 𝑑 ⊀ βˆ’ 𝑉 𝑑 ⊀ ​ 𝑉 𝑑 + 1 , βŸ‚ ​ 𝑉 𝑑 + 1 , βŸ‚ ⊀ )

β‰₯ 𝜎 π‘Ÿ ⋆ ​ ( 𝑉 𝑑 ⊀ ) βˆ’ β€– 𝑉 𝑑 ⊀ ​ 𝑉 𝑑 + 1 , βŸ‚ ​ 𝑉 𝑑 + 1 , βŸ‚ ⊀ β€–

β‰₯ 1 βˆ’ β€– 𝑉 𝑑 ⊀ ​ 𝑉 𝑑 + 1 , βŸ‚ β€– ,

where the last inequality follows from 𝜎 π‘Ÿ ⋆ ​ ( 𝑉 𝑑 ⊀ )

1 (since 𝑉 𝑑 ∈ ℝ π‘Ÿ Γ— π‘Ÿ ⋆ is orthonormal) and from that β€– 𝑉 𝑑 ⊀ ​ 𝑉 𝑑 + 1 , βŸ‚ ​ 𝑉 𝑑 + 1 , βŸ‚ ⊀ β€– ≀ β€– 𝑉 𝑑 ⊀ ​ 𝑉 𝑑 + 1 , βŸ‚ β€– . This implies that, to show 𝜎 min ​ ( 𝑉 𝑑 ⊀ ​ 𝑉 𝑑 + 1 ) β‰₯ 1 / 2 , it suffices to prove β€– 𝑉 𝑑 ⊀ ​ 𝑉 𝑑 + 1 , βŸ‚ β€– ≀ 1 / 2 .

Next we prove that β€– 𝑉 𝑑 ⊀ ​ 𝑉 𝑑 + 1 , βŸ‚ β€– ≀ 1 / 2 . Recall that by definition we have 𝑆 𝑑 + 1 ​ 𝑉 𝑑 + 1 , βŸ‚

0 . Right-multiplying both sides of (53a) by 𝑉 𝑑 + 1 , βŸ‚ , we obtain

0

( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 + 𝐸 𝑑 π‘Ž ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ) ​ 𝑆 ~ 𝑑 ​ ( 𝑉 𝑑 ⊀ ​ 𝑉 𝑑 + 1 , βŸ‚ ) + πœ‚ ​ 𝐸 𝑑 𝑏 ​ 𝑉 𝑑 + 1 , βŸ‚ ,

hence

β€– 𝑉 𝑑 ⊀ ​ 𝑉 𝑑 + 1 , βŸ‚ β€– ≀ πœ‚ ​ β€– 𝐸 𝑑 𝑏 ​ 𝑉 𝑑 + 1 , βŸ‚ β€– ​ β€– 𝑆 ~ 𝑑 βˆ’ 1 β€– ​ β€– ( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 + 𝐸 𝑑 π‘Ž ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ) βˆ’ 1 β€– .

By (54b) we have

β€– 𝐸 𝑑 𝑏 ​ 𝑉 𝑑 + 1 , βŸ‚ β€– ​ β€– 𝑆 ~ 𝑑 βˆ’ 1 β€– ≀ β€– 𝐸 𝑑 𝑏 β€– 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ≀ 1 10 ​ πœ… ,

thus it suffices to show

πœ‚ ​ β€– ( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 + 𝐸 𝑑 π‘Ž ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ) βˆ’ 1 β€– ≀ 5 ​ πœ… ,

(109)

or equivalently,

𝜎 min ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 + 𝐸 𝑑 π‘Ž ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ) β‰₯ πœ‚ 5 ​ πœ… .

(110)

To this end, we write

( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 + 𝐸 𝑑 π‘Ž ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1

( 𝐼 + πœ‚ ​ 𝐸 𝑑 π‘Ž ​ ( ( 1 βˆ’ πœ‚ ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) ) βˆ’ 1 ) ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 )

(111)

and control the two terms separately.

β€’

To control the first factor, starting from (54a) we may deduce

β€– 𝐸 𝑑 π‘Ž β€–
≀ πœ… βˆ’ 4 ​ β€– 𝑋 ⋆ β€– ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– + β€– π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 β€–

≀ πœ… βˆ’ 4 ​ β€– 𝑋 ⋆ β€– ​ 𝑐 3 ​ πœ… βˆ’ 𝐢 𝛿 / 2 ​ β€– 𝑋 ⋆ β€– + 𝑐 12 ​ πœ… βˆ’ 2 ​ 𝐢 𝛿 / 3 ​ β€– 𝑋 ⋆ β€– 2

≀ πœ… βˆ’ 2 ​ β€– 𝑋 ⋆ β€– 2 / 2

𝜎 min 2 ​ ( 𝑋 ⋆ ) / 2 ,

where the second inequality follows from (22c) and Lemma 12; the last inequality follows from choosing 𝑐 𝛿 sufficiently small (recall that 𝑐 3 , 𝑐 12 ≲ 𝑐 𝛿 / 𝑐 πœ† ) and 𝐢 𝛿 sufficiently large. Furthermore, since 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ is positive semidefinite, we have

β€– ( ( 1 βˆ’ πœ‚ ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) ) βˆ’ 1 β€– ≀ πœ‚ βˆ’ 1 ​ 𝜎 min βˆ’ 2 ​ ( Ξ£ ⋆ )

πœ‚ βˆ’ 1 ​ 𝜎 min βˆ’ 2 ​ ( 𝑋 ⋆ ) ,

hence

𝜎 min ​ ( 1 + πœ‚ ​ 𝐸 𝑑 π‘Ž ​ ( ( 1 βˆ’ πœ‚ ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) ) βˆ’ 1 )

β‰₯ 1 βˆ’ πœ‚ ​ β€– 𝐸 𝑑 π‘Ž β€– ​ β€– ( ( 1 βˆ’ πœ‚ ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) ) βˆ’ 1 β€–

β‰₯ 1 βˆ’ πœ‚ β‹… 𝜎 min 2 ​ ( 𝑋 ⋆ ) 2 β‹… πœ‚ βˆ’ 1 ​ 𝜎 min βˆ’ 2 ​ ( 𝑋 ⋆ )

1 / 2 .

(112) β€’

Now we control the second factor. By Lemma 10 we have

𝜎 min ​ ( 1 βˆ’ πœ‚ + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 )

( 1 βˆ’ πœ‚ ) ​ 𝜎 min ​ ( 𝐼 + πœ‚ 1 βˆ’ πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 )

β‰₯ ( 1 βˆ’ πœ‚ ) ​ ( β€– Ξ£ ⋆ 2 + πœ† ​ 𝐼 β€– 𝜎 min ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) ) βˆ’ 1 / 2

( 1 βˆ’ πœ‚ ) ​ ( β€– 𝑋 ⋆ β€– 2 + πœ† 𝜎 min 2 ​ ( 𝑋 ⋆ ) + πœ† ) βˆ’ 1 / 2 .

It is easy to check that the function πœ† ↦ ( π‘Ž + πœ† ) / ( 𝑏 + πœ† ) is decreasing on [ 0 , ∞ ) for π‘Ž β‰₯ 𝑏

0 , thus

β€– 𝑋 ⋆ β€– 2 + πœ† 𝜎 min 2 ​ ( 𝑋 ⋆ ) + πœ† ≀ β€– 𝑋 ⋆ β€– 2 𝜎 min 2 ​ ( 𝑋 ⋆ )

πœ… 2 ,

which implies

𝜎 min ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ) β‰₯ 1 βˆ’ πœ‚ πœ… .

(113)

Plugging (113) and (β€’ β€£ C.2.2) into (111) yields

𝜎 min ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 + 𝐸 𝑑 π‘Ž ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ) β‰₯ 1 βˆ’ πœ‚ 2 ​ πœ… β‰₯ πœ‚ 5 ​ πœ… ,

(114)

where the last inequality follows from the assumption πœ‚ ≀ 𝑐 πœ‚ . This shows (110) as desired, thereby completing the proof.

C.2.3Proof of Lemma 25

Combine (101b) and Lemma 14 to see that

β€– 𝑆 ~ 𝑑 + 1 β€–

≀ β€– 𝑆 𝑑 + 1 ​ 𝑉 𝑑 + 𝑆 𝑑 + 1 ​ 𝑉 𝑑 , βŸ‚ ​ 𝑄 β€–

≀ β€– 1 + πœ‚ ​ 𝐸 𝑑 14 β€– β‹… β€– ( 1 βˆ’ πœ‚ ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) 1 / 2 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 / 2 β€– β‹… β€– ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 β€–

≀ ( 1 + πœ‚ ​ β€– 𝐸 𝑑 14 β€– ) ​ ( ( 1 βˆ’ πœ‚ ) ​ ( β€– 𝑆 ~ 𝑑 β€– 2 + πœ† ) 1 / 2 + 4 ​ πœ‚ ​ πœ† βˆ’ 1 / 2 ​ β€– 𝑋 ⋆ β€– 2 ) ​ ( β€– 𝑆 ~ 𝑑 β€– 2 + πœ† ) βˆ’ 1 / 2 ​ β€– 𝑆 ~ 𝑑 β€–

≀ ( 1 + πœ‚ 4 ) ​ ( ( 1 βˆ’ πœ‚ ) ​ β€– 𝑆 ~ 𝑑 β€– + 4 ​ πœ‚ ​ β€– 𝑋 ⋆ β€– 2 ​ β€– 𝑆 ~ 𝑑 β€– πœ† ​ ( β€– 𝑆 ~ 𝑑 β€– 2 + πœ† ) )

≀ ( 1 βˆ’ πœ‚ 2 ) ​ β€– 𝑆 ~ 𝑑 β€– + 5 ​ πœ‚ ​ β€– 𝑋 ⋆ β€– 2 πœ† ,

(115)

where the third line follows from β€– Ξ£ ⋆ 2 + πœ† ​ 𝐼 β€– ≀ ( 1 + πœ† ) ​ β€– 𝑋 ⋆ β€– 2 ≀ 2 ​ β€– 𝑋 ⋆ β€– 2 assuming 𝑐 πœ† ≀ 1 and from the fact that the singular values of ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 are ( 𝜎 𝑗 2 ​ ( 𝑆 ~ 𝑑 ) + πœ† ) βˆ’ 1 / 2 ​ 𝜎 𝑗 ​ ( 𝑆 ~ 𝑑 ) , 𝑗

1 , … , π‘Ÿ ⋆ ,4 which is bounded by ( β€– 𝑆 ~ 𝑑 β€– 2 + πœ† ) βˆ’ 1 / 2 ​ β€– 𝑆 ~ 𝑑 β€– since 𝜎 ↦ ( 𝜎 2 + πœ† ) βˆ’ 1 / 2 ​ 𝜎 is increasing and since β€– 𝑆 ~ 𝑑 β€– is the largest singular value of 𝑆 ~ 𝑑 . In the fourth line, we used the error bound β€– 𝐸 𝑑 14 β€– ≀ 1 / 4 and the last line follows from the elementary inequalities 1 + πœ‚ / 4 ≀ ( 1 βˆ’ πœ‚ / 2 ) ​ ( 1 βˆ’ πœ‚ ) βˆ’ 1 ≀ 5 / 4 given that πœ‚ ≀ 𝑐 πœ‚ for sufficiently small constant 𝑐 πœ‚

0 . The conclusion readily follows from the above inequality and the assumption πœ† β‰₯ 1 100 ​ πœ… βˆ’ 4 ​ 𝑐 πœ† ​ 𝜎 min 2 ​ ( 𝑋 ⋆ ) .

Appendix DProofs for Phase II

This section collects the proofs for Phase II.

D.1Proof of Lemma 4

Since β€– 𝑉 𝑑 + 1 ⊀ ​ 𝑉 𝑑 β€– ≀ 1 , we have

𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 + 1 )
β‰₯ 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 + 1 ​ 𝑉 𝑑 + 1 ⊀ ​ 𝑉 𝑑 )

𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 𝑑 + 1 ​ 𝑉 𝑑 ) ,

where the second equality follows from 𝑆 𝑑 + 1

𝑆 ~ 𝑑 + 1 ​ 𝑉 𝑑 + 1 ⊀ (cf. (31)). Apply Lemma 14 with 𝑄

0 to see that

𝑆 𝑑 + 1 ​ 𝑉 𝑑

( 𝐼 + πœ‚ ​ 𝐸 𝑑 14 ) ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ) ​ 𝑆 ~ 𝑑 ,

(116)

where 𝐸 𝑑 14 ∈ ℝ π‘Ÿ ⋆ Γ— π‘Ÿ ⋆ satisfies β€– 𝐸 𝑑 14 β€– ≀ 1 200 ​ ( 𝐢 3 . π‘Ž + 1 ) 4 ​ πœ… 5 . To simplify the notation, we denote

π‘Œ 𝑑 ≔ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 ,

which allows us to write (116) as

( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 𝑑 + 1 ​ 𝑉 𝑑

( 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝐸 𝑑 14 ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) 1 / 2 ) ​ ( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( π‘Œ 𝑑 ​ π‘Œ 𝑑 ⊀ + πœ† ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 ) βˆ’ 1 ) ​ π‘Œ 𝑑 .

(117)

Note that

β€– ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝐸 𝑑 14 ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) 1 / 2 β€–

≀ β€– ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 β€– β‹… β€– ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) 1 / 2 β€– β‹… β€– 𝐸 𝑑 14 β€–

≀ πœ… ​ β€– 𝑋 ⋆ β€– βˆ’ 1 β‹… ( 2 ​ β€– 𝑋 ⋆ β€– ) β‹… β€– 𝐸 𝑑 14 β€–

≀ 2 ​ πœ… β‹… 1 200 ​ ( 𝐢 3 . π‘Ž + 1 ) 4 ​ πœ… 5 ≀ 1 / 32 ,

(118)

where in the second inequality we used πœ† ≀ 𝑐 πœ† ​ β€– 𝑀 ⋆ β€– ≀ β€– 𝑋 ⋆ β€– 2 as 𝑐 πœ† ≀ 1 , and in the third inequality we used the claimed bound of β€– 𝐸 𝑑 14 β€– . Therefore, it follows that

𝜎 min ​ ( 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝐸 𝑑 14 ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) 1 / 2 ) β‰₯ 1 βˆ’ πœ‚ / 32 .

(119)

On the other hand, using 𝜎 min ​ ( 𝐴 ​ 𝐡 ) β‰₯ 𝜎 min ​ ( 𝐴 ) ​ 𝜎 min ​ ( 𝐡 ) for any matrices 𝐴 , 𝐡 , it is obvious that

𝜎 min ​ ( ( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( π‘Œ 𝑑 ​ π‘Œ 𝑑 ⊀ + πœ† ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 ) βˆ’ 1 ) ​ π‘Œ 𝑑 ) β‰₯ ( 1 βˆ’ πœ‚ ) ​ 𝜎 min ​ ( π‘Œ 𝑑 ) ,

which in turn implies that

𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 𝑑 + 1 ​ 𝑉 𝑑 ) β‰₯ ( 1 βˆ’ πœ‚ / 32 ) ​ ( 1 βˆ’ πœ‚ ) ​ 𝜎 min ​ ( π‘Œ 𝑑 ) β‰₯ ( 1 βˆ’ 2 ​ πœ‚ ) ​ 𝜎 min ​ ( π‘Œ 𝑑 ) ,

as long as πœ‚ ≀ 𝑐 πœ‚ for some sufficiently small constant 𝑐 πœ‚ . This proves the first part of Lemma 4.

Now we move to the second part assuming 𝜎 min ​ ( π‘Œ 𝑑 ) ≀ 1 / 3 . Using the assumption πœ† ≀ 𝑐 πœ† ​ 𝜎 min ​ ( 𝑀 ⋆ ) , we see that

β€– πœ† ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 β€– ≀ 𝑐 πœ† .

Given that 𝑐 πœ† is sufficiently small (such that 𝑐 πœ† ≀ 𝑐 11 , where 𝑐 11 is the positive constant in Lemma 11), one may apply Lemma 11 with π‘Œ

π‘Œ 𝑑 and Ξ›

πœ† ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 to obtain

𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 𝑑 + 1 ​ 𝑉 𝑑 )

β‰₯ 𝜎 min ​ ( 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝐸 𝑑 14 ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) 1 / 2 ) ​ ( 1 + 1 6 ​ πœ‚ ) ​ 𝜎 min ​ ( π‘Œ 𝑑 )

β‰₯ ( i ) ​ ( 1 βˆ’ πœ‚ / 32 ) ​ ( 1 + 1 6 ​ πœ‚ ) ​ 𝜎 min ​ ( π‘Œ 𝑑 ) ​ β‰₯ ( ii ) ​ ( 1 + 1 8 ​ πœ‚ ) ​ 𝜎 min ​ ( π‘Œ 𝑑 ) ,

where (i) uses (119), and (ii) follows as long as πœ‚ ≀ 𝑐 πœ‚ for some sufficiently small constant 𝑐 πœ‚ . The desired conclusion follows.

D.2Proof of Corollary 1

We will prove a strengthened version of (25), that is

𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 ) β‰₯ 1 / 10 .

(120)

It is clear that (120) implies (25). Indeed, for each 𝑒 ∈ ℝ π‘Ÿ ⋆ , by taking 𝑣

( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) 1 / 2 ​ 𝑒 , we have

𝑒 ⊀ ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ ​ 𝑒

𝑣 ⊀ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑣 β‰₯ 1 10 ​ β€– 𝑣 β€– 2 β‰₯ 1 10 ​ 𝑒 ⊀ ​ Ξ£ ⋆ 2 ​ 𝑒 ,

which implies (25). It then boils down to establish (120).

Step 1: establishing the claim for a midpoint 𝑑 2 .

From Lemma 3 we know that

𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 1 ) β‰₯ β€– Ξ£ ⋆ 2 + πœ† ​ 𝐼 β€– βˆ’ 1 / 2 ​ 𝜎 min ​ ( 𝑆 ~ 𝑑 1 ) ​ β‰₯ ( i ) ​ ( 𝑐 πœ† + 1 ) βˆ’ 1 / 2 ​ β€– 𝑋 ⋆ β€– βˆ’ 1 β‹… 𝛼 2 / β€– 𝑋 ⋆ β€– β‰₯ 1 3 ​ ( 𝛼 / β€– 𝑋 ⋆ β€– ) 2 ,

where (i) follows from the assumption (12b) and Lemma 3, and the last inequality follows by choosing 𝑐 πœ† ≀ 1 . By the second part of Lemma 4, starting from 𝑑 1 , whenever 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 ) < 1 / 10 < 1 / 3 , it would increase exponentially with rate at least ( 1 + πœ‚ 8 ) . On the other end, it is easy to verify, given that πœ‚ ≀ 𝑐 πœ‚ is sufficiently small,

( 1 + πœ‚ 8 ) 16 πœ‚ ​ log ⁑ ( 3 10 ​ β€– 𝑋 ⋆ β€– 2 𝛼 2 ) β‰₯ 3 ​ β€– 𝑋 ⋆ β€– 2 10 ​ 𝛼 2 β‰₯ 1 10 ​ 1 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 1 ) .

Therefore, it takes at most 16 πœ‚ ​ log ⁑ ( 3 10 ​ β€– 𝑋 ⋆ β€– 2 𝛼 2 ) ≀ 𝑇 min / 16 more iterations to make 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 ) grow to at least 1 / 10 . Equivalent, for some 𝑑 2 : 𝑑 1 ≀ 𝑑 2 ≀ 𝑑 1 + 𝑇 min / 16 , we have

𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 2 ) β‰₯ 1 / 10 .

Step 2: establishing the claim for all 𝑑 ∈ [ 𝑑 2 , 𝑇 max ] .

It remains to show that (120) continues to hold for all 𝑑 ∈ [ 𝑑 2 , 𝑇 max ] . We prove this by induction on 𝑑 .

Assume that (120) holds for some 𝑑 ∈ [ 𝑑 2 , 𝑇 max βˆ’ 1 ] . We show that it will also hold for 𝑑 + 1 . We divide the proof into two cases.

Case 1.

If 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 ) ≀ 1 / 3 , we deduce from the second part of Lemma 4 that

𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 + 1 ) β‰₯ ( 1 + πœ‚ 8 ) ​ 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 ) β‰₯ 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 ) ,

which by the induction hypothesis is no less than 1 / 10 , as desired.

Case 2.

If 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 )

1 / 3 , the first part of Lemma 4 yields

𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 + 1 ) β‰₯ ( 1 βˆ’ 2 ​ πœ‚ ) ​ 𝜎 min ​ ( ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 ) β‰₯ ( 1 βˆ’ 2 ​ πœ‚ ) / 3 ,

which is greater than 1 / 10 provided πœ‚ ≀ 𝑐 πœ‚ ≀ 1 / 100 , as desired.

Combining the two cases completes the proof.

D.3Proof of Lemma 5

For simplicity, in this section we denote

Ξ“ 𝑑 ≔ Ξ£ ⋆ βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ ​ Ξ£ ⋆ βˆ’ 1 βˆ’ 𝐼

Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 .

(121)

It turns out that Lemma 5 follows naturally from the following technical lemma, whose proof is deferred to the end of this section.

Lemma 26.

For any 𝑑 : 𝑑 2 ≀ 𝑑 ≀ 𝑇 max , one has

β€– | Ξ“ 𝑑 + 1 | β€– ≀ ( 1 βˆ’ πœ‚ ) ​ β€– | Ξ“ 𝑑 | ​ β€– + πœ‚ ​ 𝐢 26 ​ πœ… 6 β€– 𝑋 ⋆ β€– 2 β€– ​ | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | β€– + 1 16 ​ πœ‚ ​ β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€– + πœ‚ ​ ( β€– 𝑂 ~ 𝑑 β€– β€– 𝑋 ⋆ β€– ) 7 / 12 ,

(122)

where 𝐢 26 ≲ 𝑐 πœ† βˆ’ 1 / 2 is some positive constant and | | | β‹… | | | can either be the Frobenius norm or the spectral norm.

From Lemma 12, we know that β€– π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 β€– ≀ β€– Ξ” 𝑑 β€– ≀ β€– 𝑋 ⋆ β€– 2 300 ​ 𝐢 26 ​ πœ… 4 as 𝑐 𝛿 is sufficiently small. Similarly, β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– ≀ β€– 𝑋 ⋆ β€– / 100 and ( β€– 𝑂 ~ 𝑑 β€– / β€– 𝑋 ⋆ β€– ) 7 / 12 ≀ 1 / 300 by Lemma 3. Applying Lemma 26 with the spectral norm, we prove Lemma 5 as desired.

Proof of Lemma 26.

We start by rewriting (53a) as

𝑆 𝑑 + 1

( ( 1 βˆ’ πœ‚ ) ​ 𝐼 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ) ​ 𝑆 ~ 𝑑 ​ 𝑉 𝑑 ⊀ + πœ‚ ​ 𝐸 𝑑 𝑔

( 𝐼 βˆ’ πœ‚ ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 + πœ‚ ​ ( Ξ£ ⋆ 2 + πœ† ​ 𝐼 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ) ​ 𝑆 ~ 𝑑 ​ 𝑉 𝑑 ⊀ + πœ‚ ​ 𝐸 𝑑 𝑔

( 𝐼 βˆ’ πœ‚ ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ) ​ 𝑆 ~ 𝑑 ​ 𝑉 𝑑 ⊀ + πœ‚ ​ 𝐸 𝑑 𝑔 ,

(123)

where

𝐸 𝑑 𝑔

𝐸 𝑑 π‘Ž ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑉 𝑑 ⊀ + 𝐸 𝑑 𝑏 .

(124)

By Corollary 1, we have 𝜎 min ​ ( 𝑆 ~ 𝑑 ) 2 β‰₯ 1 100 ​ 𝜎 min ​ ( 𝑀 ⋆ ) for 𝑑 ∈ [ 𝑑 2 , 𝑇 max ] , so

β€– ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑉 𝑑 ⊀ β€– ≀ β€– ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 / 2 β€– ​ β€– ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 / 2 ​ 𝑆 ~ 𝑑 β€– ≀ 𝜎 min βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ) ≲ 1 / 𝜎 min ​ ( 𝑋 ⋆ ) .

Combined with the error bounds (54a), (54b), we have for some universal constant 𝐢

0 that

β€– | 𝐸 𝑑 𝑔 | β€– ≀ β€– | 𝐸 𝑑 π‘Ž | ​ β€– + πœ‚ β€– ​ | 𝐸 𝑑 𝑏 | β€– ≀ 𝐢 ​ πœ… β€– 𝑋 ⋆ β€– ​ β€– | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | ​ β€– + 𝐢 ​ 𝑐 13 ​ πœ… βˆ’ 5 β€– ​ | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€– + 𝐢 ​ β€– 𝑂 ~ 𝑑 β€– 3 / 4 ​ β€– 𝑋 ⋆ β€– 1 / 4 .

(125) Step 1: deriving a recursion of Ξ“ 𝑑 .

Define

𝐴 𝑑 := ( 𝐼 βˆ’ πœ‚ ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) βˆ’ 1 ) ​ 𝑆 ~ 𝑑 ​ 𝑉 𝑑 ⊀ .

Then we can rewrite (123) as 𝐴 𝑑

𝑆 𝑑 + 1 βˆ’ πœ‚ ​ 𝐸 𝑑 𝑔 , and by rearranging 𝐴 𝑑 ​ 𝐴 𝑑 ⊀

( 𝑆 𝑑 + 1 βˆ’ πœ‚ ​ 𝐸 𝑑 𝑔 ) ​ ( 𝑆 𝑑 + 1 βˆ’ πœ‚ ​ 𝐸 𝑑 𝑔 ) ⊀ in view of (31), it follows that

𝑆 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 ⊀

𝑆 𝑑 + 1 ​ 𝑆 𝑑 + 1 ⊀

𝐴 𝑑 ​ 𝐴 𝑑 ⊀ + πœ‚ ​ ( β€– 𝑆 𝑑 + 1 β€– + β€– 𝐸 𝑑 𝑔 β€– ) ​ ( 𝐸 𝑑 𝑔 ​ 𝑄 1 + 𝑄 2 ​ 𝐸 𝑑 𝑔 ⊀ )

: 𝐴 𝑑 𝐴 𝑑 ⊀ + πœ‚ 𝐸 𝑑 𝑓

for some matrices 𝑄 1 , 𝑄 2 with β€– 𝑄 1 β€– , β€– 𝑄 2 β€– ≀ 1 . By mapping both sides of the above equation by ( β‹… ) ↦ Ξ£ ⋆ βˆ’ 1 ​ ( β‹… ) ​ Ξ£ ⋆ βˆ’ 1 βˆ’ 𝐼 , we obtain

Ξ“ 𝑑 + 1

( 𝐼 βˆ’ πœ‚ ​ Ξ“ 𝑑 ​ ( 𝐼 + Ξ“ 𝑑 + πœ† ​ Ξ£ ⋆ βˆ’ 2 ) βˆ’ 1 ) ​ ( Ξ“ 𝑑 + 𝐼 ) ​ ( 𝐼 βˆ’ πœ‚ ​ ( 𝐼 + Ξ“ 𝑑 + πœ† ​ Ξ£ ⋆ βˆ’ 2 ) βˆ’ 1 ​ Ξ“ 𝑑 ) βˆ’ 𝐼 + πœ‚ ​ Ξ£ ⋆ βˆ’ 1 ​ 𝐸 𝑑 𝑓 ​ Ξ£ ⋆ βˆ’ 1 ,

(126)

where we recall the definition of Ξ“ 𝑑 in (121).

Step 2: simplify the recursion.

Note that 𝜎 min ​ ( Ξ£ ⋆ βˆ’ 1 ​ 𝑆 ~ 𝑑 ) β‰₯ 1 / 10 implies 𝐼 + Ξ“ 𝑑 βͺ° 1 100 ​ 𝐼 . From our assumption πœ† ≀ 𝑐 πœ† ​ 𝜎 min ​ ( 𝑀 ⋆ ) , it follows that β€– πœ† ​ Ξ£ ⋆ βˆ’ 2 β€– ≀ 𝑐 πœ† ≀ 1 / 200 ≀ 1 2 ​ 𝜎 min ​ ( 𝐼 + Ξ“ 𝑑 ) , thus in virtue of Lemma 9 we have

( 𝐼 + Ξ“ 𝑑 + πœ† ​ Ξ£ ⋆ βˆ’ 2 ) βˆ’ 1

( 𝐼 + Ξ“ 𝑑 ) βˆ’ 1 + ( 𝐼 + Ξ“ 𝑑 ) βˆ’ 1 ​ ( 𝑐 πœ† ​ 𝑄 β€² ) ​ ( 𝐼 + Ξ“ 𝑑 ) βˆ’ 1 ,

for some matrix 𝑄 β€² with β€– 𝑄 β€² β€– ≀ 2 . Plugging this into (126) yields

Ξ“ 𝑑 + 1

( 𝐼 βˆ’ πœ‚ ​ Ξ“ 𝑑 ​ ( 𝐼 + Ξ“ 𝑑 ) βˆ’ 1 ) ​ ( Ξ“ 𝑑 + 𝐼 ) ​ ( 𝐼 βˆ’ πœ‚ ​ ( 𝐼 + Ξ“ 𝑑 ) βˆ’ 1 ​ Ξ“ 𝑑 ) + πœ‚ ​ 𝐸 𝑑 β„Ž + πœ‚ ​ Ξ£ ⋆ βˆ’ 1 ​ 𝐸 𝑑 𝑓 ​ Ξ£ ⋆ βˆ’ 1

( 1 βˆ’ 2 ​ πœ‚ ) ​ Ξ“ 𝑑 + πœ‚ 2 ​ Ξ“ 𝑑 2 ​ ( 1 + Ξ“ 𝑑 ) βˆ’ 1 + πœ‚ ​ 𝐸 𝑑 β„Ž + πœ‚ ​ Ξ£ ⋆ βˆ’ 1 ​ 𝐸 𝑑 𝑓 ​ Ξ£ ⋆ βˆ’ 1 ,

(127)

where the additional error term 𝐸 𝑑 β„Ž is defined by

𝐸 𝑑 β„Ž :=

Ξ“ 𝑑 ​ ( 𝐼 + Ξ“ 𝑑 ) βˆ’ 1 ​ ( 𝑐 πœ† ​ 𝑄 β€² ) ​ ( 1 βˆ’ πœ‚ ​ Ξ“ 𝑑 ​ ( 𝐼 + Ξ“ 𝑑 ) βˆ’ 1 ) + ( 1 βˆ’ πœ‚ ​ Ξ“ 𝑑 ​ ( 𝐼 + Ξ“ 𝑑 ) βˆ’ 1 ) ​ ( 𝑐 πœ† ​ 𝑄 β€² ) ​ ( 𝐼 + Ξ“ 𝑑 ) βˆ’ 1 ​ Ξ“ 𝑑

  • πœ‚ ​ Ξ“ 𝑑 ​ ( 𝐼
  • Ξ“ 𝑑 ) βˆ’ 1 ​ ( 𝑐 πœ† ​ 𝑄 β€² ) ​ ( 𝐼
  • Ξ“ 𝑑 ) βˆ’ 2 ​ ( 𝑐 πœ† ​ 𝑄 β€² ) ​ ( 𝐼
  • Ξ“ 𝑑 ) βˆ’ 1 ​ Ξ“ 𝑑 .

(128) Step 3: controlling the error terms.

We now control the error terms in (127) separately.

β€’

By (22d) we have β€– 𝑆 𝑑 + 1 β€– ≀ 𝐢 3 . π‘Ž ​ πœ… ​ β€– 𝑋 ⋆ β€– , and by controlling the right hand side of (125) using (22c), (24), and (50) in Lemma 12, it is evident that β€– 𝐸 𝑑 𝑔 β€– ≀ πœ… ​ β€– 𝑋 ⋆ β€– . Hence, the term 𝐸 𝑑 𝑓 obeys

β€– | 𝐸 𝑑 𝑓 | β€–

≀ ( 𝐢 3 . π‘Ž + 1 ) ​ πœ… 3 ​ β€– 𝑋 ⋆ β€– β‹… β€– | 𝐸 𝑑 𝑔 | β€–

≀ 𝐢 β€² ​ 𝐢 3 . π‘Ž ​ ( πœ… 4 ​ β€– | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | β€– + 𝑐 13 ​ πœ… βˆ’ 2 ​ β€– 𝑋 ⋆ β€– ​ β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€– + πœ… ​ β€– 𝑂 ~ 𝑑 β€– 3 / 4 ​ β€– 𝑋 ⋆ β€– 5 / 4 ) ,

(129)

where 𝐢 β€²

0 is again some universal constant.

β€’

Since Ξ“ 𝑑 βͺ° 1 100 ​ 𝐼 βˆ’ 𝐼

βˆ’ 99 100 ​ 𝐼 as already proved, it is easy to see that β€– ( 1 + Ξ“ 𝑑 ) βˆ’ 1 β€– ≀ 𝐢 and β€– Ξ“ 𝑑 ​ ( 1 + Ξ“ 𝑑 ) βˆ’ 1 β€– ≀ 𝐢 for some universal constant 𝐢

0 . Thus,

β€– | 𝐸 𝑑 β„Ž | β€– ≀ 2 ​ 𝑐 πœ† ​ 𝐢 ​ ( 1 + πœ‚ ​ 𝐢 ) ​ β€– 𝑄 β€² β€– β‹… β€– | Ξ“ 𝑑 | β€– + πœ‚ ​ 𝑐 πœ† 2 ​ 𝐢 4 ​ β€– 𝑄 β€² β€– 2 ​ β€– | Ξ“ 𝑑 | β€– ≀ 1 2 ​ β€– | Ξ“ 𝑑 | β€– ,

(130)

where the last line follows by using β€– 𝑄 β€² β€– ≀ 2 and by choosing 𝑐 πœ† , 𝑐 πœ‚ sufficiently small.

β€’

We still need to control πœ‚ 2 ​ Ξ“ 𝑑 2 ​ ( 1 + Ξ“ 𝑑 ) βˆ’ 1 . This can be accomplished by invoking β€– Ξ“ 𝑑 ​ ( 1 + Ξ“ 𝑑 ) βˆ’ 1 β€– ≀ 𝐢 again. In fact, we have

πœ‚ 2 ​ β€– | Ξ“ 𝑑 2 ​ ( 1 + Ξ“ 𝑑 ) βˆ’ 1 | β€– ≀ πœ‚ β‹… πœ‚ ​ β€– Ξ“ 𝑑 ​ ( 1 + Ξ“ 𝑑 ) βˆ’ 1 β€– β‹… β€– | Ξ“ 𝑑 | β€– ≀ πœ‚ β‹… πœ‚ ​ 𝐢 ​ β€– | Ξ“ 𝑑 | β€– ≀ πœ‚ 2 ​ β€– | Ξ“ 𝑑 | β€–

(131)

provided that πœ‚ ≀ 𝑐 πœ‚ is sufficiently small.

Plugging (129), (130), (131) into (127), we readily obtain

β€– | Ξ“ 𝑑 + 1 | β€–

≀ ( 1 βˆ’ 2 ​ πœ‚ ) ​ β€– | Ξ“ 𝑑 | ​ β€– + πœ‚ 2 β€– ​ | Ξ“ 𝑑 | ​ β€– + πœ‚ 2 β€– ​ | Ξ“ 𝑑 | β€– + πœ‚ ​ πœ… 2 ​ β€– 𝑋 ⋆ β€– βˆ’ 2 ​ β€– | 𝐸 𝑑 𝑓 | β€–

≀ ( 1 βˆ’ πœ‚ ) ​ β€– | Ξ“ 𝑑 | ​ β€– + πœ‚ ​ 𝐢 β€² ​ 𝐢 3 . π‘Ž ​ πœ… 4 β€– 𝑋 ⋆ β€– 2 β€– ​ | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | β€– + πœ‚ ​ 𝑐 13 ​ 𝐢 β€² ​ 𝐢 3 . π‘Ž ​ β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€– + πœ‚ ​ 𝐢 β€² ​ 𝐢 3 . π‘Ž ​ πœ… 3 ​ β€– 𝑂 ~ 𝑑 β€– 3 / 4 ​ β€– 𝑋 ⋆ β€– βˆ’ 3 / 4

≀ ( 1 βˆ’ πœ‚ ) ​ β€– | Ξ“ 𝑑 | ​ β€– + πœ‚ ​ 𝐢 26 ​ πœ… 4 β€– 𝑋 ⋆ β€– 2 β€– ​ | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | β€– + 1 16 ​ πœ‚ ​ β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€– + πœ‚ ​ ( β€– 𝑂 ~ 𝑑 β€– β€– 𝑋 ⋆ β€– ) 7 / 12 ,

where in the last line we set 𝐢 26

𝐢 β€² ​ 𝐢 3 . π‘Ž , chose 𝑐 13 sufficiently small and used (24). Finally note that 𝐢 26 ≲ 𝐢 3 . π‘Ž ≲ 𝑐 πœ† βˆ’ 1 / 2 as desired.

D.4Proof of Corollary 2

From Lemma 5, it is elementary (e.g., by induction on 𝑑 ) to show that

β€– Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 β€– ≀ ( 1 βˆ’ πœ‚ ) 𝑑 βˆ’ 𝑑 2 ​ β€– Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 2 ​ 𝑆 ~ 𝑑 2 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 β€– + 1 100 , βˆ€ 𝑑 ∈ [ 𝑑 2 , 𝑇 max ] .

(132)

Suppose for the moment that

β€– Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 2 ​ 𝑆 ~ 𝑑 2 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 β€– ≀ 𝐢 3 . π‘Ž 2 ​ πœ… 4 ,

(133)

where 𝐢 3 . π‘Ž is given in Lemma 3. Then given that πœ‚ ≀ 𝑐 πœ‚ for some sufficiently small 𝑐 πœ‚ , we have log ⁑ ( 1 βˆ’ πœ‚ ) β‰₯ βˆ’ πœ‚ / 2 . As a result, if 𝑑 3 βˆ’ 𝑑 2 β‰₯ 8 ​ log ⁑ ( 10 ​ 𝐢 3 . π‘Ž ​ πœ… ) / πœ‚ β‰₯ log ⁑ ( 𝐢 3 . π‘Ž βˆ’ 2 ​ πœ… βˆ’ 4 / 100 ) / log ⁑ ( 1 βˆ’ πœ‚ ) , we have ( 1 βˆ’ πœ‚ ) 𝑑 3 βˆ’ 𝑑 2 ≀ 𝐢 3 . π‘Ž βˆ’ 2 ​ πœ… βˆ’ 4 / 100 . When 𝐢 min is sufficiently large we may choose such 𝑑 3 which simultaneously satisfies 𝑑 3 ≀ 𝑑 2 + 𝑇 min / 16 ≀ 𝑇 max since 8 ​ log ⁑ ( 10 ​ 𝐢 3 . π‘Ž ​ πœ… ) / πœ‚ ≀ 𝐢 min 32 ​ πœ‚ ​ log ⁑ ( β€– 𝑋 ⋆ β€– / 𝛼 )

𝑇 min / 32 . Invoking (132), we obtain

β€– Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 3 ​ 𝑆 ~ 𝑑 3 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 β€– ≀ ( 𝐢 3 . π‘Ž βˆ’ 2 ​ πœ… βˆ’ 4 / 100 ) ​ ( 𝐢 3 . π‘Ž 2 ​ πœ… 4 ) + 1 100

1 50 ≀ 1 10 ,

(134)

which implies the desired bound (27).

Proof of inequality (133).

It is straightforward to verify that

β€– Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 2 ​ 𝑆 ~ 𝑑 2 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 β€– ≀ max ⁑ ( β€– Ξ£ ⋆ βˆ’ 1 ​ 𝑆 ~ 𝑑 2 β€– 2 βˆ’ 1 , 1 βˆ’ 𝜎 min 2 ​ ( Ξ£ ⋆ βˆ’ 1 ​ 𝑆 ~ 𝑑 2 ) ) ,

which combined with (22d) implies that

β€– Ξ£ ⋆ βˆ’ 1 ​ 𝑆 ~ 𝑑 2 β€– 2 βˆ’ 1 ≀ β€– Ξ£ ⋆ βˆ’ 1 β€– 2 ​ β€– 𝑆 ~ 𝑑 2 β€– 2 ≀ 𝜎 min βˆ’ 2 ​ ( 𝑋 ⋆ ) ​ 𝐢 3 . π‘Ž 2 ​ πœ… 2 ​ β€– 𝑋 ⋆ β€– 2

𝐢 3 . π‘Ž 2 ​ πœ… 4 .

In addition, by Corollary 1 we have

1 βˆ’ 𝜎 min 2 ​ ( Ξ£ ⋆ βˆ’ 1 ​ 𝑆 ~ 𝑑 2 ) ≀ 1 βˆ’ 1 10

9 10 .

Choosing 𝐢 3 . π‘Ž sufficiently large (say 𝐢 3 . π‘Ž β‰₯ 1 ) yields 𝐢 3 . π‘Ž 2 ​ πœ… 4 β‰₯ 9 / 10 , and hence the claim (133).

Appendix EProofs for Phase III

To characterize the behavior of β€– 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ β€– π–₯ , it is particularly helpful to consider the following decomposition into three error terms related to the signal term, the misalignment term, and the overparametrization term.

Lemma 27.

For all 𝑑 β‰₯ 𝑑 3 , as long as β€– Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 β€– ≀ 1 / 10 , one has

β€– 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ β€– π–₯ ≀ 4 ​ β€– 𝑋 ⋆ β€– 2 ​ ( β€– Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 β€– π–₯ + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– π–₯ ) + 4 ​ β€– 𝑋 ⋆ β€– ​ β€– 𝑂 ~ 𝑑 β€– .

Note that the overparametrization error β€– 𝑂 ~ 𝑑 β€– stays small, as stated in (22b) and (24). Therefore we only need to focus on the shrinkage of the first two terms β€– Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 β€– π–₯ + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– π–₯ , which is the focus of the lemma below.

Lemma 28.

For any 𝑑 : 𝑑 3 ≀ 𝑑 ≀ 𝑇 max , one has

β€– Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 β€– π–₯ + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– 𝑁 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 βˆ’ 1 ​ Ξ£ ⋆ β€– π–₯

≀ ( 1 βˆ’ πœ‚ 10 ) ​ ( β€– Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 β€– π–₯ + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– π–₯ ) + πœ‚ ​ ( β€– 𝑂 ~ 𝑑 β€– β€– 𝑋 ⋆ β€– ) 1 / 2 .

(135)

In particular, β€– Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 β€– ≀ 1 / 10 for all 𝑑 such that 𝑑 3 ≀ 𝑑 ≀ 𝑇 max .

We now show how Lemma 6 is implied by the above two lemmas. To begin with, we apply Lemma 28 repeatedly to obtain the following bound for all 𝑑 ∈ [ 𝑑 3 , 𝑇 max ] :

β€– Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 β€– π–₯ + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– π–₯

≀ ( 1 βˆ’ πœ‚ 10 ) 𝑑 βˆ’ 𝑑 3 ( βˆ₯ Ξ£ ⋆ βˆ’ 1 ( 𝑆 ~ 𝑑 3 𝑆 ~ 𝑑 3 ⊀ βˆ’ Ξ£ ⋆ 2 ) Ξ£ ⋆ βˆ’ 1 βˆ₯ π–₯ + βˆ₯ 𝑋 ⋆ βˆ₯ βˆ’ 1 βˆ₯ 𝑁 ~ 𝑑 3 𝑆 ~ 𝑑 3 βˆ’ 1 Ξ£ ⋆ βˆ₯ π–₯ ) + 10 max 𝑑 3 ≀ 𝜏 ≀ 𝑑 ( β€– 𝑂 ~ 𝜏 β€– β€– 𝑋 ⋆ β€– ) 1 / 2 ,

(136)

which motivates us to control the error at time 𝑑 3 .

We know from Corollary 2 that β€– Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 3 ​ 𝑆 ~ 𝑑 3 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 β€– ≀ 1 / 10 . Since Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 3 ​ 𝑆 ~ 𝑑 3 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 is a π‘Ÿ ⋆ Γ— π‘Ÿ ⋆ matrix, we have β€– Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 3 ​ 𝑆 ~ 𝑑 3 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 β€– π–₯ ≀ π‘Ÿ ⋆ / 10 . In addition, we infer from (22c) that

β€– 𝑁 ~ 𝑑 3 ​ 𝑆 ~ 𝑑 3 βˆ’ 1 ​ Ξ£ ⋆ β€– π–₯ ≀ π‘Ÿ ⋆ ​ β€– 𝑁 ~ 𝑑 3 ​ 𝑆 ~ 𝑑 3 βˆ’ 1 ​ Ξ£ ⋆ β€– ≀ π‘Ÿ ⋆ ​ 𝑐 3 ​ πœ… βˆ’ 𝐢 𝛿 / 2 ​ β€– 𝑋 ⋆ β€– ≀ π‘Ÿ ⋆ ​ β€– 𝑋 ⋆ β€– / 10 ,

as long as 𝑐 3 is sufficiently small. Combine the above two bounds to arrive at the conclusion that

β€– Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 3 ​ 𝑆 ~ 𝑑 3 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 β€– π–₯ + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– 𝑁 ~ 𝑑 3 ​ 𝑆 ~ 𝑑 3 βˆ’ 1 ​ Ξ£ ⋆ β€– π–₯ ≀ π‘Ÿ ⋆ 10 + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ π‘Ÿ ⋆ ​ β€– 𝑋 ⋆ β€– 10

π‘Ÿ ⋆ 5 .

(137)

Combining the two inequalities (136) and (137) yields for all 𝑑 ∈ [ 𝑑 3 , 𝑇 max ]

βˆ₯ Ξ£ ⋆ βˆ’ 1 ( 𝑆 ~ 𝑑 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 ) Ξ£ ⋆ βˆ’ 1 βˆ₯ π–₯ + βˆ₯ 𝑋 ⋆ βˆ₯ βˆ’ 1 βˆ₯ 𝑁 ~ 𝑑 𝑆 ~ 𝑑 βˆ’ 1 Ξ£ ⋆ βˆ₯ π–₯ ≀ 1 5 ( 1 βˆ’ πœ‚ 10 ) 𝑑 βˆ’ 𝑑 3 π‘Ÿ ⋆ + 10 max 𝑑 3 ≀ 𝜏 ≀ 𝑑 ( β€– 𝑂 ~ 𝜏 β€– β€– 𝑋 ⋆ β€– ) 1 / 2 .

We can then invoke Lemma 27 to see that

β€– 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ β€– π–₯

≀ 4 ​ β€– 𝑋 ⋆ β€– 2 5 ( 1 βˆ’ πœ‚ 10 ) 𝑑 βˆ’ 𝑑 3 π‘Ÿ ⋆ + 40 βˆ₯ 𝑋 ⋆ βˆ₯ 2 max 𝑑 3 ≀ 𝜏 ≀ 𝑑 ( β€– 𝑂 ~ 𝜏 β€– β€– 𝑋 ⋆ β€– ) 1 / 2 + 4 βˆ₯ 𝑋 ⋆ βˆ₯ βˆ₯ 𝑂 ~ 𝑑 βˆ₯

≀ ( 1 βˆ’ πœ‚ 10 ) 𝑑 βˆ’ 𝑑 3 π‘Ÿ ⋆ βˆ₯ 𝑀 ⋆ βˆ₯ + 80 βˆ₯ 𝑀 ⋆ βˆ₯ max 𝑑 3 ≀ 𝜏 ≀ 𝑑 ( β€– 𝑂 ~ 𝜏 β€– β€– 𝑋 ⋆ β€– ) 1 / 2 ,

where in the last line we use β€– 𝑂 ~ 𝑑 β€– ≀ β€– 𝑋 ⋆ β€– β€”an implication of (24). To see this, the assumption (12c) implies that 𝛼 ≀ β€– 𝑋 ⋆ β€– as long as πœ‚ ≀ 1 / 2 and 𝐢 𝛼 β‰₯ 4 , which in turn implies β€– 𝑂 ~ 𝑑 β€– ≀ 𝛼 2 / 3 ​ β€– 𝑋 ⋆ β€– 1 / 3 ≀ β€– 𝑋 ⋆ β€– . This completes the proof for the first part of Lemma 6 with 𝑐 6

1 / 10 .

For the second part of Lemma 6, notice that

8 𝑐 6 βˆ’ 1 max 𝑑 3 ≀ 𝜏 ≀ 𝑇 max ( βˆ₯ 𝑂 ~ 𝜏 βˆ₯ / βˆ₯ 𝑋 ⋆ βˆ₯ ) 1 / 2 ≀ 1 2 ( 𝛼 β€– 𝑋 ⋆ β€– ) 1 / 3

by (24), thus

β€– 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ β€– π–₯ ≀ ( 1 βˆ’ 𝑐 6 ​ πœ‚ ) 𝑑 βˆ’ 𝑑 3 ​ π‘Ÿ ⋆ ​ β€– 𝑀 ⋆ β€– + 1 2 ​ ( 𝛼 β€– 𝑋 ⋆ β€– ) 1 / 3

for 𝑑 3 ≀ 𝑑 ≀ 𝑇 max . There exists some iteration number 𝑑 4 : 𝑑 3 ≀ 𝑑 4 ≀ 𝑑 3 + 2 𝑐 6 ​ πœ‚ ​ log ⁑ ( β€– 𝑋 ⋆ β€– / 𝛼 ) ≀ 𝑑 3 + 𝑇 min / 16 such that

( 1 βˆ’ 𝑐 6 ​ πœ‚ ) 𝑑 4 βˆ’ 𝑑 3 ≀ ( 𝛼 β€– 𝑋 ⋆ β€– ) 2 ≀ 1 2 ​ π‘Ÿ ⋆ ​ ( 𝛼 β€– 𝑋 ⋆ β€– ) 1 / 3 ,

where the last inequality is due to (12c). It is then clear that 𝑑 4 has the property claimed in the lemma.

E.1Proof of Lemma 27

Starting from (51), we may deduce

β€– 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ β€– π–₯

≀ β€– 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 β€– π–₯ + 2 ​ β€– 𝑆 ~ 𝑑 β€– ​ β€– 𝑁 ~ 𝑑 β€– π–₯ + β€– 𝑁 ~ 𝑑 β€– ​ β€– 𝑁 ~ 𝑑 β€– π–₯ + β€– 𝑂 ~ 𝑑 β€– ​ β€– 𝑂 ~ 𝑑 β€– π–₯

≀ β€– 𝑋 ⋆ β€– 2 ​ ( β€– Ξ£ ⋆ βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ ​ Ξ£ ⋆ βˆ’ 1 βˆ’ 𝐼 β€– π–₯ + 2 ​ β€– Ξ£ ⋆ βˆ’ 1 ​ 𝑆 ~ 𝑑 β€– 2 ​ β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– π–₯ + 𝑛 ​ ( β€– 𝑂 ~ 𝑑 β€– β€– 𝑋 ⋆ β€– ) 2 )

≀ 4 ​ β€– 𝑋 ⋆ β€– 2 ​ ( β€– Ξ£ ⋆ βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ ​ Ξ£ ⋆ βˆ’ 1 βˆ’ 𝐼 β€– π–₯ + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– π–₯ + β€– 𝑂 ~ 𝑑 β€– β€– 𝑋 ⋆ β€– ) ,

(138)

where the penultimate line used β€– 𝑂 ~ 𝑑 β€– π–₯ ≀ 𝑛 ​ β€– 𝑂 ~ 𝑑 β€– , and the last line follows from β€– Ξ£ ⋆ βˆ’ 1 ​ 𝑆 ~ 𝑑 β€– 2

β€– Ξ£ ⋆ βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ ​ Ξ£ ⋆ βˆ’ 1 β€– ≀ 1 + β€– Ξ£ ⋆ βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ ​ Ξ£ ⋆ βˆ’ 1 βˆ’ 𝐼 β€– ≀ 2 (recall that β€– Ξ£ ⋆ βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ ​ Ξ£ ⋆ βˆ’ 1 βˆ’ 𝐼 β€– ≀ 1 / 10 by assumption) and from (24).

E.2Proof of Lemma 28

Recall the definition of Ξ“ 𝑑 from (121):

Ξ“ 𝑑 ≔ Ξ£ ⋆ βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ ​ Ξ£ ⋆ βˆ’ 1 βˆ’ 𝐼 .

Fix any 𝑑 ∈ [ 𝑑 3 , 𝑇 max ] , if (135) were true for all 𝜏 ∈ [ 𝑑 3 , 𝑑 ] , taking into account that β€– 𝑂 ~ 𝜏 β€– / β€– 𝑋 ⋆ β€– ≀ 1 / 10000 for all 𝜏 ∈ [ 𝑑 3 , 𝑇 max ] by (24), we could show by induction that β€– Ξ“ 𝜏 β€– ≀ 1 / 10 for all 𝜏 ∈ [ 𝑑 3 , 𝑑 ] . Thus it suffices to assume β€– Ξ“ 𝑑 β€– ≀ 1 / 10 and prove (135).

Apply Lemma 26 with Frobenius norm to obtain

β€– Ξ“ 𝑑 + 1 β€– π–₯ ≀ ( 1 βˆ’ πœ‚ ) ​ β€– Ξ“ 𝑑 β€– π–₯ + πœ‚ ​ 𝐢 26 ​ πœ… 4 β€– 𝑋 ⋆ β€– 2 ​ β€– π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 β€– π–₯ + 1 16 ​ πœ‚ ​ β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– π–₯ + πœ‚ ​ ( β€– 𝑂 ~ 𝑑 β€– β€– 𝑋 ⋆ β€– ) 7 / 12 ,

(139)

In addition, Lemma 24 tells us that

β€– 𝑁 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 βˆ’ 1 ​ Ξ£ ⋆ β€– π–₯ ≀ ( 1 βˆ’ πœ‚ 3 ​ ( β€– 𝑍 𝑑 β€– + πœ‚ ) ) ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– π–₯ + πœ‚ ​ 𝐢 24 ​ πœ… 6 𝑐 πœ† ​ β€– 𝑋 ⋆ β€– ​ β€– π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 β€– π–₯ + πœ‚ ​ ( β€– 𝑂 ~ 𝑑 β€– 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ) 2 / 3 ​ β€– 𝑋 ⋆ β€– ,

where 𝑍 𝑑

Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) ​ Ξ£ ⋆ βˆ’ 1 . It is easy to check that β€– 𝑍 𝑑 β€– ≀ 1 + β€– Ξ“ 𝑑 β€– + 𝑐 πœ† ≀ 2 as β€– Ξ“ 𝑑 β€– ≀ 1 / 10 and 𝑐 πœ† is sufficiently small. In addition, one has 𝜎 min ​ ( 𝑆 ~ 𝑑 ) 2 β‰₯ ( 1 βˆ’ β€– Ξ“ 𝑑 β€– ) ​ 𝜎 min ​ ( 𝑋 ⋆ ) 2 and β€– 𝑂 ~ 𝑑 β€– / 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ≀ ( 2 ​ πœ… ) βˆ’ 24 . Combine these relationships together to arrive at

β€– 𝑁 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 βˆ’ 1 ​ Ξ£ ⋆ β€– π–₯ ≀ ( 1 βˆ’ πœ‚ 8 ) ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– π–₯ + πœ‚ ​ 𝐢 24 ​ πœ… 6 𝑐 πœ† ​ β€– 𝑋 ⋆ β€– ​ β€– π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 β€– π–₯ + 1 2 ​ πœ‚ ​ β€– 𝑋 ⋆ β€– ​ ( β€– 𝑂 ~ 𝑑 β€– β€– 𝑋 ⋆ β€– ) 7 / 12 .

(140)

Summing up (139), (140), we obtain

β€– Ξ“ 𝑑 + 1 β€– π–₯ + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– 𝑁 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 βˆ’ 1 ​ Ξ£ ⋆ β€– π–₯

≀ ( 1 βˆ’ πœ‚ 8 ) ​ ( β€– Ξ“ 𝑑 β€– π–₯ + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– π–₯ ) + πœ‚ ​ 2 ​ ( 𝐢 24 + 𝐢 26 ​ 𝑐 πœ† ) ​ πœ… 8 𝑐 πœ† ​ β€– 𝑋 ⋆ β€– 2 ​ β€– π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 β€– π–₯ + 2 ​ πœ‚ ​ ( β€– 𝑂 ~ 𝑑 β€– β€– 𝑋 ⋆ β€– ) 7 / 12 .

(141)

This is close to our desired conclusion, but we would need to eliminate β€– π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 β€– π–₯ . To this end we observe

β€– π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 β€– π–₯

≀ π‘Ÿ ⋆ ​ β€– Ξ” 𝑑 β€–

≀ 8 ​ 𝛿 ​ π‘Ÿ ⋆ ​ ( β€– 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 β€– π–₯ + β€– 𝑆 ~ 𝑑 β€– ​ β€– 𝑁 ~ 𝑑 β€– π–₯ + 𝑛 ​ β€– 𝑂 ~ 𝑑 β€– 2 )

≀ 16 ​ 𝑐 𝛿 ​ πœ… βˆ’ 4 ​ β€– 𝑋 ⋆ β€– 2 ​ ( β€– Ξ“ 𝑑 β€– π–₯ + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– π–₯ + ( β€– 𝑂 ~ 𝑑 β€– β€– 𝑋 ⋆ β€– ) 2 / 3 ) ,

where the first line follows from π‘ˆ ⋆ being of rank π‘Ÿ ⋆ , the second line follows from Lemma 12, and the last line follows from (10) and from controlling the sum inside the brackets in a similar way as (138).

The conclusion follows from plugging the above inequality into (141), noting that 𝑐 𝛿 can be chosen sufficiently small and that β€– 𝑂 ~ 𝑑 β€– / β€– 𝑋 ⋆ β€– is sufficiently small due to (24).

E.3Proof of Proposition 2

Recall that in the proof of Lemma 23 (Appendix C.2.1), we have shown

β€– 𝑂 ~ 𝑑 β€– ≀ β€– 𝑂 ~ 𝑑 β€– + πœ‚ ​ β€– 𝑁 𝑑 + 1 ​ 𝑉 𝑑 ​ ( 𝑆 𝑑 + 1 ​ 𝑉 𝑑 ) βˆ’ 1 β€– β‹… β€– 𝐸 𝑑 𝑏 β€– + πœ‚ ​ β€– 𝐸 𝑑 𝑑 β€– .

(142)

This, along with all the conclusions in Section 3 (Lemma 3, Lemma 4, Lemma 6) and in the proof, hold for all 𝑑 ≀ 𝑇 max . However, it is clear from the proof that these continue to hold for 𝑑 ≀ 𝜏 , where 𝜏 is the minimal number such that

β€– 𝑂 ~ 𝜏 + 1 β€–

𝛼 7 / 10 ​ β€– 𝑋 ⋆ β€– 3 / 10 ,

(143)

cf. (24). In other words, β€– 𝑂 ~ 𝑑 β€– ≀ 𝛼 7 / 10 ​ β€– 𝑋 ⋆ β€– 3 / 10 for all 𝑑 ≀ 𝜏 . By Lemma 6 extended to the stopping time 𝜏 , we have for 𝑑 4 ≀ 𝑑 ≀ 𝜏 that

β€– 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ β€– π–₯ ≀ 𝛼 1 / 3 ​ β€– 𝑋 ⋆ β€– 5 / 3 .

(144)

We recall that Lemma 6 was derived from Lemma 28. Following the same derivation, this time controlling the term

β€– Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 β€– π–₯ + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– π–₯

directly using lemma 28 instead of passing to β€– 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ β€– , we find that for 𝑑 4 ≀ 𝑑 ≀ 𝜏 , the following stronger conclusion holds:

β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– ≀ ( 𝛼 β€– 𝑋 ⋆ β€– ) 1 / 3 .

(145)

Back to the recursive inequality (142), We bound each terms, this time using (98), (71) and a similar bound for 𝐸 𝑑 𝑑 , to obtain for all 𝑑 4 ≀ 𝑑 ≀ 𝜏 that:

β€– 𝑂 ~ 𝑑 + 1 β€–

≀ β€– 𝑂 ~ 𝑑 β€– + 𝐢 ​ πœ‚ ​ πœ… 𝐢 ​ β€– 𝑋 ⋆ β€– βˆ’ 1 ​ ( β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 ​ Ξ£ ⋆ β€– + β€– 𝑂 ~ 𝑑 β€– ) ​ β€– 𝑂 ~ 𝑑 β€–

≀ β€– 𝑂 ~ 𝑑 β€– + 𝐢 ​ πœ‚ ​ πœ… 𝐢 ​ [ ( 𝛼 β€– 𝑋 ⋆ β€– ) 1 / 3 + ( 𝛼 β€– 𝑋 ⋆ β€– ) 7 / 10 ] ​ β€– 𝑂 ~ 𝑑 β€–

≀ ( 1 + πœ‚ ​ ( 𝛼 β€– 𝑋 ⋆ β€– ) 3 / 10 ) ​ β€– 𝑂 ~ 𝑑 β€–

where 𝐢

0 is a universal constant; the second line follows from (145) and that β€– 𝑂 ~ 𝑑 β€– ≀ 𝛼 7 / 10 ​ β€– 𝑋 ⋆ β€– 3 / 10 for 𝑑 ≀ 𝜏 , and the last line follows from (12c).

By induction on 𝑑 , it is easy to see

β€– 𝑂 ~ 𝜏 + 1 β€–

≀ ( 1 + πœ‚ ​ ( 𝛼 β€– 𝑋 ⋆ β€– ) 3 / 10 ) 𝜏 βˆ’ 𝑇 max ​ β€– 𝑂 ~ 𝑇 max β€–

≀ ( 1 + πœ‚ ​ ( 𝛼 β€– 𝑋 ⋆ β€– ) 3 / 10 ) 𝜏 βˆ’ 𝑇 max ​ 𝛼 3 / 4 ​ β€– 𝑋 ⋆ β€– 1 / 4 ,

where the last inequality follows from (24). Plug this back into (143), we readily obtain

𝜏 βˆ’ 𝑇 max β‰₯ 𝑐 ​ log ⁑ ( β€– 𝑋 ⋆ β€– 𝛼 ) log ⁑ ( 1 + πœ‚ ​ ( 𝛼 β€– 𝑋 ⋆ β€– ) 3 / 10 ) β‰₯ 2 ​ 𝑐 ​ log ⁑ ( β€– 𝑋 ⋆ β€– 𝛼 ) πœ‚ ​ ( 𝛼 / β€– 𝑋 ⋆ β€– ) 3 / 10 β‰₯ ( β€– 𝑋 ⋆ β€– 𝛼 ) 3 / 10 ,

where 𝑐

3 4 βˆ’ 7 10

0 is a universal constant, and the last two inequalities follow from (12a) and (12c). This completes the proof.

Appendix FProofs for the noisy and the approximate low-rank settings

Both Theorem 4 and Theorem 5 can be viewed as special cases of the following theorem.

Theorem 6.

Assume the iterates 𝑋 𝑑 of ScaledGD( πœ† ) obeys

𝑋 𝑑 + 1

𝑋 𝑑 βˆ’ πœ‚ ​ ( π’œ βˆ— ​ π’œ ​ ( 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ ) βˆ’ 𝐸 ) ​ 𝑋 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 ,

(146)

for some matrix 𝐸 ∈ ℝ 𝑛 Γ— 𝑛 , where 𝑀 ⋆

𝑋 ⋆ ​ 𝑋 ⋆ ⊀ ∈ ℝ 𝑛 Γ— 𝑛 is a positive semidefinite matrix of rank π‘Ÿ ⋆ , 𝑋 ⋆ ∈ ℝ 𝑛 Γ— π‘Ÿ ⋆ . Assume further that

β€– 𝐸 β€– ≀ 𝑐 𝜎 ​ πœ… βˆ’ 𝐢 𝜎 ​ β€– 𝑀 ⋆ β€–

(147)

for some sufficiently small universal constant 𝑐 𝜎

0 and some sufficiently large universal constant 𝐢 𝜎

0 . Then the following holds with high probability (with respect to the realization of the random initialization 𝐺 ). Under Assumptions 1 and 2, there exist universal constants 𝐢 min

0 , 𝐢 6

0 , such that for some 𝑇 ≀ 𝑇 min ≔ 𝐢 min πœ‚ ​ log ⁑ β€– 𝑋 ⋆ β€– 𝛼 , the iterates of (146) obey

β€– 𝑋 𝑇 ​ 𝑋 𝑇 ⊀ βˆ’ 𝑀 ⋆ β€–

≀ max ⁑ ( πœ€ ​ β€– 𝑀 ⋆ β€– , 𝐢 6 ​ πœ… 4 ​ β€– π‘ˆ ⋆ ⊀ ​ 𝐸 β€– ) ,

β€– 𝑋 𝑇 ​ 𝑋 𝑇 ⊀ βˆ’ 𝑀 ⋆ β€– π–₯

≀ max ⁑ ( πœ€ ​ β€– 𝑀 ⋆ β€– , 𝐢 6 ​ πœ… 4 ​ β€– π‘ˆ ⋆ ⊀ ​ 𝐸 β€– π–₯ ) .

The proof is postponed to Appendix G. The rest of this appendix is devoted to showing how to deduce Theorem 4 and Theorem 5 from Theorem 6.

F.1Proof of Theorem 4

In the noisy setting, the update rule (14) of ScaledGD( πœ† ) can be written as

𝑋 𝑑 + 1

𝑋 𝑑 βˆ’ πœ‚ ​ ( π’œ βˆ— ​ π’œ ​ ( 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ ) βˆ’ 𝐸 ) ​ 𝑋 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 ,

(148)

where

𝐸 ≔ π’œ βˆ— ​ ( πœ‰ )

βˆ‘ 𝑖

1 π‘š πœ‰ 𝑖 ​ 𝐴 𝑖 .

(149)

We use the following classical lemma to show that the matrix 𝐸 defined above fulfills the assumption of Theorem 6.

Lemma 29.

Under Assumption 1, the following holds with probability at least 1 βˆ’ 2 ​ exp ⁑ ( βˆ’ 𝑐 ​ 𝑛 ) .

β€– 𝐸 β€– ≀ 8 ​ 𝜎 ​ 𝑛 , β€– π‘ˆ ⋆ ⊀ ​ 𝐸 β€– π–₯ ≀ 8 ​ 𝜎 ​ 𝑛 ​ π‘Ÿ ⋆ .

Proof.

The first inequality can be found in candes2010noisymc, Lemma 1.1. The second inequality can be deduced from the first one as follows. Note that π‘ˆ ⋆ ⊀ ​ 𝐸 has rank at most π‘Ÿ ⋆ , one has β€– π‘ˆ ⋆ ⊀ ​ 𝐸 β€– π–₯ ≀ π‘Ÿ ⋆ ​ β€– π‘ˆ ⋆ ⊀ ​ 𝐸 β€– ≀ π‘Ÿ ⋆ ​ β€– 𝐸 β€– ≀ 8 ​ 𝜎 ​ 𝑛 ​ π‘Ÿ ⋆ , as desired. ∎

The conclusion of Theorem 4 follows immediately by conditioning on the event that the inequalities in Lemma 29 hold, and then invoking Theorem 6.

F.2Proof of Theorem 5

In the approximately low-rank setting, the update rule of ScaledGD( πœ† ) can be written as

𝑋 𝑑 + 1

𝑋 𝑑 βˆ’ πœ‚ ​ ( π’œ βˆ— ​ π’œ ​ ( 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 π‘Ÿ ⋆ ) βˆ’ 𝐸 ) ​ 𝑋 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 ,

(150)

where

𝐸 ≔ π’œ βˆ— ​ π’œ ​ ( 𝑀 π‘Ÿ ⋆ β€² ) .

(151)

Recall that we assumed π’œ follows the Gaussian design in Theorem 5. One may show that the matrix 𝐸 defined above fulfills the assumption of Theorem 6 using random matrix theory, detailed below.

Lemma 30.

Under the assumptions on π’œ and π‘š in Theorem 5, the following holds with probability at least 1 βˆ’ 2 ​ exp ⁑ ( βˆ’ 𝑐 ​ 𝑛 ) .

β€– 𝐸 β€– ≀ 2 ​ β€– 𝑀 π‘Ÿ ⋆ β€² β€– + 16 ​ 𝑛 π‘š ​ β€– 𝑀 π‘Ÿ ⋆ β€² β€– π–₯ , β€– π‘ˆ ⋆ ⊀ ​ 𝐸 β€– π–₯ ≀ 16 ​ β€– 𝑀 π‘Ÿ ⋆ β€² β€– π–₯ .

Proof.

For the first inequality, we use a standard covering argument. Let β„‹ be a 1 / 4 -net of π•Š 𝑛 βˆ’ 1 , which can be chosen to satisfy | β„‹ | ≀ 9 𝑛 . It is well known that

β€– π’œ βˆ— ​ π’œ ​ ( 𝑀 π‘Ÿ ⋆ β€² ) β€–

sup 𝑣 ∈ π•Š 𝑛 βˆ’ 1 | ⟨ 𝑣 , π’œ βˆ— ​ π’œ ​ ( 𝑀 π‘Ÿ ⋆ β€² ) ​ 𝑣 ⟩ | ≀ 2 ​ sup 𝑣 ∈ β„‹ | ⟨ 𝑣 , π’œ βˆ— ​ π’œ ​ ( 𝑀 π‘Ÿ ⋆ β€² ) ​ 𝑣 ⟩ | .

(152)

Note that ⟨ 𝑣 , π’œ βˆ— ​ π’œ ​ ( 𝑀 π‘Ÿ ⋆ β€² ) ​ 𝑣 ⟩ is an order- 2 Gaussian chaos, which can be bounded by standard methods (see e.g. candes2010noisymc), yielding

| ⟨ 𝑣 , π’œ βˆ— ​ π’œ ​ ( 𝑀 π‘Ÿ ⋆ β€² ) ​ 𝑣 ⟩ βˆ’ ⟨ 𝑣 , 𝑀 π‘Ÿ ⋆ β€² ​ 𝑣 ⟩ |

| ⟨ 𝑣 , π’œ βˆ— ​ π’œ ​ ( 𝑀 π‘Ÿ ⋆ β€² ) ​ 𝑣 ⟩ βˆ’ 𝔼 ​ ⟨ 𝑣 , π’œ βˆ— ​ π’œ ​ ( 𝑀 π‘Ÿ ⋆ β€² ) ​ 𝑣 ⟩ | ≀ 8 ​ 𝑛 π‘š ​ β€– 𝑀 π‘Ÿ ⋆ β€² β€– π–₯

with probability at least 1 βˆ’ 2 ​ exp ⁑ ( βˆ’ 4 ​ 𝑛 ) . The desired inequality then follows from (152) and a union bound.

For the second inequality, we first note that the random vector π’œ ​ ( 𝑀 π‘Ÿ ⋆ β€² ) ∈ ℝ π‘š is Gaussian with law 𝒩 ​ ( 0 , 1 π‘š ​ β€– 𝑀 π‘Ÿ ⋆ β€² β€– π–₯ 2 ​ 𝐼 ) . A standard Gaussian concentration inequality implies β€– π’œ ​ ( 𝑀 π‘Ÿ ⋆ β€² ) β€– ≀ 2 ​ β€– 𝑀 π‘Ÿ ⋆ β€² β€– π–₯ with probability at least 1 βˆ’ 2 ​ exp ⁑ ( βˆ’ π‘š / 2 ) . To bound β€– π‘ˆ ⋆ ⊀ ​ π’œ βˆ— ​ π’œ ​ ( 𝑀 π‘Ÿ ⋆ β€² ) β€– π–₯ , the next step is to control the operator norm of π‘ˆ ⋆ ⊀ ​ π’œ βˆ— as an operator on the following spaces:

π‘ˆ ⋆ ⊀ π’œ βˆ— : ( ℝ π‘š , β„“ 2 ) β†’ ( ℝ π‘Ÿ ⋆ Γ— 𝑛 , βˆ₯ β‹… βˆ₯ π–₯ ) ⏟ ≕ β„³ .

In this sense, we may see that π‘ˆ ⋆ ⊀ ​ π’œ βˆ— is a Gaussian operator, since the matrix form of this operator is a ( π‘Ÿ ⋆ ​ 𝑛 ) Γ— π‘š matrix whose 𝑖 -th column is the vectorization of π‘ˆ ⋆ ⊀ ​ 𝐴 𝑖 , which is i.i.d. Gaussian as 𝐴 𝑖 is. Assume the covariance of such a column is Ξ› 2 ∈ ℝ ( π‘Ÿ ⋆ ​ 𝑛 ) Γ— ( π‘Ÿ ⋆ ​ 𝑛 ) , then the matrix form of π‘ˆ ⋆ ⊀ ​ π’œ βˆ— has the same distribution as Ξ› ​ 𝐺 , where 𝐺 is a ( π‘Ÿ ⋆ ​ 𝑛 ) Γ— π‘š random matrix with i.i.d. standard Gaussian entries. Again, a standard bound in random matrix theory (c.f. (30a)) implies that β€– 𝐺 β€– ≀ 4 ​ ( π‘š + π‘Ÿ ⋆ ​ 𝑛 ) with probability at least 1 βˆ’ exp ⁑ ( 𝑐 ​ π‘š ) , given π‘š β‰₯ 𝐢 ​ 𝑛 ​ π‘Ÿ ⋆ as assumed in Theorem 5. Conditioning on this event, we have

β€– π‘ˆ ⋆ ⊀ ​ π’œ βˆ— β€– ≀ 4 ​ ( π‘š + π‘Ÿ ⋆ ​ 𝑛 ) ​ β€– Ξ› β€– .

To compute β€– Ξ› β€– , note that since Ξ› ​ 𝐺 has the same distribution as the matrix form of π‘ˆ ⋆ ⊀ ​ π’œ βˆ— , we have

β€– 𝔼 ​ ( π‘ˆ ⋆ ⊀ ​ π’œ βˆ— ​ π’œ ​ π‘ˆ ⋆ ) β€– β„³

β€– 𝔼 ​ ( Ξ› ​ 𝐺 ​ 𝐺 ⊀ ​ Ξ› ) β€–

β€– Ξ› ​ ( π‘š ​ 𝐼 ) ​ Ξ› β€–

π‘š ​ β€– Ξ› β€– 2 ,

where the norm βˆ₯ β‹… βˆ₯ β„³ denotes the operator norm for operators on β„³ . But 𝔼 ​ ( π’œ βˆ— ​ π’œ )

ℐ , thus 𝔼 ​ ( π‘ˆ ⋆ ⊀ ​ π’œ βˆ— ​ π’œ ​ π‘ˆ ⋆ )

π‘ˆ ⋆ ⊀ ​ π‘ˆ ⋆

𝐼 is the identity operator, hence β€– 𝔼 ​ ( π‘ˆ ⋆ ⊀ ​ π’œ βˆ— ​ π’œ ​ π‘ˆ ⋆ ) β€– β„³

1 . Plugging this into the above identity, we find β€– Ξ› β€–

1 / π‘š . These together imply

β€– π‘ˆ ⋆ ⊀ ​ π’œ βˆ— β€– ≀ 4 ​ ( π‘š + π‘Ÿ ⋆ ​ 𝑛 ) β‹… 1 π‘š

4 ​ ( 1 + π‘Ÿ ⋆ ​ 𝑛 π‘š )

with probability at least 1 βˆ’ 2 ​ exp ⁑ ( βˆ’ 𝑐 ​ π‘š ) . The last quantity is less than 8 by the assumption π‘š β‰₯ 𝐢 ​ 𝑛 ​ π‘Ÿ ⋆ in Theorem 5. Therefore

β€– π‘ˆ ⋆ ⊀ ​ π’œ βˆ— ​ π’œ ​ ( 𝑀 π‘Ÿ ⋆ β€² ) β€– π–₯ ≀ β€– π‘ˆ ⋆ ⊀ ​ π’œ βˆ— β€– β‹… β€– π’œ ​ ( 𝑀 π‘Ÿ ⋆ β€² ) β€– ≀ 8 β‹… 2 ​ β€– 𝑀 π‘Ÿ ⋆ β€² β€– π–₯

16 ​ β€– 𝑀 π‘Ÿ ⋆ β€² β€– π–₯

with probability at least 1 βˆ’ exp ⁑ ( βˆ’ 𝑐 ​ π‘š ) , as desired. ∎

The conclusion of Theorem 5 follows immediately by conditioning on the event that the inequalities in Lemma 30 hold, and then invoking Theorem 6 with 𝑀 ⋆ substituted by 𝑀 π‘Ÿ ⋆ .

Appendix GProof of Theorem 6

The proof is based on a reduction to the noiseless setting. We begin with two heuristic observations that connect the generalized setting with the noiseless one, and make these observations formal later.

Observation 1: Phase I approximates power method for π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) + 𝐸 .

As in the noiseless setting, in the first few iterations we expect β€– 𝑋 𝑑 β€– to remain small, thus the update equation (14) can be approximated by

𝑋 𝑑 + 1 β‰ˆ ( 𝐼 + πœ‚ ​ ( π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) + 𝐸 ) ) ​ 𝑋 𝑑 .

This coincides with the update equation of power method for π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) + 𝐸 . Recall that in the noiseless setting, the first phase is also akin to power method, albeit for π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) . The key observation is that π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) + 𝐸 enjoys all the same properties of π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) that were required to establish Lemma 19. In fact, the only property of π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) used in the proof of Lemma 19 is

β€– ( π’œ βˆ— ​ π’œ βˆ’ ℐ ) ​ 𝑀 ⋆ β€– ≲ 𝑐 𝛿 ​ πœ… βˆ’ 2 ​ 𝐢 𝛿 / 3 ,

but by the assumption (147), π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) + 𝐸 also satisfies

β€– π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) + 𝐸 βˆ’ 𝑀 ⋆ β€– ≀ β€– ( π’œ βˆ— ​ π’œ βˆ’ ℐ ) ​ 𝑀 ⋆ β€– + β€– 𝐸 β€– ≲ 𝑐 𝛿 ​ πœ… βˆ’ 2 ​ 𝐢 𝛿 / 3 .

Thus all conclusion of Lemma 19 remains valid in the generalized setting.

Observation 2: In Phase II and III, the update equation has the same form as that in the noiseless setting.

Set

Ξ” 𝑑 β€²

Ξ” 𝑑 βˆ’ 𝐸 ,

then the update equation in the generalized setting can be expressed as

𝑋 𝑑 + 1

𝑋 𝑑 βˆ’ πœ‚ ​ ( 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ ) ​ 𝑋 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 + πœ‚ ​ Ξ” 𝑑 β€² ​ 𝑋 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 ,

which has the same form with the noiseless update equation (63), if we replace Ξ” 𝑑 there by Ξ” 𝑑 β€² . In the proof of Phase II, the only property of Ξ” 𝑑 we used is (50), which still holds for Ξ” 𝑑 β€² since β€– 𝐸 β€– is small. Thus the proof can be simply carried over to the generalized setting of Theorem 6. Moreover, in the proof of Phase III, the only places that involve controlling Ξ” 𝑑 in a different manner than (50) are  (122) and (140). These equations require us to control β€– | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | β€– for some unitarily invariant norm | | | β‹… | | | . If we replace Ξ” 𝑑 by Ξ” 𝑑 β€² , we can bound in these equations that

β€– | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 β€² | β€– ≀ β€– | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | β€– + β€– | π‘ˆ ⋆ ⊀ ​ 𝐸 | β€– .

Since any unitarily invariant | | | β‹… | | | is bounded by the operator norm up to a multiplicative constant5 (depending on the rank of the matrix), we may control β€– | π‘ˆ ⋆ ⊀ ​ 𝐸 | β€– using the assumption (147). Then we may combine (122) and (140) (assuming (140) also holds with the Frobenius norm replaced by | | | β‹… | | | ) to obtain

β€– | Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 | β€– + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– | 𝑁 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 βˆ’ 1 ​ Ξ£ ⋆ | β€–

≀ ( 1 βˆ’ πœ‚ 10 ) ​ ( β€– | Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 | β€– + β€– | 𝑋 ⋆ β€– βˆ’ 1 ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€– ) + πœ‚ ​ 𝐢 ​ πœ… 4 ​ β€– | π‘ˆ ⋆ ⊀ ​ 𝐸 | β€– + πœ‚ ​ ( β€– 𝑂 ~ 𝑑 β€– β€– 𝑋 ⋆ β€– ) 1 / 2 .

(153)

The conclusion of the theorem would immediately follow from the above inequality combined with Lemma 29 and Lemma 27, by taking | | | β‹… | | | to be the operator norm and the Frobenius norm.

Based on these observations, we formally state below the generalizations of key lemmas in the three phases required to prove Theorem 6. Most of them have identical proofs to their noiseless counterparts, and in such cases the proofs will be omitted. The few of them that require a slightly modified proof will be discussed in full detail.

G.1Generalization of Phase I

Our goal is to prove Lemma 3 in the generalized setting.

Lemma 31.

The conclusions of Lemma 3, along with its corollaries (3) and (24), still hold in the setting of Theorem 6.

As in the proof in the noiseless setting, this lemma is proved if we can prove the two parts of it respectively: the base case, where we show that there exists some 𝑑 1 ≀ 𝑇 min / 16 such that (21) holds and that (22) hold with 𝑑

𝑑 1 , and the induction step, where we show that (22) continues to hold for 𝑑 ∈ [ 𝑑 1 , 𝑇 max ] .

G.1.1Establishing the base case

We first show that Lemma 19 still holds in the generalized setting.

Lemma 32.

Under the same setting as Theorem 6, we have for some 𝑑 1 ≀ 𝑇 min / 16 such that (21) holds and that (22) hold with 𝑑

𝑑 1 .

We prove this result in a slightly more general setting. We consider a general symmetric matrix 𝑀 ^ ∈ ℝ 𝑛 Γ— 𝑛 , and set

𝑋 ^ 𝑑

( 𝐼 + πœ‚ πœ† ​ 𝑀 ^ ) 𝑑 ​ 𝑋 0 , 𝑑

0 , 1 , 2 , β‹―

We also denote

𝑠 𝑗 ≔ 𝜎 𝑗 ​ ( 𝐼 + πœ‚ πœ† ​ 𝑀 ^ )

1 + πœ‚ πœ† ​ 𝜎 𝑗 ​ ( 𝑀 ^ ) , 𝑗

1 , 2 , … , 𝑛

The treatment of the noiseless setting in Appendix C corresponds to the special case 𝑀 ^

π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) . In the generalized setting, we choose 𝑀 ^

π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) + 𝐸 . The following two lemmas are generalized from the lemmas in Appendix C, but have verbatim proofs as those, which are therefore omitted.

Lemma 33 (Generalization of Lemma 20).

Suppose that πœ† β‰₯ 1 100 ​ πœ… βˆ’ 4 ​ 𝑐 πœ† ​ 𝜎 min 2 ​ ( 𝑋 ⋆ ) . For any πœƒ ∈ ( 0 , 1 ) , there exists a large enough constant 𝐾

𝐾 ​ ( πœƒ , 𝑐 πœ† , 𝐢 𝐺 )

0 such that the following holds. As long as 𝛼 obeys

log ⁑ β€– 𝑋 ⋆ β€– 𝛼 β‰₯ 𝐾 max ⁑ ( πœ‚ , πœ… βˆ’ 2 ) ​ log ⁑ ( 2 ​ πœ… ​ 𝑛 ) β‹… ( 1 + log ⁑ ( 1 + πœ‚ πœ† ​ β€– 𝑀 ^ β€– ) ) ,

(154)

one has for all 𝑑 ≀ 1 πœƒ ​ πœ‚ ​ log ⁑ ( πœ… ​ 𝑛 ) :

βˆ₯ 𝑋 𝑑 βˆ’ 𝑋 ^ 𝑑 βˆ₯

≀ 𝑑 ​ ( 1 + πœ‚ πœ† ​ βˆ₯ 𝑀 ^ βˆ₯ ) 𝑑 ​ 𝛼 2 β€– 𝑋 ⋆ β€– .

(155)

Moreover, βˆ₯ 𝑋 𝑑 βˆ₯ ≀ βˆ₯ 𝑋 ⋆ βˆ₯ for all such 𝑑 .

Lemma 34 (Generalization of Lemma 21).

There exists some small universal constant 𝑐 34

0 such that the following hold. Assume that for some 𝛾 ≀ 𝑐 34 ,

βˆ₯ 𝑀 ^ βˆ’ 𝑀 ⋆ βˆ₯ ≀ 𝛾 ​ 𝜎 min 2 ​ ( 𝑋 ⋆ ) ,

(156)

and furthermore,

πœ™ ≔ 𝛼 ​ β€– 𝐺 β€– ​ 𝑠 π‘Ÿ ⋆ + 1 𝑑 + β€– 𝑋 𝑑 βˆ’ 𝑋 ^ 𝑑 β€– 𝛼 ​ 𝜎 min ​ ( π‘ˆ ^ ⊀ ​ 𝐺 ) ​ 𝑠 π‘Ÿ ⋆ 𝑑 ≀ 𝑐 34 ​ πœ… βˆ’ 2 .

(157)

Then for some universal 𝐢 34

0 the following hold:

𝜎 min ​ ( 𝑆 ~ 𝑑 )

β‰₯ 𝛼 4 ​ 𝜎 min ​ ( π‘ˆ ^ ⊀ ​ 𝐺 ) ​ 𝑠 π‘Ÿ ⋆ 𝑑 ,

(158a)

β€– 𝑂 ~ 𝑑 β€–

≀ 𝐢 34 ​ πœ™ ​ 𝛼 ​ 𝜎 min ​ ( π‘ˆ ^ ⊀ ​ 𝐺 ) ​ 𝑠 π‘Ÿ ⋆ 𝑑 ,

(158b)

β€– π‘ˆ ⋆ , βŸ‚ ⊀ ​ π‘ˆ 𝑋 ~ 𝑑 β€–

≀ 𝐢 34 ​ ( 𝛾 + πœ™ ) ,

(158c)

where 𝑋 ~ 𝑑 ≔ 𝑋 𝑑 ​ 𝑉 𝑑 ∈ ℝ 𝑛 Γ— π‘Ÿ ⋆ .

We are now ready to prove Lemma 32.

Proof of Lemma 32.

Recall that the generalized setting corresponds to 𝑀 ^

π’œ βˆ— ​ π’œ ​ ( 𝑀 ⋆ ) + 𝐸 . The proof is mostly identical to the proof of Lemma 19. Similar to that proof, we first need to verify the two assumptions in Lemma 34. The rest of the proof goes exactly the same, thus is omitted here.

Verifying assumption (156).

By the RIP in (9), Lemma 8, the condition of 𝛿 in (10), and the assumption (147), we have

β€– 𝑀 ^ βˆ’ 𝑀 ⋆ β€–

βˆ₯ ( ℐ βˆ’ π’œ βˆ— ​ π’œ ) ​ ( 𝑀 ⋆ ) + 𝐸 βˆ₯

≀ π‘Ÿ ⋆ ​ 𝛿 ​ β€– 𝑀 ⋆ β€– + 𝑐 𝜎 ​ πœ… βˆ’ 𝐢 𝜎 ​ β€– 𝑀 ⋆ β€–

≀ 𝑐 𝛿 ​ πœ… βˆ’ ( 𝐢 𝛿 βˆ’ 2 ) ​ 𝜎 min 2 ​ ( 𝑋 ⋆ ) + 𝑐 𝜎 ​ πœ… βˆ’ ( 𝐢 𝜎 βˆ’ 2 ) ​ 𝜎 min 2 ​ ( 𝑋 ⋆ )

≕ 𝛾 ​ 𝜎 min 2 ​ ( 𝑋 ⋆ ) .

(159)

Here 𝛾

𝑐 𝛿 ​ πœ… βˆ’ ( 𝐢 𝛿 βˆ’ 2 ) + 𝑐 𝜎 ​ πœ… βˆ’ ( 𝐢 𝜎 βˆ’ 2 ) ≀ 𝑐 21 , as 𝑐 𝛿 and 𝑐 𝜎 are assumed to be sufficiently small.

Verifying assumption (157).

By Weyl’s inequality and (G.1.1), we have

| 𝑠 𝑗 βˆ’ 1 βˆ’ πœ‚ πœ† ​ 𝜎 𝑗 ​ ( 𝑀 ⋆ ) | ≀ πœ‚ πœ† ​ βˆ₯ 𝑀 ^ βˆ’ 𝑀 ⋆ βˆ₯ ≀ πœ‚ πœ† ​ 𝛾 ​ 𝜎 min 2 ​ ( 𝑋 ⋆ ) ≀ 100 ​ ( 𝑐 𝛿 + 𝑐 𝜎 ) 𝑐 πœ† ​ πœ‚ ,

where the last inequality follows from the condition πœ† β‰₯ 1 100 ​ 𝑐 πœ† ​ 𝜎 min 2 ​ ( 𝑋 ⋆ ) . Furthermore, using the condition πœ† ≀ 𝑐 πœ† ​ 𝜎 min 2 ​ ( 𝑋 ⋆ ) assumed in (12b), the above bound implies that, for some 𝐢

𝐢 ​ ( 𝑐 πœ† , 𝑐 𝜎 , 𝑐 𝛿 )

0 ,

𝑠 1

≀ 1 + πœ‚ πœ† ​ β€– 𝑀 ⋆ β€– + 100 ​ ( 𝑐 𝛿 + 𝑐 𝜎 ) 𝑐 πœ† ​ πœ‚ ≀ 1 + 𝐢 ​ πœ‚ ​ πœ… 6 ,

(160a)

𝑠 π‘Ÿ ⋆

β‰₯ 1 + πœ‚ πœ† ​ 𝜎 min 2 ​ ( 𝑋 ⋆ ) βˆ’ 100 ​ ( 𝑐 𝛿 + 𝑐 𝜎 ) 𝑐 πœ† ​ πœ‚ β‰₯ 1 + πœ‚ 2 ​ πœ† / 𝜎 min 2 ​ ( 𝑋 ⋆ ) ,

(160b)

𝑠 π‘Ÿ ⋆

≀ 1 + πœ‚ πœ† ​ 𝜎 min 2 ​ ( 𝑋 ⋆ ) + 100 ​ ( 𝑐 𝛿 + 𝑐 𝜎 ) 𝑐 πœ† ​ πœ‚ ≀ 1 + 2 ​ πœ‚ πœ† / 𝜎 min 2 ​ ( 𝑋 ⋆ ) ,

(160c)

𝑠 π‘Ÿ ⋆ + 1

≀ 1 + 100 ​ ( 𝑐 𝛿 + 𝑐 𝜎 ) 𝑐 πœ† ​ πœ‚ ≀ 1 + πœ‚ 4 ​ 𝑐 πœ† ,

(160d)

where we use the fact that 𝜎 π‘Ÿ ⋆ + 1 ​ ( 𝑀 ⋆ )

0 , and 𝑐 𝛿 + 𝑐 𝜎 ≀ 1 / 400 . The rest of the verification is the same as the verification of (81) in the proof of Lemma 19. ∎

G.1.2Establishing the induction step

Following the proof of the noiseless setting, we would like to show that Lemmas 23, 24, 25 still hold in the generalized setting, which in turn relies entirely on Lemmas 13, 14, 15. Since Lemma 14 and Lemma 15 are both corollaries of Lemma 13, it suffices to prove the generalization of Lemma 13 in the generalized setting.

Lemma 35 (Generalization of Lemma 13).

Assume the update equation of 𝑋 𝑑 has the following form (cf. (63)):

𝑋 𝑑 + 1

𝑋 𝑑 βˆ’ πœ‚ ​ ( 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ ) ​ 𝑋 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 + πœ‚ ​ Ξ” 𝑑 β€² ​ 𝑋 𝑑 ​ ( 𝑋 𝑑 ⊀ ​ 𝑋 𝑑 + πœ† ​ 𝐼 ) βˆ’ 1 ,

where Ξ” 𝑑 β€² ∈ ℝ 𝑛 Γ— 𝑛 is some symmetric matrix satisfying β€– Ξ” 𝑑 β€² β€– ≀ 𝑐 12 ​ πœ… βˆ’ 2 ​ 𝐢 𝛿 / 3 ​ β€– 𝑋 ⋆ β€– 2 . For any 𝑑 such that 𝑆 ~ 𝑑 is invertible and (22) holds, the equations (53a) and (53b) hold, where error terms are bounded by (54b)–(54d) and the following modifications of (54a) and (54e):

β€– | 𝐸 𝑑 π‘Ž | β€–

≀ 2 ​ 𝑐 3 ​ πœ… βˆ’ 4 ​ β€– 𝑋 ⋆ β€– β‹… β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | ​ β€– + 2 β€– ​ | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 β€² | β€– ,

(161a)

β€– | 𝐸 𝑑 𝑒 | β€–

≀ 2 ​ β€– | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 β€² | β€– + 𝑐 12 ​ πœ… βˆ’ 5 ​ β€– 𝑋 ⋆ β€– β‹… β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€– .

(161b)

The proof is verbatim to Lemma 13. Note that the noiseless setting corresponds to the special case Ξ” 𝑑 β€²

Ξ” 𝑑

( ℐ βˆ’ π’œ βˆ— ​ π’œ ) ​ ( 𝑀 ⋆ ) , while the generalized setting corresponds to Ξ” 𝑑 β€²

Ξ” 𝑑 βˆ’ 𝐸

( ℐ βˆ’ π’œ βˆ— ​ π’œ ) ​ ( 𝑀 ⋆ ) βˆ’ 𝐸 . To show that Lemma 35 is applicable to the generalized setting we need to verify that this choice of Ξ” 𝑑 β€² guarantees the smallness of β€– Ξ” 𝑑 β€² β€– , which is proved in the following lemma.

Lemma 36 (Generalization of (50) in Lemma 12).

Under the same setting as Theorem 6, for any 𝑑 such that (22) holds, we have

β€– Ξ” 𝑑 β€² β€– ≀ 𝑐 12 ​ πœ… βˆ’ 2 ​ 𝐢 𝛿 / 3 ​ β€– 𝑋 ⋆ β€– 2 .

Proof.

Combining (52) in the proof of Lemma 12 the assumption β€– 𝐸 β€– ≀ 𝑐 𝜎 ​ πœ… βˆ’ 𝐢 𝜎 ​ β€– 𝑀 ⋆ β€– in (147), we obtain

β€– Ξ” 𝑑 β€² β€–

≀ 16 ​ 𝛿 ​ π‘Ÿ ⋆ ​ πœ… 2 ​ ( 𝐢 3 . π‘Ž 2 + 1 ) ​ β€– 𝑋 ⋆ β€– 2 + 𝑐 𝜎 ​ πœ… βˆ’ 𝐢 𝜎 ​ β€– 𝑋 ⋆ β€– 2

≀ ( 16 ​ 𝑐 𝛿 ​ πœ… βˆ’ 𝐢 𝛿 + 2 ​ ( 𝐢 3 . π‘Ž 2 + 1 ) 2 + 𝑐 𝜎 ​ πœ… βˆ’ 𝐢 𝜎 ) ​ β€– 𝑋 ⋆ β€– 2

≀ 𝑐 12 ​ πœ… βˆ’ 2 ​ 𝐢 𝛿 / 3 ​ β€– 𝑋 ⋆ β€– 2 ,

if we choose 𝐢 𝜎 β‰₯ 𝐢 𝛿 , 𝑐 𝜎 ≀ 𝑐 𝛿 , and note that 𝑐 12

32 ​ ( 𝐢 3 . π‘Ž + 1 ) 2 ​ 𝑐 𝛿 as defined in Lemma 12 (please refer to the argument after (52) for details). ∎

With these fundamental results in hand we can follow the same arguments as in the noiseless case to prove the following generalization of the lemmas in Appendix C.2.

Lemma 37.

The conclusions of Lemmas 23, 25 still hold in the setting of Theorem 6. Moreover, the following modification of Lemma 24 holds in the setting of Theorem 6. For any 𝑑 such that (22) holds, setting 𝑍 𝑑

Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) ​ Ξ£ ⋆ βˆ’ 1 , there exists some universal constant 𝐢 24

0 such that

β€– | 𝑁 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 βˆ’ 1 ​ Ξ£ ⋆ | β€– ≀ ( 1 βˆ’ πœ‚ 3 ​ ( β€– 𝑍 𝑑 β€– + πœ‚ ) ) ​ β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | ​ β€– + πœ‚ ​ 𝐢 24 ​ πœ… 6 𝑐 πœ† ​ β€– 𝑋 ⋆ β€– β€– ​ | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 β€² | β€– + πœ‚ ​ ( β€– 𝑂 ~ 𝑑 β€– 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ) 1 / 2 ​ β€– 𝑋 ⋆ β€– .

(162)

In particular, if 𝑐 3

100 ​ 𝐢 24 ​ ( 𝐢 3 . π‘Ž + 1 ) 4 ​ 𝑐 𝛿 / 𝑐 πœ† , then β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– ≀ 𝑐 3 ​ πœ… βˆ’ 𝐢 𝛿 / 2 ​ β€– 𝑋 ⋆ β€– implies β€– 𝑁 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 βˆ’ 1 ​ Ξ£ ⋆ β€– ≀ 𝑐 3 ​ πœ… βˆ’ 𝐢 𝛿 / 2 ​ β€– 𝑋 ⋆ β€– .

By the arguments following Lemma 25, the above results are sufficient to prove the induction step, thereby completing the proof of Lemma 3 in the generalized setting.

G.2Generalization of Phase II

We will prove Lemma 4 and Lemma 5, the main results of Phase II, in the generalized setting.

Lemma 38.

The conclusions of Lemma 4 and Lemma 5, along with Corollary 1 and Corollary 2, still hold under the generalized setting of Theorem 6.

Tracking the proof of Phase II in Appendix D, one may verify that all proofs there hold verbatim in the generalized setting, with Lemma 35 in place of Lemma 13 (the proof also used Lemmas 14, 15, which are corollaries of Lemma 13, hence hold in the generalized setting given Lemma 35), except for Lemma 26, which should be substituted by the following generalization:

Lemma 39.

Under the same setting as Theorem 6, for any 𝑑 : 𝑑 2 ≀ 𝑑 ≀ 𝑇 max , one has

β€– | Ξ“ 𝑑 + 1 | β€– ≀ ( 1 βˆ’ πœ‚ ) ​ β€– | Ξ“ 𝑑 | ​ β€– + πœ‚ ​ 𝐢 26 ​ πœ… 4 β€– 𝑋 ⋆ β€– 2 β€– ​ | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 β€² | β€– + 1 16 ​ πœ‚ ​ β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€– + πœ‚ ​ ( β€– 𝑂 ~ 𝑑 β€– β€– 𝑋 ⋆ β€– ) 7 / 12 ,

(163)

where 𝐢 26 ≲ 𝑐 πœ† βˆ’ 1 / 2 is some positive constant and | | | β‹… | | | can either be the Frobenius norm or the spectral norm.

The proof is identical to that of Lemma 26, thus is omitted here. Following the proof in Appendix D, these generalized results are sufficient to prove Lemma 38, thereby completing the proof of Phase II in the generalized setting.

G.3Generalization of Phase III

Our goal is to prove the following modification of Lemma 6 in the generalized setting.

Lemma 40 (Generalization of Lemma 6).

Under the same setting as Theorem 6, there exists some universal constant 𝑐 40

0 such that for any 𝑑 : 𝑑 3 ≀ 𝑑 ≀ 𝑇 max , with | | | β‹… | | | taken to be the operator norm βˆ₯ β‹… βˆ₯ or the Frobenius norm βˆ₯ β‹… βˆ₯ π–₯ , we have

| | | 𝑋 𝑑 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ | | | ≀ ( 1 βˆ’ 𝑐 40 πœ‚ ) 𝑑 βˆ’ 𝑑 3 π‘Ÿ ⋆ βˆ₯ 𝑀 ⋆ βˆ₯ + 𝑐 40 βˆ’ 1 πœ… 4 | | | π‘ˆ ⋆ ⊀ 𝐸 | | | + 8 𝑐 40 βˆ’ 1 βˆ₯ 𝑀 ⋆ βˆ₯ max 𝑑 3 ≀ 𝜏 ≀ 𝑑 ( β€– 𝑂 ~ 𝜏 β€– β€– 𝑋 ⋆ β€– ) 1 / 2 .

(164)

In particular, there exists an iteration number 𝑑 4 : 𝑑 3 ≀ 𝑑 4 ≀ 𝑑 3 + 𝑇 min / 16 such that for any 𝑑 ∈ [ 𝑑 4 , 𝑇 max ] , we have

β€– | 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ | β€– ≀ max ⁑ ( 𝛼 1 / 3 ​ β€– 𝑋 ⋆ β€– 5 / 3 , 𝑐 40 βˆ’ 1 ​ πœ… 4 ​ β€– | π‘ˆ ⋆ ⊀ ​ 𝐸 | β€– ) ≀ max ⁑ ( πœ€ ​ β€– 𝑀 ⋆ β€– , 𝑐 40 βˆ’ 1 ​ πœ… 4 ​ β€– | π‘ˆ ⋆ ⊀ ​ 𝐸 | β€– ) .

(165)

Here, πœ€ and 𝛼 are as stated in Theorem 2.

As in Appendix E, this will be accomplished by decomposing the error β€– | 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ | β€– using Lemma 27, and then control the components in the decomposition using Lemma 28. It is easy to check that the proof of Lemma 27 applies without modification to the generalized setting, and in fact works with the Frobenius norm replaced by any unitarily invariant norm. This leads to the following generalization.

Lemma 41 (Generalization of Lemma 27).

Under the same setting as Theorem 6, for all 𝑑 β‰₯ 𝑑 3 , as long as β€– Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 β€– ≀ 1 / 10 , one has

β€– | 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ | β€– ≀ 4 ​ β€– 𝑋 ⋆ β€– 2 ​ ( β€– | Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 | β€– + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€– ) + 4 ​ β€– 𝑋 ⋆ β€– ​ β€– 𝑂 ~ 𝑑 β€– .

It remains to prove the generalization of Lemma 28, stated below.

Lemma 42 (Generalization of Lemma 28).

Under the same setting as Theorem 6, there exists some universal constant 𝐢 42

0 such that for any 𝑑 : 𝑑 3 ≀ 𝑑 ≀ 𝑇 max , with | | | β‹… | | | taken to be the operator norm βˆ₯ β‹… βˆ₯ or the Frobenius norm βˆ₯ β‹… βˆ₯ π–₯ , one has

β€– | Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 | β€– + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– | 𝑁 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 βˆ’ 1 ​ Ξ£ ⋆ | β€–

≀ ( 1 βˆ’ πœ‚ 10 ) ​ ( β€– | Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 | β€– + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€– ) + πœ‚ ​ 𝐢 42 ​ πœ… 4 β€– 𝑋 ⋆ β€– 2 ​ β€– | π‘ˆ ⋆ ⊀ ​ 𝐸 | β€– + πœ‚ ​ ( β€– 𝑂 ~ 𝑑 β€– β€– 𝑋 ⋆ β€– ) 1 / 2 .

(166)

In particular, β€– Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 β€– ≀ 1 / 10 for all 𝑑 such that 𝑑 3 ≀ 𝑑 ≀ 𝑇 max .

We are prepared to formally prove Lemma 40. Similar to the noiseless setting, we apply Lemma 42 repeatedly to obtain the following bound for all 𝑑 ∈ [ 𝑑 3 , 𝑇 max ] :

β€– | Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 | β€– + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€–

≀ ( 1 βˆ’ πœ‚ 10 ) 𝑑 βˆ’ 𝑑 3 ​ ( β€– | Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 3 ​ 𝑆 ~ 𝑑 3 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 | β€– + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– | 𝑁 ~ 𝑑 3 ​ 𝑆 ~ 𝑑 3 βˆ’ 1 ​ Ξ£ ⋆ | β€– )

  • 10 ​ 𝐢 42 ​ πœ… 4 β€– 𝑋 ⋆ β€– 2 | | | π‘ˆ ⋆ ⊀ 𝐸 | | |
  • 10 max 𝑑 3 ≀ 𝜏 ≀ 𝑑 ( β€– 𝑂 ~ 𝜏 β€– β€– 𝑋 ⋆ β€– ) 1 / 2 ,

(167)

which motivates us to control the error at time 𝑑 3 . With the same arguments as in the noiseless setting (cf. Equation (137) in Appendix E), we obtain

β€– Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 3 ​ 𝑆 ~ 𝑑 3 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 β€– π–₯ + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– 𝑁 ~ 𝑑 3 ​ 𝑆 ~ 𝑑 3 βˆ’ 1 ​ Ξ£ ⋆ β€– π–₯ ≀ π‘Ÿ ⋆ 5 .

Since the operator norm of a matrix is always less than or equal to the Frobenius norm of it, the above inequality also holds if the Frobenius norm is replaced by the operator norm. Recalling that in this lemma, | | | β‹… | | | is taken to be either the operator norm or the Frobenius norm, we have shown

β€– | Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 3 ​ 𝑆 ~ 𝑑 3 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 | β€– + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– | 𝑁 ~ 𝑑 3 ​ 𝑆 ~ 𝑑 3 βˆ’ 1 ​ Ξ£ ⋆ | β€– ≀ π‘Ÿ ⋆ 5 .

(168)

Combining the two inequalities (167) and (168) yields for all 𝑑 ∈ [ 𝑑 3 , 𝑇 max ]

β€– | Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 ) ​ Ξ£ ⋆ βˆ’ 1 | β€– + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€–

≀ 1 5 ( 1 βˆ’ πœ‚ 10 ) 𝑑 βˆ’ 𝑑 3 π‘Ÿ ⋆ + 10 ​ 𝐢 42 ​ πœ… 4 β€– 𝑋 ⋆ β€– 2 | | | π‘ˆ ⋆ ⊀ 𝐸 | | | + 10 max 𝑑 3 ≀ 𝜏 ≀ 𝑑 ( β€– 𝑂 ~ 𝜏 β€– β€– 𝑋 ⋆ β€– ) 1 / 2 .

We can then invoke Lemma 41 to see that

β€– | 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ | β€–

≀ 4 ​ β€– 𝑋 ⋆ β€– 2 5 ( 1 βˆ’ πœ‚ 10 ) 𝑑 βˆ’ 𝑑 3 π‘Ÿ ⋆ + 10 𝐢 42 πœ… 4 | | | π‘ˆ ⋆ ⊀ 𝐸 | | | + 40 βˆ₯ 𝑋 ⋆ βˆ₯ 2 max 𝑑 3 ≀ 𝜏 ≀ 𝑑 ( β€– 𝑂 ~ 𝜏 β€– β€– 𝑋 ⋆ β€– ) 1 / 2 + 4 βˆ₯ 𝑋 ⋆ βˆ₯ βˆ₯ 𝑂 ~ 𝑑 βˆ₯

≀ ( 1 βˆ’ πœ‚ 10 ) 𝑑 βˆ’ 𝑑 3 π‘Ÿ ⋆ βˆ₯ 𝑀 ⋆ βˆ₯ + 10 𝐢 42 πœ… 4 | | | π‘ˆ ⋆ ⊀ 𝐸 | | | + 80 βˆ₯ 𝑀 ⋆ βˆ₯ max 𝑑 3 ≀ 𝜏 ≀ 𝑑 ( β€– 𝑂 ~ 𝜏 β€– β€– 𝑋 ⋆ β€– ) 1 / 2 ,

where in the last line we use β€– 𝑂 ~ 𝑑 β€– ≀ β€– 𝑋 ⋆ β€– β€”an implication of (24), which holds in the generalized setting by Lemma 31. To see this, the assumption (12c) implies that 𝛼 ≀ β€– 𝑋 ⋆ β€– as long as πœ‚ ≀ 1 / 2 and 𝐢 𝛼 β‰₯ 4 , which in turn implies β€– 𝑂 ~ 𝑑 β€– ≀ 𝛼 2 / 3 ​ β€– 𝑋 ⋆ β€– 1 / 3 ≀ β€– 𝑋 ⋆ β€– . This completes the proof for the first part of Lemma 40 with 𝑐 40

1 / ( 10 ​ 𝐢 42 ) .

For the second part of Lemma 40, notice that

8 𝑐 40 βˆ’ 1 max 𝑑 3 ≀ 𝜏 ≀ 𝑇 max ( βˆ₯ 𝑂 ~ 𝜏 βˆ₯ / βˆ₯ 𝑋 ⋆ βˆ₯ ) 1 / 2 ≀ 1 2 ( 𝛼 β€– 𝑋 ⋆ β€– ) 1 / 3

by (24), thus

β€– | 𝑋 𝑑 ​ 𝑋 𝑑 ⊀ βˆ’ 𝑀 ⋆ | β€– ≀ ( 1 βˆ’ 𝑐 6 ​ πœ‚ ) 𝑑 βˆ’ 𝑑 3 ​ π‘Ÿ ⋆ ​ β€– 𝑀 ⋆ ​ β€– + 𝑐 40 βˆ’ 1 ​ πœ… 4 | ​ β€– π‘ˆ ⋆ ⊀ ​ 𝐸 β€– | + 1 2 ​ ( 𝛼 β€– 𝑋 ⋆ β€– ) 1 / 3

for 𝑑 3 ≀ 𝑑 ≀ 𝑇 max . There exists some iteration number 𝑑 4 : 𝑑 3 ≀ 𝑑 4 ≀ 𝑑 3 + 2 𝑐 40 ​ πœ‚ ​ log ⁑ ( β€– 𝑋 ⋆ β€– / 𝛼 ) ≀ 𝑑 3 + 𝑇 min / 16 such that

( 1 βˆ’ 𝑐 6 ​ πœ‚ ) 𝑑 4 βˆ’ 𝑑 3 ≀ ( 𝛼 β€– 𝑋 ⋆ β€– ) 2 ≀ 1 2 ​ π‘Ÿ ⋆ ​ ( 𝛼 β€– 𝑋 ⋆ β€– ) 1 / 3 ,

where the last inequality is due to (12c). It is then clear that 𝑑 4 has the property claimed in the lemma.

G.3.1Proof of Lemma 42

The idea is the same as the proof of Lemma 28. Fix any 𝑑 ∈ [ 𝑑 3 , 𝑇 max ] , if (166) were true for all 𝜏 ∈ [ 𝑑 3 , 𝑑 ] , taking into account that β€– 𝑂 ~ 𝜏 β€– / β€– 𝑋 ⋆ β€– ≀ 1 / 10000 for all 𝜏 ∈ [ 𝑑 3 , 𝑇 max ] by (24) (which still holds in the generalized setting according to Lemma 32), we could show by induction that β€– Ξ“ 𝜏 β€– ≀ 1 / 10 for all 𝜏 ∈ [ 𝑑 3 , 𝑑 ] . Thus it suffices to assume β€– Ξ“ 𝑑 β€– ≀ 1 / 10 and prove (166).

Apply Lemma 39 to obtain

β€– | Ξ“ 𝑑 + 1 | β€– ≀ ( 1 βˆ’ πœ‚ ) ​ β€– | Ξ“ 𝑑 | ​ β€– + πœ‚ ​ 𝐢 26 ​ πœ… 4 β€– 𝑋 ⋆ β€– 2 β€– ​ | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 β€² | β€– + 1 16 ​ πœ‚ ​ β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€– + πœ‚ ​ ( β€– 𝑂 ~ 𝑑 β€– β€– 𝑋 ⋆ β€– ) 7 / 12 ,

(169)

In addition, Lemma 37 tells us that

β€– | 𝑁 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 βˆ’ 1 ​ Ξ£ ⋆ | β€– ≀ ( 1 βˆ’ πœ‚ 3 ​ ( β€– 𝑍 𝑑 β€– + πœ‚ ) ) ​ β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | ​ β€– + πœ‚ ​ 𝐢 24 ​ πœ… 4 𝑐 πœ† ​ β€– 𝑋 ⋆ β€– β€– ​ | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 β€² | β€– + πœ‚ ​ ( β€– 𝑂 ~ 𝑑 β€– 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ) 2 / 3 ​ β€– 𝑋 ⋆ β€– ,

where 𝑍 𝑑

Ξ£ ⋆ βˆ’ 1 ​ ( 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ + πœ† ​ 𝐼 ) ​ Ξ£ ⋆ βˆ’ 1 . It is easy to check that β€– 𝑍 𝑑 β€– ≀ 1 + β€– Ξ“ 𝑑 β€– + 𝑐 πœ† ≀ 2 as β€– Ξ“ 𝑑 β€– ≀ 1 / 10 and 𝑐 πœ† is sufficiently small. In addition, one has 𝜎 min ​ ( 𝑆 ~ 𝑑 ) 2 β‰₯ ( 1 βˆ’ β€– Ξ“ 𝑑 β€– ) ​ 𝜎 min ​ ( 𝑋 ⋆ ) 2 and β€– 𝑂 ~ 𝑑 β€– / 𝜎 min ​ ( 𝑆 ~ 𝑑 ) ≀ ( 2 ​ πœ… ) βˆ’ 24 . Combine these relationships together to arrive at

β€– | 𝑁 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 βˆ’ 1 ​ Ξ£ ⋆ | β€– ≀ ( 1 βˆ’ πœ‚ 8 ) ​ β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | ​ β€– + πœ‚ ​ 𝐢 24 ​ πœ… 2 𝑐 πœ† ​ β€– 𝑋 ⋆ β€– β€– ​ | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 β€² | β€– + 1 2 ​ πœ‚ ​ β€– 𝑋 ⋆ β€– ​ ( β€– 𝑂 ~ 𝑑 β€– β€– 𝑋 ⋆ β€– ) 7 / 12 .

(170)

Summing up (139), (140), we obtain

β€– | Ξ“ 𝑑 + 1 | β€– + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– | 𝑁 ~ 𝑑 + 1 ​ 𝑆 ~ 𝑑 + 1 βˆ’ 1 ​ Ξ£ ⋆ | β€–

≀ ( 1 βˆ’ πœ‚ 8 ) ​ ( β€– | Ξ“ 𝑑 | β€– + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€– ) + πœ‚ ​ 2 ​ ( 𝐢 24 + 𝐢 26 ​ 𝑐 πœ† ) ​ πœ… 6 𝑐 πœ† ​ β€– 𝑋 ⋆ β€– 2 ​ β€– | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 β€² | β€– + 2 ​ πœ‚ ​ ( β€– 𝑂 ~ 𝑑 β€– β€– 𝑋 ⋆ β€– ) 7 / 12 .

≀ ( 1 βˆ’ πœ‚ 8 ) ​ ( β€– | Ξ“ 𝑑 | β€– + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€– ) + πœ‚ ​ 2 ​ ( 𝐢 24 + 𝐢 26 ​ 𝑐 πœ† ) ​ πœ… 8 𝑐 πœ† ​ β€– 𝑋 ⋆ β€– 2 ​ ( β€– | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | β€– + β€– | π‘ˆ ⋆ ⊀ ​ 𝐸 | β€– ) + 2 ​ πœ‚ ​ ( β€– 𝑂 ~ 𝑑 β€– β€– 𝑋 ⋆ β€– ) 7 / 12 .

(171)

This is close to our desired conclusion, but we would need to eliminate β€– | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | β€– . To this end we shall need the following lemma.

Lemma 43.

If | | | β‹… | | | is taken to be the operator norm βˆ₯ β‹… βˆ₯ or the Frobenius norm βˆ₯ β‹… βˆ₯ π–₯ , under the same setting as Lemma 42, one has

β€– | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | β€– ≀ 32 ​ 𝑐 𝛿 ​ πœ… βˆ’ 6 ​ β€– 𝑋 ⋆ β€– 2 ​ ( β€– | Ξ“ 𝑑 | β€– + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– | 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ | β€– + ( β€– 𝑂 ~ 𝑑 β€– β€– 𝑋 ⋆ β€– ) 2 / 3 ) .

(172)

Return to the proof of Lemma 42. The conclusion follows from applying the above lemma to the term β€– | π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 | β€– in  (171), noting that 𝑐 𝛿 can be chosen sufficiently small such that

2 ​ ( 𝐢 24 + 𝐢 26 ​ 𝑐 πœ† ) 𝑐 πœ† β‹… 32 ​ 𝑐 𝛿 < 1 16 ,

and that β€– 𝑂 ~ 𝑑 β€– / β€– 𝑋 ⋆ β€– is sufficiently small due to (24), which still holds in the generalized setting in virtue of Lemma 32.

G.3.2Proof of Lemma 43

Observe that π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 has rank at most π‘Ÿ ⋆ , thus

β€– π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 β€– ≀ β€– Ξ” 𝑑 β€– , β€– π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 β€– π–₯ ≀ π‘Ÿ ⋆ ​ β€– Ξ” 𝑑 β€– .

On the other hand, from Lemma 12, we know

β€– Ξ” 𝑑 β€–

≀ 8 ​ 𝛿 ​ ( β€– 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ βˆ’ Ξ£ ⋆ 2 β€– π–₯ + β€– 𝑆 ~ 𝑑 β€– ​ β€– 𝑁 ~ 𝑑 β€– π–₯ + 𝑛 ​ β€– 𝑂 ~ 𝑑 β€– 2 )

≀ 16 ​ 𝑐 𝛿 ​ π‘Ÿ ⋆ βˆ’ 1 / 2 ​ πœ… βˆ’ 4 ​ β€– 𝑋 ⋆ β€– 2 ​ ( β€– Ξ“ 𝑑 β€– π–₯ + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– π–₯ + ( β€– 𝑂 ~ 𝑑 β€– β€– 𝑋 ⋆ β€– ) 2 / 3 )

≀ 32 ​ 𝑐 𝛿 ​ πœ… βˆ’ 4 ​ β€– 𝑋 ⋆ β€– 2 ​ ( β€– Ξ“ 𝑑 β€– + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– + ( β€– 𝑂 ~ 𝑑 β€– β€– 𝑋 ⋆ β€– ) 2 / 3 )

where the penultimate line follows from (10) and from controlling the sum inside the brackets in a similar way as (138), and the last line follows from Ξ“ 𝑑

Ξ£ ⋆ βˆ’ 1 ​ 𝑆 ~ 𝑑 ​ 𝑆 ~ 𝑑 ⊀ ​ Ξ£ ⋆ βˆ’ 1 βˆ’ 𝐼 being a matrix of rank at most π‘Ÿ ⋆ + 1 , which implies β€– Ξ“ 𝑑 β€– π–₯ ≀ π‘Ÿ ⋆ + 1 ​ β€– Ξ“ 𝑑 β€– , and similarly β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– π–₯ ≀ π‘Ÿ ⋆ ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– . The conclusion then follows from bounding β€– π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 β€– and β€– π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 β€– π–₯ separately. We have

β€– π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 β€– ≀ β€– Ξ” 𝑑 β€– ≀ 32 ​ 𝑐 𝛿 ​ πœ… βˆ’ 4 ​ β€– 𝑋 ⋆ β€– 2 ​ ( β€– Ξ“ 𝑑 β€– + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– + ( β€– 𝑂 ~ 𝑑 β€– β€– 𝑋 ⋆ β€– ) 2 / 3 ) ,

and

β€– π‘ˆ ⋆ ⊀ ​ Ξ” 𝑑 β€– π–₯
≀ π‘Ÿ ⋆ ​ β€– Ξ” 𝑑 β€–

≀ π‘Ÿ ⋆ β‹… 16 ​ 𝑐 𝛿 ​ π‘Ÿ ⋆ βˆ’ 1 / 2 ​ πœ… βˆ’ 4 ​ β€– 𝑋 ⋆ β€– 2 ​ ( β€– Ξ“ 𝑑 β€– π–₯ + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– π–₯ + ( β€– 𝑂 ~ 𝑑 β€– β€– 𝑋 ⋆ β€– ) 2 / 3 )

16 ​ 𝑐 𝛿 ​ πœ… βˆ’ 4 ​ β€– 𝑋 ⋆ β€– 2 ​ ( β€– Ξ“ 𝑑 β€– π–₯ + β€– 𝑋 ⋆ β€– βˆ’ 1 ​ β€– 𝑁 ~ 𝑑 ​ 𝑆 ~ 𝑑 βˆ’ 1 ​ Ξ£ ⋆ β€– π–₯ + ( β€– 𝑂 ~ 𝑑 β€– β€– 𝑋 ⋆ β€– ) 2 / 3 ) .

Combining the above two inequalities together proves that (172) holds with | | | β‹… | | | taken to be the operator norm or the Frobenius norm.

G.4Proof of Theorem 6

Combining Lemma 32, Lemma 38 and Lemma 40, the final 𝑑 4 given by Lemma 40 is no more than 4 Γ— 𝑇 min / 16 ≀ 𝑇 min / 2 , thus (165) holds for all 𝑑 ∈ [ 𝑇 min / 2 , 𝑇 max ] , in particular, for some 𝑇 ≀ 𝑇 min . Plugging in (165) the bound for β€– | π‘ˆ ⋆ ⊀ ​ 𝐸 | β€– given by Lemma 29 when | | | β‹… | | | is taken to be the operator norm βˆ₯ β‹… βˆ₯ or the Frobenius norm βˆ₯ β‹… βˆ₯ π–₯ , we obtain the conclusion as desired.

Report Issue Report Issue for Selection Generated by L A T E xml Instructions for reporting errors

We are continuing to improve HTML versions of papers, and your feedback helps enhance accessibility and mobile support. To report errors in the HTML that will help us improve conversion and rendering, choose any of the methods listed below:

Click the "Report Issue" button. Open a report feedback form via keyboard, use "Ctrl + ?". Make a text selection and click the "Report Issue for Selection" button near your cursor. You can use Alt+Y to toggle on and Alt+Shift+Y to toggle off accessible reporting links at each section.

Our team has already identified the following issues. We appreciate your time reviewing and reporting rendering errors we may not have found yet. Your efforts will help us improve the HTML versions for all readers, because disability should not be a barrier to accessing research. Thank you for your continued support in championing open access for all.

Have a free development cycle? Help support accessibility at arXiv! Our collaborators at LaTeXML maintain a list of packages that need conversion, and welcome developer contributions.

Xet Storage Details

Size:
255 kB
Β·
Xet hash:
9366f78b5ecc8125a6502dca523f1c2ebd48775a57f3317d8e9937d34500eabd

Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.