Unsolved

Polynomial-time warm initialization at statistically optimal thresholds

Sourced from the work of Wanteng Ma, Dong Xia

§ Problem Statement

Setup

Let m2m\ge 2 and dimensions d1,,dmNd_1,\dots,d_m\in\mathbb N. Let TRd1××dmT^\star\in\mathbb R^{d_1\times\cdots\times d_m} be an unknown low-multilinear-rank tensor with mode-kk rank rkr_k, and define

d:=k=1mdk,d:=maxk[m]jkdj.d:=\sum_{k=1}^m d_k,\qquad d^\star:=\max_{k\in[m]}\prod_{j\ne k} d_j.

Observe i.i.d. samples

(ωi,Yi),ωiUnif([d1]××[dm]),Yi=Tωi+ξi,(\omega_i,Y_i),\qquad \omega_i\sim\mathrm{Unif}([d_1]\times\cdots\times[d_m]),\qquad Y_i=T^\star_{\omega_i}+\xi_i,

where ξi\xi_i are mean-zero noise with scale σ\sigma (e.g., Gaussian/sub-Gaussian). For mode-kk matricization Mk(T)\mathcal M_k(T^\star), define

λmin:=mink[m]σrk ⁣(Mk(T)).\lambda_{\min}:=\min_{k\in[m]}\sigma_{r_k}\!\big(\mathcal M_k(T^\star)\big).

A statistically optimal scaling regime considered in Ma & Xia (2024) (up to absolute constants and polylogarithmic factors) is

nd,λminσddn.n\gtrsim d,\qquad \frac{\lambda_{\min}}{\sigma}\gtrsim\sqrt{\frac{d^\star d}{n}}.

Ma et al. (2024) require a warm-initialization condition (Ma & Xia (2024)). Specifically, for an initializer T~\widetilde T built on an independent data split, there exists an absolute constant C1>0C_1>0 such that

T~TC1σdlogdn\|\widetilde T-T^\star\|_{\ell_\infty}\le C_1\,\sigma\sqrt{\frac{d\log d}{n}}

with high probability (the paper states probability at least 1d3m1-d^{-3m} in the corresponding assumption).

Unsolved Problem

Under the statistical scaling above, does there exist a randomized algorithm running in polynomial time in problem size that outputs, with high probability, an initializer satisfying this warm-start accuracy guarantee?

§ Discussion

Loading discussion…

§ Significance & Implications

This is a central bottleneck for achieving statistically optimal inference without oracle/computationally intractable initialization. A positive result would shrink a key statistical-to-computational gap; a negative result would formalize a computational barrier in tensor inference. See Ma & Xia (2024).

§ Known Partial Results

  • Ma et al. (2024): The paper proves that if a sufficiently accurate initializer is available, debiasing plus one-step power iteration yields asymptotic normality and optimal variance. It also gives a computationally intractable constrained least-squares initializer achieving warm start under minimal sample size (with data splitting), and polynomial-time guarantees only under stronger computational conditions.

§ References

[1]

Statistical Inference in Tensor Completion: Optimal Uncertainty Quantification and Statistical-to-Computational Gaps

Wanteng Ma, Dong Xia (2024)

Annals of Statistics (Future Papers / in press record)

📍 Section 7.1, discussion linking Condition (30) to initialization of (9) (arXiv v2 updated Nov 1, 2024).

Definitive in-press listing in Annals of Statistics Future Papers; preprint version is arXiv:2410.11225v2.

§ Tags