Remove the Polylogarithmic Gap to Exact Minimax Optimality
Sourced from the work of Kaizheng Wang
§ Problem Statement
Setup
Let be an input space and a symmetric positive-semidefinite kernel with RKHS and feature map (so ). Observe independent datasets:
where target labels in the target sample are unobserved.
This setup follows Wang (2023).
Assume a common regression function under covariate shift:
for some , with allowed. Define
For , kernel ridge regression on source labels is
Assume there exist constants such that almost surely, noise is conditionally -sub-Gaussian given , and either almost surely or is strongly sub-Gaussian in with proxy . Define
and effective sample size
The source proves (up to polylogarithmic factors) a high-probability upper bound of the form
where are eigenvalues of .
Unsolved Problem
Remove the remaining polylogarithmic gap. Determine whether one can (i) construct an estimator with a matching upper bound without extra polylog factors under the same assumptions, or (ii) prove a matching lower bound showing those log factors are unavoidable for this formulation.
§ Discussion
§ Significance & Implications
The paper's minimax discussion indicates matching rates up to polylogarithmic factors in the effective-sample-size regime. Resolving whether the remaining log gap is removable would sharpen the statistical optimality picture for pseudo-label-based adaptation and clarify whether the gap is methodological or intrinsic.
§ Known Partial Results
Ma et al. (2023): In arXiv:2302.10160v4, Section 4, Theorem 4.1 ("Excess risk") proves non-asymptotic high-probability excess-risk bounds driven by the effective sample size . Corollary 4.1 gives explicit rates (finite-rank, exponential, polynomial spectral regimes) that are minimax-optimal up to polylogarithmic factors under the paper's assumptions. Section 4, Remark 3 ("Optimality and adaptivity") relates sharpness to Theorem 2 of Ma-Pathak-Wainwright (2023) in bounded-likelihood-ratio settings.
§ References
Pseudo-Labeling for Kernel Ridge Regression under Covariate Shift
Kaizheng Wang (2023)
Annals of Statistics (accepted; forthcoming)
📍 Section 4: definition of effective sample size (Eq. (4.1) in the PDF), Theorem 4.1 ("Excess risk"), Corollary 4.1, and Remark 3 ("Optimality and adaptivity"); Section 7 for broader future directions.
Primary source for the minimax-up-to-polylog results. Year is recorded by arXiv first posting (2023), while v4 reflects a later revision.
Optimally tackling covariate shift in RKHS-based nonparametric regression
Cong Ma, Reese Pathak, Martin J Wainwright (2023)
The Annals of Statistics
📍 Theorem 2 (minimax lower bound under bounded likelihood-ratio assumptions), as cited in Wang's Section 4 optimality remark.
Lower-bound benchmark cited by Wang for sharpness comparisons in bounded-likelihood-ratio regimes.