Tight Convergence of SGD in Constant Dimension
Posed by Tomer Koren et al. (2020)
§ Problem Statement
Setup
Fix a dimension that is a constant independent of the iteration budget . Let be a nonempty closed convex set with finite Euclidean diameter . Let be convex and attain its minimum on , and fix some minimizer . Assume access to a stochastic first-order oracle such that, for each query point , it returns a random vector satisfying (i) unbiasedness: , and (ii) bounded second moment: for a known constant .
Projected stochastic gradient descent (SGD) starts from some and iterates
where is Euclidean projection and is a stepsize schedule that may depend on but not on additional unknown properties of beyond the model above. The performance measure is the expected last-iterate suboptimality , where the expectation is over the oracle randomness.
Unsolved Problem
Characterize the minimax-optimal dependence on of the worst-case expected last-iterate suboptimality achievable by projected SGD over the above problem class when is fixed. In particular, determine whether there exist constants (depending only on ) and a stepsize schedule such that for all ,
and whether this characterization already holds in the one-dimensional case .
§ Discussion
§ Significance & Implications
SGD is typically deployed without iterate averaging, so last-iterate guarantees govern the algorithm's actual output. A tight minimax characterization in fixed (especially one) dimension would resolve a basic mismatch between existing upper and lower bounds for in the standard stochastic subgradient oracle model, clarifying whether last-iterate SGD fundamentally matches the canonical stochastic-optimization rate in low dimensions or is intrinsically slower.
§ Known Partial Results
Koren et al. (2020): Koren and Segal (COLT 2020 open problem) highlight that, when the dimension is a constant independent of , known upper and lower bounds for the expected last-iterate suboptimality of projected SGD do not match.
Koren et al. (2020): They emphasize that this gap persists even for .
Koren et al. (2020): They provide evidence in the one-dimensional case consistent with a last-iterate rate.
Koren et al. (2020): They conjecture that the same last-iterate rate should hold for any constant dimension .
§ References
Open Problem: Tight Convergence of SGD in Constant Dimension
Tomer Koren, Shahar Segal (2020)
Conference on Learning Theory (COLT), PMLR 125
📍 Open-problem note in COLT proceedings.
Open Problem: Tight Convergence of SGD in Constant Dimension (PDF)
Tomer Koren, Shahar Segal (2020)
Conference on Learning Theory (COLT), PMLR 125
📍 Proceedings PDF.