Running time complexity of accelerated $\ell_1$-regularized PageRank
Posed by Kimon Fountoulakis et al. (2022)
§ Problem Statement
Setup
Let be a (possibly directed) graph with and let be a row-stochastic random-walk matrix (so and ). Fix a teleportation parameter and a preference distribution with and . The Personalized PageRank (PPR) vector is the unique solution of
For a sparsity/regularization parameter , define the -regularized PPR objective
and let . Consider computing an approximate minimizer by first-order composite optimization methods, e.g. guaranteeing a prescribed accuracy such as for a chosen .
The COLT 2022 open-problem note reports that the best currently analyzed method for this task is (non-accelerated) proximal gradient, with total running time (up to logarithmic factors) and, importantly, a bound that does not depend on .
Unsolved Problem
Determine the worst-case total running time complexity of an accelerated proximal gradient method (e.g. a Nesterov/FISTA-type scheme) applied to under a comparable accuracy criterion. In particular:
- Can one prove an -independent worst-case bound of the form
thereby achieving a improvement over ?
- Failing that, can one at least prove that acceleration cannot be asymptotically worse in the worst case than , or else exhibit a family of instances for which accelerated proximal gradient is provably worse (in worst-case total running time) than the non-accelerated proximal gradient method?
§ Discussion
§ Significance & Implications
The appeal of -regularized PPR is that it targets sparse/thresholded solutions and admits algorithms whose proven total running time can be independent of the ambient graph size , which is critical in large-scale network settings. The currently best analyzed complexity scales poorly as , a regime emphasized as practically relevant in the COLT 2022 note. Establishing (or refuting) an accelerated -independent guarantee around would clarify whether one can obtain a principled speedup for this widely used regularized PPR primitive without sacrificing the key property of size-independent worst-case bounds.
§ Known Partial Results
Fountoulakis et al. (2022): Prior work (as summarized in the COLT 2022 open-problem note) motivates -regularization for PPR as a mechanism that automatically thresholds small PPR probabilities, and relates this to early termination, yielding approximate (sparse) PPR computations.
Fountoulakis et al. (2022): The COLT 2022 note states that the fastest currently analyzed method for computing -regularized PPR is a proximal gradient method with total running time .
Fountoulakis et al. (2022): A key feature emphasized in the note is that this guarantee does not depend on (the dimension/graph size), which is treated as a prerequisite for modern large-scale networks.
Fountoulakis et al. (2022): A natural idea is to apply acceleration (e.g. FISTA-type) to proximal gradient, with the goal of improving the dependence on to .
Fountoulakis et al. (2022): The note cautions that the existing proximal-gradient analysis in the cited prior work does not directly carry over to the accelerated variant.
Fountoulakis et al. (2022): The note reports empirical evidence that accelerated proximal gradient can reduce total running time in practice, but leaves open whether acceleration has a provable worst-case improvement, or could even be worse in the worst case.
§ References
Open Problem: Running time complexity of accelerated $\ell_1$-regularized PageRank
Kimon Fountoulakis, Shenghao Yang (2022)
Conference on Learning Theory (COLT), PMLR 178
📍 Open-problem note in COLT proceedings.
Open Problem: Running time complexity of accelerated $\ell_1$-regularized PageRank (PDF)
Kimon Fountoulakis, Shenghao Yang (2022)
Conference on Learning Theory (COLT), PMLR 178
📍 Proceedings PDF.