Is There an Online Learning Algorithm That Learns Whenever Online Learning Is Possible?
Posed by Steve Hanneke (2021)
§ Problem Statement
Setup
Let be a measurable space. An (online) binary classification protocol proceeds in rounds : an instance is revealed to the learner, the learner outputs a prediction , and then the label is revealed. The learner may be randomized, i.e., it is specified by a sequence of measurable maps such that [ \hat Y_t = f_t\bigl(X_{1:t-1},Y_{1:t-1},X_t;U\bigr), ] where denotes internal randomness independent of the instance sequence .
Assume realizability: there exists an unknown measurable target function such that for all . Define the cumulative number of mistakes through time by
Fix an (arbitrary, possibly random) instance sequence . An algorithm is
-
weakly universally consistent under if for every measurable , as (expectation over and any randomness of ), and
-
strongly universally consistent under if for every measurable , almost surely.
Say weak (respectively, strong) universal online learning is possible under if there exists at least one algorithm that is weakly (respectively, strongly) universally consistent under . Call an algorithm optimistically weakly (respectively, strongly) universal if it is weakly (respectively, strongly) universally consistent under every for which weak (respectively, strong) universal online learning is possible.
Unsolved Problem
Does there exist an optimistically universal online learning algorithm (in the weak sense, the strong sense, or both)?
Open Problem 2 (characterizing when universal online learning is possible): For a (possibly random) instance sequence and , write for the set of observed instances. For any sequence of pairwise disjoint measurable sets , define
Let be the set of all such that for every pairwise disjoint , . Let be the set of all such that for every pairwise disjoint , almost surely. Is exactly the set of instance sequences under which weak universal online learning is possible? Is exactly the set of instance sequences under which strong universal online learning is possible?
§ Discussion
§ Significance & Implications
The question asks whether online learnability in the realizable, universal sense (sublinear 0/1 mistakes simultaneously for all measurable targets ) admits a single, process-agnostic learner that succeeds whenever any learner could succeed on that same instance sequence . A positive answer would produce a canonical "no-extra-assumptions" online classifier and isolate the minimal process property that governs universal sublinear mistake rates; a negative answer would show an irreducible gap between (i) existence of some universally consistent learner tailored to a given and (ii) existence of a single strategy that automatically adapts to every learnable .
§ Known Partial Results
Hanneke (2021): formulates weak and strong universal online learning for realizable binary classification over arbitrary measurable spaces and poses the existence of an optimistically universal learner.
Hanneke (2021): proposes the criteria and (defined via sublinear growth in the number of disjoint measurable sets hit by ) as candidate characterizations of when weak/strong universal online learning is possible.
§ References
Steve Hanneke (2021)
Conference on Learning Theory (COLT), PMLR 134
📍 Open-problem note in COLT proceedings.
Steve Hanneke (2021)
Conference on Learning Theory (COLT), PMLR 134
📍 Proceedings PDF.