Do you pay for Privacy in Online learning?
Posed by Amartya Sanyal et al. (2022)
§ Problem Statement
Setup
Fix an instance domain and a binary hypothesis class . In the realizable online (mistake-bound) model, an adversary chooses a length- labeled sequence such that there exists with for all . An (interactive, possibly randomized) learner operates for : it observes , outputs a prediction , then observes , incurring a mistake if . Let the transcript (output) of on be the full prediction sequence (a random variable over the learner's internal randomness).
Adjacency: Two labeled sequences of the same length are adjacent if they differ in exactly one position, i.e., there exists a unique with and for all .
Online differential privacy (transcript DP): is -differentially private if for all adjacent and all measurable events over transcripts,
Mistake-bound learnability: is (non-privately) learnable in the mistake-bound sense if there exists a finite and an online algorithm such that for every realizable sequence (any ), the expected number of mistakes of is at most .
Online-DP mistake-bound learnability: is online-DP-learnable if for every and there exists an online algorithm that is -DP (as above) and has a finite mistake bound on every realizable sequence (uniformly over ).
Unsolved Problem
Problem 2022. Is privacy for free in this setting? Concretely, is the set of hypothesis classes that admit a finite (non-private) mistake bound exactly the set that admit a finite -DP mistake bound for all ? If the sets coincide, what (tight) quantitative relationship is unavoidable between the optimal non-private mistake bound and the optimal private mistake bound as a function of ?
§ Discussion
§ Significance & Implications
The mistake-bound model gives a sharp, distribution-free notion of online learnability (finite number of errors on every realizable adversarial sequence). Requiring -differential privacy for the entire prediction transcript asks whether one can guarantee this form of sequential learnability while limiting how much any single labeled example can influence the released sequence of predictions. An equivalence would yield a structural characterization: every realizably learnable class would remain learnable under transcript-level DP (possibly with an explicit dependence of on ). A separation would exhibit an intrinsic privacy cost in online prediction by identifying a class with a finite non-private mistake bound but no finite DP mistake bound (for some fixed privacy parameters), pinpointing a concrete barrier for private sequential learning.
§ Known Partial Results
Sanyal et al. (2022): Sanyal and Ramponi (COLT 2022) explicitly pose the question of whether realizable mistake-bound online learnability coincides with online transcript-level -differentially private learnability.
Sanyal et al. (2022): Their open-problem note formulates the privacy requirement as differential privacy of the full prediction transcript under single-example adjacency in the labeled sequence.
Sanyal et al. (2022): The note (as provided here) motivates the question and asks for a qualitative equivalence and a quantitative relationship between the optimal non-private and private mistake bounds, but does not claim an equivalence theorem or a separating counterexample.
§ References
Open Problem: Do you pay for Privacy in Online learning?
Amartya Sanyal, Giorgia Ramponi (2022)
Conference on Learning Theory (COLT), PMLR 178
📍 Open-problem note in COLT proceedings.
Open Problem: Do you pay for Privacy in Online learning? (PDF)
Amartya Sanyal, Giorgia Ramponi (2022)
Conference on Learning Theory (COLT), PMLR 178
📍 Proceedings PDF.