Unsolved

Do you pay for Privacy in Online learning?

Posed by Amartya Sanyal et al. (2022)

§ Problem Statement

Setup

Fix an instance domain X\mathcal{X} and a binary hypothesis class H{0,1}X\mathcal{H} \subseteq \{0,1\}^{\mathcal{X}}. In the realizable online (mistake-bound) model, an adversary chooses a length-TT labeled sequence S=((x1,y1),,(xT,yT))S=((x_1,y_1),\dots,(x_T,y_T)) such that there exists hHh^*\in\mathcal{H} with yt=h(xt)y_t=h^*(x_t) for all t[T]t\in[T]. An (interactive, possibly randomized) learner A\mathcal{A} operates for t=1,,Tt=1,\dots,T: it observes xtx_t, outputs a prediction y^t{0,1}\hat y_t\in\{0,1\}, then observes yty_t, incurring a mistake if y^tyt\hat y_t\neq y_t. Let the transcript (output) of A\mathcal{A} on SS be the full prediction sequence A(S)=(y^1,,y^T)\mathcal{A}(S)=(\hat y_1,\dots,\hat y_T) (a random variable over the learner's internal randomness).

Adjacency: Two labeled sequences S,SS,S' of the same length TT are adjacent if they differ in exactly one position, i.e., there exists a unique t[T]t\in[T] with (xt,yt)(xt,yt)(x_t,y_t)\neq (x'_t,y'_t) and (xs,ys)=(xs,ys)(x_s,y_s)=(x'_s,y'_s) for all sts\neq t.

Online differential privacy (transcript DP): A\mathcal{A} is (ε,δ)(\varepsilon,\delta)-differentially private if for all adjacent S,SS,S' and all measurable events EE over transcripts,

Pr[A(S)E]eεPr[A(S)E]+δ.\Pr[\mathcal{A}(S)\in E] \le e^{\varepsilon}\Pr[\mathcal{A}(S')\in E] + \delta.

Mistake-bound learnability: H\mathcal{H} is (non-privately) learnable in the mistake-bound sense if there exists a finite M(H)M(\mathcal{H}) and an online algorithm A\mathcal{A} such that for every realizable sequence (any TT), the expected number of mistakes of A\mathcal{A} is at most M(H)M(\mathcal{H}).

Online-DP mistake-bound learnability: H\mathcal{H} is online-DP-learnable if for every ε>0\varepsilon>0 and δ[0,1)\delta\in[0,1) there exists an online algorithm Aε,δ\mathcal{A}_{\varepsilon,\delta} that is (ε,δ)(\varepsilon,\delta)-DP (as above) and has a finite mistake bound Mε,δ(H)M_{\varepsilon,\delta}(\mathcal{H}) on every realizable sequence (uniformly over TT).

Unsolved Problem

Problem 2022. Is privacy for free in this setting? Concretely, is the set of hypothesis classes that admit a finite (non-private) mistake bound exactly the set that admit a finite (ε,δ)(\varepsilon,\delta)-DP mistake bound for all ε>0,δ[0,1)\varepsilon>0,\delta\in[0,1)? If the sets coincide, what (tight) quantitative relationship is unavoidable between the optimal non-private mistake bound M(H)M(\mathcal{H}) and the optimal private mistake bound Mε,δ(H)M_{\varepsilon,\delta}(\mathcal{H}) as a function of (ε,δ)(\varepsilon,\delta)?

§ Discussion

Loading discussion…

§ Significance & Implications

The mistake-bound model gives a sharp, distribution-free notion of online learnability (finite number of errors on every realizable adversarial sequence). Requiring (ε,δ)(\varepsilon,\delta)-differential privacy for the entire prediction transcript asks whether one can guarantee this form of sequential learnability while limiting how much any single labeled example can influence the released sequence of predictions. An equivalence would yield a structural characterization: every realizably learnable class would remain learnable under transcript-level DP (possibly with an explicit dependence of Mε,δ(H)M_{\varepsilon,\delta}(\mathcal{H}) on (ε,δ)(\varepsilon,\delta)). A separation would exhibit an intrinsic privacy cost in online prediction by identifying a class with a finite non-private mistake bound but no finite DP mistake bound (for some fixed privacy parameters), pinpointing a concrete barrier for private sequential learning.

§ Known Partial Results

  • Sanyal et al. (2022): Sanyal and Ramponi (COLT 2022) explicitly pose the question of whether realizable mistake-bound online learnability coincides with online transcript-level (ε,δ)(\varepsilon,\delta)-differentially private learnability.

  • Sanyal et al. (2022): Their open-problem note formulates the privacy requirement as differential privacy of the full prediction transcript under single-example adjacency in the labeled sequence.

  • Sanyal et al. (2022): The note (as provided here) motivates the question and asks for a qualitative equivalence and a quantitative relationship between the optimal non-private and private mistake bounds, but does not claim an equivalence theorem or a separating counterexample.

§ References

[1]

Open Problem: Do you pay for Privacy in Online learning?

Amartya Sanyal, Giorgia Ramponi (2022)

Conference on Learning Theory (COLT), PMLR 178

📍 Open-problem note in COLT proceedings.

[2]

Open Problem: Do you pay for Privacy in Online learning? (PDF)

Amartya Sanyal, Giorgia Ramponi (2022)

Conference on Learning Theory (COLT), PMLR 178

📍 Proceedings PDF.

§ Tags