Better Differentially Private Learning Algorithms with Margin Guarantees
Posed by Raef Bassily et al. (2022)
§ Problem Statement
Setup
Fix sample size , privacy parameters and , a margin parameter , and a distribution over with . Let .
(Margin losses.) For any measurable score function define the (population) - error and its empirical counterpart
For define the -margin losses
For linear prediction, with , define the -hinge loss and
(Differential privacy.) A randomized algorithm is -differentially private if for all neighboring datasets differing in exactly one example and all measurable ,
Write for bounds hiding polylogarithmic factors in the relevant parameters.
Unsolved Problem
Differentially private learners achieving confidence-margin (dimension-independent) generalization guarantees with improved computational or size-dependence, in two settings:
- Faster DP margin-based learning for linear and kernel classifiers. Assume and consider
The COLT 2022 open-problem note describes an -DP algorithm outputting such that, with high probability over and the algorithm randomness,
with running time about . For kernel classification with a continuous, positive definite, shift-invariant kernel with RKHS and class , the note describes an -DP algorithm outputting such that, with high probability,
with running time about .
Question 1 (computational): Do there exist -DP algorithms for these linear and kernel problems that achieve essentially the same margin-based guarantees while having strictly better running time (as a polynomial in and ) than for linear predictors and for the above kernel approach?
- Neural networks: DP margin guarantees with no explicit size dependence. Let be a family of feed-forward neural networks of depth on inputs in , and let be those networks whose weight matrices are each bounded in Frobenius norm by . Suppose each hidden layer has width (so total size scales with ). The note describes a pure -DP algorithm outputting such that, with high probability,
which is independent of the input dimension but depends explicitly on .
Question 2 (size dependence): Is it possible to design a DP learning algorithm for with a margin-based generalization guarantee that has no explicit dependence on network size, e.g. achieving (with high probability)
§ Discussion
§ Significance & Implications
Margin-based guarantees are a central route to dimension-independent generalization bounds in classification. Under differential privacy, the COLT 2022 open-problem note highlights that existing approaches either incur substantial computational cost (even when the statistical guarantee is favorable) or, for neural networks, achieve dimension independence at the price of an explicit dependence on network size. Progress on Question 1 would turn the best-known DP margin guarantees for linear and kernel methods into algorithms with genuinely scalable polynomial-time dependence on and . Progress on Question 2 would clarify whether DP learning for neural networks can attain margin bounds controlled by norm/geometry parameters (e.g., ) without an explicit width/parameter-count term, which would align DP generalization guarantees more closely with the kind of size-insensitive margin theory studied in non-private settings.
§ Known Partial Results
Bassily et al. (2022): The COLT 2022 open-problem note (Bassily, Mohri, Suresh) formulates DP confidence-margin learning as a route to dimension-independent guarantees and identifies two main gaps: computational efficiency for linear/kernel methods and network-size dependence for neural networks.
Bassily et al. (2022): For linear predictors with and , the note describes an -DP learner achieving a high-probability guarantee of the form with running time about .
Bartlett et al. (2017): For shift-invariant kernels with RKHS norm constraint , the note describes an -DP learner achieving an analogous margin-based guarantee with running time about .
Bartlett et al. (2017): For depth- feed-forward neural networks with Frobenius-norm-bounded weight matrices (bound ) and width , the note describes a pure -DP learner that is independent of the input dimension but whose bound includes explicit dependence on (e.g., terms scaling like and ).
Bassily et al. (2022): The note poses as open whether the explicit network-size dependence in such DP margin guarantees is inherent, and whether comparable guarantees can be achieved with substantially better computational complexity for linear/kernel margin-based learners.
§ References
Open Problem: Better Differentially Private Learning Algorithms with Margin Guarantees
Raef Bassily, Mehryar Mohri, Ananda Theertha Suresh (2022)
Conference on Learning Theory (COLT), PMLR 178
📍 Open-problem note in COLT proceedings.
Open Problem: Better Differentially Private Learning Algorithms with Margin Guarantees (PDF)
Raef Bassily, Mehryar Mohri, Ananda Theertha Suresh (2022)
Conference on Learning Theory (COLT), PMLR 178
📍 Proceedings PDF.
Spectrally-normalized margin bounds for neural networks
Peter L. Bartlett, Dylan J. Foster, Matus J. Telgarsky (2017)
📍 Mentioned in the open-problem note as a non-private margin-bound comparator.