§ Problem Statement
Setup
Let be an instance space and a (finite or infinite) label set. For an integer , write . A -list hypothesis is a function , with - list loss on an example given by . For a distribution over , define population loss . For an i.i.d. sample , define empirical loss .
A (possibly improper) -list learning rule is a map . Its expected error at sample size on is
A distribution is realizable by a (standard, -list) concept class if there exists with . The (realizable) PAC learning curve of is
where the infimum ranges over all -list learning rules . For a fixed marginal distribution over , define by and , and the fixed-marginal learning curve
For concept classes and , define their product (direct sum) class by
For , write ( times). If is a distribution on , write for the product distribution on .
Unsolved Problem
Whether joint learning on product classes can beat the naive linear direct-sum bound, and to characterize how key learnability quantities behave under , in the following concrete forms.
-
Realizable direct-sum (worst-case distributions): For every and , one has . Does there exist and such that as ?
-
Realizable direct-sum (fixed marginal): For every , , and , one has . Does there exist such that as ?
-
Uniform convergence under products: With
determine the dependence of on and on .
- Agnostic direct sums: With
determine the dependence of on and on .
-
Minimal list size under products: For a standard class , let be the minimal such that is realizable -list PAC learnable (or if no such exists). Given , determine tight bounds on as a function of and .
-
Compressibility and products: A finite-size -list sample compression scheme for consists of a compressor and reconstructor such that for every finite labeled sample realizable by , the reconstructor outputs a -list hypothesis consistent with (i.e. for all ) from a bounded-size message produced by the compressor. Let be the minimal such that is -list compressible. Given , determine (or tightly bound) the minimal for which is realizable -list PAC learnable as a function of and .
§ Discussion
§ Significance & Implications
Direct-sum questions in learning theory ask whether learning labeled prediction problems that factor across coordinates (i.e. a product class ) necessarily costs Theta() times the samples needed for one task, or whether joint learning can provably achieve a sublinear-in- improvement in worst-case expected excess error. A sharp answer would pin down when multitask/product structure only yields the trivial union-bound scaling versus when it yields genuine sample reuse, and it would also clarify how foundational quantities (realizable and agnostic learning curves, uniform convergence rates, and minimal list size/compressibility parameters) behave under forming direct-sum/product concept classes.
§ Known Partial Results
Hanneke et al. (2024): Trivial direct-sum upper bounds hold for learning curves: for every and , ; and for every fixed marginal , .
Hanneke et al. (2024): For minimal list size, the COLT 2024 note states general multiplicative-type bounds for any classes :
Hanneke et al. (2024): For specific multiclass dimensions, the note records partial product rules: for Natarajan dimension ,
Hanneke et al. (2024): Using list Daniely-Shwartz (DS) dimension inequalities as stated in the note, non-learnability can tensorize: if is not -list learnable and is not -list learnable, then is not -list learnable.
Hanneke et al. (2024): The note highlights a route to linear-in- lower bounds for realizable direct sums via DS dimension: it states linear growth of in and relates this (via known DS-based lower bounds on realizable learning curves) to linear-in- lower bounds on up to constants and scaling, pointing to the importance of pinning down the tight relationship between and .
Hanneke et al. (2024): For compression versus learnability, the note records that -list compression can be strictly stronger than -list learnability (i.e. there exist -list learnable classes with no finite -list compression scheme), motivating the open question about how interacts with compressibility-based parameters.
§ References
Open problem: Direct Sums in Learning Theory
Steve Hanneke, Shay Moran, Waknine Tom (2024)
Conference on Learning Theory (COLT), PMLR 247
📍 Open-problem note in COLT proceedings.
Open problem: Direct Sums in Learning Theory (PDF)
Steve Hanneke, Shay Moran, Waknine Tom (2024)
Conference on Learning Theory (COLT), PMLR 247
📍 Proceedings PDF.