Buckets:

|
download
raw
199 kB

Title: Realizable Learning is All You Need

URL Source: https://arxiv.org/html/2111.04746

Markdown Content: Back to arXiv

This is experimental HTML to improve accessibility. We invite you to report rendering errors. Use Alt+Y to toggle on accessible reporting links and Alt+Shift+Y to toggle off. Learn more about this project and help improve conversions.

Why HTML? Report Issue Back to Abstract Download PDF 1Introduction 2Distribution Family Classification 3Related Work 4Preliminaries 5The Core Reduction: Finite Label Classes 6Four Modification Archetypes 7Doubly Bounded Loss 8Robust Learning 9Partial PAC-Learning 10Uniform Stability 11Statistical Query Model 12Fairness 13Notions of Coverability License: CC BY 4.0 arXiv:2111.04746v4 [cs.LG] 03 Feb 2024 \ThCSauthor

[ucsd-cs]Max Hopkinsnmhopkin@ucsd.edu \ThCSauthor[ucsd-cs,ucsd-math]Daniel M. Kanedakane@eng.ucsd.edu \ThCSauthor[ucsd-cs]Shachar Lovettslovett@cs.ucsd.edu \ThCSauthor[yale]Gaurav Mahajangaurav.mahajan@yale.edu \ThCSaffil[ucsd-cs]Department of Computer Science and Engineering, UCSD, California, CA 92092 \ThCSaffil[ucsd-math]Department of Mathematics, UCSD, California, CA 92092 \ThCSaffil[yale]Department of Computer Science, Yale University, Connecticut, CT 06511 \ThCSshorttitleRealizable Learning is All You Need \ThCSyear2024 \ThCSarticlenum2 \ThCSreceivedSep 28, 2022 \ThCSrevisedNov 21, 2023 \ThCSacceptedDec 11, 2023 \ThCSpublishedFeb 3, 2024 \ThCSkeywordsPAC-Learning, Realizable Learning, Agnostic Learning, Blackbox Reductions \ThCSdoi10.46298/theoretics.24.2 \ThCSshortnamesM. Hopkins, D. M. Kane, S. Lovett and G. Mahajan \ThCSthanksInvited article from COLT 2022.

Realizable Learning is All You Need Abstract

The equivalence of realizable and agnostic learnability is a fundamental phenomenon in learning theory. With variants ranging from classical settings like PAC learning and regression to recent trends such as adversarially robust learning, itโ€™s surprising that we still lack a unified theory; traditional proofs of the equivalence tend to be disparate, and rely on strong model-specific assumptions like uniform convergence and sample compression.

In this work, we give the first model-independent framework explaining the equivalence of realizable and agnostic learnability: a three-line blackbox reduction that simplifies, unifies, and extends our understanding across a wide variety of settings. This includes models with no known characterization of learnability such as learning with arbitrary distributional assumptions and more general loss functions, as well as a host of other popular settings such as robust learning, partial learning, fair learning, and the statistical query model.

More generally, we argue that the equivalence of realizable and agnostic learning is actually a special case of a broader phenomenon we call property generalization: any desirable property of a learning algorithm (e.g. noise tolerance, privacy, stability) that can be satisfied over finite hypothesis classes extends (possibly in some variation) to any learnable hypothesis class.

1Introduction

The equivalence of realizable and agnostic learnability in Valiantโ€™s Probably Approximately Correct (PAC) model [58] is one of the best known results in learning theory, and numbers among its most surprising. Given a set ๐‘‹ and a family of binary classifiers ๐ป , the result states that the ability to learn a classifier โ„Ž โˆˆ ๐ป from examples of the form ( ๐‘ฅ , โ„Ž โข ( ๐‘ฅ ) ) is in fact sufficient for something much stronger: given samples from any distribution ๐ท over ๐‘‹ ร— { 0 , 1 } , it is possible to learn the best approximation to ๐ท in ๐ป . This surprising equivalence stems from a classical result of Vapnik and Chervonenkis (VC) [60], and independently Blumer, Ehrenfeucht, Haussler, and Warmuth (BEHW) [19] and Haussler [36], who equate both the former model (known as realizable learning) and the latter model (known as agnostic learning) to a strong property of pairs ( ๐‘‹ , ๐ป ) called uniform convergence.1

VC, BEHW, and Hausslerโ€™s result was certainly a breakthrough in its own right, but its proof technique is too indirect to reveal any deeper connections between realizable and agnostic learning beyond the PAC setting. Further, recent years have seen both theory and practice shift away not only from this original formalization, but more generally from the โ€œuniform convergence equals learnabilityโ€ paradigm, often in favor of distributional or data-dependent assumptions like margin that are more applicable to the real world. The inability of VC, BEHW, and Hausslerโ€™s proof technique to generalize to such scenarios raises a fundamental question: is the equivalence of realizable and agnostic learning a fundamental property of learnability, or simply a happy coincidence derived from the original PAC framework?

In the 30 years since these works, a mountain of evidence has amassed in favor of the former: almost every reasonable variant of learning shares some sort of similar equivalence. This includes a long list of popular settings such as regression [9], distribution-dependent learning [17], multi-class learning [29], robust learning [52], online learning [16], private learning [13, 3], and partial learning [47, 5]. Whatโ€™s more, the uniform convergence paradigm fails miserably in most of these models. In the distribution-dependent model, for instance, it is easy to build classes which are trivially learnable (even with one sample!) but completely fail to satisfy uniform convergence [17]. On the other hand, models such as private learning give well-known examples where uniform convergence fails to imply learnability [6]. In spite of this, we are really no closer today to a general understanding of this phenomenon than we were in the early 90s. Much like Vapnik and Chervonenkis [60], Blumer, Ehrenfeucht, Haussler, and Warmuth [19], and Hausslerโ€™s [36] proofs, the above works often use indirect methods and tend to rely on powerful model-dependent assumptions.

In this work, we aim to offer a generic, unifying theory by way of the first direct reduction from agnostic to realizable learning. Unlike any previous work, our reduction is blackbox, relies on no additional assumptions, and, perhaps most importantly, is incredibly simple. In fact, the basic algorithm can be stated in three lines.

Input: Realizable PAC-Learner ๐’œ , Unlabeled Sample Oracle ๐’ช ๐‘ˆ , Labeled Sample Oracle ๐’ช ๐ฟ Algorithm:

1.

Draw an unlabeled sample ๐‘† ๐‘ˆ โˆผ ๐’ช ๐‘ˆ , and labeled sample ๐‘† ๐ฟ โˆผ ๐’ช ๐ฟ .

2.

Run ๐’œ over all possible labelings of ๐‘† ๐‘ˆ to get:

๐ถ โข ( ๐‘† ๐‘ˆ ) โข \coloneqq โข { ๐’œ โข ( ๐‘† ๐‘ˆ , โ„Ž โข ( ๐‘† ๐‘ˆ ) ) | โ„Ž โˆˆ ๐ป | ๐‘† ๐‘ˆ } .

3.

Return the hypothesis in ๐ถ โข ( ๐‘† ๐‘ˆ ) with lowest empirical error over ๐‘† ๐ฟ .

Algorithm 1 Agnostic to Realizable Reduction

This basic reduction simplifies and unifies classic results such as VC [60], BEHW [19], and Hausslerโ€™s [36] distribution-free equivalence and Benedek and Itaiโ€™s [17] analogous result in the distribution-dependent setting2 up to log factors. More importantly, because Algorithm 1 doesnโ€™t rely on model-dependent properties like uniform convergence, it extends to learning regimes without known characterizations. One such example is the notoriously difficult distribution-family model, in which the adversary is given a restricted family of distributions ๐’Ÿ along with the pair ( ๐‘‹ , ๐ป ) . The distribution-family model has no finitary characterization [44],3 yet Algorithm 1 can still be used to show that the realizable and agnostic settings are equivalent.

Unfortunately, while Algorithm 1 does avoid any significant blowup in sample complexity, it is inherently computationally inefficient. In fact, this is necessary unless ๐‘ƒ

๐‘ โข ๐‘ƒ . There are many basic classes (e.g. halfspaces) which are easy to learn in the realizable model, but NP or cryptographically hard in the agnostic setting (see e.g. [33]). As such we focus in this work only on information theoretic considerations, though building computationally efficient reductions in restricted settings remains an interesting avenue of research.

The core of Algorithm 1 lies in an equivalence between PAC learning and a new form of randomized covering of independent interest we call a non-uniform cover. In contrast to more classical notions, a non-uniform cover is a distribution over subsets of hypotheses that covers any fixed hypothesis in the class with high probability, but may fail to cover all hypotheses simultaneously. The connection between supervised learning and non-uniform covering is inherent in Algorithm 1, where Steps 1 and 2 turn the realizable learner ๐’œ into a non-uniform cover ๐ถ โข ( ๐‘† ๐‘ˆ ) , and Step 3 uses the cover to perform agnostic learning. At a high level, this process works because the adversary does not see the randomness inherent to Steps 1 and 2, and therefore cannot detect or exploit which hypotheses in the class will fail to be well-estimated in the process.

In fact, this connection has many broader implications within the theory of supervised learning. For one, the method is not inherently restricted to the agnostic setting. Algorithm 1 achieves agnostic learning for general classes by applying an agnostic learner for finite classes in Step 3; replacing this with learners satisfying other properties (e.g. the exponential mechanism for privacy) leads to other families of reductions to the realizable setting. At a high level, this can be summarized by the following informal โ€˜guiding principleโ€™:

Guiding Principle (Property Generalization).

If there is a (sample-efficient) algorithm with property ๐‘ƒ over finite classes, then Algorithm 1 gives a (sample-efficient) learner with property ๐‘ƒ over any โ€˜learnableโ€™ class.

We stress that the above is a guide, not a theorem, and indeed often requires modification or weakening of the desired property for a given application. For instance, when Step 3 is replaced with a private algorithm, the result is a semi-private learner for general classes, a weakened model allowing the use of a small amount of public unlabeled data [14, 2].

Algorithm 1 also provides a unified framework for many settings beyond the PAC-model. This includes basic extensions such as general loss functions4 and the distribution-family model, but also more involved modifications such as partial learning, robust learning, or even the statistical query model. Moreover, in some of these settings removing reliance on setting-specific assumptions like uniform convergence actually quantitatively improves the sample complexity; such is the case for the semi-private model, where we use this fact to achieve information-theoretically optimal unlabeled sample complexity for the first time.

Finally, we note there are a few settings where Algorithm 1 runs into issues, especially discrete infinite settings such as infinite multi-class classification and properties such as privacy that require more careful data handling. We leave the extension of our method to these settings as an intriguing open problem.

1.1Proof Overview

Before moving to more detailed discussion of our results, we briefly overview the proof that Algorithm 1 is an agnostic learner. At its most basic, the idea is simply to observe that the set ๐ถ โข ( ๐‘† ๐‘ˆ ) almost certainly contains a โ€˜near-optimalโ€™ hypothesis โ„Ž ~ โˆˆ ๐ป . Since ๐ถ โข ( ๐‘† ๐‘ˆ ) has bounded size, standard uniform convergence for finite classes promises Step 3 outputs a hypothesis close to โ„Ž ~ , which is therefore itself โ€˜near-optimalโ€™ as desired.

Taking a step back, the key observation in this process is really that ๐ถ โข ( ๐‘† ๐‘ˆ ) โ€˜acts like a coverโ€™ of ๐ป in the following weak sense we call non-uniform covering:

Definition 1.1 (Non-uniform Cover (Informal Definition 5.1)).

Let ( ๐‘‹ , ๐ป ) be a hypothesis class, ๐ท a marginal distribution over ๐‘‹ , and ๐ถ a random variable over the power set ๐‘ƒ โข ( ๐ป ) . We call ๐ถ a non-uniform ( ๐œ€ , ๐›ฟ ) -cover of ๐ป with respect to ๐ท if:

โˆ€ โ„Ž โˆˆ ๐ป : Pr ๐ถ [ โˆƒ โ„Ž โ€ฒ โˆˆ ๐ถ : Pr ๐‘ฅ โˆผ ๐ท [ โ„Ž โ€ฒ ( ๐‘ฅ ) โ‰  โ„Ž ( ๐‘ฅ ) ] โ‰ค ๐œ€ ] โ‰ฅ 1 โˆ’ ๐›ฟ .

In other words, for every fixed hypothesis โ„Ž in the class, ๐ถ is very likely to contain a hypothesis close to โ„Ž . This differs from the standard covering which enforces that ๐ถ simultaneously cover every โ„Ž โˆˆ ๐ป (equivalent to pushing the โ€˜ โˆ€ โ€™ quantifier into the probability above). In Section 13, we show that the latter (which is the standard in the literature) is strictly stronger and requires more samples to generate. This is critical in our application to semi-private learning where we cut the additional samples required to generate a full cover and thereby achieve the optimal sample complexity.

It is left to observe that ๐ถ โข ( ๐‘† ๐‘ˆ ) in Algorithm 1 is actually a non-uniform cover, but this is essentially immediate from the definition of realizable learning! Realizable learning promises that for every fixed hypothesis โ„Ž โˆˆ ๐ป , when ๐ด receives samples labeled by โ„Ž it outputs a hypothesis close to โ„Ž with high probability. ๐ถ โข ( ๐‘† ๐‘ˆ ) is generated by running ๐ด across all โ„Ž โˆˆ ๐ป , so this guarantee exactly translates to the above. We refer the reader to Sections 2 and 5 for a formal explanation of this argument.

1.2Beyond PAC Learning

Algorithm 1 is an extremely flexible framework for proving agnostic to realizable reductions in supervised learning. In this section, we informally overview the many extended models studied in this paper. The most basic setting in which Algorithm 1 applies beyond the standard model is learning under distributional assumptions (or formally what we call the โ€œdistribution-family modelโ€, see Sections 2 and 4). Standard techniques using combinatorial dimensions require that the learner ๐ด works over every distribution (the โ€œdistribution-freeโ€ setting), while modern algorithms frequently only work under some niceness conditions on the distribution. On the other hand, it is easy to observe that the process described above works under arbitrary distributional assumptionsโ€”as long as ๐ด is a realizable learner over a distribution ๐ท , then Algorithm 1 is an agnostic learner whenever the data is marginally distributed as ๐ท . This leads to the following corollary:

Theorem 1.2 (Distribution Family Model (Informal Theorem 5.4)).

The sample complexity of agnostic learning a class ( ๐‘‹ , ๐ป ) in the distribution family model is at most:

๐‘š โข ( ๐œ€ , ๐›ฟ ) โ‰ค ๐‘‚ โข ( ๐‘› โข ( ๐œ€ , ๐›ฟ ) + log โก ( 1 / ๐›ฟ ) ๐œ€ 2 ) ,

where ๐‘› โข ( ๐œ€ , ๐›ฟ ) is the realizable sample complexity of ( ๐‘‹ , ๐ป ) .

In the finite VC setting, this can be improved to roughly ๐‘‘ โข log โก ( ๐‘‘ ๐œ€ ) + log โก ( 1 ๐›ฟ ) ๐œ€ 2 for VC dimension ๐‘‘ , which is optimal up to a log factor.

Algorithm 1 also covers many other supervised settings in the literature that diverge from the PAC model in more substantial ways, including general loss functions, constraints on the learner, and properties beyond agnostic learning. For simplicity we cover these here informally, avoiding exact definitions and sample complexities (which are generally similar to the above), and give formal details in the main body.

1.2.1General Loss Functions

Perhaps the most natural extension of the PAC framework is to loss functions beyond binary classification. Here data points are labeled by an arbitrary label space ๐‘Œ and error is measured with respect to a generic loss function โ„“ : ๐‘Œ ร— ๐‘Œ โ†’ โ„ โ‰ฅ 0 , for instance we may take ๐‘Œ

โ„ and โ„“ โข ( ๐‘ฆ , ๐‘ฆ โ€ฒ )

( ๐‘ฆ โˆ’ ๐‘ฆ โ€ฒ ) 2 , the square loss. In general, agnostic and realizable learning are actually not equivalent in this setting, even for nice loss functions (see Proposition 6.1). The issue is that over infinite label spaces it is possible to encode hypotheses into the labels with infinite precision, making the class trivial to learn realizably, but impossible even with the smallest amount of noise (which wipes out the encoding). We show this type of counter-example is essentially the only barrier to agnostic learnability.

Somewhat more formally, we call a class ๐ป discretely learnable if for every ๐œ€

0 , there exists an ๐œ€ -discretization5 of ๐ป that is learnable up to ๐‘‚ โข ( ๐œ€ ) error. Discrete learnability can informally be thought of as a very weak type of noise tolerance that essentially acts only to rule out the above construction. We prove discrete and agnostic learnability are equivalent under weak conditions on the loss function.

Theorem 1.3 (General Loss (Informal Theorems 7.2 and 6.6)).

If โ„“ is a (probably) upper bounded loss function6 and either:

1.

โ„“ is bounded from below

2.

โ„“ satisfies a ๐ถ -approximate triangle inequality

then discrete and agnostic learnability are equivalent under โ„“ up to polynomial factors.

We remark that in the latter case, the learner only achieves ๐ถ โ‹… ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ error and that this is tight (Proposition 6.8). To our knowledge, these results are new even to the distribution-free setting, where such an equivalence was only known for bounded Lipschitz [9, 62] or binary-valued [15, 29] loss functions. In the models below, similar guarantees hold under the above assumptions on loss. We omit the exact dependence on โ„“ for simplicity and discuss where relevant in the main body.

Another well-studied setting for loss functions beyond standard classification is adversarial robustness. Robust learning is an extension of the PAC model introduced to handle adversarial perturbations at test time by taking a maximum of the loss function over specified โ€˜perturbation setsโ€™ around the test point. We give a modification of Algorithm 1 in the robust setting that handles general loss and distributional assumptions.

Theorem 1.4 (Robust Classification (Informal Theorem 8.5)).

Robust realizable and robust agnostic learning are equivalent up to polynomial factors.

Note this is completely independent of the perturbation sets. In the classification setting, Theorem 1.4 generalizes recent work giving such an equivalence in the distribution-free model [52, 50], though the sample complexity of our algorithm suffers an extra factor of ๐œ€ โˆ’ 1 in this special case.

Finally, another recent setting that works with a sort of โ€˜modifiedโ€™ loss function is the learning of partial functions, capturing โ€˜data-dependentโ€™ models such as halfspaces with geometric margin where. Here the functions in ๐ป are allowed to be โ€˜undefinedโ€™ on some portions of ๐‘‹ , and the loss is fixed to 1 under any undefined point no matter the response of the algorithm.

Theorem 1.5 (Partial functions (Informal Theorem 9.5)).

Realizable and agnostic learning of partial functions are equivalent up to polynomial factors.

In the distribution-free setting, a variant of this equivalence is known via compression schemes, but the above is new under distributional assumptions and more general loss (again at the cost of an extra ๐œ€ โˆ’ 1 factor).

1.2.2Constrained Models

Another frequent modification of the PAC setting is to impose constraints either on how the algorithm ๐ด uses the data, or on various properties of the output classifier itself. In this section, we cover application to three such examples: fairness, stability, and statistical queries.

We start with fairness, where the goal is to output a hypothesis with low error conditional on โ€˜treating similar individuals similarlyโ€™ under a fixed metric on the data space.

Theorem 1.6 (Fair Learning (Informal Theorem 12.3)).

Realizable and agnostic fair learning are equivalent up to polynomial factors.

Formally, the above result holds in Rothblum and Yonaโ€™s [63] popular โ€˜Probably Approximately Correct and Fairโ€™ (PACF)-learning. To our knowledge it is the first such reduction even in the distribution-free setting, as Rothblum and Yona [63] study the agnostic case directly.

Another popular constraint in the literature is algorithmic stability, where one roughly imposes that running the algorithm twice should produce similar results. We study a simple variant known as uniform stability [24] (closely related to โ€˜private predictionโ€™ [32]) which promises that for every fixed point in the space ๐‘ฅ the output distribution of ๐ด on ๐‘ฅ is similar when ๐ด is run on neighboring datasets.

Theorem 1.7 (Uniform Stability (Informal Theorem 10.2)).

Realizable and agnostic learning are equivalent under uniform stability up to polynomial factors.

This is known in the distribution-free setting by VC arguments [32, 24], but is new for general loss and under distributional assumptions. Unlike prior examples the sample complexity matches prior techniques.

Another way to constrain the algorithm ๐ด is through the way it interacts with data. Perhaps the most popular example of this is the statistical query model, where ๐ด is constrained to approximating general population statistics (such algorithms are then guaranteed to satisfy certain nice properties such as noise-tolerance and privacy). We give a variant of our reduction for the statistical query setting as well.

Theorem 1.8 (Statistical Query Model (Informal Theorem 11.2)).

Realizable and agnostic learning are equivalent in the statistical query model up to polynomial factors.

Variants of such a result are known via combinatorial characterization in the fixed distribution setting [57, 33], but new in our general setup.

1.2.3Beyond Agnostic Learning

In the introduction, we claimed Algorithm 1 can be used to build a learner satisfying any โ€œfinitely-satisfiableโ€ property, not just agnostic learning. In fact, one of the examples above already displays this property: Theorem 1.7 does not require the base learner to be uniformly stable. Instead, we apply a uniformly stable agnostic learner to ๐ถ โข ( ๐‘† ๐‘ˆ ) in Step 3 of Algorithm 1 and โ€˜liftโ€™ uniform stability from finite to infinite classes. We give two further applications of this strategy: to malicious noise, and (semi)-private learning.

Kearns and Liโ€™s [41] malicious noise is a model for data contamination where one is given a faulty sample oracle ๐’ช ๐‘€ โข ( โ‹… ) which (with some small probability) returns a completely adversarial pair ( ๐‘ฅ , ๐‘ฆ ) , and otherwise draws from a true โ€˜ground truthโ€™ distribution. A learner is said to be tolerant to malicious noise if it achieves standard PAC guarantees while drawing from ๐’ช ๐‘€ โข ( โ‹… ) . Like agnostic learning, tolerance to malicious noise is known to be achievable for finite classes [41]. As a result, (a slight modification of) Algorithm 1 gives a blackbox reduction from learning with malicious noise to realizable learning.

Theorem 1.9 (Realizable โ†’ Malicious (Informal Theorem 6.11)).

Realizable learning and learning with malicious noise are equivalent up to polynomial factors.

This extends the original result of [41] to the distribution-family and general loss settings.

Our final application of Algorithm 1 is to the ubiquitous notion of differential privacy. Informally, an algorithm is said to be differentially private if its output is not susceptible to small changes in the underlying sample (see Section 6.3). Privacy is a strong condition, and even relaxed notions require finite Littlestone dimension in the distribution-free setting [6], ruling out a direct reduction from realizable learning. However, Algorithm 1 does recover a weaker variant known as semi-privacy [14] where the learner is allowed to use a few โ€˜publicโ€™ unlabeled samples, but must otherwise maintain privacy with respect to the data.

Theorem 1.10 (Realizable โ†’ Semi-Private (Informal Theorem 6.21)).

Realizable and semi-private learning are equivalent up to polynomial factors.

This generalizes Beimel, Nissim, and Stemmerโ€™s [14] equivalence in the distribution-free setting and extends to general loss. Perhaps most interesting is that in this case, due to relying only on non-uniform covering over more classical uniform convergence arguments, Algorithm 1 actually gives gives a quantitative improvement over prior methods.

Corollary 1.11 (Semi-Optimal Semi-Private Learning (Informal Corollary 6.23)).

Let ( ๐‘‹ , ๐ป ) be a class with VC-dimension ๐‘‘ . Then ( ๐‘‹ , ๐ป ) is ๐›ผ -semi-private, agnostically learnable in

๐‘š ๐‘๐‘ข๐‘ โข ( ๐œ€ , ๐›ฟ , ๐›ผ ) โ‰ค ๐‘‚ โข ( ๐‘‘ + log โก ( 1 / ๐›ฟ ) ๐œ€ )

unlabeled (public) samples, and

๐‘š ๐‘๐‘Ÿ๐‘– โข ( ๐œ€ , ๐›ฟ , ๐›ผ ) โ‰ค ๐‘‚ โข ( ๐‘‘ โข log โก ( 1 / ๐œ€ ) + log โก ( 1 / ๐›ฟ ) ๐œ€ โ‹… min โก { ๐œ€ , ๐›ผ } )

labeled (private) samples.

For fixed ๐‘‘ and ๐›ฟ , the public sample complexity of Corollary 1.11 is tight and resolves a conjecture of Alon, Bassily, and Moran [2] who gave the corresponding lower bound.7 The private complexity is off by a log factor from known lower bounds [22], and it remains an interesting question whether this can be fixed.

Theorem 1.10 and Corollary 1.11 are also robust to light distribution shift between the public and private databases. This problem, called covariate shift, is a commonly observed in machine learning practice and is especially of concern in privacy where a distribution over โ€œopt-inโ€ public users could easily differ from the overall distribution of private data. We discuss covariate shift in more depth in Section 6.4.

1.3Modification Archetypes

Finally before moving to formal presentation of our methods and results, we briefly overview the four generic modifications of Algorithm 1 used in the above extensions, and outline where they can be found formally in the main body of the paper.

Discretization.

Weโ€™ll start by discussing our main technique to extend Algorithm 1 to infinite label spaces. The basic idea is simple: since we cannot afford to run our learner over all possible labelings of ๐‘† ๐‘ˆ , we instead run the learner over labelings coming from some discretization of the class. As long as we have access to a learner for the discretization, we can then use the same arguments covered in Section 2 to prove various occurrences of property generalization. We formalize these notions more generally in Section 6.1, where we use the technique to prove Theorem 1.3. Discretization can also be used to handle learning models such as the statistical query setting which output real-valued query responses (see Section 11).

Subsampling.

Another core limitation to Algorithm 1 is access to clean unlabeled data. Algorithm 1 works by running a realizable learner over a representative set of unlabeled data, but, in practice, such data may often be corrupted, and data-dependent assumptions such as margin might mean that the optimal hypothesis isnโ€™t even well-defined on this set. We handle cases like these by a simple sub-sampling procedure: instead of running our realizable learner over labelings of ๐‘† ๐‘ˆ , we run the learner over all labelings of all subsets of ๐‘† ๐‘ˆ . As long as ๐‘† ๐‘ˆ contains some amount of uncorrupted data, this subsampling procedure will find it and we can maintain the guarantees discussed in Section 2. We use this technique to prove property generalization for models such as robust learning (see Theorem 1.4 and Section 8), partial learning (see Section 9), and malicious noise (see Theorem 1.9 and Section 6.2).

Replacing the Finite Learner.

In the introduction, we proposed a general paradigm (guiding principle) called property generalization: that a variant of any learning property which holds for finite classes should in fact hold for any โ€œlearnableโ€ class in the base model. The main idea relies on replacing Step 3 of Algorithm 1 (which, as stated, is an empirical risk minimization process) with a generic learner for finite classes with the desired property. For noise-tolerance properties such as agnostic and malicious noise, empirical risk minimization works. Properties such as privacy or stability, however, require a different finite learner. To prove Theorem 1.10, for example, we replace the ERM process in Algorithm 1 with the exponential mechanism [49]. We use a similar strategy in Section 10 to prove an analogous result for uniform stability.

Replacing the Base Learner.

Finally we note a very basic modification of Algorithm 1 that allows us to extend property generalization beyond the PAC setting: simply replace the input realizable PAC learner with a realizable learner in the desired model. This is usually combined with one of the techniques above depending on the specific application, e.g. to prove property generalization for robust learning and the statistical query model. The same idea can also be used to analyze semi-private learning with covariate shift (see Section 6.4) and property generalization for fair learning (see Section 12).

2Distribution Family Classification

Since all of our results are derived from variants of Algorithm 1, it is instructive to start by considering its basic analysis in our simplest non-trivial setting: distribution-family classification. We remark that this section is entirely pedalogical, and the results are subsumed in Section 5.

The distribution-family model captures learnability with arbitrary distributional assumptions, a well-studied relaxation of PAC learning in practice where worst-case distributional assumptions are often too strong, and encompasses both the distribution-free and distribution-dependent PAC settings. Unlike these models, however, we cannot assume uniform convergence in the distribution family setting, due to a classical example of Benedek and Itai [17].

Proposition 2.1 (Benedek and Itai [17]).

There exists a PAC-learnable class ( ๐ท , ๐‘‹ , ๐ป ) over binary labels and classification loss without the uniform convergence property.

Proof 2.2.

Let ๐‘‹

[ 0 , 1 ] , ๐ท be the uniform distribution over ๐‘‹ , ๐‘Œ

{ 0 , 1 } , and ๐ป consist of all indicator functions for finite sets ๐‘† โŠ‚ ๐‘‹ , as well as for ๐‘‹ itself. It is not hard to see that ( ๐ท , ๐‘‹ , ๐ป ) is realizably PAC-learnable by the following scheme in only a single sample: if the learner draws a sample labeled 1 , output the all 1 โ€™s function. Otherwise, output all 0 s. When the adversary has chosen a finite set, with probability 1 the learner draws a sample labeled 0 , and outputs a hypothesis with 0 error (since the finite set has measure 0 ). If the adversary chooses the all 1 โ€™s function, the learner will always output the all 1 โ€™s function.

On the other hand, it is clear that when the adversary chooses the all 1 โ€™s function, no matter how many samples the learner draws, there will exist a hypothesis in the class that is poorly approximated by the sample. Namely the hypothesis whose support is given by the support of the sample itself has empirical measure 1 , but true measure 0 . As a result, this class fails to have the uniform convergence property despite its learnability.

This simple example rules out many classical techniques in supervised learning, including uniform convergence based reductions (e.g. [8, 14, 2]). Furthermore, since the distribution-family setting has no finitary characterization, we also cannot hope to take the traditional approach [60, 17, 36, 9] of proving equivalence of realizable and agnostic models by simultaneous characterization.

With this motivation in mind, letโ€™s define distribution-family learning more formally. Let ๐‘‹ be a set (called the instance space), ๐‘Œ

{ 0 , 1 } the set of binary labels, ๐’Ÿ a family of distributions over ๐‘‹ , and ๐ป

{ โ„Ž : ๐‘‹ โ†’ ๐‘Œ } a family of binary classifiers. A tuple ( ๐’Ÿ , ๐‘‹ , ๐ป ) is said to be realizably learnable if there exists an algorithm8 ๐’œ and a function ๐‘› โข ( ๐œ€ , ๐›ฟ ) such that for every ๐œ€ , ๐›ฟ

0 , every choice of distribution ๐ท โˆˆ ๐’Ÿ , and every hypothesis โ„Ž โˆˆ ๐ป , ๐’œ outputs a good classifier with high probability on samples of size ๐‘› โข ( ๐œ€ , ๐›ฟ ) :

Pr ๐‘† โˆผ ๐ท ๐‘› โข ( ๐œ€ , ๐›ฟ ) โก [ err ๐ท ร— โ„Ž โข ( ๐’œ โข ( ๐‘† , โ„Ž โข ( ๐‘† ) ) ) โ‰ค ๐œ€ ] โ‰ฅ 1 โˆ’ ๐›ฟ ,

where err ๐ท ร— โ„Ž โข ( โ‹… ) is commonly called the error or risk of โ„Ž :

๐‘’ โข ๐‘Ÿ โข ๐‘Ÿ ๐ท ร— โ„Ž โข ( โ„Ž โ€ฒ )

Pr ๐‘ฅ โˆผ ๐ท โข [ โ„Ž โ€ฒ โข ( ๐‘ฅ ) โ‰  โ„Ž โข ( ๐‘ฅ ) ] .

Likewise, a tuple ( ๐’Ÿ , ๐‘‹ , ๐ป ) is said to be agnostically learnable if there exists an algorithm ๐’œ which for every distribution ๐ท over ๐‘‹ ร— ๐‘Œ whose marginal ๐ท ๐‘‹ โˆˆ ๐’Ÿ outputs โ„Ž โ€ฒ close to the best hypothesis in ๐ป with probability 1 โˆ’ ๐›ฟ :

Pr ๐‘† โˆผ ๐ท ๐‘› โข ( ๐œ€ , ๐›ฟ ) โก [ err ๐ท โข ( ๐’œ โข ( ๐‘† ) ) โ‰ค ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ ] โ‰ฅ 1 โˆ’ ๐›ฟ ,

where ๐‘‚ โข ๐‘ƒ โข ๐‘‡

inf โ„Ž โˆˆ ๐ป { err ๐ท โข ( โ„Ž ) } is the error of the best hypothesis in the class9 and the risk err ๐ท โข ( โ‹… ) is similarly defined:

err ๐ท โข ( โ„Ž โ€ฒ )

Pr ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท โข [ โ„Ž โ€ฒ โข ( ๐‘ฅ ) โ‰  ๐‘ฆ ] .

With this in mind, we can now state the most basic application of Algorithm 1: the equivalence of agnostic and realizable learning for distribution-family classification.

Theorem 2.3 (Realizable โ†’ Agnostic (Distribution-Family Classification)).

Let ๐’œ be a realizable learner for ( ๐’Ÿ , ๐‘‹ , ๐ป ) using ๐‘› โข ( ๐œ€ , ๐›ฟ ) samples. Then Algorithm 1 is an agnostic learner for ( ๐’Ÿ , ๐‘‹ , ๐ป ) using:

๐‘š ๐‘ˆ โข ( ๐œ€ , ๐›ฟ ) โ‰ค ๐‘› โข ( ๐œ€ / 2 , ๐›ฟ / 2 )

unlabeled samples, and

๐‘š ๐ฟ โข ( ๐œ€ , ๐›ฟ ) โ‰ค ๐‘‚ โข ( ๐‘› โข ( ๐œ€ / 2 , ๐›ฟ / 2 ) + log โก ( 1 / ๐›ฟ ) ๐œ€ 2 )

labeled samples. Moreover if ( ๐‘‹ , ๐ป ) has finite VC dimension ๐‘‘ , Algorithm 1 needs only

๐‘š ๐ฟ โข ( ๐œ€ , ๐›ฟ ) โ‰ค ๐‘‚ โข ( ๐‘‘ โข log โก ( 1 / ๐œ€ ) + log โก ( 1 / ๐›ฟ ) ๐œ€ 2 )

labeled samples.

Along with its novelty in the distribution-family setting (where no such equivalence was known), it is worth noting that in the distribution-free setting, Theorem 2.3 actually recovers the same sample complexity bound as standard analysis of uniform convergence (though it remains off by a log factor from the optimal bound given by chaining [45]). We also note that while unlabeled sample complexity is not usually considered separately from labeled complexity in the PAC setting, this will become a useful distinction in semi-supervised extensions considered later in the work. As such, it is instructive to keep the complexities separate for the time being.

With this out of the way, letโ€™s prove Theorem 2.3. The analysis breaks naturally into two parts, corresponding respectively to Step 2 and Step 3 of Algorithm 1. In the first part, weโ€™ll show that ๐ถ โข ( ๐‘† ๐‘ˆ ) , the set of outputs corresponding to running the realizable learner ๐’œ across all possible labelings of the unlabeled sample ๐‘† ๐‘ˆ , is in some sense a โ€œgood approximationโ€ of the class ๐ป . More formally, the crucial observation is that for any choice of the adversaryโ€™s distribution, ๐ถ โข ( ๐‘† ๐‘ˆ ) will (almost) always contain a hypothesis close to the optimal solution.

Claim 1.

For any distribution ๐ท over ๐‘‹ ร— ๐‘Œ whose marginal ๐ท ๐‘‹ โˆˆ ๐’Ÿ , with probability 1 โˆ’ ๐›ฟ / 2 , there exists โ„Ž โ€ฒ โˆˆ ๐ถ โข ( ๐‘† ๐‘ˆ ) which is within ๐œ€ / 2 of the optimal risk:

๐‘’ โข ๐‘Ÿ โข ๐‘Ÿ ๐ท โข ( โ„Ž โ€ฒ ) โ‰ค ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ / 2 .

Once we have this claim, the second step is to show that Step 3, an empirical risk minimization process on ๐ถ โข ( ๐‘† ๐‘ˆ ) , gives the desired agnostic learner. This actually follows from standard arguments. In particular, given a hypothesis โ„Ž โˆˆ ๐ถ โข ( ๐‘† ๐‘ˆ ) , let

๐‘’ โข ๐‘Ÿ โข ๐‘Ÿ ๐‘† ๐ฟ โข ( โ„Ž )

Pr ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐‘† ๐ฟ โก [ โ„Ž โข ( ๐‘ฅ ) โ‰  ๐‘ฆ ]

denote its empirical risk with respect to ๐‘† ๐ฟ . Since ๐ถ โข ( ๐‘† ๐‘ˆ ) is finite, a standard Chernoff+Union bound gives that with probability at least 1 โˆ’ ๐›ฟ / 2 , the empirical risk of every hypothesis in ๐ถ โข ( ๐‘† ๐‘ˆ ) with respect to ๐‘† ๐ฟ is close to its true risk. Then as long as ๐‘† ๐ฟ is sufficiently large, empirical risk minimization returns a solution with at most ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ error with high probability (weโ€™ll formalize this in a moment).

It remains to prove Claim 1. The key observation lies in an equivalence between realizable PAC-learning and a weak type of randomized covering: for any fixed โ„Ž โˆˆ ๐ป , ๐ถ โข ( ๐‘† ๐‘ˆ ) contains a hypothesis close to โ„Ž with high probability.

Lemma 2.4.

For any distribution ๐ท over ๐‘‹ ร— ๐‘Œ with marginal ๐ท ๐‘‹ โˆˆ ๐’Ÿ and any โ„Ž โˆˆ ๐ป , with probability 1 โˆ’ ๐›ฟ / 2 , there exists โ„Ž โ€ฒ โˆˆ ๐ถ โข ( ๐‘† ๐‘ˆ ) which is within ๐œ€ / 2 of โ„Ž in classification distance:

Pr ๐‘ฅ โˆผ ๐ท ๐‘‹ โก [ โ„Ž โ€ฒ โข ( ๐‘ฅ ) โ‰  โ„Ž โข ( ๐‘ฅ ) ] โ‰ค ๐œ€ / 2 .

Proof 2.5.

The proof is essentially immediate from the definition of realizable PAC-learning. ๐’œ promises that for any โ„Ž โˆˆ ๐ป and ๐ท โˆˆ ๐’Ÿ , a 1 โˆ’ ๐›ฟ / 2 fraction of labeled samples ( ๐‘† , โ„Ž โข ( ๐‘† ) ) โˆผ ๐ท ๐‘› โข ( ๐œ€ / 2 , ๐›ฟ / 2 ) satisfy

๐‘’๐‘Ÿ๐‘Ÿ ๐ท ร— โ„Ž โข [ ๐’œ โข ( ๐‘† , โ„Ž โข ( ๐‘† ) ) ]

Pr ๐ท โข [ โ„Ž โ€ฒ โข ( ๐‘ฅ ) โ‰  โ„Ž โข ( ๐‘ฅ ) ] โ‰ค ๐œ€ / 2 ,

where โ„Ž โ€ฒ

๐’œ โข ( ๐‘† , โ„Ž โข ( ๐‘† ) ) . Since ๐ถ โข ( ๐‘† ๐‘ˆ ) contains ๐’œ โข ( ๐‘† ๐‘ˆ , โ„Ž โข ( ๐‘† ๐‘ˆ ) ) for every โ„Ž โˆˆ ๐ป by definition, the result follows.

More generally, we call such objects non-uniform covers.

Definition 2.6 (Non-uniform Cover (Informal Definition 5.1)).

Let ( ๐‘‹ , ๐ป ) be a class over label space ๐‘Œ , ๐ท a marginal distribution over ๐‘‹ , and ๐ถ a random variable over the power set ๐‘ƒ โข ( ๐ป ) . We call ๐ถ a non-uniform ( ๐œ€ , ๐›ฟ ) -cover of ๐ป with respect to ๐ท if for every โ„Ž โˆˆ ๐ป :

Pr ๐ถ โก [ โˆƒ โ„Ž โ€ฒ โˆˆ ๐ถ : Pr ๐‘ฅ โˆผ ๐ท โก [ โ„Ž โ€ฒ โข ( ๐‘ฅ ) โ‰  โ„Ž โข ( ๐‘ฅ ) ] โ‰ค ๐œ€ ] โ‰ฅ 1 โˆ’ ๐›ฟ .

Note that Lemma 2.4 (and in general non-uniform covering) does not imply that ๐ถ โข ( ๐‘† ๐‘ˆ ) contains hypotheses close to every โ„Ž โˆˆ ๐ป simultaneously. This stronger object is called a uniform cover and takes provably more samples to construct (see Section 13). In our case, a non-uniform cover is sufficient. Since the guarantee holds for every fixed โ„Ž โˆˆ ๐ป , it must hold in particular for the optimal hypothesis โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ , so ๐ถ โข ( ๐‘† ๐‘ˆ ) contains some โ„Ž โ€ฒ within ๐œ€ / 2 of optimal. Letโ€™s now formalize these ideas and put everything together to prove Theorem 2.3.

Proof 2.7 (Proof of Theorem 2.3).

Let ๐ท be the adversaryโ€™s distribution over ๐‘‹ ร— ๐‘Œ , and let โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โˆˆ ๐ป be a hypothesis achieving the optimal error. By Lemma 2.4, with probability 1 โˆ’ ๐›ฟ / 2 , ๐ถ โข ( ๐‘† ๐‘ˆ ) contains a hypothesis โ„Ž โ€ฒ such that:

Pr ๐‘ฅ โˆผ ๐ท ๐‘‹ โก [ โ„Ž โ€ฒ โข ( ๐‘ฅ ) โ‰  โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข ( ๐‘ฅ ) ] โ‰ค ๐œ€ / 2 .

This implies Claim 1 (that ๐ถ โข ( ๐‘† ๐‘ˆ ) contains a hypothesis with error at most ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ / 2 ) since

๐‘’ โข ๐‘Ÿ โข ๐‘Ÿ ๐ท โข ( โ„Ž โ€ฒ )

Pr ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท โข [ โ„Ž โ€ฒ โข ( ๐‘ฅ ) โ‰  ๐‘ฆ ]

โ‰ค Pr ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท โข [ โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข ( ๐‘ฅ ) โ‰  ๐‘ฆ ] + Pr ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท โข [ โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข ( ๐‘ฅ ) โ‰  โ„Ž โ€ฒ โข ( ๐‘ฅ ) ]

โ‰ค ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ / 2 .

We can now use standard empirical risk minimization bounds on ๐ถ โข ( ๐‘† ๐‘ˆ ) to find a hypothesis with error at most ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ . Chernoff and union bounds imply that with probability at least 1 โˆ’ ๐›ฟ / 2 , the empirical risk of every hypothesis in ๐ถ โข ( ๐‘† ๐‘ˆ ) on a sample of size ๐‘‚ โข ( log โก ( | ๐ถ โข ( ๐‘† ๐‘ˆ ) | / ๐›ฟ ) ๐œ€ 2 ) is at most ๐œ€ / 4 away from its true error. Since โ„Ž โ€ฒ has error at most ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ / 2 , its empirical risk is at most ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + 3 โข ๐œ€ / 4 , and by the above guarantee any hypothesis in ๐ถ โข ( ๐‘† ๐‘ˆ ) with empirical risk at most ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + 3 โข ๐œ€ / 4 has true error at most ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ .

Putting everything together, we have that with probability 1 โˆ’ ๐›ฟ over the entire process, the empirical risk minimizer of ๐ถ โข ( ๐‘† ๐‘ˆ ) has error at most ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ as desired. The sample complexity bounds follow from noting that | ๐ถ โข ( ๐‘† ๐‘ˆ ) | is at most 2 ๐‘› โข ( ๐œ€ / 2 , ๐›ฟ / 2 ) , and at most ( ๐‘’ โ‹… ๐‘› โข ( ๐œ€ / 2 , ๐›ฟ / 2 ) ๐‘‘ ) ๐‘‘ if the class has VC dimension ๐‘‘ . The sample complexity bound for the latter case then follows by plugging in the standard bound for distribution-free classification: ๐‘› โข ( ๐œ€ / 2 , ๐›ฟ / 2 ) โ‰ค ๐‘‚ โข ( ๐‘‘ โข log โก ( 1 / ๐œ€ ) + log โก ( 1 / ๐›ฟ ) ๐œ€ ) .

3Related Work

Agnostic learning is a very widely studied model across learning theory, and works across many different sub-areas have noted model specific equivalences with realizable learning. Here weโ€™ll survey a few representative examples, and discuss how they relate and differ from our approach.

3.1Beyond Binary Classification Uniform Convergence and Multiclass Classification.

It is well known that the uniform convergence equals learnability paradigm continues to hold for 0 / 1 -valued loss functions over constant-size label spaces [19, 15, 59, 54], and that agnostic and realizable learning are equivalent as a result. On the other hand, Daniely, Sabato, Ben-Devid, and Shalev-Shwartz [25] showed this is no longer the case as the number of labels grows large. In this regime, even basic multi-class learning is no longer equivalent to uniform convergence, so the connection between realizable and agnostic learning becomes non-trivial. A few years later, David, Moran, and Yehuyadoff (DMY) [29] proved the equivalence nevertheless holds in the infinite multi-class setting through the weaker sample compression equals learnability paradigm. While more general than the uniform convergence paradigm, their proof remains model-specific and fails in many of the settings we consider, e.g. partial learning [5].

Discretization and General Loss Functions.

Basic forms of discretization were also considered back in the mid 90s in work on characterizing the learnability of real-valued functions. In a seminal work, Bartlett, Long, and Williamson (BLW) [9] proved that a scale-sensitive measure introduced by Kearns and Schapire [42] called fat-shattering dimension characterizes learnability under bounded Lipschitz loss functions.10 BLW use a basic form of discrete learning (called quantization) to prove that fat-shattering dimension is a necessary condition, and use uniform convergence to prove sufficiency. We give a similar argument as BLW in the necessary direction, but show that uniform convergence is not necessary for the equivalence to hold, and instead use Algorithm 1 to appeal directly to discrete learnability. This allows us to extend BLWโ€™s result across a much more general set of loss functions and scenarios without strong model-specific assumptions.

3.2Semi-supervised, Active, and Semi-Private Learning

Our reduction hinges on combining a realizable learner with unlabeled data to cut down the number of potential hypotheses in our class. The use of unlabeled samples to this effect is one of the core ideas in the field of semi-supervised learning [8, 65]. Here, it is usually additionally assumed that the function to be learnt has some relation (or โ€˜compatibilityโ€™) to the underlying data distribution, for example it might have large margin on unlabelled data as in Transductive SVM [38] or redundant sufficient information as in Co-training [18, 28]. In their seminal work on the topic, Balcan and Blum [8] employed a similar strategy to Algorithm 1 in which they draw an unlabeled sample ๐‘† ๐‘ˆ and select hypotheses consistent with each possible labeling based upon compatibility. They argue via uniform convergence that this results in a uniform cover, and then use empirical risk minimization to select a good hypothesis in the cover. It is worth noting that around the same time a similar strategy independently found use in the online learning literature in work of Ben-David, Pal, and Shalev-Schwarz [16] who simulated the so-called โ€˜standard optimal algorithmโ€™ (SOA) over a sequence of examples and applied weighted majority [46] over the resulting set of hypotheses to obtain an agnostic online learner.

Similar strategies have also found use in the related active learning literature. Hanneke and Yang [35] use the same technique to build a cover from unlabeled samples (adding one hypothesis consistent with each possible labeling), and then apply active (adaptive) query algorithms to learn the best hypothesis in the cover in as few labeled samples as possible. This generalized earlier work of Dasgupta [27], who assumed a priori that the cover was known to the learner ahead of time. Most recently, the approach has seen use in the study of semi-private learning. In their original work on the model, Beimel, Nissim, and Stemmer [14] again apply the same trick for building a uniform cover, but then find the best hypothesis privately via the exponential mechanism (similar to our proof of Theorem 1.10). The analysis of this strategy was later improved by Alon, Bassily, and Moran (ABM) [2].

The above works differ from ours in two crucial senses. First, each work focuses solely on developing an algorithm for their specific framework (rather than working to understand a more general equivalence or reduction between settings). In this sense, one can view each of these prior results as a specific instance of our general framework where the โ€œbase learnerโ€ in our reduction is restricted to be an empirical risk minimizer (or SOA in the online setting), and a problem-specific learner for the relevant property (online, agnostic, active, or private) is then applied over the resulting cover. Second, and perhaps most importantly, these previous works all rely fundamentally on uniform convergence.11 This means that their algorithms break down as soon as one moves away from the original PAC model (even to say the basic distribution-dependent setting), and can also lead to sub-optimal sample complexity bounds. In the analysis of semi-private learning, for instance, we show that avoiding uniform convergence leads to asymptotically better bounds, actually resolving the public sample complexity of the model altogether. Indeed, one can show that building a uniform cover requires asymptotically more unlabeled samples than a non-uniform one, and therefore cannot result in optimal semi-supervised algorithms.12

3.3Non-Uniform Covering and Probabilistic Representations

Covering techniques have long been used in learning theory, and while almost all prior works focus on uniform notions (where all hypotheses are covered simultaneously), there is one notable exception. In 2013, Beimel, Nissim, and Stemmer (BNS) [12] introduced probabilistic representations, a strong randomized form of covering used to characterize pure differentially private learning. In the language of our work, given a class ( ๐‘‹ , ๐ป ) , a probabilistic representation is a distribution over subsets of ๐ป which is a non-uniform cover simultaneously over all distributions of ๐‘‹ . BNS prove that private learning is equivalent to the existence of a probabilistic representations for the class. Equivalently, this can be thought of as the ability to build a non-uniform cover without access to the underlying distribution at all. On the other hand, we are interested in the much weaker setting where a non-uniform cover can be built from a bounded number of samples from the distribution (and crucially argue that this is equivalent to realizable learning). Thus in a sense, our core connection between realizable learning and non-uniform covering can be thought of as an analog of BNSโ€™ characterization of private learning by probabilistic representations.

Paper Organization

The main body of this paper is split into two main portions. In the first we cover preliminaries (Section 4), the base reduction for all finitely supported loss functions (Section 5), and discuss in detail the four main modification archetypes to Algorithm 1: discretization (Section 6.1), sub-sampling (Section 6.2), replacing ERM (Section 6.3), and changing the base model (Section 6.4), covering extensions to infinite label classes, malicious noise, semi-private learning, and covariate shift respectively. In the remainder of the paper we cover applications of these modified versions to doubly-bounded loss (Section 7), robust learning (Section 8), partial learning (Section 9), uniformly-stable learning (Section 10), the statistical query model (Section 11), and fair learning (Section 12), and discuss further connections of non-uniform covers to previous notions of covering (Section 13). Each of the latter sections are self-contained and the reader is encouraged to skip directly to any model they wish to see directly.

4Preliminaries

Before moving to a more formal discussion of our results, weโ€™ll cover the most basic learning models discussed in this work: standard (distribution-free) PAC-learning and distribution-family PAC-learning. Extended models we consider beyond these (e.g. malicious noise, robust learning, partial learning, etc.) will instead be introduced in their respective sections.

4.1PAC-Learning

Weโ€™ll start by reviewing the seminal PAC-learning model of Valiant [58] and Vapnik and Chervonenkis [60]. We start with a few core definitions for the setting of general loss. Let ๐‘‹ be an arbitrary set called the instance space (e.g. โ„ ๐‘‘ ), ๐‘Œ a set called the label space (e.g. { 0 , 1 } ), and ๐ป a family of labelings of ๐‘‹ by ๐‘Œ (that is a family of functions of the form โ„Ž : ๐‘‹ โ†’ ๐‘Œ ). Given a class ( ๐‘‹ , ๐ป ) , it will often be useful to consider its growth function ฮ  ๐ป โข ( ๐‘› ) which measures the maximum size of ๐ป when restricted to a sample of size ๐‘› :

ฮ  ๐ป ( ๐‘› )

max โ„Ž โˆˆ ๐ป , ๐‘† โˆˆ ๐‘‹ ๐‘› ( | ๐ป | ๐‘† | ) ) .

We note that the growth function is trivially bounded by | ๐‘Œ | ๐‘› , but one can often give stronger bounds when ( ๐‘‹ , ๐ป ) satisfies some finite combinatorial dimension (e.g. VC-dimension for the binary case).

While PAC-learning is sometimes used to refer only to classification, we will study the model under general loss functions. With that in mind, we call a function โ„“ : ๐‘Œ ร— ๐‘Œ โ†’ โ„ โ‰ฅ 0 a loss function if โ„“ โข ( ๐‘ฆ , ๐‘ฆ )

0 for all ๐‘ฆ โˆˆ ๐‘Œ . We say a loss โ„“ satisfies the identity of indiscernibles if โ„“ โข ( ๐‘ฆ 1 , ๐‘ฆ 2 )

0 iff ๐‘ฆ 1

๐‘ฆ 2 . Given any distribution ๐ท over ๐‘‹ ร— ๐‘Œ and loss โ„“ , the risk of a labeling โ„Ž : ๐‘‹ โ†’ ๐‘Œ with respect to ๐ท and โ„“ is its expected loss:

err ๐ท , โ„“ โข ( โ„Ž )

๐”ผ ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท โข [ โ„“ โข ( โ„Ž โข ( ๐‘ฅ ) , ๐‘ฆ ) ] .

The goal of learning is generally to find a classifier โ„Ž โˆˆ ๐ป that minimizes risk. More formally, there are two commonly studied variants of this problem. The original formulation, now called realizable learning, assumes the existence of a hypothesis in ๐ป with no loss.

Definition 4.1 ((Realizable) PAC-learning).

We say ( ๐‘‹ , ๐ป , โ„“ ) is realizable PAC-learnable if there exists an algorithm ๐’œ and function ๐‘› โข ( ๐œ€ , ๐›ฟ ) such that for all ๐œ€ , ๐›ฟ > 0 and distributions ๐ท over ๐‘‹ ร— ๐‘Œ such that min โ„Ž โˆˆ ๐ป โก ๐‘’๐‘Ÿ๐‘Ÿ ๐ท , โ„“ โข ( โ„Ž )

0 :

Pr ๐‘† โˆผ ๐ท ๐‘› โข ( ๐œ€ , ๐›ฟ ) โก [ ๐‘’๐‘Ÿ๐‘Ÿ ๐ท , โ„“ โข ( ๐’œ โข ( ๐‘† ) )

๐œ€ ] โ‰ค ๐›ฟ .

๐’œ is called proper if it outputs only labels in ๐ป .

Perhaps a more realistic variant of PAC-learning is to drop this restriction on the adversary, and let them choose an arbitrary distribution over ๐‘‹ ร— ๐‘Œ . This model, introduced by Haussler [36] and Kearns, Schapire, and Sellie [41], is known as agnostic learning.

Definition 4.2 ((Agnostic) PAC-learning).

We say ( ๐‘‹ , ๐ป , โ„“ ) is agnostic PAC-learnable if there exists an algorithm ๐’œ and function ๐‘› โข ( ๐œ€ , ๐›ฟ ) such that for all ๐œ€ , ๐›ฟ

0 and distributions ๐ท over ๐‘‹ ร— ๐‘Œ :

Pr ๐‘† โˆผ ๐ท ๐‘› โข ( ๐œ€ , ๐›ฟ ) โก [ ๐‘’๐‘Ÿ๐‘Ÿ ๐ท , โ„“ โข ( ๐’œ โข ( ๐‘† ) )

๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ ] โ‰ค ๐›ฟ ,

where ๐‘‚ โข ๐‘ƒ โข ๐‘‡

min โ„Ž โˆˆ ๐ป โก { ๐‘’๐‘Ÿ๐‘Ÿ ๐ท , โ„“ โข ( โ„Ž ) } .

For some settings covered in this work, it will turn out that reaching ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ error is too stringent of a condition. However, we will show in these cases that it is sometimes possible to maintain a weaker guarantee and learn up to ๐‘ โ‹… ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ error for some constant ๐‘

1 . We call such classes ๐‘ -agnostic learnable.

Finally, we note that for simplicity when โ„“ is the standard โ€œclassification error:โ€

โ„“ โข ( ๐‘ฆ 1 , ๐‘ฆ 2 )

{ 0
if  โข ๐‘ฆ 1

๐‘ฆ 2

1

else ,

weโ€™ll simply write ( ๐‘‹ , ๐ป ) to mean ( ๐‘‹ , ๐ป , โ„“ ) . Realizable and Agnostic Learning are well studied under many basic loss functions including binary classification, where both models are known to be characterized by a combinatorial parameter called VC-dimension.

4.2Learning Under Distribution Families

The standard PAC-models described above are often called distribution-free due to the fact that no assumptions are made on the marginal distribution over ๐‘‹ . In practice, however, this is usually too worst-case an assumption. We often expect distributions in nature to be โ€œniceโ€ in some way, or at least somewhat restricted. This is reflected in the fact that popular machine learning algorithms usually significantly outperform the PAC-modelโ€™s worst-case generalization bounds. Indeed such niceness assumptions have long been popular in learning theory as well, where conditions such as tail bounds or anti-concentration are frequently used to build efficient algorithms.

These ideas are captured more generally by a simple (but notoriously difficult) extension to the PAC framework originally proposed by Benedek and Itai [17], where the adversary is restricted to picking from a fixed, known set of distributions.

Definition 4.3 ((Realizable) Distribution-Family PAC-learning).

Let ๐‘‹ be an instance space and ๐’Ÿ a family of distributions over ๐‘‹ . We say ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) is realizable PAC-learnable if there exists an algorithm ๐’œ and function ๐‘› โข ( ๐œ€ , ๐›ฟ ) such that for all ๐œ€ , ๐›ฟ

0 and distributions ๐ท over ๐‘‹ ร— ๐‘Œ satisfying:

1.

The marginal ๐ท ๐‘‹ โˆˆ ๐’Ÿ ,

2.

min โ„Ž โˆˆ ๐ป โก ๐‘’๐‘Ÿ๐‘Ÿ ๐ท , โ„“ โข ( โ„Ž )

0 ,

we have

Pr ๐‘† โˆผ ๐ท ๐‘› โข ( ๐œ€ , ๐›ฟ ) โก [ ๐‘’๐‘Ÿ๐‘Ÿ ๐ท , โ„“ โข ( ๐’œ โข ( ๐‘† ) )

๐œ€ ] โ‰ค ๐›ฟ .

Agnostic learning is defined similarly. The adversary must still choose a marginal distribution in ๐’Ÿ , but the conditional labeling can be arbitrary.

Definition 4.4 (Agnostic Distribution-Family PAC-learning).

We say ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) is agnostic PAC-learnable if there exists an algorithm ๐’œ and function ๐‘› โข ( ๐œ€ , ๐›ฟ ) such that for all ๐œ€ , ๐›ฟ

0 and distributions ๐ท over ๐‘‹ ร— ๐‘Œ satisfying:

1.

The marginal ๐ท ๐‘‹ โˆˆ ๐’Ÿ ,

๐’œ outputs a good hypothesis with high probability:

Pr ๐‘† โˆผ ๐ท ๐‘› โข ( ๐œ€ , ๐›ฟ ) โก [ ๐‘’๐‘Ÿ๐‘Ÿ ๐ท , โ„“ โข ( ๐’œ โข ( ๐‘† ) )

๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ ] โ‰ค ๐›ฟ

where ๐‘‚ โข ๐‘ƒ โข ๐‘‡

min โ„Ž โˆˆ ๐ป โก { ๐‘’๐‘Ÿ๐‘Ÿ ๐ท , โ„“ โข ( โ„Ž ) } .

The weaker ๐‘ -agnostic learning is defined analogously with ๐‘‚ โข ๐‘ƒ โข ๐‘‡ replaced by ๐‘ โ‹… ๐‘‚ โข ๐‘ƒ โข ๐‘‡ . Unlike the standard model, very little is known about distribution-family learnability. A number of works have given partial characterizations of the model [17, 31, 43, 61], but it was recently shown by Lechner and Ben-David [44] that no general finitary characterization is possible.

5The Core Reduction: Finite Label Classes

In this section, we give a more detailed exposition of our main reduction as covered in Section 2 for the general setting of arbitrary loss on constant size label spaces, matching lower bounds, and additional discussion of non-uniform covers. As mentioned previously, since there is no combinatorial characterization of learnability in the distribution-family model [44], standard techniques [19, 15, 59, 54] cannot be used.

Before jumping into our reduction proper, it is worth re-iterating why we canโ€™t simply take the approach of prior works and rely on uniform convergence, a strong condition which promises that on a large enough sample, the empirical error of every hypothesis will be close to its true error. While uniform convergence was a very popular technique in the early years of learning, practitioners have since moved away from the paradigm which fails to capture learning rates seen in practice [64, 53]. Indeed it soon became clear that the technique failed to capture even basic theoretical models such as the distribution-dependent setting (as discussed in Proposition 2.1). In later sections, we will even see distribution-free models where uniform convergence fails, such as the Partial PAC model [47, 5] which captures realistic scenarios such as learning with margin. Since even the most basic modifications of PAC-learning fail to satisfy uniform convergence, it is clear we need to move beyond the condition to gain a more general understanding of the common phenomenon of equivalence between learning models.

Instead of relying on uniform convergence, our core observation is an equivalence between learning and sample access to a combinatorial object we call a non-uniform cover.

Definition 5.1 (Non-uniform Cover).

Let ( ๐‘‹ , ๐ป ) be a class over label space ๐‘Œ and ๐ฟ ๐‘‹ , ๐‘Œ denote the family of all labelings from ๐‘‹ to ๐‘Œ . If ๐ถ is a random variable over the power set ๐‘ƒ โข ( ๐ฟ ๐‘‹ , ๐‘Œ ) and ๐‘‘ : ๐ฟ ๐‘‹ , ๐‘Œ ร— ๐ฟ ๐‘‹ , ๐‘Œ โ†’ โ„ โ‰ฅ 0 is a โ€œdistanceโ€ function between labelings, we call ๐ถ a non-uniform ( ๐œ€ , ๐›ฟ ) -cover of ๐ป with respect to ๐‘‘ if for all โ„Ž โˆˆ ๐ป :

Pr ๐‘‡ โˆผ ๐ถ โก [ โˆƒ โ„Ž โ€ฒ โˆˆ ๐‘‡ : ๐‘‘ โข ( โ„Ž โ€ฒ , โ„Ž ) โ‰ค ๐œ€ ] โ‰ฅ 1 โˆ’ ๐›ฟ .

We call ๐ถ bounded if its support lies entirely on subsets of size at most some ๐‘˜ โˆˆ โ„• , and we call the smallest such ๐‘˜ its size.

Non-uniform covers share a close connection to several notions of covering used throughout the learning literature such as uniform covers [2] and fractional covers [4]. We discuss these connections in more detail in Section 13. For the moment, we note only that previous works using the strictly stronger notion of uniform covering necessarily lose factors in the sample complexity as a result. We discuss this further in Section 6.3.

In Section 2, we argued (at least implicitly) that once we have sampling access to a bounded non-uniform cover, agnostic learnability follows from standard arguments. Namely since a sample ๐‘‡ has bounded size and is guaranteed to contain a concept โ€œcloseโ€ to optimal, it suffices to run empirical risk minimization over about log โก ( | ๐‘‡ | / ๐›ฟ ) / ๐œ€ 2 samples. The key to our reduction therefore boils down to turning blackbox access to a realizable PAC-learner into sampling access to some relevant non-uniform cover. This is given by Step 2 of Algorithm 1, which we rewrite here as a subroutine called LearningToCover.

Input: Hypothesis Class ๐ป , Realizable PAC-Learner ๐’œ , Unlabeled Sample size ๐‘† ๐‘ˆ . Algorithm:

1.

Run ๐’œ over all possible labelings of ๐‘† ๐‘ˆ by ๐ป .

2.

Return the set of responses ๐ถ โข ( ๐‘† ๐‘ˆ ) :

๐ถ โข ( ๐‘† ๐‘ˆ ) โข \coloneqq โข { ๐’œ โข ( ๐‘† ๐‘ˆ , โ„Ž โข ( ๐‘† ๐‘ˆ ) ) | โ„Ž โˆˆ ๐ป | ๐‘† ๐‘ˆ } .

Algorithm 2 LearningToCover( ๐ป , ๐’œ , ๐‘† ๐‘ˆ )

In fact, we already argued in Section 2 that LearningToCover gives sampling access to a non-uniform cover, but we will re-write the result here in this formulation for convenience.

Lemma 5.2 (Core Lemma: Realizable Learning Implies Non-Uniform Covering).

Let ๐’œ be an algorithm that ( ๐œ€ , ๐›ฟ ) -PAC learns a class ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) in ๐‘›

๐‘› โข ( ๐œ€ , ๐›ฟ ) samples. Then for any ๐ท โˆˆ ๐’Ÿ , running LearningToCover on ๐‘† ๐‘ˆ โˆผ ๐ท ๐‘› returns a sample from a size ฮ  ๐ป โข ( ๐‘› ) , non-uniform ( ๐œ€ , ๐›ฟ ) -cover with respect to the standard distance between hypotheses:

๐‘‘ โข ( โ„Ž , โ„Ž โ€ฒ )

๐”ผ ๐ท โข [ โ„“ โข ( โ„Ž โ€ฒ โข ( ๐‘ฅ ) , โ„Ž โข ( ๐‘ฅ ) ) ] .

Proof 5.3.

The proof is essentially immediate from the definition of realizable PAC-learning. ๐’œ promises that for any โ„Ž โˆˆ ๐ป and ๐ท โˆˆ ๐’Ÿ , a 1 โˆ’ ๐›ฟ fraction of labeled samples ( ๐‘† ๐‘ˆ , โ„Ž โข ( ๐‘† ๐‘ˆ ) ) โˆผ ๐ท ๐‘› โข ( ๐œ€ , ๐›ฟ ) satisfy

๐‘’๐‘Ÿ๐‘Ÿ ๐ท ร— โ„Ž , โ„“ โข ( ๐’œ โข ( ๐‘† ๐‘ˆ , โ„Ž โข ( ๐‘† ๐‘ˆ ) ) )

๐”ผ ๐ท โข [ โ„“ โข ( โ„Ž โ€ฒ โข ( ๐‘ฅ ) , โ„Ž โข ( ๐‘ฅ ) ) ] โ‰ค ๐œ€ ,

where โ„Ž โ€ฒ

๐’œ โข ( ๐‘† ๐‘ˆ , โ„Ž โข ( ๐‘† ๐‘ˆ ) ) . Since ๐ถ โข ( ๐‘† ๐‘ˆ ) contains ๐’œ โข ( ๐‘† ๐‘ˆ , โ„Ž โข ( ๐‘† ๐‘ˆ ) ) for every โ„Ž by definition, the result follows.

This means that as long as we have blackbox access to a realizable PAC-learner and unlabeled samples from the adversaryโ€™s distribution, we can simulate access to a non-uniform cover. Letโ€™s now formalize our previous intuition that this is sufficient to turn a realizable learner into an agnostic one for any finite label class. We will generalize this result to doubly-bounded loss in Section 7, but it is instructive to consider the setting of finite ๐‘Œ first.

Theorem 5.4 (Realizable โ†’ Agnostic (Finite Label Classes)).

Let ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) be any class on a finite label space ๐‘Œ with loss function โ„“ : ๐‘Œ โ†’ ๐‘Œ satisfying the identity of indiscernibles. Then Algorithm 1 is an agnostic learner with sample complexity:

๐‘š โข ( ๐œ€ , ๐›ฟ ) โ‰ค ๐‘› โข ( ๐œ‚ โ„“ โข ๐œ€ , ๐›ฟ / 2 ) + ๐‘‚ โข ( log โก ( ฮ  ๐ป โข ( ๐‘› โข ( ๐œ‚ โ„“ โข ๐œ€ , ๐›ฟ / 2 ) ) ๐›ฟ ) ๐œ€ 2 )

where ๐œ‚ โ„“ โ‰ฅ ฮฉ โข ( min ๐‘Ž โ‰  ๐‘ โก ( โ„“ โข ( ๐‘Ž , ๐‘ ) ) max ๐‘Ž โ‰  ๐‘ โก ( โ„“ โข ( ๐‘Ž , ๐‘ ) ) ) is a constant depending only on โ„“ .

Proof 5.5.

Let ๐’œ be the promised realizable learner for ( ๐‘‹ , ๐ป , ๐’Ÿ , โ„“ ) with sample complexity ๐‘› โข ( ๐œ€ , ๐›ฟ ) . Run LearningToCover with parameters ๐œ€ โ€ฒ

๐œ‚ โ„“ โข ๐œ€ and ๐›ฟ โ€ฒ

๐›ฟ / 2 . We argue that the output contains some โ„Ž โ€ฒ such that ๐‘’๐‘Ÿ๐‘Ÿ ๐ท , โ„“ โข ( โ„Ž โ€ฒ ) โ‰ค ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ / 2 . Since ๐ถ โข ( ๐‘† ๐‘ˆ ) is finite and โ„“ is upper bounded, a standard Chernoff bound gives that choosing an empirical risk minimizer from ๐ถ โข ( ๐‘† ๐‘ˆ ) based on ๐‘‚ โข ( log โก ( | ๐ถ โข ( ๐‘† ๐‘ˆ ) | ๐›ฟ ) ๐œ€ 2 ) additional samples gives the desired learner. The sample complexity then follows immediately from the fact that ๐ถ โข ( ๐‘† ๐‘ˆ ) contains one hypothesis for every labeling of the sample, and is therefore bounded by the growth function ฮ  ๐ป โข ( ๐‘› ) .

To see why ๐ถ โข ( ๐‘† ๐‘ˆ ) has this property, recall that for any โ„Ž โˆˆ ๐ป , Lemma 5.2 states that ๐ถ โข ( ๐‘† ๐‘ˆ ) contains some โ„Ž โ€ฒ such that:

๐”ผ ๐‘ฅ โˆผ ๐ท ๐‘‹ โข [ โ„“ โข ( โ„Ž โ€ฒ โข ( ๐‘ฅ ) , โ„Ž โข ( ๐‘ฅ ) ) ] โ‰ค ๐œ‚ โ„“ โข ๐œ€ .

Because we assume that โ„“ โข ( ๐‘Ž , ๐‘ )

0 iff ๐‘Ž

๐‘ , this actually implies a stronger relationโ€” โ„Ž and โ„Ž โ€ฒ must be close in classification error:

Pr ๐‘ฅ โˆผ ๐ท ๐‘‹ โก [ โ„Ž โข ( ๐‘ฅ ) โ‰  โ„Ž โ€ฒ โข ( ๐‘ฅ ) ] โ‰ค ๐œ‚ โ„“ โข ๐œ€ min ๐‘Ž โ‰  ๐‘ โก ( โ„“ โข ( ๐‘Ž , ๐‘ ) ) .

Let โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โˆˆ ๐ป be an optimal hypothesis, and let โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โ€ฒ denote its corresponding close hypothesis as promised above in the output of LearningToCover. Then by the above, we have that:

๐‘’๐‘Ÿ๐‘Ÿ ๐ท , โ„“ โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โ€ฒ )

๐”ผ ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท โข [ โ„“ โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โ€ฒ โข ( ๐‘ฅ ) , ๐‘ฆ ) ]

๐”ผ ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท โข [ โ„“ โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โ€ฒ โข ( ๐‘ฅ ) , ๐‘ฆ ) โˆ’ โ„“ โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข ( ๐‘ฅ ) , ๐‘ฆ ) + โ„“ โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข ( ๐‘ฅ ) , ๐‘ฆ ) ]

๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐”ผ ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท โข [ โ„“ โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โ€ฒ โข ( ๐‘ฅ ) , ๐‘ฆ ) โˆ’ โ„“ โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข ( ๐‘ฅ ) , ๐‘ฆ ) ]

โ‰ค ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐‘ƒ โข ๐‘Ÿ ๐ท โข [ โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข ( ๐‘ฅ ) โ‰  โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โ€ฒ โข ( ๐‘ฅ ) ] โข max ๐‘Ž โ‰  ๐‘ โก ( โ„“ โข ( ๐‘Ž , ๐‘ ) )

โ‰ค ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ / 2

where we have used the assumption that we set ๐œ‚ โ„“

๐‘ โข min ๐‘Ž โ‰  ๐‘ โก ( โ„“ โข ( ๐‘Ž , ๐‘ ) ) max ๐‘Ž โ‰  ๐‘ โก ( โ„“ โข ( ๐‘Ž , ๐‘ ) ) for some universal ๐‘

0 .

Itโ€™s worth spending a moment discussing our only assumption on the loss function โ„“ , that it satisfies the identity of indiscernibles. This is not only a natural assumption for most cases in practice (that mislabeling has non-zero error), it is theoretically justified as well: realizable and agnostic learning arenโ€™t necessarily equivalent for โ„“ without this property, even in the distribution-free setting.

Proposition 5.6 (Identity of Indiscernibles Lower Bound).

There exists a realizably learnable class ( ๐‘‹ , ๐ป , โ„“ ) over a finite label space ๐‘Œ which is not agnostically learnable.

Proof 5.7.

Let the instance space ๐‘‹

โ„• be the set of natural numbers, the label space ๐‘Œ

{ 0 , 1 } 2 . We consider the hypothesis class ๐ป with all functions which output the first bit as 0 , that is:

๐ป

{ โ„Ž : โ„Ž โข ( ๐‘ฅ )

( 0 , โ‹… ) โข โˆ€ ๐‘ฅ โˆˆ ๐‘‹ } .

Furthermore, we define the loss function โ„“ : ๐‘Œ ร— ๐‘Œ โ†’ { 0 , 1 , ๐‘ } as

โ„“ โข ( ( ๐‘ 1 , ๐‘Ÿ 1 ) , ( ๐‘ 2 , ๐‘Ÿ 2 ) )

{ 0
๐‘ 1

๐‘ 2

1
๐‘ 1 โ‰  ๐‘ 2 โข ๐‘Ž๐‘›๐‘‘ โข ๐‘Ÿ 1

๐‘Ÿ 2

๐‘

otherwise.

Note that ( ๐‘‹ , ๐ป , โ„“ ) is trivially learnable in the realizable setting simply by returning any โ„Ž โˆˆ ๐ป . On the other hand, we will show it is only ๐‘‚ โข ( ๐‘ ) -agnostically learnable. First, notice that for any labelling ๐‘“ : ๐‘‹ โ†’ ๐‘Œ , there exists a hypothesis โ„Ž โˆˆ ๐ป which matches ๐‘“ on the second bit, and therefore for any marginal ๐ท over ๐‘‹ :

๐‘‚ โข ๐‘ƒ โข ๐‘‡ โ‰ค ๐‘’ โข ๐‘Ÿ โข ๐‘Ÿ ๐ท , ๐‘“ โข ( โ„Ž ) โ‰ค 1 .

As a result, it suffices to show that for every ๐‘š โˆˆ โ„• and (randomized) algorithm ๐’œ using ๐‘š samples there exists a labeling ๐‘“ : ๐‘‹ โ†’ ๐‘Œ and marginal distribution ๐ท ๐‘‹ such that

๐”ผ ๐‘† โˆผ ๐ท ๐‘‹ ๐‘š โข ๐”ผ ๐‘ฅ โˆผ ๐ท ๐‘‹ โข [ โ„“ โข ( ๐’œ โข ( ๐‘† , ๐‘“ โข ( ๐‘† ) ) โข ( ๐‘ฅ ) , ๐‘“ โข ( ๐‘ฅ ) ) ] โ‰ฅ ๐‘ / 12 .

(1)

As long as this holds Markovโ€™s inequality gives that every algorithm must have error at least ฮฉ โข ( ๐‘ ) with constant probability.

For simplicity, we will restrict our attention in the rest of the proof to the marginal distribution ๐ท ๐‘‹ which is uniform over the set [ ๐‘˜ ] for some natural number ๐‘˜ we will fix later. To prove Equation 1, by Yaoโ€™s minimax principle it is enough to prove there is a distribution ๐œ‡ over functions ๐‘“ : [ ๐‘˜ ] โ†’ ๐‘Œ such that any deterministic algorithm ๐’œ has expected loss at least ๐‘ / 12 over ๐œ‡ :

๐”ผ ๐‘“ โˆผ ๐œ‡ โข ๐”ผ ๐‘† โˆผ ๐ท ๐‘‹ ๐‘š โข ๐”ผ ๐‘ฅ โˆผ ๐ท ๐‘‹ โข [ โ„“ โข ( ๐’œ โข ( ๐‘† , ๐‘“ โข ( ๐‘† ) ) โข ( ๐‘ฅ ) , ๐‘“ โข ( ๐‘ฅ ) ) ]

๐‘ / 12 .

We now show that the above holds for ๐œ‡ being uniform over all functions from [ ๐‘˜ ] to ๐‘Œ for any ๐‘˜

2 โข ๐‘š . Here, we have that

๐”ผ ๐‘ฅ โˆผ ๐ท ๐‘‹ ๐”ผ ๐‘“ โˆผ ๐œ‡ ๐”ผ ๐‘† โˆผ ๐ท ๐‘‹ ๐‘š [ โ„“ ( ๐’œ ( ๐‘† , ๐‘“ ( ๐‘† ) ) ( ๐‘ฅ ) , ๐‘“ ( ๐‘ฅ ) ) ) ] โ‰ฅ ๐”ผ ๐‘ฅ โˆผ ๐ท ๐‘‹ [ Pr ๐‘† โˆผ ๐ท ๐‘‹ ๐‘š [ ๐‘ฅ โˆ‰ ๐‘† ] โ‹… ๐‘ / 4 ] ,

where the last step follows from noting that for any value ( ๐‘Ž , ๐‘ ) that ๐’œ โข ( ๐‘† , ๐‘“ โข ( ๐‘† ) ) assigns to ๐‘ฅ โˆ‰ ๐‘† , ๐‘“ โข ( ๐‘ฅ ) will be ( 1 โˆ’ ๐‘Ž , 1 โˆ’ ๐‘ ) with probability 1 / 4 incurring a loss of ๐‘ . The result then follows by noting that for every ๐‘ฅ โˆˆ [ ๐‘˜ ] :

Pr ๐‘† โˆผ ๐ท ๐‘‹ ๐‘š โก [ ๐‘ฅ โˆ‰ ๐‘† ]

( 1 โˆ’ 1 ๐‘˜ ) ๐‘š โ‰ฅ 1 / ๐‘’

since 1 โˆ’ ๐‘ฅ โ‰ฅ exp โก { โˆ’ ๐‘ฅ / ( 1 โˆ’ ๐‘ฅ ) } and ๐‘˜

2 โข ๐‘š . Therefore, we get that

๐”ผ ๐‘ฅ โˆผ ๐ท ๐‘‹ โข ๐”ผ ๐‘“ โˆผ ๐œ‡ โข ๐”ผ ๐‘† โˆผ ๐ท ๐‘‹ ๐‘š โข [ โ„“ โข ( ๐’œ โข ( ๐‘† , ๐‘“ โข ( ๐‘† ) ) โข ( ๐‘ฅ ) , ๐‘“ โข ( ๐‘ฅ ) ) ] โ‰ฅ ๐‘ / 4 โข ๐‘’ ,

which completes the proof.

Note that this bound holds even if ๐’œ is allowed to be improper. It is worth noting that if we are willing to increase the size of ๐‘Œ , the learnerโ€™s error in this bound can actually be increased all the way to ๐‘ , the maximum possible (see Proposition 6.1). This is in fact tight, as we will show that any loss function like the above satisfying a ๐‘ -approximate triangle inequality can be ๐‘ -agnostically learned (that is learned to within ๐‘ โ‹… ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ error).

6Four Modification Archetypes 6.1Discretization: Infinite Label Classes

In the previous section, we showed that our base reduction characterizes the equivalence of realizable and agnostic learning for loss functions satisfying the identity of indiscernibles for all finite label classes. In this section, we discuss a technique called discretization that allows our reduction to extend this result to infinite label classes. Normally, itโ€™s clear that when ๐‘Œ is infinite our standard reduction will fail: since the total number of possible labelings of a finite sample may be infinite, LearningToCover may output an infinite set. In fact, this is more than a technical barrier: realizable and agnostic learning simply arenโ€™t equivalent for infinite label classes.

Proposition 6.1.

Let โ„“ be any continuous loss function (in the first variable) over โ„ satisfying the identity of indiscernibles. Then there exists a class ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) which is realizably learnable but not agnostically learnable.

Proof 6.2.

Let ๐‘‹

โ„• and ๐‘Œ

[ 0 , 2 ] . To construct our class, we first consider the set of all boolean functions over ๐‘‹ with finite support. Each function ๐‘“ in this class may equivalently be thought of as a binary string in { 0 , 1 } * . Denote the corresponding decimal value of this string in [ 0 , 1 ] by ๐‘  ๐‘“ . To construct ๐ป , for every boolean function ๐‘“ : โ„• โ†’ { 0 , 1 } with finite support, include in ๐ป the function โ„Ž ๐‘“ โข ( ๐‘ฅ )

๐‘“ โข ( ๐‘ฅ ) + ๐‘  ๐‘“ . Note that ( ๐‘‹ , ๐ป ) is clearly realizably learnable under any distribution family ๐’Ÿ and any loss function, since a single sample always uniquely determines โ„Ž ๐‘“ . On the other hand, adding even the smallest amount of noise erases this unique identification, making the class impossible to learn.

More formally, let ๐’Ÿ be the family of all distributions. By the continuity of โ„“ and the fact that โ„“ โข ( 0 , 0 )

โ„“ โข ( 1 , 1 )

0 , for all ๐œ€ > 0 notice that there exists ๐›พ

๐›พ โข ( ๐œ€ )

0 such that max 0 โ‰ค ๐›พ โ€ฒ โ‰ค ๐›พ โก { โ„“ โข ( ๐›พ โ€ฒ , 0 ) , โ„“ โข ( 1 + ๐›พ โ€ฒ , 1 ) } < ๐œ€ . Let ๐‘› ๐›พ โˆˆ โ„• be the index of the first non-zero digit in the binary representation of ๐›พ . The idea is to note that beyond these first ๐‘› ๐›พ coordinates, our class if within ๐œ€ of an arbitrary boolean function. More formally, notice that for any distribution ๐ท and boolean function ๐‘“ which is 0 on [ ๐‘› ๐›พ ] , we have that ๐‘‚๐‘ƒ๐‘‡ ๐ป โข ( ๐‘“ ) โข \coloneqq โข min โ„Ž โˆˆ ๐ป โก { ๐‘’๐‘Ÿ๐‘Ÿ ๐ท ร— ๐‘“ , โ„“ โข ( โ„Ž ) } โ‰ค ๐œ€ . The bound then follows from the fact that such arbitrary functions are not learnable.

In more detail, Yaoโ€™s minimax principle states that it is sufficient to show that for any potential sample complexity ๐‘š โข ( ๐œ€ , ๐›ฟ ) , there exists a randomized strategy for the adversary such that no deterministic learner can achieve ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐‘ accuracy with constant probability for some constant ๐‘ > 0 . To this end, consider the following strategy: the adversary chooses the uniform distribution over [ ๐‘› ๐›พ , ๐‘› ๐›พ + 2 โข ๐‘š โข ( ๐œ€ , ๐›ฟ ) ] , and a binary function on that interval uniformly at random (recall ๐‘‚ โข ๐‘ƒ โข ๐‘‡ is at most ๐œ€ for every such function). Since the learner can only see 1 / 2 of the mass, any strategy must be incorrect on half of the remaining points in expectation. In particular conditioned on any sample, the expected loss of any predicted labeling of an unseen point is at least โ„“ min-err

min ๐‘ฆ โˆˆ [ 0 , 2 ] โข { โ„“ โข ( ๐‘ฆ , 0 ) + โ„“ โข ( ๐‘ฆ , 1 ) 2 } (since each unseen label appears with 1 / 2 probability conditioned on the learnerโ€™s sample). The total expected loss of any strategy is then at least โ„“ min-err / 2 , which is bounded away from 0 . Setting ๐œ€ and ๐‘ sufficiently small then gives the desired result.

Proposition 6.1 relies crucially on the fact that the adversary can erase a significant amount of information with a very small label perturbation. In the rest of this section, weโ€™ll discuss a technique for modifying our reduction that shows this is essentially the only barrier between realizable and agnostic learning (at least for a broad class of loss functions). The key is to require a slightly stronger notion of learnability based upon discretization.

Definition 6.3 (Discretization).

We say ( ๐’Ÿ , ๐‘‹ , ๐ป โ€ฒ , โ„“ ) is an ๐œ€ -discretization of ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) if the following three conditions hold:

1.

๐ป โ€ฒ is probably bounded. That is for all ๐‘› โˆˆ โ„• , ๐›ฟ

0 , and ๐ท โˆˆ ๐’Ÿ there exists a bound ๐‘š โข ( ๐‘› , ๐›ฟ ) โˆˆ โ„• such that:

Pr ๐‘† โˆผ ๐ท ๐‘› [ | ๐ผ ๐‘š ( ๐ป โ€ฒ | ๐‘† ) | โ‰ค ๐‘š ( ๐‘› , ๐›ฟ ) ] โ‰ฅ 1 โˆ’ ๐›ฟ .

2.

๐ป โ€ฒ point-wise13 ฮต -covers ๐ป with respect to โ„“ . That is for all for all โ„Ž โˆˆ ๐ป , there exists โ„Ž โ€ฒ โˆˆ ๐ป โ€ฒ satisfying:

โˆ€ ๐‘ฅ โˆˆ ๐‘‹ : โ„“ โข ( โ„Ž โ€ฒ โข ( ๐‘ฅ ) , โ„Ž โข ( ๐‘ฅ ) ) โ‰ค ๐œ€ .

3.

๐ป โ€ฒ is always useful. That is for all โ„Ž โ€ฒ โˆˆ ๐ป โ€ฒ , there exists โ„Ž โˆˆ ๐ป such that:

โˆ€ ๐‘ฅ โˆˆ ๐‘‹ : โ„“ โข ( โ„Ž โ€ฒ โข ( ๐‘ฅ ) , โ„Ž โข ( ๐‘ฅ ) ) โ‰ค ๐œ€ .

Note that most realistic settings have reasonable discretizations (e.g. it is enough to have some Lipshitz-like condition and a weak tail-bound on the loss). We now define a basic notion of learnability based on discretization which essentially serves to rule out adversarial constructions in the vein of Proposition 6.1.

Definition 6.4 (Discretely-learnable).

We say ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) is discretely-learnable with sample complexity ๐‘› โข ( ๐œ€ , ๐›ฟ ) if there is some constant ๐‘ 1

0 such that for all ๐œ€ , ๐›ฟ

0 there exists an ๐œ€ -discretization ๐ป ๐œ€ which is ( ๐‘ 1 โข ๐œ€ , ๐›ฟ ) -PAC-learnable in at most ๐‘› โข ( ๐œ€ , ๐›ฟ ) samples. We call the learner proper if it outputs hypotheses in ๐ป .

Weโ€™ll prove that discrete and agnostic learnability are equivalent as long as the loss satisfies an approximate triangle inequality.

Definition 6.5 (Approximate pseudometric).

We call a loss function โ„“ : ๐‘Œ ร— ๐‘Œ โ†’ โ„ โ‰ฅ 0 a ๐‘ -approximate pseudometric if for all triples ๐‘ฆ 1 , ๐‘ฆ 2 , ๐‘ฆ 3 โˆˆ ๐‘Œ :

โ„“ โข ( ๐‘ฆ 1 , ๐‘ฆ 3 ) โ‰ค ๐‘ โข ( โ„“ โข ( ๐‘ฆ 1 , ๐‘ฆ 2 ) + โ„“ โข ( ๐‘ฆ 2 , ๐‘ฆ 3 ) ) .

Approximate pseudometrics are natural choices for loss functions in practice and capture a broad set of scenarios including finite-range losses and standard setups such as โ„“ ๐‘ -regression, and have seen some previous study in the literature [23]. By modifying the first step of our reduction to take discretization into account and leveraging the approximate triangle inequality in the second, we prove that discrete learnability and ๐‘ -agnostic learnability are equivalent under ๐‘ -approximate pseudometrics.

Theorem 6.6.

Let โ„“ : ๐‘Œ ร— ๐‘Œ โ†’ โ„ โ‰ฅ 0 be a bounded ๐‘ -approximate pseudometric. Then the following are equivalent for all ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) :

1.

( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) is discretely-learnable.

2.

( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) is ๐‘ -agnostically learnable.

Proof 6.7.

The proof is similar to Theorem 5.4. We first show the forward direction. Assume ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) is discretely-learnable. Fix ๐œ€ โ€ฒ

๐œ€ 4 โข ๐‘ 2 โข ๐‘ 1 (where ๐‘ 1 is the constant from Definition 6.4), and let ๐ป ๐œ€ โ€ฒ be a learnable ๐œ€ โ€ฒ -discretization of ๐ป . We argue that running LearningToCover on ๐ป ๐œ€ โ€ฒ gives the desired agnostic learner. Since โ„“ is bounded, it is sufficient to prove that ๐ถ โข ( ๐‘† ๐‘ˆ ) contains a hypothesis โ„Ž โ€ฒ such that ๐‘’๐‘Ÿ๐‘Ÿ ๐ท , โ„“ โข ( โ„Ž โ€ฒ ) โ‰ค ๐‘ โ‹… ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ / 2 . Empirical risk minimization then works as in the finite case.

Let โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โˆˆ ๐ป be an optimal hypothesis. Since ๐ป ๐œ€ โ€ฒ is a discretization of ๐ป , there exists โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ๐œ€ โ€ฒ โˆˆ ๐ป ๐œ€ โ€ฒ such that:

โˆ€ ๐‘ฅ โˆˆ ๐‘‹ : โ„“ โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข ( ๐‘ฅ ) , โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ๐œ€ โ€ฒ โข ( ๐‘ฅ ) ) < ๐œ€ โ€ฒ .

Further, by the guarantees of discrete learnability, with probability at least 1 โˆ’ ๐›ฟ / 2 there exists โ„Ž โ€ฒ โˆˆ ๐ถ โข ( ๐‘† ๐‘ˆ ) such that close to โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ๐œ€ โ€ฒ in the following sense:

๐”ผ ๐‘ฅ โˆผ ๐ท ๐‘‹ โข [ โ„“ โข ( โ„Ž โ€ฒ โข ( ๐‘ฅ ) , โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ๐œ€ โ€ฒ โข ( ๐‘ฅ ) ) ] โ‰ค ๐‘ 1 โข ๐œ€ โ€ฒ

๐œ€ 4 โข ๐‘ 2

Plugging in the previous observation and applying our approximate triangle inequality, we get that โ„Ž โ€ฒ is close to โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ in the following sense:

๐”ผ ๐‘ฅ โˆผ ๐ท ๐‘‹ โข [ โ„“ โข ( โ„Ž โ€ฒ โข ( ๐‘ฅ ) , โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข ( ๐‘ฅ ) ) ]

โ‰ค ๐‘ โข ( ๐”ผ ๐‘ฅ โˆผ ๐ท ๐‘‹ โข [ โ„“ โข ( โ„Ž โ€ฒ โข ( ๐‘ฅ ) , โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ๐œ€ โ€ฒ โข ( ๐‘ฅ ) ) ] + ๐”ผ ๐‘ฅ โˆผ ๐ท ๐‘‹ โข [ โ„“ โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ๐œ€ โ€ฒ โข ( ๐‘ฅ ) , โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข ( ๐‘ฅ ) ) ] )

โ‰ค ๐œ€ 2 โข ๐‘

The final step is to transfer from the marginal ๐ท ๐‘‹ to the full joint distribution of the adversary, which follows immediately from a similar application of the approximate triangle inequality. This is the only step that loses a factor in the OPT term:

๐‘’๐‘Ÿ๐‘Ÿ ๐ท , โ„“ โข ( โ„Ž โ€ฒ )

๐”ผ ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท โข [ โ„“ โข ( โ„Ž โ€ฒ โข ( ๐‘ฅ ) , ๐‘ฆ ) ]

โ‰ค ๐‘ โข ( ๐”ผ ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท โข [ โ„“ โข ( โ„Ž โ€ฒ โข ( ๐‘ฅ ) , โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข ( ๐‘ฅ ) ) ] + ๐”ผ ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท โข [ โ„“ โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข ( ๐‘ฅ ) , ๐‘ฆ ) ] )

โ‰ค ๐‘ โ‹… ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ / 2

as desired.

We now prove the reverse direction, which is essentially immediate. Assume the existence of a ๐‘ -agnostic learner for ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) . Given a discretization ๐ป ๐œ€ , we want to show ( ๐’Ÿ , ๐‘‹ , ๐ป ๐œ€ , โ„“ ) is learnable to within ๐‘ 1 โข ๐œ€ error for some ๐‘ 1

0 . This is achieved simply by running the agnostic learner on ( ๐’Ÿ , ๐‘‹ , ๐ป ๐œ€ , โ„“ ) . Since ๐ป ๐œ€ is โ€œalways usefulโ€, every โ„Ž โˆˆ ๐ป ๐œ€ is ๐œ€ -close to some โ„Ž โ€ฒ โˆˆ ๐ป in the sense that:

โˆ€ ๐‘ฅ โˆˆ ๐‘‹ : โ„“ โข ( โ„Ž โ€ฒ โข ( ๐‘ฅ ) , โ„Ž โข ( ๐‘ฅ ) ) โ‰ค ๐œ€ .

In particular, this means that for any choice of โ„Ž by the adversary there exists โ„Ž โ€ฒ โˆˆ ๐ป with low error:

๐‘’๐‘Ÿ๐‘Ÿ ๐ท , โ„“ โข ( โ„Ž โ€ฒ )

๐”ผ โข [ โ„“ โข ( โ„Ž โ€ฒ โข ( ๐‘ฅ ) , โ„Ž โข ( ๐‘ฅ ) ) ] โ‰ค ๐œ€ .

As a result, running the ๐‘ -agnostic learner for ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) returns a hypothesis of at most ( ๐‘ + 1 ) โข ๐œ€ error with high probability.

It is worth noting that bounded loss is not really necessary for Theorem 6.6. More generally we can require that ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) is โ€œfinitely learnableโ€ in the sense that for all finite subsets ๐ป โ€ฒ โŠ‚ ๐ป , ( ๐’Ÿ , ๐‘‹ , ๐ป โ€ฒ , โ„“ ) is agnostically learnable. When โ„“ is bounded, this is true for any finite class by empirical risk minimization. When โ„“ is unbounded, we can still get finite learnability if โ„“ โข ( โ„Ž โข ( โ‹… ) , โ‹… ) satisfies some concentration bounds [48] (such classes are discretizable since one can simply truncate the loss). It is also possible to achieve improved convergence rates by using algorithms other than empirical risk minimization for the finite learner (for example a generalization of the median-of-means estimator [37]), but our reduction will suffer an additional factor of ๐œ€ over applying such techniques directly due to the size of the non-uniform cover.

It is also worth noting that various modifications to the definition of loss (e.g. defining loss between hypotheses rather than on ๐‘Œ directly) will continue to work with the above. Similarly, there are various cases when one can get better than ๐‘ โ‹… ๐‘‚ โข ๐‘ƒ โข ๐‘‡ accuracy for a ๐‘ -approximate pseudometric, generally by instead optimizing over some surrogate loss function. For instance, if a simple transformation of the loss gives a ๐‘ โ€ฒ -approximate pseudometric for ๐‘ โ€ฒ < ๐‘ , then one can generally learn up to ๐‘ โ€ฒ โ‹… ๐‘‚ โข ๐‘ƒ โข ๐‘‡ .14 As an example, note that while square loss โ„“ 2 โข ( ๐‘ฅ , ๐‘ฆ )

( ๐‘ฅ โˆ’ ๐‘ฆ ) 2 is a 2-approximate pseudometric, taking err ๐ท , โ„“ 2 gives a true metric between hypotheses. As a result, as long as ๐‘‚ โข ๐‘ƒ โข ๐‘‡ is bounded, we can get truly agnostic learning by optimizing err ๐ท , โ„“ 2 instead. This strategy works for any polynomial loss, such as โ„“ ๐‘ โข ( ๐‘ฅ , ๐‘ฆ )

| ๐‘ฅ โˆ’ ๐‘ฆ | ๐‘ .

On the other hand, outside of these special cases, Theorem 6.6 is tight: there exist ๐‘ -approximate pseudometric loss functions which cannot be ๐‘ โ€ฒ -agnostically learned for any ๐‘ โ€ฒ < ๐‘ . The argument is similar to Proposition 6.1, but requires a bit more care.

Proposition 6.8.

There exists a discretely-learnable class over a ๐‘ -approximate pseudometric that is not ๐‘ โ€ฒ -agnostically learnable for any ๐‘ โ€ฒ < ๐‘ .

Proof 6.9.

The proof is similar to Proposition 5.6. We consider the same instance space ๐‘‹

โ„• and hypothesis class ๐ป :

๐ป

{ โ„Ž : โ„Ž โข ( ๐‘ฅ )

( 0 , โ‹… ) โข โˆ€ ๐‘ฅ โˆˆ ๐‘‹ } .

The loss function โ„“ : ๐‘Œ ร— ๐‘Œ โ†’ { 0 , 1 , ๐‘ } is also the same, but extended to the larger domain ๐‘Œ

โ„• 2 :

โ„“ โข ( ( ๐‘ 1 , ๐‘Ÿ 1 ) , ( ๐‘ 2 , ๐‘Ÿ 2 ) )

{ 0
๐‘ 1

๐‘ 2

1
๐‘ 1 โ‰  ๐‘ 2 โข ๐‘Ž๐‘›๐‘‘ โข ๐‘Ÿ 1

๐‘Ÿ 2

๐‘

๐‘œ๐‘กโ„Ž๐‘’๐‘Ÿ๐‘ค๐‘–๐‘ ๐‘’ .

As before, note that ( ๐‘‹ , ๐ป , โ„“ ) is trivially realizably learnable by always returning any โ„Ž โˆˆ ๐ป , โ„“ is a ๐‘ -psuedometric by definition, and for any labeling ๐‘“ : ๐‘‹ โ†’ ๐‘Œ there exists โ„Ž โˆˆ ๐ป such that for all distributions ๐ท :

๐‘‚ โข ๐‘ƒ โข ๐‘‡ โ‰ค ๐‘’ โข ๐‘Ÿ โข ๐‘Ÿ ๐ท , ๐‘“ โข ( โ„Ž ) โ‰ค 1 .

We now show that the class ( ๐‘‹ , ๐ป , โ„“ ) is only ๐‘ -agnostically learnable. Since ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โ‰ค 1 , it suffices to show that for every ๐‘š โˆˆ โ„• , large enough ๐‘› โˆˆ โ„• , and randomized algorithm ๐’œ on ๐‘š samples, there exists a labeling ๐‘“ : ๐‘‹ โ†’ ๐‘Œ and a marginal distribution ๐ท ๐‘‹ such that:

๐”ผ ๐‘† โˆผ ๐ท ๐‘‹ ๐‘š โข ๐”ผ ๐‘ฅ โˆผ ๐ท ๐‘‹ โข [ โ„“ โข ( ๐’œ โข ( ๐‘† , ๐‘“ โข ( ๐‘† ) ) โข ( ๐‘ฅ ) , ๐‘“ โข ( ๐‘ฅ ) ) ] โ‰ฅ ( 1 โˆ’ 1 ๐‘› ) 3 โข ๐‘ .

(2)

For ๐‘› โ‰ฅ 1 1 โˆ’ ( 1 โˆ’ ( ๐‘ โˆ’ ๐‘ โ€ฒ ) / ๐‘ ) 1 / 3 , applying Markovโ€™s inequality to Equation 2 implies that ๐’œ has error at least ๐‘ โ€ฒ with constant probability.

For simplicity, we now restrict our attention to the marginal ๐ท ๐‘‹ which is uniform over the set [ ๐‘˜ ] for some ๐‘˜ โˆˆ โ„• to be fixed. By Yaoโ€™s minimax principle, its enough to prove that there exists a distribution ๐œ‡ over functions ๐‘“ : [ ๐‘˜ ] โ†’ [ ๐‘› ] 2 such that any deterministic algorithm ๐’œ the following holds

๐”ผ ๐‘“ โˆผ ๐œ‡ โข ๐”ผ ๐‘† โˆผ ๐ท ๐‘‹ ๐‘š โข ๐”ผ ๐‘ฅ โˆผ ๐ท ๐‘‹ โข [ โ„“ โข ( ๐’œ โข ( ๐‘† , ๐‘“ โข ( ๐‘† ) ) โข ( ๐‘ฅ ) , ๐‘“ โข ( ๐‘ฅ ) ) ] โ‰ฅ ( 1 โˆ’ 1 ๐‘› ) 3 โข ๐‘ .

We now show that the above holds for ๐œ‡ being uniform over all functions from [ ๐‘˜ ] to [ ๐‘› ] 2 when ๐‘˜

2 โข ๐‘š ln โก ( ๐‘› / ( ๐‘› โˆ’ 1 ) ) . Similar to Proposition 5.6, we have that

๐”ผ ๐‘ฅ โˆผ ๐ท ๐‘‹ โข ๐”ผ ๐‘“ โˆผ ๐œ‡ โข ๐”ผ ๐‘† โˆผ ๐ท ๐‘‹ ๐‘š โข [ โ„“ โข ( ๐’œ โข ( ๐‘† , ๐‘“ โข ( ๐‘† ) ) โข ( ๐‘ฅ ) , ๐‘“ โข ( ๐‘ฅ ) ) ] โ‰ฅ ๐”ผ ๐‘ฅ โˆผ ๐ท ๐‘‹ โข [ Pr ๐‘† โˆผ ๐ท ๐‘‹ ๐‘š โก [ ๐‘ฅ โˆ‰ ๐‘† ] โ‹… ( 1 โˆ’ 1 ๐‘› ) 2 โ‹… ๐‘ ] ,

since no matter the assignment ๐’œ gives to ๐‘ฅ โˆ‰ ๐‘† , it will be wrong on both coordinates with probability ( 1 โˆ’ 1 / ๐‘› ) 2 over the randomness of ๐œ‡ . The result follows by noting that for every ๐‘ฅ โˆˆ [ ๐‘˜ ]

Pr ๐‘† โˆผ ๐ท ๐‘‹ ร— ๐‘“ โข ( ๐‘‹ ) ๐‘š โก [ ( ๐‘ฅ , โ‹… ) โˆ‰ ๐‘† ]

( 1 โˆ’ 1 ๐‘˜ ) ๐‘š โ‰ฅ 1 โˆ’ 1 ๐‘›

since 1 โˆ’ ๐‘ฅ โ‰ฅ exp โก { โˆ’ ๐‘ฅ / ( 1 โˆ’ ๐‘ฅ ) } , and we have assumed ๐‘˜

2 โข ๐‘š ln โก ( ๐‘› / ( ๐‘› โˆ’ 1 ) ) and ๐‘›

1 . Therefore, we get that

๐”ผ ๐‘ฅ โˆผ ๐ท ๐‘‹ โข ๐”ผ ๐‘“ โˆผ ๐œ‡ โข ๐”ผ ๐‘† โˆผ ๐ท ๐‘‹ ๐‘š โข [ โ„“ โข ( ๐’œ โข ( ๐‘† , ๐‘“ โข ( ๐‘† ) ) โข ( ๐‘ฅ ) , ๐‘“ โข ( ๐‘ฅ ) ) ] โ‰ฅ ( 1 โˆ’ 1 ๐‘› ) 3 โข ๐‘

which completes the proof.

6.2Sub-sampling: Malicious Noise

Now that weโ€™ve seen how to handle practical problems like regression over infinite label spaces, weโ€™ll discuss a technique that helps handle data corruption and data-dependent assumptions: sub-sampling. The main idea is as follows. Say that the original unlabeled sample we draw is, in some sense, partially corrupted: perhaps an adversary has changed some fraction of examples (malicious noise), or some portion of the sample is un-realizable for a concept in the class (robust and partial learning). In either case, there generally exists a core subset of โ€œcleanโ€ samples that we can use to recover the guarantees of LearningToCover. Since we cannot necessarily identify these, the idea is to run LearningToCover over enough subsets of the unlabeled sample that we find a clean subsample with high probability. In this section weโ€™ll discuss the application of this technique in detail to Kearns and Liโ€™s [41] well-studied malicious noise model. In Sections 8 and 9, we discuss further applications of the method to adversarially robust and partial learning.

To start, letโ€™s recall the standard malicious noise model. In this variant of PAC learning, instead of having access to the standard sample oracle from the adversaryโ€™s distribution ๐ท over ๐‘‹ ร— ๐‘Œ , we have access to a malicious oracle ๐’ช ๐‘€ โข ( โ‹… ) which, with probability ๐œ‚ , outputs an adversarially chosen pair ( ๐‘ฅ , ๐‘ฆ ) , and otherwise samples from ๐ท as usual.

Definition 6.10 (PAC-learning with Malicious Noise).

We call ( ๐‘‹ , ๐ป , ๐’Ÿ , โ„“ ) (agnostically) ( ๐œ€ , ๐›ฟ ) -learnable with malicious noise at rate ๐œ‚

๐œ‚ โข ( ๐œ€ ) if there exists an algorithm ๐’œ and function ๐‘š

๐‘š ๐‘š๐‘Ž๐‘™ โข ( ๐œ€ , ๐›ฟ ) such that for all ๐œ€ , ๐›ฟ

0 , and distributions ๐ท over ๐‘‹ ร— ๐‘Œ satisfying:

1.

The marginal ๐ท ๐‘‹ โˆˆ ๐’Ÿ ,

๐’œ outputs a good hypothesis with high probability over samples drawn from the malicious oracle of size ๐‘› โข ( ๐œ€ , ๐›ฟ ) :

Pr ๐‘† โˆผ ๐’ช ๐‘€ โข ( โ‹… ) ๐‘š โก [ ๐‘’๐‘Ÿ๐‘Ÿ ๐ท , โ„“ โข ( ๐’œ โข ( ๐‘† ) )

๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ ] โ‰ค ๐›ฟ .

where ๐‘‚ โข ๐‘ƒ โข ๐‘‡

min โ„Ž โˆˆ ๐ป โก { ๐‘’๐‘Ÿ๐‘Ÿ ๐ท , โ„“ โข ( โ„Ž ) } .

In other words, malicious noise essentially gives a worst-case formalization of the idea that an ๐œ‚ -fraction of the learnerโ€™s data is (adversarial) garbage.

Letโ€™s now formalize the argument above: modifying LearningToCover to run over subsamples gives a sample-efficient algorithm for learning with malicious noise. For readability, weโ€™ll (somewhat informally) restate the algorithm with this change.

Input: Realizable PAC-Learner ๐’œ , Accuracy Parameter ๐œ€ < 1 / 2 , Noise Parameter ๐œ‚ < ๐œ€ 1 + ๐œ€ , Unlabeled Sample Oracle ๐’ช ๐‘ˆ , Labeled Sample Oracle ๐’ช ๐ฟ Algorithm:

1.

Draw an unlabeled sample ๐‘† ๐‘ˆ โˆผ ๐’ช ๐‘ˆ , and labeled sample ๐‘† ๐ฟ โˆผ ๐’ช ๐ฟ .

2.

Run LearningToCover over all ๐‘† โˆˆ ( ๐‘† ๐‘ˆ ) ๐œ‚ โ€ฒ โข \coloneqq โข { ๐‘† โІ ๐‘† ๐‘ˆ : | ๐‘† |

โŒŠ ( 1 โˆ’ ๐œ‚ โ€ฒ ) โข | ๐‘† ๐‘ˆ | โŒ‹ } , where:

๐œ‚ โ€ฒ

3 โข ๐œ‚ + ๐œ€ / ( 1 โˆ’ ๐œ€ ) 4 .

3.

Return the hypothesis in ๐ถ โข ( ๐‘† ๐‘ˆ ) โข \coloneqq โข โ‹ƒ ๐‘† โˆˆ ( ๐‘† ๐‘ˆ ) ๐œ‚ โ€ฒ โข ๐ถ โข ( ๐‘† ) with lowest empirical error over ๐‘† ๐ฟ .

Algorithm 3 Malicious to Realizable Reduction

We now prove that Algorithm 3 gives an (agnostic) learner that is tolerant to malicious noise.

Theorem 6.11.

Let ( ๐’Ÿ , ๐‘‹ , ๐ป ) be realizably PAC-learnable with sample complexity ๐‘› โข ( ๐œ€ , ๐›ฟ ) . Then for any ๐œ‚ < ๐œ€ 1 + ๐œ€ , Algorithm 3 is an agnostic learner for ( ๐’Ÿ , ๐‘‹ , ๐ป ) tolerant to ๐œ‚ malicious noise. Furthermore letting ฮ”

๐œ€ 1 + ๐œ€ โˆ’ ๐œ‚ and ๐›ฝ

( 1 + ๐œ‚ ฮ” ) โข log โก ( 1 ๐œ‚ ) , its sample complexity is at most:

๐‘š ๐‘š๐‘Ž๐‘™ โข ( ๐œ€ , ๐›ฟ ) โ‰ค ๐‘‚ โข ( ๐›ฝ โข ๐‘› โข ( ฮ” 4 , ๐›ฟ 4 ) ฮ” + log โก ( ฮ  ๐ป โข ( ๐‘‚ โข ( ๐‘› โข ( ฮ” 2 , ๐›ฟ 2 ) + ๐œ‚ 2 โข log โก ( 1 / ๐›ฟ ) ฮ” 2 ) ) ) + log โก ( 1 / ๐›ฟ ) ฮ” 2 + ๐›ฝ โข ๐œ‚ 2 โข log โก ( 1 / ๐›ฟ ) ฮ” 3 )

where weโ€™ve assumed ๐œ€ < 1 / 2 for simplicity.

Proof 6.12.

To start, weโ€™ll review for completeness a fairly standard analysis of empirical risk minimization under malicious noise. Assume for the moment that the output of LearningToCover, ๐ถ โข ( ๐‘† ๐‘ˆ ) , contains a hypothesis โ„Ž โ€ฒ satisfying ๐‘’ โข ๐‘Ÿ โข ๐‘Ÿ โข ( โ„Ž โ€ฒ ) โ‰ค ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐›ฝ 1 . Say we draw ๐‘€ labeled samples for the ERM step, and an ๐œ‚ โ€ฒ

๐œ‚ + ๐›ฝ 2 fraction are corrupted by the adversary. For large enough ๐‘€ , we can assume by a Chernoff and union bound that the empirical loss of every hypothesis returned by LearningToCover is at most some ๐›ฝ 3 away from its true loss on the un-corrupted portion of ๐‘€ (we will make all these assumptions formal in a moment). Given these facts, notice that the empirical loss of โ„Ž โ€ฒ on ๐‘€ is at most:

๐‘’ โข ๐‘Ÿ โข ๐‘Ÿ ๐‘€ โข ( โ„Ž โ€ฒ ) โ‰ค ( 1 โˆ’ ๐œ‚ โ€ฒ ) โข ( ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐›ฝ 1 + ๐›ฝ 3 ) + ๐œ‚ โ€ฒ .

On the other hand, the empirical error of any โ„Ž ๐œ€ โˆˆ ๐ป whose true error is greater than ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ is at least:

๐‘’ โข ๐‘Ÿ โข ๐‘Ÿ ๐‘€ โข ( โ„Ž )

( 1 โˆ’ ๐œ‚ โ€ฒ ) โข ( ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ โˆ’ ๐›ฝ 3 ) .

To ensure that our ERM works, it is enough to show that for any such โ„Ž , ๐‘’ โข ๐‘Ÿ โข ๐‘Ÿ ๐‘€ โข ( โ„Ž ) > ๐‘’ โข ๐‘Ÿ โข ๐‘Ÿ ๐‘€ โข ( โ„Ž โ€ฒ ) . A simple calculation shows that this is satisfied as long as ๐›ฝ 1 + 2 โข ๐›ฝ 3 โ‰ค ๐œ€ and ๐›ฝ 1 + ๐›ฝ 2 + 2 โข ๐›ฝ 3 โ‰ค ฮ” . Setting ๐›ฝ 1

๐›ฝ 2

๐›ฝ 3

ฮ” / 4 gives the desired result.

It is left to argue that our assumptions above hold with high probability. First, note that by a Chernoff bound, the probability that ๐œ‚ โ€ฒ โ‰ฅ ๐œ‚ + ฮ” / 4 on a set of ๐‘€ samples is at most ๐‘’ โˆ’ ๐‘ โข ( ฮ” ๐œ‚ ) 2 โข ๐‘€ for some ๐‘

0 , so this occurs with high probability as long as ๐‘€ โ‰ฅ ฮฉ โข ( log โก ( 1 / ๐›ฟ ) โข ๐œ‚ 2 / ฮ” 2 ) . Similarly, the empirical error of every hypothesis in ๐ถ โข ( ๐‘† ๐‘ˆ ) on the remaining clean samples will be within ฮ” / 4 of its true error as long as ๐‘€ โ‰ฅ ฮฉ โข ( log โก ( | ๐ถ โข ( ๐‘† ๐‘ˆ ) | / ๐›ฟ ) โข ๐œ‚ 2 / ฮ” 2 ) .

It is then left to show that ๐ถ โข ( ๐‘† ๐‘ˆ ) contains a hypothesis of error at most ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ฮ” / 4 . To show this, it is enough to ensure that we run LearningToCover over a clean subsample of size at least ๐‘› โข ( ฮ” / 4 , ๐›ฟ / 4 ) with high probability. If we draw | ๐‘† ๐‘ˆ |

๐‘‚ โข ( ๐‘› โข ( ฮ” / 4 , ๐›ฟ / 4 ) 1 โˆ’ ๐œ‚ โ€ฒ + log โก ( 1 / ๐›ฟ ) โข ๐œ‚ 2 ฮ” 2 ) unlabeled samples, a similar Chernoff bound to the above promises that at most an ๐œ‚ โ€ฒ fraction are corrupted with high probability, and therefore that at least ๐‘› โข ( ฮ” / 4 , ๐›ฟ / 4 ) remain un-corrupted. Running LearningToCover over all subsets of size ( 1 โˆ’ ๐œ‚ โ€ฒ ) โข | ๐‘† ๐‘ข | then gives the desired result. The sample complexity bound follows from choosing ๐‘€ large enough to satisfy the above conditions along with the fact that | ๐ถ โข ( ๐‘† ๐‘ˆ ) | โ‰ค ( ๐‘› ( 1 โˆ’ ๐œ‚ โ€ฒ ) โข ๐‘› ) โข ฮ  ๐ป โข ( ๐‘› ) .

A few remarks are in order. First, the error tolerance of Theorem 6.11 is tight. In their original introduction of malicious noise, Kearns and Li [41] proved that for most non-trivial concept classes, no PAC-learner can be tolerate ๐œ€ 1 + ๐œ€ malicious noise. Second, we note that Kearns and Li also give a reduction procedure from a base learner which achieves better dependence on ฮ” above, however it allows the learner to query specifically for positive or negative examples which is a stronger setup than our model. Finally, we note in the case of finite VC, the above is off from known (randomized) bounds by a factor of roughly ฮ” 2 [21]. On the other hand, Theorem 6.11 has the advantage of extending far beyond the VC setting, including to the general loss models discussed in Section 6.1. The proof remains mostly the same (with optimal error tolerance differing slightly), and is omitted.

Since our agnostic model restricts the adversary to choosing a distribution whose marginal lies in the original family, Theorem 6.11 provides the first insight on robustness against an adversary who can corrupt the underlying data as well as the labels. One might wonder whether this result can be pushed further: is it possible to be robust against an adversary who can corrupt the marginal over ๐‘‹ in some stronger sense? Unfortunately, the answer is no: malicious noise is necessarily the most distributional corruption we can handle. Letโ€™s look at two basic lower bounds to see why. First, weโ€™ll consider an adversary who can remove a portion of the learnerโ€™s sample.

Proposition 6.13.

For any ๐›ฟ

0 , there exists a class ( ๐’Ÿ , ๐‘‹ , ๐ป ) which is realizably PAC-learnable, but not learnable under an adversary who can remove a ๐›ฟ fraction of the learnerโ€™s sample.

Proof 6.14.

This follows from a result of Dudley, Kulkarni, Richardson, and Zeitouni [31] that there exists an unlearnable class ( ๐’Ÿ , ๐‘‹ , ๐ป ) such that for some ๐‘› โข ( ๐œ€ , ๐›ฟ ) , ( ๐ท , ๐‘‹ , ๐ป ) is learnable in ๐‘› โข ( ๐œ€ , ๐›ฟ ) samples for every ๐ท โˆˆ ๐’Ÿ . The lower bound then follows simply from adding an extra unique identifying point ๐‘ฅ ๐ท to ๐‘‹ for every distribution ๐ท , and modifying each ๐ท โˆˆ ๐’Ÿ to have ฮ˜ โข ( ๐›ฟ ) support on ๐‘ฅ ๐ท . This modified class is clearly learnable, since after drawing ๐‘‚ โข ( 1 / ๐›ฟ ) samples, the learner will draw ๐‘ฅ ๐ท and identify the distribution ๐ท with good probability. However, the class is not learnable under an adversary who removes points, since with high probability the adversary can completely remove any mention of ๐‘ฅ ๐ท from the learnerโ€™s sample, reducing to the original unlearnable class ( ๐’Ÿ , ๐‘‹ , ๐ป ) .

An adversary who can add samples is similarly powerful. Consider a model in which the adversary, after choosing a marginal distribution and labeling, sees the learnerโ€™s sample ahead of time and may add an additional ๐›พ fraction of correctly labeled elements.15 While the realizable setting is largely unaffected by such an adversary (as any non-optimal hypothesis will still see some bad example), even near-trivial concept classes cannot be agnostically learned.

Proposition 6.15.

There exists a class ( ๐‘‹ , ๐ป ) which for any ๐›พ

0 is realizably but not agnostically learnable under an adversary who can add a ๐›พ fraction of correctly labeled points to the learnerโ€™s sample.

Proof 6.16.

Let ๐‘‹

{ ๐‘ฅ , ๐‘ฅ 1 , ๐‘ฅ 2 } and ๐ป

{ โ„Ž 1 , โ„Ž 2 } be any class such that โ„Ž 1 โข ( ๐‘ฅ )

โ„Ž 2 โข ( ๐‘ฅ ) , but โ„Ž 1 โข ( ๐‘ฅ ๐‘– ) โ‰  โ„Ž 2 โข ( ๐‘ฅ ๐‘– ) for ๐‘–

1 , 2 . In the realizable setting, note that a single labeled example on ๐‘ฅ 1 or ๐‘ฅ 2 exactly determines the hypothesis. As long as there is less than 1 โˆ’ ๐œ€ mass on ๐‘ฅ , the learner will draw such a sample after ๐‘‚ โข ( 1 / ๐œ€ ) samples with good probability. Further, if the mass on ๐‘ฅ is at least 1 โˆ’ ๐œ€ , then โ„Ž 1 and โ„Ž 2 are both valid outputs. As a result, any ERM is a valid PAC-learner. Since adding correctly labeled examples can only help this learner, the class remains realizably learnable under an adversary who can add an arbitrary number of clean samples.

In the agnostic setting, consider an adversary who chooses a labeling ๐‘“ such that ๐‘“ โข ( ๐‘ฅ )

โ„Ž 1 โข ( ๐‘ฅ )

โ„Ž 2 โข ( ๐‘ฅ ) , ๐‘“ โข ( ๐‘ฅ 1 )

โ„Ž 1 โข ( ๐‘ฅ 1 ) , and ๐‘“ โข ( ๐‘ฅ 2 )

โ„Ž 2 โข ( ๐‘ฅ 2 ) . The optimal hypothesis โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ is then decided by the amount of mass on ๐‘ฅ 1 and ๐‘ฅ 2 in the marginal distribution. Namely if the adversary chooses a distribution ๐ท over { ๐‘ฅ , ๐‘ฅ 1 , ๐‘ฅ 2 } , the optimal error is min โก { ๐ท โข ( ๐‘ฅ 1 ) , ๐ท โข ( ๐‘ฅ 2 ) } . The idea is then to note that for ๐›พ โ€ฒ โ‰ค ๐›พ / 4 , the learner cannot distinguish between the two following distributions:

๐ท 1 โข ( ๐‘ฅ 1 )

๐‘ 1 โข ๐›พ โ€ฒ , ๐ท 1 โข ( ๐‘ฅ 2 )

( 1 โˆ’ ๐‘ 1 ) โข ๐›พ โ€ฒ

and

๐ท 2 โข ( ๐‘ฅ 1 )

( 1 โˆ’ ๐‘ 1 ) โข ๐›พ โ€ฒ , ๐ท 2 โข ( ๐‘ฅ 2 )

๐‘ 1 โข ๐›พ โ€ฒ

where 1 / 2

๐‘ 1

0 is some small constant. Informally, if the two distributions are indistinguishable, any learner will always incur error of around ( 1 โˆ’ ๐‘ 1 ) โข ๐›พ โ€ฒ 2 , whereas OPT is ๐‘ 1 โข ๐›พ โ€ฒ for both distributions.

Letโ€™s now give the formal argument. By Yaoโ€™s Minimax Principle it is enough to prove there is a strategy over distributions such that any deterministic learner has high error. In particular, if we can prove that the expected error is at least 3 โ‹… ๐‘‚ โข ๐‘ƒ โข ๐‘‡ , then Pr โก [ ๐‘’๐‘Ÿ๐‘Ÿ๐‘œ๐‘Ÿ โ‰ฅ 2 โ‹… ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ] โ‰ฅ ๐‘‚ โข ๐‘ƒ โข ๐‘‡ . Since OPT is just some constant ๐‘ 1 โข ๐›พ โ€ฒ (dependent only on ๐›พ ), this is sufficient to prove the result. Moving on, consider the strategy in which the adversary chooses the labeling described above, and chooses each marginal ( ๐ท 1 or ๐ท 2 ) with probability 1 / 2 . Weโ€™ll break our analysis into two cases dependent on the sample complexity of the learner. If the learner uses ๐‘‚ โข ( 1 / ๐›พ โ€ฒ ) examples, then there is a constant probability of drawing a sample only consisting of the point ๐‘ฅ . Let ๐‘“ โ€ฒ be the hypothesis returned by the deterministic learner on this sample. By construction, ๐‘“ โ€ฒ must disagree with either โ„Ž 1 or โ„Ž 2 on ๐‘ฅ 1 or ๐‘ฅ 2 . Assume ๐‘“ โ€ฒ differs on โ„Ž 1 โข ( ๐‘ฅ 1 ) (the other cases will follow similarly). When the distribution is ๐ท 2 , ๐‘“ โ€ฒ has error at least ( 1 โˆ’ ๐‘ 1 ) โข ๐›พ โ€ฒ . Since this occurs with constant probability independent of the choice of ๐‘ 1 , choosing ๐‘ 1 sufficiently small leads to an expected error of at least 3 โข ๐‘ 1 โข ๐›พ โ€ฒ as desired.

On the other hand, when there are ๐‘›

ฮฉ โข ( 1 / ๐›พ โ€ฒ ) samples, we claim that the adversary can force the following sample to occur with constant probability: 2 โข ๐›พ โ€ฒ โข ๐‘› instances of ๐‘ฅ 1 and ๐‘ฅ 2 , and ( 1 + ๐›พ ) โข ๐‘› โˆ’ 4 โข ๐›พ โ€ฒ โข ๐‘› instances of ๐‘ฅ . This follows from the fact that for the appropriate choice of constant for ๐‘› , a Chernoff bound gives that both ๐‘ฅ 1 and ๐‘ฅ 2 occur at most 2 โข ๐›พ โ€ฒ โข ๐‘› times with constant probability. Since the adversary is allowed to add ๐›พ โข ๐‘› โ‰ฅ 4 โข ๐›พ โ€ฒ โข ๐‘› arbitrary examples, they can add instances of ๐‘ฅ 1 , ๐‘ฅ 2 , and ๐‘ฅ until the above sample is achieved. The remainder of the argument is then the same as the previous case, as any learner response on this sample will incur similarly high expected error.

It is also reasonable to consider distributional corruption in the semi-supervised setting, where the unlabeled and labeled data-sets might have different underlying distributions. We discuss this model in Section 6.4.

6.3Replacing ERM: Semi-Private Learning

So far we have focused on property generalization for two forms of noise-toleranceโ€”agnostic learning and learning with malicious noise. In this section, weโ€™ll show how to use Algorithm 1 to generalize a broader spectrum of finitely-satisfiable properties through replacing the ERM process with a generic finite learner with the desired property. Our prototypical example will be privacy, which is well known to be finitely-satisfiable via McSherry and Talwarโ€™s [49] exponential mechanism. To start, weโ€™ll cover a few basic privacy definitions.

Definition 6.17 (Differential Privacy).

A learning algorithm ๐’œ is said to be ๐›ผ -differentially private if for all neighboring inputs ๐‘† , ๐‘† โ€ฒ which differ on a single example:

Pr โก [ ๐’œ โข ( ๐‘† ) โˆˆ ๐‘‡ ] โ‰ค ๐‘’ ๐›ผ โข Pr โก [ ๐’œ โข ( ๐‘† โ€ฒ ) โˆˆ ๐‘‡ ] ,

for all measurable events ๐‘‡ in the range of ๐’œ .

The exponential mechanism is one of the most widely used techniques in privacy. Informally, the algorithm allows for differentially private selection of a โ€œgoodโ€ choice from a finite set of objects (potential hypotheses in our case). More formally, let ๐‘  : ( ๐‘‹ ร— ๐‘Œ ) * ร— ๐ป โ†’ โ„ be a โ€œscoreโ€ function, and define โ€œsensitivityโ€ ฮ” ๐‘  to be

ฮ” ๐‘ 

max โ„Ž โˆˆ ๐ป โก max ๐‘† , ๐‘† โ€ฒ โก | ๐‘  โข ( ๐‘† , โ„Ž ) โˆ’ ๐‘  โข ( ๐‘† โ€ฒ , โ„Ž ) |

where ๐‘† , ๐‘† โ€ฒ are two neighbouring datasets. The exponential mechanism selects an item with a good score with high probability, while maintaining privacy.

Definition 6.18 (Exponential Mechanism [49]).

The exponential mechanism ๐‘€ ๐ธ on inputs ๐‘† , ๐ป , ๐‘  with privacy parameter ๐›ผ selects and outputs โ„Ž โˆˆ ๐ป with probability

exp โก ( ๐›ผ โข ๐‘  โข ( ๐‘† , โ„Ž ) 2 โข ฮ” ) โˆ‘ โ„Ž โ€ฒ โˆˆ ๐ป exp โก ( ๐›ผ โข ๐‘  โข ( ๐‘† , โ„Ž โ€ฒ ) 2 โข ฮ” ) .

It is well known that the exponential mechanism leads to a private learner for finite hypothesis classes under bounded loss.

Theorem 6.19 (Theorem 3.4 [39]).

Let ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) be a finite class with a bounded loss function. Then the sample complexity of ๐›ผ -differentially private learning ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) is at most:

๐‘› ๐‘ โข ๐‘Ÿ โข ๐‘– โข ( ๐›ผ , ๐œ€ , ๐›ฟ ) โ‰ค ๐‘‚ โข ( log โก ( | ๐ป | ๐›ฟ ) โข max โก { ๐œ€ โˆ’ 2 , ๐œ€ โˆ’ 1 โข ๐›ผ โˆ’ 1 } ) .

We note that [39, Theorem 3.4] only considers classification loss, but the extension to bounded loss is immediate. Unfortunately, even with the power of the exponential mechanism, privacy is a very restrictive condition in the general PAC framework, since weโ€™re most often interested in infinite hypothesis sets. Indeed even improper private learning requires finiteness of a highly restrictive measure known as representation dimension [12], which can be infinite for classes of VC dimension 1 . As a result, the past decade has seen the introduction of a number of weaker, more practical definitions of privacy. In this section weโ€™ll focus on a model introduced in 2013 by Beimel, Nissim, and Stemmer [14] called semi-private learning.

Definition 6.20 (Semi-Private Learning).

We call a class ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) semi-private PAC-Learnable if there exists an algorithm ๐’œ and two functions ๐‘› ๐‘๐‘ข๐‘

๐‘› ๐‘๐‘ข๐‘ โข ( ๐œ€ , ๐›ฟ , ๐›ผ ) and ๐‘› ๐‘๐‘Ÿ๐‘–

๐‘› ๐‘๐‘Ÿ๐‘– โข ( ๐œ€ , ๐›ฟ , ๐›ผ ) such that for all ๐œ€ , ๐›ฟ

0 and distributions ๐ท over ๐‘‹ ร— ๐‘Œ whose marginal ๐ท ๐‘‹ is in ๐’Ÿ , ๐’œ satisfies the following:

1.

๐’œ outputs a good hypothesis with high probability:

Pr ๐‘† ๐‘ˆ โˆผ ๐ท ๐‘‹ ๐‘› ๐‘๐‘ข๐‘ , ๐‘† ๐ฟ โˆผ ๐ท ๐‘› ๐‘๐‘Ÿ๐‘– โก [ ๐‘’๐‘Ÿ๐‘Ÿ ๐ท , โ„“ โข ( ๐’œ โข ( ๐‘† ๐‘ˆ , ๐‘† ๐ฟ ) )

๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ ] โ‰ค ๐›ฟ .

2.

๐’œ is semi-private. That is for all ๐‘† ๐‘ˆ โˆˆ ๐‘‹ ๐‘› ๐‘๐‘ข๐‘ :

๐’œ โข ( ๐‘† ๐‘ˆ , โ‹… ) โข  is  โข ๐›ผ โข -differentially private .

In other words, semi-private learning offers a model for applications where labeled data is sensitive, but some (perhaps opt-in) users might not care about their participation itself being released. Unlike standard private learning, distribution-free semi-private classification is known to be characterized by VC dimension, just like realizable PAC-learning [14]. The best sample complexity bounds are due to Alon, Bassily, and Moran (ABM) [2], who use uniform convergence to build a uniform cover for ๐ป from unlabeled data, and then apply the exponential mechanism to the resulting cover.

Due to their reliance on uniform convergence, ABMโ€™s techniques fail in the more general settings we consider. Further, their use of uniform covers results in sub-optimal public sample complexity even for distribution-free classification. We prove in Section 13 that these objects require asymptotically more samples than non-uniform covers (at least in the distribution-family model), and therefore cannot be used to achieve optimal semi-private learning. We circumvent both of these issues by appealing directly to a realizable learner to build a weaker non-uniform cover. For readability, we first restate the algorithm here.

Input: Realizable PAC-Learner ๐’œ , Unlabeled Sample Oracle ๐’ช ๐‘ˆ , Labeled Sample Oracle ๐’ช ๐ฟ Algorithm:

1.

Draw an unlabeled sample ๐‘† ๐‘ˆ โˆผ ๐’ช ๐‘ˆ , and labeled sample ๐‘† ๐ฟ โˆผ ๐’ช ๐ฟ .

2.

Run LearningToCover over ๐‘† ๐‘ˆ to get ๐ถ โข ( ๐‘† ๐‘ˆ ) .

3.

Return the hypothesis in ๐ถ โข ( ๐‘† ๐‘ˆ ) given by applying the exponential mechanism with respect to ๐‘† ๐ฟ .

Algorithm 4 Semi-Private to Realizable Reduction

We prove that Algorithm 4 gives a semi-private agnostic learner in the distribution-family setting.

Theorem 6.21 (PAC-learning implies Semi-Private Learning).

Let โ„“ : ๐‘Œ ร— ๐‘Œ โ†’ โ„ โ‰ฅ 0 be a bounded ๐‘ -approximate pseudometric. Then the following are equivalent for all triples ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) :

1.

( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) is discretely-learnable

2.

( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) is ๐‘ -agnostically, semi-private learnable.

Proof 6.22.

The proof is essentially the same as Theorem 6.6. The only difference in the argument is to replace the generic ERM learner over the output of LearningToCover with the exponential mechanism [39].

Letโ€™s now take a look at what Theorem 6.21 implies about the special case of distribution-free classification.

Corollary 6.23.

Let ( ๐‘‹ , ๐ป ) be a class of VC-dimension ๐‘‘ with sample complexity ๐‘› โข ( ๐œ€ , ๐›ฟ ) . The sample complexity of (agnostic) semi-private learning ( ๐‘‹ , ๐ป ) is at most:

๐‘› ๐‘๐‘ข๐‘ โข ( ๐œ€ , ๐›ฟ , ๐›ผ ) โ‰ค ๐‘› โข ( ๐œ€ / 2 , ๐›ฟ / 2 )

and

๐‘› ๐‘๐‘Ÿ๐‘– โข ( ๐œ€ , ๐›ฟ , ๐›ผ ) โ‰ค ๐‘‚ โข ( ( ๐‘‘ โข log โก ( ๐‘› โข ( ๐œ€ / 2 , ๐›ฟ / 2 ) ๐‘‘ ) + log โก ( 1 ๐›ฟ ) ) โข max โก { ๐œ€ โˆ’ 2 , ๐œ€ โˆ’ 1 โข ๐›ผ โˆ’ 1 } ) .

Further, the sample complexity of improperly (agnostic) semi-private learning ( ๐‘‹ , ๐ป ) is

๐‘› ๐‘๐‘ข๐‘ โข ( ๐œ€ , ๐›ฟ , ๐›ผ ) โ‰ค ๐‘‚ โข ( ๐‘‘ + log โก ( 1 / ๐›ฟ ) ๐œ€ )

and

๐‘› ๐‘๐‘Ÿ๐‘– โข ( ๐œ€ , ๐›ฟ , ๐›ผ ) โ‰ค ๐‘‚ โข ( ( ๐‘‘ โข log โก ( 1 ๐œ€ ) + log โก ( 1 ๐›ฟ ) ) โข max โก { ๐œ€ โˆ’ 2 , ๐œ€ โˆ’ 1 โข ๐›ผ โˆ’ 1 } ) .

Proof 6.24.

LearningToCover uses ๐‘› โข ( ๐œ€ / 2 , ๐›ฟ / 2 ) samples by definition. In the improper case, Hanneke showed that ๐‘› โข ( ๐œ€ / 2 , ๐›ฟ / 2 ) โ‰ค ๐‘‚ โข ( ๐‘‘ + log โก ( 1 / ๐›ฟ ) ๐œ€ ) . Since the class has VC dimension ๐‘‘ , the size of the resulting cover is at most ( ๐‘’ โ‹… ๐‘› โข ( ๐œ€ / 2 , ๐›ฟ / 2 ) / ๐‘‘ ) ๐‘‘ , and the private sample complexity bound then follows from Theorem 6.19.

This improves over the recent upper bound of ABM who showed that

๐‘› pub โข ( ๐œ€ , ๐›ฟ , ๐›ผ ) โ‰ค ๐‘‚ โข ( ๐‘‘ โข log โก ( 1 / ๐œ€ ) + log โก ( 1 / ๐›ฟ ) ๐œ€ ) .

In fact, for constant ๐‘‘ and ๐›ฟ , Corollary 6.23 completely resolves the unlabeled sample complexity of semi-private learning, as ABM [2] prove a matching lower bound.

Theorem 6.25 (Public Lower Bound [2]).

Every class that is not privately learnable requires at least ฮฉ โข ( 1 / ๐œ€ ) unlabeled samples to learn in the semi-private model under classification error.

On the other hand, we note that the private sample complexity remains off by a log factor from the best known lower bounds of Chaudhari and Hsu [22].

Theorem 6.26 (Private Lower Bound [22]).

There exist classes of VC dimension ๐‘‚ โข ( 1 ) which require at least:

๐‘› ๐‘๐‘Ÿ๐‘– โข ( ๐œ€ , ๐›ฟ , ๐›ผ ) โ‰ฅ ฮฉ โข ( max โก { ๐œ€ โˆ’ 2 , ๐œ€ โˆ’ 1 โข ๐›ผ โˆ’ 1 } )

private samples to learn.

While we have now resolved the public sample complexity of improper learning, it remains an interesting open problem for the proper regime where certain adversarial examples are known to require an extra log โก ( 1 / ๐œ€ ) factor in the standard PAC sample complexity [26, 34]. We conjecture that Theorem 6.21 should still be tight in this setting: namely that the unlabeled semi-private sample complexity should always be at least the realizable PAC sample complexity.

Conjecture 6.27.

Let ( ๐‘‹ , ๐ป ) be a hypothesis class which is not privately learnable. Then the realizable sample complexity of ( ๐‘‹ , ๐ป ) lower bounds the unlabeled sample complexity of proper semi-private learning:

๐‘› ๐‘๐‘ข๐‘ โข ( ๐œ€ , 1 / 2 ) โ‰ฅ ฮฉ โข ( ๐‘› โข ( ๐œ€ , 1 / 2 ) ) .

6.4Changing the Base Model: Covariate Shift

One issue with semi-supervised models like semi-private learning is that, in practice, the distribution over unlabeled data probably wonโ€™t match the labeled data exactly. In this section, weโ€™ll talk about a final modification to our reduction to tackle such scenarios and more generally to extend property generalization beyond the realizable PAC setting: replacing the base learner. In fact, we already saw this strategy used to a lesser extent in Section 6.1, where we replaced our standard realizable base learner with a discrete learner. Here weโ€™ll look at an application in which we assume our initial learner is robust to covariate shift [56], meaning that even if the distribution underlying the data shifts between train and test time, the algorithm will continue to perform well. This stronger assumption will allow us to build semi-private learners that can handle corruption between the public and private databases. To start, letโ€™s formalize covariate shift in the distribution-family model.

Definition 6.28 (Covariate Shift).

Let ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) be any class, and for every ๐œ€

0 let ๐ถ ๐œ€ be a โ€œcovariate-shiftโ€ function that maps every ๐ท โˆˆ ๐’Ÿ to some family of distributions over ๐‘‹ . Given any distribution ๐ท โˆˆ ๐’Ÿ and any โ„Ž โˆˆ ๐ป , let the error of a potential labeling be given by its worst-case error over ๐ถ ๐œ€ โข ( ๐ท ) , that is:

CS-err ๐ท ร— โ„Ž , โ„“ , ๐œ€ โข ( ๐‘ )

max ๐ท โ€ฒ โˆˆ ๐ถ ๐œ€ โข ( ๐ท ) โก { ๐”ผ ๐‘ฅ โˆผ ๐ท โ€ฒ โข [ โ„“ โข ( ๐‘ โข ( ๐‘ฅ ) , โ„Ž โข ( ๐‘ฅ ) ) ] } .

We say that ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) is realizably learnable under covariate shift ๐ถ

{ ๐ถ ๐œ€ } if there exists an algorithm ๐’œ and function ๐‘›

๐‘› โข ( ๐œ€ , ๐›ฟ ) such that for all ๐œ€ , ๐›ฟ

0 , ๐ท โˆˆ ๐’Ÿ , and โ„Ž โˆˆ ๐ป :

Pr ๐‘† โˆผ ๐ท ๐‘› โก [ CS-err ๐ท ร— โ„Ž , โ„“ , ๐œ€ โข ( ๐’œ โข ( ๐‘† , โ„Ž โข ( ๐‘† ) ) )

๐œ€ ] โ‰ค ๐›ฟ .

We call such a learner robust to covariate shift.

We emphasize that in the above definition, the covariate shift family scales with the error parameter ๐œ€ . This is a bit different than Shimodairaโ€™s original definition [56], but is a natural choice in our context since we consider algorithms which only use access to the original source distribution (sometimes called โ€œconservative domain adaptationโ€ [30]). In this setting, weโ€™d expect that as we demand higher accuracy, the amount of covariate shift we can tolerate will decrease. Indeed in the agnostic model, itโ€™s clear this scaling is necessary by a similar argument to Proposition 6.15.

The key observation to apply learning under covariate shift in our reduction is simply to notice that the non-uniform cover output by LearningtoCover must contain a hypothesis close to optimal under any shifted distribution in ๐ถ ๐œ€ โข ( ๐ท ) . This can then be used to analyze any semi-supervised model where the marginal of the labeled distribution may be corrupted from ๐ท to any distribution in ๐ถ ๐œ€ โข ( ๐ท ) . In this section, weโ€™ll again focus on the setting of semi-private learning. First, letโ€™s formalize what it means to be semi-private learnable under covariate shift.

Definition 6.29 (Semi-Private Learning under Covariate Shift).

We call a class ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) semi-private (agnostically) PAC-Learnable under covariate shift ๐ถ

{ ๐ถ ๐œ€ } if there exists an algorithm ๐’œ and two functions ๐‘› ๐‘๐‘ข๐‘

๐‘› ๐‘๐‘ข๐‘ โข ( ๐œ€ , ๐›ฟ , ๐›ผ ) and ๐‘› ๐‘๐‘Ÿ๐‘–

๐‘› ๐‘๐‘Ÿ๐‘– โข ( ๐œ€ , ๐›ฟ , ๐›ผ ) such that for all ๐œ€ , ๐›ฟ

0 and distributions ๐ท ๐‘‹ โˆˆ ๐’Ÿ and ๐ท โ€ฒ over ๐‘‹ ร— ๐‘Œ whose marginal ๐ท ๐‘‹ โ€ฒ โˆˆ ๐ถ ๐œ€ โข ( ๐ท ๐‘‹ ) , ๐’œ satisfies the following:

1.

๐’œ outputs a good hypothesis over ๐ท โ€ฒ with high probability:

Pr ๐‘† ๐‘ˆ โˆผ ๐ท ๐‘‹ ๐‘› ๐‘๐‘ข๐‘ , ๐‘† ๐‘ โˆผ ๐ท โ€ฒ โฃ ๐‘› ๐‘๐‘Ÿ๐‘– โก [ ๐‘’๐‘Ÿ๐‘Ÿ ๐ท โ€ฒ , โ„“ โข ( ๐’œ โข ( ๐‘† ๐‘ˆ , ๐‘† ๐‘ ) )

๐‘‚ โข ๐‘ƒ โข ๐‘‡ โ€ฒ + ๐œ€ ] โ‰ค ๐›ฟ ,

where ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โ€ฒ

min โ„Ž โˆˆ ๐ป โก [ ๐‘’๐‘Ÿ๐‘Ÿ ๐ท โ€ฒ , โ„“ โข ( โ„Ž ) ] is the minimum error over the shifted distribution.

2.

๐’œ is semi-private. That is for all ๐‘† ๐‘ˆ โˆˆ ๐‘‹ ๐‘› ๐‘๐‘ข๐‘ :

๐’œ โข ( ๐‘† ๐‘ˆ , โ‹… ) โข  is  โข ๐›ผ โข -differentially private

In other words, weโ€™d like to recover a near-optimal hypothesis even when the marginal distribution over private data is shifted from the public data. This is a realistic scenario in practice, since the distribution of โ€œopt-inโ€ users is likely different from the marginal over the total population. Weโ€™ll show that this issue is solvable in the semi-private setting as long as the analogous issue in the non-private setting (distribution shift between train and test time) can be resolved.

Theorem 6.30.

Let ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) be any class of a finite label space with loss function satisfying the identity of indiscernibles which is realizably learnable under covariate shift ๐ถ

{ ๐ถ ๐œ€ } . Then ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) is semi-private, agnostically learnable under covariate shift ๐ถ with sample complexity:

๐‘š โข ( ๐œ€ , ๐›ฟ ) โ‰ค ๐‘› โข ( ๐œ‚ โ„“ โข ๐œ€ , ๐›ฟ / 2 ) + ๐‘‚ โข ( log โก ( ฮ  ๐ป โข ( ๐‘› โข ( ๐œ‚ โ„“ โข ๐œ€ , ๐›ฟ / 2 ) ) ๐›ฟ ) ๐œ€ 2 )

where we recall ๐œ‚ โ„“ โ‰ฅ ฮฉ โข ( min ๐‘Ž โ‰  ๐‘ โก ( โ„“ โข ( ๐‘Ž , ๐‘ ) ) max ๐‘Ž โ‰  ๐‘ โก ( โ„“ โข ( ๐‘Ž , ๐‘ ) ) ) is a constant depending only on โ„“ .

Proof 6.31.

The proof is essentially the same as Theorem 5.4. The only difference is to note that for all marginal distributions ๐ท ๐‘‹ โˆˆ ๐’Ÿ and choices of shift ๐ท ๐‘‹ โ€ฒ โˆˆ ๐ถ ๐œ€ โข ( ๐ท ๐‘‹ ) , the output of LearningToCover has some โ„Ž โ€ฒ satisfying

๐”ผ ๐‘ฅ โˆผ ๐ท ๐‘‹ โ€ฒ โข [ โ„“ โข ( โ„Ž โ€ฒ โข ( ๐‘ฅ ) , โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โ€ฒ โข ( ๐‘ฅ ) ) ] โ‰ค ๐œ‚ โ„“ โข ๐œ€ ,

where โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โ€ฒ is some optimal hypothesis over the shifted distribution ๐ท โ€ฒ . As in the standard analysis, this is guaranteed by including the output of the realizable learner on โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โ€ฒ (robustness to covariate shift promises this output is close under ๐ท ๐‘‹ โ€ฒ with high probability). The remainder of the argument is exactly as in Theorem 6.6, with the exception of working over the shifted distribution ๐ท ๐‘‹ โ€ฒ instead of ๐ท ๐‘‹ .

We note that this result can also be extended to the more general loss functions discussed in Section 6.1 without too much difficulty with the appropriate definition of discrete learnability under covariate shift.

Since Theorem 6.30 is a bit abstract, letโ€™s take a look at one concrete application. Given a class ( ๐‘‹ , ๐ป ) , the class-dependent total variation distance is a metric on distributions measuring the worst case distance across elements of ๐ป โข ฮ” โข ๐ป โข \coloneqq โข { โ„Ž โข ฮ” โข โ„Ž โ€ฒ : โ„Ž , โ„Ž โ€ฒ โˆˆ ๐ป } :

๐‘‡ โข ๐‘‰ ๐ป โข ฮ” โข ๐ป โข ( ๐ท , ๐ท โ€ฒ ) โข \coloneqq โข max โ„Ž โˆˆ ๐ป โข ฮ” โข ๐ป โก { | ๐ท โข ( โ„Ž ) โˆ’ ๐ท โ€ฒ โข ( โ„Ž ) | } .

It is not hard to see that any realizable learner is robust to ๐‘‚ โข ( ๐œ€ ) covariate shift in ๐‘‡ โข ๐‘‰ ๐ป โข ฮ” โข ๐ป distance. We can then apply Theorem 6.30 to build a robust semi-private learner.

Corollary 6.32.

Let ( ๐‘‹ , ๐ป ) be a class of VC-dimension ๐‘‘ , and for every ๐œ€

0 and distribution ๐ท over ๐‘‹ , define a coviarate shift function:

๐ถ ๐œ€ โข ( ๐ท ) โข \coloneqq โข { ๐ท โ€ฒ : ๐‘‡ โข ๐‘‰ ๐ป โข ฮ” โข ๐ป โข ( ๐ท , ๐ท โ€ฒ ) โ‰ค ๐œ€ / 2 } .

Then ( ๐‘‹ , ๐ป ) is semi-private learnable under covariate shift ๐ถ

{ ๐ถ ๐œ€ } in only

๐‘› ๐‘๐‘ข๐‘ โข ( ๐œ€ , ๐›ฟ , ๐›ผ ) โ‰ค ๐‘‚ โข ( ๐‘‘ + log โก ( 1 / ๐›ฟ ) ๐œ€ )

unlabeled samples and

๐‘› ๐‘๐‘Ÿ๐‘– โข ( ๐œ€ , ๐›ฟ , ๐›ผ ) โ‰ค ๐‘‚ โข ( ( ๐‘‘ โข log โก ( 1 ๐œ€ ) + log โก ( 1 ๐›ฟ ) ) โข max โก { ๐œ€ โˆ’ 2 , ๐œ€ โˆ’ 1 โข ๐›ผ โˆ’ 1 } )

labeled samples.

Finally, we note again that our original learner in these results is robust to covariate shift despite having no access to samples from the new distribution. Unfortunately, this model does come with fairly strong lower bounds regarding the type of covariate shifts to which it is possible to be robust [30]. One solution to this problem is to consider a relaxed variant called (non-conservative) domain adaptation, where the learner additionally has access to a small number of unlabeled samples from the test-time distribution. It is certainly possible to define an analog in the semi-private setting, but naively the use of unlabeled data from the private distribution breaks our reduction since privacy wonโ€™t be preserved. We leave as an open question whether any sort of PAC-learner in the non-conservative model could imply semi-private learners with stronger robustness to covariate shift. Some progress has been made in this direction recently by Bassily, Moran, and Nandi [11] for distribution-free classification of halfspaces.

7Doubly Bounded Loss

In this section we discuss a natural generalization of loss functions over finite label classes we call doubly bounded loss: for all distinct ๐‘ฆ , ๐‘ฆ โ€ฒ โˆˆ ๐‘Œ , we require โ„“ โข ( ๐‘ฆ , ๐‘ฆ โ€ฒ ) โˆˆ [ ๐‘Ž , ๐‘ ] for some ๐‘ โ‰ฅ ๐‘Ž

0 . This is trivially satisfied by any loss function on a finite label class that satisfies the identity of indiscernibles.

Definition 7.1 (Doubly Bounded Loss).

We say โ„“ : ๐‘Œ โ†’ ๐‘Œ is ( ๐‘Ž , ๐‘ ) -bounded if there exist ๐‘ โ‰ฅ ๐‘Ž

0 for any distinct ๐‘ฆ , ๐‘ฆ โ€ฒ โˆˆ ๐‘Œ :

โ„“ โข ( ๐‘ฆ , ๐‘ฆ โ€ฒ ) โˆˆ [ ๐‘Ž , ๐‘ ] .

As discussed in Section 6.1, since we now allow ๐‘Œ to be infinite, we need to work with discrete learnability instead of realizable learnability. We can use a slight modification to the discretization technique in Theorem 6.6 to prove the equivalence of discrete and agnostic learnability for doubly-bounded loss functions. Note that this is stronger than our guarantee for ๐‘ -approximate pseudometrics, which only gives ๐‘ -agnostic learnability.

Theorem 7.2.

Let โ„“ : ๐‘Œ ร— ๐‘Œ โ†’ โ„ โ‰ฅ 0 be an ( ๐‘Ž , ๐‘ ) -bounded loss function. Then for any class ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) the following are equivalent:

1.

( ๐‘‹ , ๐ป , ๐’Ÿ , โ„“ ) is (properly) discretely-learnable

2.

( ๐‘‹ , ๐ป , ๐’Ÿ , โ„“ ) is (properly) agnostically learnable.

Proof 7.3.

The proof that agnostic learnability implies discrete learnability is the same as in Theorem 6.6, so we focus only the forward direction. Assume ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) is discretely-learnable. Fix ๐œ€ โ€ฒ

๐‘Ž โข ๐œ€ 4 โข ๐‘ , and let ๐ป ๐œ€ โ€ฒ be a learnable ๐œ€ โ€ฒ -discretization of ๐ป . We argue that running LearningToCover on ๐ป ๐œ€ โ€ฒ (using the promised discrete learner) gives the desired agnostic learner. As before, it is sufficient to prove that ๐ถ โข ( ๐‘† ๐‘ˆ ) contains a hypothesis โ„Ž โ€ฒ such that ๐‘’๐‘Ÿ๐‘Ÿ ๐ท , โ„“ โข ( โ„Ž โ€ฒ ) โ‰ค ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ / 2 . Since โ„“ is upper bounded, standard empirical risk minimization arguments then give the desired result.

To see why ๐ถ โข ( ๐‘† ๐‘ˆ ) has this property, recall from Lemma 5.2 that for any โ„Ž โˆˆ ๐ป ๐œ€ โ€ฒ , with probability at least 1 โˆ’ ๐›ฟ / 2 there exists โ„Ž โ€ฒ โˆˆ ๐ถ โข ( ๐‘† ๐‘ˆ ) that is ๐œ€ โ€ฒ -close to โ„Ž in the following sense:

๐”ผ ๐‘ฅ โˆผ ๐ท ๐‘‹ โข [ โ„“ โข ( โ„Ž โ€ฒ โข ( ๐‘ฅ ) , โ„Ž โข ( ๐‘ฅ ) ) ] โ‰ค ๐‘Ž โข ๐œ€ 4 โข ๐‘ .

Because โ„“ is ๐‘Ž -lower bounded, this implies that โ„Ž โ€ฒ must be close to โ„Ž in classification error:

Pr ๐‘ฅ โˆผ ๐ท ๐‘‹ โก [ โ„Ž โข ( ๐‘ฅ ) โ‰  โ„Ž โ€ฒ โข ( ๐‘ฅ ) ] โ‰ค ๐œ€ 4 โข ๐‘ ,

and since the loss is bounded by ๐‘ , the risk of โ„Ž โ€ฒ cannot be much more than of โ„Ž :

๐‘’๐‘Ÿ๐‘Ÿ ๐ท , โ„“ โข ( โ„Ž โ€ฒ ) โ‰ค ๐‘’๐‘Ÿ๐‘Ÿ ๐ท , โ„“ โข ( โ„Ž ) + ๐œ€ / 4 .

Let โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โˆˆ ๐ป be an optimal hypothesis. Since ๐ป ๐œ€ โ€ฒ

๐œ€ โ€ฒ -covers ๐ป , there exists โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ๐œ€ โ€ฒ โˆˆ ๐ป ๐œ€ โ€ฒ such that:

๐‘’๐‘Ÿ๐‘Ÿ ๐ท , โ„“ โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ๐œ€ โ€ฒ ) โˆ’ ๐‘’๐‘Ÿ๐‘Ÿ ๐ท , โ„“ โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ) โ‰ค ๐œ€ โ€ฒ .

Let โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โ€ฒ โฃ ๐œ€ โ€ฒ โˆˆ ๐ป ๐œ€ โ€ฒ denote the output of the base learner ๐’œ on the labeling given by โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ๐œ€ โ€ฒ . Then by the above, we have that:

๐‘’๐‘Ÿ๐‘Ÿ ๐ท , โ„“ โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โ€ฒ โฃ ๐œ€ โ€ฒ )

๐”ผ ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท โข [ โ„“ โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โ€ฒ โฃ ๐œ€ โ€ฒ โข ( ๐‘ฅ ) , ๐‘ฆ ) ]

โ‰ค ๐”ผ ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท โข [ โ„“ โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ๐œ€ โ€ฒ โข ( ๐‘ฅ ) , ๐‘ฆ ) ] + ๐œ€ / 4

โ‰ค ๐”ผ ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท โข [ โ„“ โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ๐œ€ โ€ฒ โข ( ๐‘ฅ ) , ๐‘ฆ ) โˆ’ โ„“ โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข ( ๐‘ฅ ) , ๐‘ฆ ) + โ„“ โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข ( ๐‘ฅ ) , ๐‘ฆ ) ] + ๐œ€ / 4

๐‘’๐‘Ÿ๐‘Ÿ ๐ท , โ„“ โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ๐œ€ โ€ฒ ) โˆ’ ๐‘’๐‘Ÿ๐‘Ÿ ๐ท , โ„“ โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ) + ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ / 4

โ‰ค ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ / 2

as desired.

It is worth noting that the upper bound on the loss can be removed if the adversary is restricted to choosing a marginal over ๐‘Œ which is weakly concentrated.

8Robust Learning

Robust learning is an extension to the PAC setting that models an adversary with the power to perturb examples at test time. In practice, this corresponds to the fact that weโ€™d like our predictors to be stable to small amounts of adversarial noiseโ€”this could range anywhere from a sticker on a stop-sign tricking a self-driving car, to completely imperceptible perturbations that totally fool standard classifiers. The latter was famously demonstrated by Athalye, Engstrom, Ilyas, and Kwok [7], who showed how to generate such perturbations and provided the classic example of tricking a standard ImageNet classifier into thinking a turtle was a gun. Their seminal work caused an explosion of both practical and theoretical research in the area.16

Formally, adversarial robustness is modeled simply by changing the error function to be the maximum error over some pre-defined set of neighboring perturbations.

Definition 8.1 (Robust Loss).

Let ๐‘‹ be an instance space and ๐’ฐ : ๐‘‹ โ†’ ๐‘ƒ โข ( ๐‘‹ ) a โ€œperturbation functionโ€ mapping elements to a set of possible perturbations. Given a loss function โ„“ : ๐‘Œ ร— ๐‘Œ โ†’ โ„ โ‰ฅ 0 , the robust loss of a concept โ„Ž : ๐‘‹ โ†’ ๐‘Œ with respect to a distribution ๐ท over ๐‘‹ ร— ๐‘Œ is:

R-err ๐’ฐ , ๐ท โข ( โ„Ž )

๐”ผ ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท โข [ max ๐‘ฅ โ€ฒ โˆˆ ๐’ฐ โข ( ๐‘ฅ ) โก โ„“ โข ( โ„Ž โข ( ๐‘ฅ โ€ฒ ) , ๐‘ฆ ) ] .

In other words, a hypothesis with low robust loss performs well even against an adversary who can perturb ๐‘ฅ to any โ€œnearby pointโ€ (i.e. any ๐‘ฅ โ€ฒ โˆˆ ๐’ฐ โข ( ๐‘ฅ ) ). Standard realizable and agnostic Robust PAC-learning are then simply defined by replacing the standard error function with the robust error function. Robust learning in the distribution-family model does require one extra twist: we need to make sure that each hypothesis in the class actually has a corresponding distribution over which it is realizable. To this end, we introduce a basic notion of closure for distribution families.

Definition 8.2 (Robust Closure).

Let ๐’Ÿ be a set of distributions over an instance space ๐‘‹ and ๐ป a concept class. Given any concept โ„Ž , let ๐‘‹ โ„Ž denote the set of points in ๐‘‹ on which โ„Ž has 0 robust loss with respect to itself, that is:

๐‘‹ โ„Ž โข \coloneqq โข { ๐‘ฅ โˆˆ ๐‘‹ : โˆ€ ๐‘ฅ โ€ฒ โˆˆ ๐’ฐ โข ( ๐‘ฅ ) , โ„“ โข ( โ„Ž โข ( ๐‘ฅ โ€ฒ ) , โ„Ž โข ( ๐‘ฅ ) )

0 } .

For notational simplicity, let ๐ท | โ„Ž denote the restriction ๐ท | ๐‘‹ โ„Ž . The robust closure of ๐’Ÿ under ๐ป is:

๐’Ÿ ๐ป โข \coloneqq โข ๐’Ÿ โˆช โ‹ƒ ๐ท โˆˆ ๐’Ÿ , โ„Ž โˆˆ ๐ป ๐ท | โ„Ž .

In the robust distribution-family model, it only really makes sense to define realizable learnability over the robust closure of ๐’Ÿ , since otherwise there may be hypotheses in the class that are not realizable with respect to any distribution in ๐’Ÿ and cannot be chosen by the adversary at all. With this in mind, letโ€™s formalize this model.

Definition 8.3 ((Realizable) Distribution-Family Robust PAC Learning).

A class ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) is Robustly PAC-learnable in the realizable setting with respect to perturbation function ๐’ฐ if there exists an algorithm ๐’œ and function ๐‘› โข ( ๐œ€ , ๐›ฟ ) such that for all ๐œ€ , ๐›ฟ

0 and distributions ๐ท over ๐‘‹ ร— ๐‘Œ satisfying:

1.

The marginal ๐ท ๐‘‹ โˆˆ ๐’Ÿ ๐ป ,

2.

min โ„Ž โˆˆ ๐ป โก R-err ๐’ฐ , ๐ท โข ( โ„Ž )

0 ,

then:

Pr ๐‘† โˆผ ๐ท ๐‘› โข ( ๐œ€ , ๐›ฟ ) โก [ R-err ๐’ฐ , ๐ท โข ( ๐’œ โข ( ๐‘† ) )

๐œ€ ] โ‰ค ๐›ฟ .

Agnostic learnability is defined similarly, but since the adversary is unrestricted, there is no longer any need to take the robust closure.

Definition 8.4 ((Agnostic) Distribution-Family Robust PAC Learning).

A class ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) is Robustly PAC-learnable in the agnostic setting with respect to perturbation function ๐’ฐ if there exists an algorithm ๐’œ and function ๐‘› โข ( ๐œ€ , ๐›ฟ ) such that for all ๐œ€ , ๐›ฟ

0 and distributions ๐ท over ๐‘‹ ร— ๐‘Œ satisfying ๐ท ๐‘‹ โˆˆ ๐’Ÿ , then:

Pr ๐‘† โˆผ ๐ท ๐‘› โข ( ๐œ€ , ๐›ฟ ) โก [ R-err ๐’ฐ , ๐ท โข ( ๐’œ โข ( ๐‘† ) )

๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ ] โ‰ค ๐›ฟ ,

where ๐‘‚ โข ๐‘ƒ โข ๐‘‡

min โ„Ž โˆˆ ๐ป โก { R-err ๐’ฐ , ๐ท โข ( โ„Ž ) } .

We note that different works consider different models of access to the perturbation set ๐’ฐ as well (e.g. assuming ๐’ฐ is known to the learner [52], or has some type of oracle access [51, 50]). Our reduction requires fairly weak access to ๐’ฐ โ€”it is enough to be able to estimate the empirical robust loss of a hypothesis โ„Ž over any finite sample ๐‘† โŠ‚ ๐‘‹ . With this in mind, letโ€™s now prove realizable and agnostic robust learning are equivalent in the distribution-family model. Weโ€™ll focus on the special case of (multi-class) classification, and start by re-stating our modified algorithm for simplicity of presentation.

Input: Realizable Robust PAC-Learner ๐’œ Algorithm:

1.

Draw an unlabeled sample ๐‘† ๐‘ˆ , and labeled sample ๐‘† ๐ฟ .

2.

Run ๐’œ over all possible subsets and labelings of ๐‘† ๐‘ˆ to get:

๐ถ โข ( ๐‘† ) โข \coloneqq โข { ๐’œ โข ( ๐‘† , โ„Ž โข ( ๐‘† ) ) | ๐‘† โІ ๐‘† ๐‘ˆ , โ„Ž โˆˆ ๐ป | ๐‘† } .

3.

Return the hypothesis in ๐ถ โข ( ๐‘† ) with lowest empirical robust error over ๐‘† ๐ฟ .

Algorithm 5 Agnostic to Realizable Reduction Robust Setting Theorem 8.5.

If ( ๐’Ÿ , ๐‘‹ , ๐ป ) is robustly PAC-learnable in the realizable setting with sample complexity ๐‘› โข ( ๐œ€ , ๐›ฟ ) , then Algorithm 5 robustly learns ( ๐’Ÿ , ๐‘‹ , ๐ป ) in the agnostic setting in:

๐‘š ๐‘ˆ โข ( ๐œ€ , ๐›ฟ ) โ‰ค ๐‘‚ โข ( max ๐œ‡ โˆˆ [ 0 , 1 โˆ’ ๐œ€ ] โก { ๐‘› โข ( ๐œ€ / ( 2 โข ( 1 โˆ’ ๐œ‡ ) ) , ๐›ฟ / 3 ) 1 โˆ’ ๐œ‡ } )

unlabeled samples, and

๐‘š ๐ฟ โข ( ๐œ€ , ๐›ฟ ) โ‰ค ๐‘‚ โข ( ๐‘š ๐‘ˆ โข ( ๐œ€ , ๐›ฟ ) + log โก ( 1 / ๐›ฟ ) ๐œ€ 2 )

labeled samples.

Proof 8.6.

The proof is similar to Theorem 5.4, but like malicious noise (Theorem 6.11), requires the use of subsampling. The key issue is that for a distribution ๐ท over ๐‘‹ ร— ๐‘Œ , the optimal hypothesis โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ may not be realizable with respect to ๐ท ๐‘‹ , that is we may have:

R-err ๐‘ˆ , ๐ท ๐‘‹ ร— โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ )

0 .

As a result, our realizable learner (and therefore LearningToCover) has no guarantees over this distribution. On the other hand, our realizable learner does have good guarantees over the restricted marginal ๐ท ๐‘‹ | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ . We can then fix the above issue by running LearningToCover of all subsets of ๐‘† including its restriction to ๐‘‹ โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ . We will see that this essentially simulates running the realizable learner on the realizable restriction ๐ท ๐‘‹ | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ and recovers the desired guarantees.

Letโ€™s take a look at this more formally. As in our previous arguments it is enough to prove that ๐ถ โข ( ๐‘† ) contains a hypothesis โ„Ž โ€ฒ with robust error at most ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ / 2 , since a standard Chernoff bound tells us that ๐‘‚ โข ( log โก ( | ๐ถ โข ( ๐‘† ) | / ๐›ฟ ) / ๐œ€ 2 ) labeled examples are enough to estimate the robust loss of every hypothesis in ๐ถ โข ( ๐‘† ) with high probability. We note that this is the only step in our reduction which requires access to the perturbation set ๐’ฐ .

It is left to show that ๐ถ โข ( ๐‘† ) satisfies this property. Let ๐ท | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ denote the restriction of ๐ท to ๐‘‹ โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ , the points in ๐‘‹ on which โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ has 0 robust loss with respect to itself. Let ๐ท ยฏ | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ be the restriction to the complement, that is ๐‘‹ โˆ– ๐‘‹ โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ . The idea is to decompose our analysis into two separate parts over ๐ท | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ and ๐ท ยฏ | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ . With this in mind, let ๐œ‡ * denote the mass of ๐ท ๐‘‹ on ๐‘‹ โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ , and let ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โ€ฒ denote the robust error of โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ over ๐ท | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ . Since we are restricting our attention to classification error, notice that we can decompose ๐‘‚ โข ๐‘ƒ โข ๐‘‡ as:

๐‘‚ โข ๐‘ƒ โข ๐‘‡

R-err ๐’ฐ , ๐ท โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ )

Pr ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท [ โˆƒ ๐‘ฅ โ€ฒ โˆˆ ๐’ฐ ( ๐‘ฅ ) : โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ( ๐‘ฅ โ€ฒ ) โ‰  ๐‘ฆ ]

๐œ‡ * Pr ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท ยฏ | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ [ โˆƒ ๐‘ฅ โ€ฒ โˆˆ ๐’ฐ ( ๐‘ฅ ) : โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ( ๐‘ฅ โ€ฒ ) โ‰  ๐‘ฆ ]

+ ( 1 โˆ’ ๐œ‡ * ) Pr ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ [ โˆƒ ๐‘ฅ โ€ฒ โˆˆ ๐’ฐ ( ๐‘ฅ ) : โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ( ๐‘ฅ โ€ฒ ) โ‰  ๐‘ฆ ]

๐œ‡ * + ( 1 โˆ’ ๐œ‡ * ) โข ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โ€ฒ ,

where the last step follows from noting that by definition for all ๐‘ฅ in the support of ๐ท ยฏ | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ , โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ is not constant on ๐’ฐ โข ( ๐‘ฅ ) . To get a function within ๐œ€ / 2 robust loss of ๐‘‚ โข ๐‘ƒ โข ๐‘‡ , we claim it is sufficient to prove ๐ถ โข ( ๐‘† ) contains some โ„Ž within robust error ๐œ€ / ( 2 โข ( 1 โˆ’ ๐œ‡ * ) ) of โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ over ๐ท | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ , that is some โ„Ž satisfying:

Pr ๐‘ฅ โˆผ ๐ท ๐‘‹ | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โก [ โˆƒ ๐‘ฅ โ€ฒ โˆˆ ๐’ฐ โข ( ๐‘ฅ ) : โ„Ž โข ( ๐‘ฅ โ€ฒ ) โ‰  โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข ( ๐‘ฅ โ€ฒ ) ] โ‰ค ๐œ€ / ( 2 โข ( 1 โˆ’ ๐œ‡ * ) ) .

(3)

This follows from a similar analysis to the above. Namely, letting ๐‘… โข ( โ„Ž , ๐‘ฅ , ๐‘ฆ ) denote the event

๐‘… โข ( โ„Ž , ๐‘ฅ , ๐‘ฆ ) โข \coloneqq โข โˆƒ ๐‘ฅ โ€ฒ โˆˆ ๐’ฐ โข ( ๐‘ฅ ) : โ„Ž โข ( ๐‘ฅ โ€ฒ ) โ‰  ๐‘ฆ

for notational simplicity, we can break down R-err ๐’ฐ , ๐ท โข ( โ„Ž ) as:

R-err ๐’ฐ , ๐ท โข ( โ„Ž )

๐œ‡ * โข Pr ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท ยฏ | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข [ ๐‘… โข ( โ„Ž , ๐‘ฅ , ๐‘ฆ ) ] + ( 1 โˆ’ ๐œ‡ * ) โข Pr ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข [ ๐‘… โข ( โ„Ž , ๐‘ฅ , ๐‘ฆ ) ]

โ‰ค ๐œ‡ * + ( 1 โˆ’ ๐œ‡ * ) โข Pr ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข [ ๐‘… โข ( โ„Ž , ๐‘ฅ , ๐‘ฆ ) ]

โ‰ค ๐œ‡ * + ( 1 โˆ’ ๐œ‡ * ) โข Pr ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข [ ๐‘… โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ , ๐‘ฅ , ๐‘ฆ ) ]

+ ( 1 โˆ’ ๐œ‡ * ) Pr ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ [ โˆƒ ๐‘ฅ โ€ฒ โˆˆ ๐’ฐ ( ๐‘ฅ ) : โ„Ž ( ๐‘ฅ โ€ฒ ) โ‰  โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ( ๐‘ฅ โ€ฒ ) ]

โ‰ค ๐œ‡ * + ( 1 โˆ’ ๐œ‡ * ) โข ( ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โ€ฒ + ๐œ€ 2 โข ( 1 โˆ’ ๐œ‡ * ) )

๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ / 2 .

It remains to prove that ๐ถ โข ( ๐‘† ) contains a hypothesis satisfying Equation 3 with high probability. To see this, note that by definition of realizable robust learning, on a labeled sample ( ๐‘† , โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข ( ๐‘† ) ) โˆผ ๐ท | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ร— โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ of size ๐‘› โข ( ๐œ€ / ( 2 โข ( 1 โˆ’ ๐œ‡ * ) ) , ๐›ฟ / 3 ) our learner will output โ„Ž satisfying:

Pr ๐‘ฅ โˆผ ๐ท ๐‘‹ | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โก [ โˆƒ ๐‘ฅ โ€ฒ โˆˆ ๐’ฐ โข ( ๐‘ฅ ) : โ„Ž โข ( ๐‘ฅ โ€ฒ ) โ‰  โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข ( ๐‘ฅ ) ] โ‰ค ๐œ€ / ( 2 โข ( 1 โˆ’ ๐œ‡ * ) )

with probability at least 1 โˆ’ ๐›ฟ / 3 . To get Equation 3, it is then enough to note that โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ is constant on ๐’ฐ โข ( ๐‘ฅ ) for all ๐‘ฅ in the support of ๐ท ๐‘‹ | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ by definition.

The idea is now to draw a large enough unlabeled sample such that with probability at least 1 โˆ’ ๐›ฟ / 3 , the restriction to ๐‘‹ โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ is at least this size. By a Chernoff bound, it is enough to draw ๐‘ 1 โข ๐‘› โข ( ๐œ€ / ( 1 โˆ’ ๐œ‡ * ) , ๐›ฟ / 3 ) 1 โˆ’ ๐œ‡ * points to achieve this for some large enough constant ๐‘ 1

0 .17 Since we do not know ๐œ‡ * , weโ€™ll need to draw ๐‘ 1 โข max ๐œ‡ โˆˆ [ 0 , 1 โˆ’ ๐œ€ ] โก { ๐‘› โข ( ๐œ€ / ( 1 โˆ’ ๐œ‡ ) , ๐›ฟ / 3 ) 1 โˆ’ ๐œ‡ } points to ensure this property holds (if ๐œ‡ * โ‰ฅ 1 โˆ’ ๐œ€ , note that any hypothesis gives a valid solution). By a union bound we have that this overall process succeeds with probability at least 1 โˆ’ 2 โข ๐›ฟ / 3 , and outputting the hypothesis in ๐ถ โข ( ๐‘† ) with the lowest empirical robust risk then succeeds with probability 1 โˆ’ ๐›ฟ as desired.

Theorem 8.5 can be extended to many of the generic property generalization results in the main body, including approximate pseudometric loss, malicious noise, and semi-private learning, though the exact parameters may be somewhat weaker (e.g. learning over non-binary loss may incur additional factors and lead to ๐‘ -agnostic rather than truly agnostic learning). This comes at the cost of an extra factor of ๐œ€ โˆ’ 1 over reductions in the finite VC setting [52].

9Partial PAC-Learning

Partial PAC-learning is an extension of the standard PAC model to functions that are only defined on a certain portion of the input. Originally introduced by Long [47] and recently developed in greater depth by Alon, Hanneke, Holzman, and Moran (AHHM) [5], this model allows for the theoretical formalization of popular data-dependent assumptions such as margin that have no known analog in the PAC model. Combined with the distribution-family framework, this captures a significant portion of learning assumptions studied in both theory and practice (e.g. learning halfspaces with margin and distributional tail bounds). Letโ€™s formalize this model, starting with partial functions.

Definition 9.1 (Partial Function).

Let ๐‘‹ be an instance space and ๐‘Œ a label space. A partial function is a labeling ๐‘“ : ๐‘‹ โ†’ ๐‘Œ โˆช { * } , where elements labeled โ€œ * โ€ are thought as of undefined. The support of ๐‘“ , denoted ๐‘ ๐‘ข๐‘๐‘ โข ( ๐‘“ ) , is the set of elements ๐‘ฅ โˆˆ ๐‘‹ s.t.  ๐‘“ โข ( ๐‘ฅ ) โ‰  * .

Standard Partial PAC-learning is defined much like the standard model with the simple modification that โ€œ * โ€ labels are always considered to be incorrect. As a result, in the realizable case, when the adversary selects a particular partial function ๐‘“ , their marginal distribution over the instance space ๐‘‹ must be restricted to lying on supp โข ( ๐‘“ ) . This makes formalizing data-dependent assumptions easy. If one wanted to consider halfspaces with margin ๐›พ for instance, one simply labels every point within ๐›พ of the decision boundary as โ€œ * .โ€ Interestingly, much like the distribution-family setting, Partial-PAC learning falls outside both the uniform convergence and the sample compression paradigm [1]. AHHM also show a dramatic failure of empirical risk minimization: not only does naively applying an ERM to the partial class fail, it will also fail on any total extension of the class. Despite the lack of these standard tools, both Long and AHHM were able to show that distribution-free classification of partial classes is still controlled by VC dimension, and as a result that the equivalence of realizable and agnostic learnability extends to this setting. In this section, weโ€™ll discuss how a variant of our reduction shows that this result extends to the distribution-family model, extended loss function, and to properties beyond agnostic learning.

In the distribution-family model, formalizing realizable learnability requires some slight changes from the standard model, since we need to make sure our hypotheses are actually realizable over some distribution in the family (this is automatic in the distribution-free setting). To this end, we introduce a basic notion of closure for distribution families.

Definition 9.2 (Partial Closure).

Let ๐’Ÿ be a set of distributions over an instance space ๐‘‹ and ๐ป a concept class. Given any concept โ„Ž , and distribution ๐ท over ๐‘‹ , let ๐ท | โ„Ž denote the restriction ๐ท | ๐‘ ๐‘ข๐‘๐‘ โข ( ๐‘“ ) . The partial closure of ๐’Ÿ under ๐ป is:

๐’Ÿ ๐ป โข \coloneqq โข ๐’Ÿ โˆช โ‹ƒ ๐ท โˆˆ ๐’Ÿ , โ„Ž โˆˆ ๐ป ๐ท | โ„Ž .

In the realizable model it makes more sense to work with the closure of ๐’Ÿ than ๐’Ÿ itself, since otherwise the class ๐ป may contain hypotheses that cannot be realized over any distribution, and therefore cannot be accessed by the adversary at all. For simplicity, weโ€™ll also restrict our attention to (multi-class) classification where the label space ๐‘Œ

[ ๐‘š ] , and recall that the loss of any undefined point is always 1 .

Definition 9.3 ((Realizable) Distribution-Family Partial PAC Learning).

A partial class ( ๐’Ÿ , ๐‘‹ , ๐ป ) is PAC-learnable in the realizable setting if there exists an algorithm ๐’œ and function ๐‘› โข ( ๐œ€ , ๐›ฟ ) such that for all ๐œ€ , ๐›ฟ

0 and distributions ๐ท over ๐‘‹ ร— ๐‘Œ satisfying:

1.

The marginal ๐ท ๐‘‹ โˆˆ ๐’Ÿ ๐ป ,

2.

min โ„Ž โˆˆ ๐ป โก ๐‘’๐‘Ÿ๐‘Ÿ ๐ท โข ( โ„Ž )

0 ,

then:

Pr ๐‘† โˆผ ๐ท ๐‘› โข ( ๐œ€ , ๐›ฟ ) โก [ ๐‘’๐‘Ÿ๐‘Ÿ ๐ท โข ( ๐’œ โข ( ๐‘† ) )

๐œ€ ] โ‰ค ๐›ฟ ,

where the error ๐‘’๐‘Ÿ๐‘Ÿ ๐ท โข ( โ„Ž ) is standard classification error:

๐‘’๐‘Ÿ๐‘Ÿ ๐ท โข ( โ„Ž )

Pr ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท โก [ โ„Ž โข ( ๐‘ฅ ) โ‰  ๐‘ฆ ] .

Agnostic learnability is defined analogously, but since the adversary is unrestricted, there is no need to move to the closure of ๐’Ÿ .

Definition 9.4 ((Agnostic) Distribution-Family Partial PAC Learning).

A partial class ( ๐’Ÿ , ๐‘‹ , ๐ป ) is PAC-learnable in the agnostic setting if there exists an algorithm ๐’œ and function ๐‘› โข ( ๐œ€ , ๐›ฟ ) such that for all ๐œ€ , ๐›ฟ

0 and distributions ๐ท over ๐‘‹ ร— ๐‘Œ satisfying ๐ท ๐‘‹ โˆˆ ๐’Ÿ , then:

Pr ๐‘† โˆผ ๐ท ๐‘› โข ( ๐œ€ , ๐›ฟ ) โก [ ๐‘’๐‘Ÿ๐‘Ÿ ๐ท โข ( ๐’œ โข ( ๐‘† ) )

๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ ] โ‰ค ๐›ฟ .

The issue with our standard reduction strategy for partial functions is that in the agnostic model, the adversaryโ€™s marginal distribution over ๐‘‹ might have support outside of supp โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ) , which causes LearningToCover to lose its guarantee of outputting a non-uniform cover. This can be dealt with by a variant of our subsampling technique. If we run LearningToCover over all subsamples of the unlabeled sample ๐‘† ๐‘ˆ , one of these subsamples must match the support of โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ . This is in fact the same strategy used for adversarial robustness in Section 8, but we will include the algorithm again to make this section self-contained.

Input: Realizable Robust PAC-Learner ๐’œ Algorithm:

1.

Draw an unlabeled sample ๐‘† ๐‘ˆ , and labeled sample ๐‘† ๐ฟ .

2.

Run ๐’œ over all possible subsets and labelings of ๐‘† ๐‘ˆ to get:

๐ถ โข ( ๐‘† ) โข \coloneqq โข { ๐’œ โข ( ๐‘† , โ„Ž โข ( ๐‘† ) ) | ๐‘† โІ ๐‘† ๐‘ˆ , โ„Ž โˆˆ ๐ป | ๐‘† } .

3.

Return the hypothesis in ๐ถ โข ( ๐‘† ) with lowest empirical error over ๐‘† ๐ฟ .

Algorithm 6 Agnostic to Realizable Reduction Partial PAC Setting Theorem 9.5.

If ( ๐’Ÿ , ๐‘‹ , ๐ป ) is a realizably PAC-learnable partial class with sample complexity ๐‘› โข ( ๐œ€ , ๐›ฟ ) , then Algorithm 6 agnostically learns ( ๐’Ÿ , ๐‘‹ , ๐ป ) in

๐‘š ๐‘ˆ โข ( ๐œ€ , ๐›ฟ ) โ‰ค ๐‘‚ โข ( max ๐œ‡ โˆˆ [ 0 , 1 โˆ’ ๐œ€ ] โก { ๐‘› โข ( ๐œ€ / ( 2 โข ( 1 โˆ’ ๐œ‡ ) ) , ๐›ฟ / 3 ) 1 โˆ’ ๐œ‡ } )

unlabeled samples, and

๐‘š ๐ฟ โข ( ๐œ€ , ๐›ฟ ) โ‰ค ๐‘‚ โข ( ๐‘š ๐‘ˆ โข ( ๐œ€ , ๐›ฟ ) + log โก ( 1 / ๐›ฟ ) ๐œ€ 2 )

labeled samples.

Proof 9.6.

The proof is essentially the same as for Theorem 8.5, but we repeat it here for completeness. As always, it is enough to prove that ๐ถ โข ( ๐‘† ) (from Algorithm 6) contains a hypothesis โ„Ž โ€ฒ with error at most ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ / 2 . The key issue with our standard reduction is that the optimal hypothesis โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ may be undefined on certain examples in the unlabeled sample ๐‘† ๐‘ˆ . By running over all subsamples of ๐‘† ๐‘ˆ , we in essence simulate pulling samples only from the support of โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ , which is enough to get the desired guarantee.

More formally, let ๐ท | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ be the restriction of ๐ท to ๐‘ ๐‘ข๐‘๐‘ โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ) , and ๐ท ยฏ | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ the restriction to its complement ๐‘‹ โˆ– ๐‘ ๐‘ข๐‘๐‘ โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ) . The idea is to decompose our analysis into two separate parts over ๐ท | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ and ๐ท ยฏ | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ . With this in mind, let ๐œ‡ * denote the mass of ๐ท ๐‘‹ on the undefined portion of โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ , and let ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โ€ฒ denote the error of โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ over ๐ท | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ . Since we have restricted our attention to classification, notice that we can decompose ๐‘‚ โข ๐‘ƒ โข ๐‘‡ as:

๐‘’๐‘Ÿ๐‘Ÿ ๐ท โข ( โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ )

Pr ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท โข [ โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข ( ๐‘ฅ ) โ‰  ๐‘ฆ ]

๐œ‡ * โข Pr ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท ยฏ | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข [ โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข ( ๐‘ฅ ) โ‰  ๐‘ฆ ] + ( 1 โˆ’ ๐œ‡ * ) โข Pr ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข [ โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข ( ๐‘ฅ ) โ‰  ๐‘ฆ ]

๐œ‡ * + ( 1 โˆ’ ๐œ‡ * ) โข ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โ€ฒ .

Weโ€™d like to prove that ๐ถ โข ( ๐‘† ) contains a hypothesis โ„Ž within ๐œ€ / 2 error of optimal. We claim it is sufficient to show that ๐ถ โข ( ๐‘† ) contains a hypothesis within ๐œ€ / ( 2 โข ( 1 โˆ’ ๐œ‡ * ) ) classification distance of โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ , since:

๐‘’๐‘Ÿ๐‘Ÿ ๐ท โข ( โ„Ž )

๐œ‡ * โข ๐”ผ ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท ยฏ | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข [ โ„Ž โข ( ๐‘ฅ ) โ‰  ๐‘ฆ ] + ( 1 โˆ’ ๐œ‡ * ) โข ๐”ผ ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข [ โ„Ž โข ( ๐‘ฅ ) โ‰  ๐‘ฆ ]

โ‰ค ๐œ‡ * + ( 1 โˆ’ ๐œ‡ * ) โข ๐”ผ ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข [ โ„Ž โข ( ๐‘ฅ ) โ‰  ๐‘ฆ ]

โ‰ค ๐œ‡ * + ( 1 โˆ’ ๐œ‡ * ) โข ( ๐”ผ ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข [ โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข ( ๐‘ฅ ) โ‰  ๐‘ฆ ] + ๐œ€ 2 โข ( 1 โˆ’ ๐œ‡ * ) )

๐œ‡ * + ( 1 โˆ’ ๐œ‡ * ) โข ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โ€ฒ + ๐œ€ 2

๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ / 2 ,

where the third line follows from the fact that โ„Ž โ€ฒ and โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ only differ on a ๐œ€ 2 โข ( 1 โˆ’ ๐œ‡ * ) fraction of inputs over ๐ท | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ .

It is left to argue that ๐ถ โข ( ๐‘† ) contains such a hypothesis โ„Ž . Recall that on a labeled sample ( ๐‘† , โ„Ž โข ( ๐‘† ) ) โˆผ ๐ท | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ร— โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ of size ๐‘› โข ( ๐œ€ / ( 2 โข ( 1 โˆ’ ๐œ‡ * ) ) , ๐›ฟ / 3 ) , LearningToCover will contain โ„Ž that is ๐œ€ / ( 2 โข ( 1 โˆ’ ๐œ‡ * ) ) -close to โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ in classification error over ๐ท | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ with probability at least 1 โˆ’ ๐›ฟ / 3 . The idea is then to draw a large enough unlabeled sample such that with probability at least 1 โˆ’ ๐›ฟ / 3 , the restriction of the sample to ๐ท | โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ is at least this size (since we run over every subsample, we will always hit this restriction). By a Chernoff bound, it is enough to draw ๐‘ 1 โข ๐‘› โข ( ๐œ€ / ( 2 โข ( 1 โˆ’ ๐œ‡ * ) ) , ๐›ฟ / 3 ) 1 โˆ’ ๐œ‡ * points to achieve this for some large enough constant ๐‘ 1

0 .18 Since we do not know ๐œ‡ * , weโ€™ll need to draw ๐‘ 1 โข max ๐œ‡ โˆˆ [ 0 , 1 โˆ’ ๐œ€ ] โก { ๐‘› โข ( ๐œ€ / ( 1 โˆ’ ๐œ‡ ) , ๐›ฟ / 3 ) 1 โˆ’ ๐œ‡ } points to ensure this property holds (if ๐œ‡ * โ‰ฅ 1 โˆ’ ๐œ€ , note that any hypothesis gives a valid solution). By a union bound we have that this overall process succeeds with probability at least 1 โˆ’ 2 โข ๐›ฟ / 3 , and outputting the hypothesis in ๐ถ โข ( ๐‘† ) with the lowest empirical risk then succeeds with probability 1 โˆ’ ๐›ฟ as desired.

Like Theorem 8.5, Theorem 9.5 can be extended to many of the generic property generalization results in the main body, including approximate pseudometric loss, malicious noise, and semi-private learning, though it may experience some degradation of parameters (e.g.  ๐‘ -agnostic rather than truly agnostic learning) depending on how the loss of โ€œ * โ€ values are formalized in these settings. Again this comes at the cost of an additional factor of ๐œ€ over known reductions based on sample compression in the finite VC regime [5].

10Uniform Stability

Uniform stability, originally introduced by Bousquet and Elisseeff [20], is a useful algorithmic property that is closely tied to both generalization and privacy. Informally, an algorithm ๐’œ is said to be uniformly stable if for all elements ๐‘ฅ โˆˆ ๐‘‹ , the probability that ๐’œ changes its output on ๐‘ฅ over neighboring datasets is small.

Definition 10.1 (Uniform Stability).

A learning algorithm is said to be ๐›ผ -uniformly stable if for all neighboring inputs ๐‘† , ๐‘† โ€ฒ which differ on a single example, all ๐‘ฅ โˆˆ ๐‘‹ , and all ๐‘ฆ โˆˆ ๐‘Œ :

Pr โก [ ๐’œ โข ( ๐‘† ) โข ( ๐‘ฅ )

๐‘ฆ ] โ‰ค Pr โก [ ๐’œ โข ( ๐‘† โ€ฒ ) โข ( ๐‘ฅ )

๐‘ฆ ] + ๐›ผ .

Uniform stability can also be thought of as a variant of private prediction [32], which protects against adversaries who have restricted access to a model only through prediction responses on individual points (this is often the case in practice since it is common to release APIs with query access rather than full models). Like semi-privacy, this definition has the benefit of maintaining practicality in a reasonable range of circumstances while weakening the stringent requirements of standard private learning. Indeed, it is well known that in the distribution-free classification setting, uniformly stable learning and private prediction are both possible for any class with finite VC dimension [55, 32, 24]. Unsurprisingly, these previous works (at least those working in the agnostic model), rely on uniform convergence and uniform covers. Weโ€™ll show these can be replaced with a variant of our standard reduction. The argument is otherwise similar to the proof in [24].

Theorem 10.2.

Let ( ๐’Ÿ , ๐‘‹ , ๐ป ) be a realizably learnable class with sample complexity ๐‘› โข ( ๐œ€ , ๐›ฟ ) . Then there exists an ๐›ผ -uniformly stable, ๐›ผ -semi private algorithm that agnostically learns ( ๐’Ÿ , ๐‘‹ , ๐ป ) in only

๐‘š ๐‘ˆ โข ( ๐œ€ , ๐›ฟ , ๐›ผ ) โ‰ค ๐‘‚ โข ( ๐‘› โข ( ๐œ€ / 2 , ๐›ฟ / 2 ) ๐›ผ )

unlabeled samples, and

๐‘š ๐ฟ โข ( ๐œ€ , ๐›ฟ , ๐›ผ ) โ‰ค ๐‘‚ โข ( log โก ( ฮ  ๐ป โข ( ๐‘› โข ( ๐œ€ , ๐›ฟ ) ) ) min โก { ๐›ผ โข ๐œ€ , ๐œ€ 2 } )

labeled samples.

Proof 10.3.

The proof boils down to a standard subsampling trick first noted by [55]. Instead of drawing our standard unlabeled sample of size ๐‘› โข ( ๐œ€ / 2 , ๐›ฟ / 2 ) , we draw a sample of size ๐‘› โข ( ๐œ€ / 2 , ๐›ฟ / 2 ) 2 โข ๐›ผ and run LearningToCover over a random ๐›ผ / 2 fraction of the sample. This ensures that swapping out any individual sample can only effect the result with probability ๐›ผ / 2 . Since this subsample is of size ๐‘› โข ( ๐œ€ / 2 , ๐›ฟ / 2 ) , LearningToCover keeps its standard guarantees and the output ๐ถ โข ( ๐‘† ๐‘ˆ ) has a hypothesis within ๐œ€ / 2 of optimal with probability 1 โˆ’ ๐›ฟ / 2 . We can now apply the exponential mechanism with privacy parameter ๐›ผ / 4 , which ensures the algorithm is ๐›ผ / 2 -uniformly stable with respect to the labeled sample as well. The sample complexity bounds come from standard analysis of the exponential mechanism and the size of ๐ถ โข ( ๐‘† ๐‘ˆ ) . Semi-privacy comes for free due to our use of the exponential mechanism.

As in previous sections, Theorem 10.2 can be extended to any of the generic property generalization results in the main body, including for instance ๐‘ -approximate pseudometric loss, malicious noise, and robustness to covariate shift. In the VC setting, the complexity essentially matches the best known bounds in the literature, which are given by a similar reduction using uniform convergence [24].

11Statistical Query Model

Kearnsโ€™ [40] statistical query model is a popular modification of PAC learning where the sample oracle is replaced with the ability to ask noisy statistical questions about the data.

Definition 11.1 (Realizable SQ-learning).

Given a distribution ๐ท over ๐‘‹ and โ„Ž โˆˆ ๐ป , let ๐‘† โข ๐‘‡ โข ๐ด โข ๐‘‡ โข ( ๐ท , โ„Ž ) be an oracle which upon input of a function ๐œ“ : ๐‘‹ ร— ๐‘Œ โ†’ [ โˆ’ 1 , 1 ] and tolerance ๐œ โˆˆ โ„ โ‰ฅ 0 may output any estimate of the expectation of ๐œ“ up to ๐œ error, that is:

๐‘† โข ๐‘‡ โข ๐ด โข ๐‘‡ โข ( ๐ท , โ„Ž ) โข ( ๐œ“ , ๐œ ) โˆˆ ๐”ผ ๐‘ฅ โˆผ ๐ท โข [ ๐œ“ โข ( ๐‘ฅ , โ„Ž โข ( ๐‘ฅ ) ) ] ยฑ ๐œ .

We call a class ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) SQ-learnable if for all ๐œ€ > 0 , there exists some tolerance ๐œ

๐œ โข ( ๐œ€ ) , query complexity ๐‘› โข ( ๐œ€ , ๐œ ) , and an algorithm ๐’œ such that for all ๐ท โˆˆ ๐’Ÿ and โ„Ž โˆˆ ๐ป , ๐’œ achieves ๐œ€ error in at most ๐‘› โข ( ๐œ€ , ๐œ ) oracle calls to ๐‘† โข ๐‘‡ โข ๐ด โข ๐‘‡ โข ( ๐ท , โ„Ž ) with tolerance at worst ๐œ .19

Agnostic learning is then defined analogously where ๐ท , โ„Ž is replaced with a generic distribution over ๐‘‹ ร— ๐‘Œ whose marginal lies in ๐’Ÿ . We can use a basic form of discretization to prove property generalization in the SQ model.

Theorem 11.2.

Let โ„“ be a ๐‘ -approximate pseudometric and ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) a realizably SQ-learnable class with query complexity ๐‘› โข ( ๐œ€ , ๐œ ) . Then ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) is ๐‘ -agnostically SQ-learnable up to ๐œ€ + ๐œ error in ( 1 / ๐œ ) ๐‘› โข ( ๐œ€ , ๐œ ) statistical queries of tolerance at worst ๐œ .

Proof 11.3.

The idea is similar to our discretization in Theorem 6.6. The realizable SQ-learner ๐’œ makes some finite ๐‘› โข ( ๐œ€ , ๐œ ) queries. Let ๐ถ ๐’œ denote the set of outputs of ๐’œ when fed every possible combination of responses from the discretized set { โˆ’ 1 , โˆ’ 1 + 2 โข ๐œ , โ€ฆ , 1 โˆ’ 2 โข ๐œ , 1 } . For every ๐ท โˆˆ ๐’Ÿ and โ„Ž โˆˆ ๐ป , one of these combinations must be a valid query response in the realizable model, so ๐ถ ๐’œ covers ( ๐’Ÿ , ๐‘‹ , ๐ป , โ„“ ) . By the same arguments of Theorem 6.6, ๐ถ ๐’œ must contains a hypothesis with error at most ๐‘ โ‹… ๐‘‚ โข ๐‘ƒ โข ๐‘‡ + ๐œ€ . Since we can directly compute the loss of every element in ๐ถ ๐’œ up to ๐œ error in the SQ model simply by querying the loss function, this gives the desired result in | ๐ถ ๐’œ |

( 1 / ๐œ ) ๐‘› โข ( ๐œ€ , ๐œ ) queries.

We note that while our reduction in this model experiences exponential blowup in the number of queries, this should really be thought of as corresponding to a blow up in run-time instead of โ€œsample complexityโ€ in the standard sense (which corresponds more closely to ๐œ ).

12Fairness

Recent years have seen rising interest in an algorithmic property called fairness. Informally, fairness tries to tackle the issue that โ€œwell-performingโ€ classifiers in the standard sense may actually be discriminatory against certain individuals or subgroups. We will consider a form of fair leaning introduced by Rothblum and Yona [63] called Probably Approximately Correct and Fair (PACF) learning. Their definition is based off of a notion of fairness that ensures that similar individuals are treated similarly with respect to a fixed metric.

Definition 12.1 (Metric Fairness).

Let ๐‘‘ : ๐‘‹ ร— ๐‘‹ โ†’ โ„ โ‰ฅ 0 be a similarity measure on ๐‘‹ and ๐ท a distribution over ๐‘‹ . A classifier โ„Ž : ๐‘‹ โ†’ ๐‘Œ ๐‘œ๐‘ข๐‘ก is called ( ๐›ผ , ๐›พ ) -fair with respect to ๐‘‘ and ๐ท if h acts similarly on most similar individuals:

Pr ๐‘ฅ , ๐‘ฅ โ€ฒ โˆผ ๐ท โก [ | โ„Ž โข ( ๐‘ฅ ) โˆ’ โ„Ž โข ( ๐‘ฅ โ€ฒ ) |

๐‘‘ โข ( ๐‘ฅ , ๐‘ฅ โ€ฒ ) + ๐›พ ] โ‰ค ๐›ผ .

We note that the output space ๐‘Œ ๐‘œ๐‘ข๐‘ก may differ from the label space ๐‘Œ in general learning problems.

In fact, this definition only really makes sense when the output classifier โ„Ž is allowed to be real-valued (as this allows for some flexibility in the | โ„Ž โข ( ๐‘ฅ ) โˆ’ โ„Ž โข ( ๐‘ฅ โ€ฒ ) | term). As such, when considering settings such as binary classification where ๐‘Œ

{ 0 , 1 } is discrete, Rothblum and Yonaโ€™s [63] initial formalization considers returning probabilistic classifiers with ๐‘Œ out

[ 0 , 1 ] . Here โ„Ž โข ( ๐‘ฅ )

๐‘ฆ โˆˆ [ 0 , 1 ] is taken to be the probability of the label being 1 . The error of a probabilistic classifier โ„Ž with respect to any distribution ๐ท over ๐‘‹ ร— { 0 , 1 } is then given by its expected โ„“ 1 distance:

๐‘’ โข ๐‘Ÿ โข ๐‘Ÿ ๐ท โข ( โ„Ž )

๐”ผ ( ๐‘ฅ , ๐‘ฆ ) โˆผ ๐ท โข [ | โ„Ž โข ( ๐‘ฅ ) โˆ’ ๐‘ฆ | ] .

For simplicity weโ€™ll focus in this section on this same regime extended to the distribution-family model.

In broad strokes, the goal of Fair PAC learning is to output a fair classifier satisfying standard PAC guarantees. Practically this requires a few modifications. First, since there may be no fair classifier satisfying these guarantees, we will only require our output to be as good as the best fair classifier. Second, we will actually allow some slack in the fairness parameters, which Rothblum and Yona [63] show is a practical way to ensure that fair learnability remains possible across a broad range of classes.

Definition 12.2 (PACF-learning20  [63]).

We say ( ๐’Ÿ , ๐‘‹ , ๐ป ) is (agnostically) ( ๐›ผ , ๐›พ ) -PACF-learnable with respect to a similarity metric ๐‘‘ : ๐‘‹ ร— ๐‘‹ โ†’ ๐‘Œ if there exists an algorithm ๐’œ and function ๐‘›

๐‘› โข ( ๐œ€ , ๐œ€ ๐›ผ , ๐œ€ ๐›พ , ๐›ฟ ) such that for all ๐œ€ , ๐œ€ ๐›ผ , ๐œ€ ๐›พ , ๐›ฟ

0 , and distributions ๐ท over ๐‘‹ ร— ๐‘Œ such that ๐ท ๐‘‹ โˆˆ ๐’Ÿ , ๐’œ โข ( ๐‘† ) satisfies the following guarantees with probability 1 โˆ’ ๐›ฟ over samples ๐‘† of size ๐‘› :

1.

๐’œ โข ( ๐‘† ) is accurate:

๐‘’๐‘Ÿ๐‘Ÿ ๐ท , โ„“ โข ( ๐’œ โข ( ๐‘† ) ) โ‰ค ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ๐›ผ , ๐›พ + ๐œ€

2.

๐’œ โข ( ๐‘† ) is ( ๐›ผ + ๐œ€ ๐›ผ , ๐›พ + ๐œ€ ๐›พ ) -fair.

Here ๐‘‚ โข ๐‘ƒ โข ๐‘‡ ๐›ผ , ๐›พ is the optimal error of any ( ๐›ผ , ๐›พ ) -fair classifier, that is:

๐‘‚ โข ๐‘ƒ โข ๐‘‡ ๐‘Ž , ๐‘ โข \coloneqq โข min โ„Ž โˆˆ ๐ป ๐ท ๐‘‹ , ๐‘Ž , ๐‘ ๐‘‘ โก { ๐‘’ โข ๐‘Ÿ โข ๐‘Ÿ ๐ท , โ„“ โข ( โ„Ž ) } ,

and

๐ป ๐ท ๐‘‹ , ๐‘Ž , ๐‘ ๐‘‘

{ โ„Ž โˆˆ ๐ป : โ„Ž โข  is  ( ๐‘Ž , ๐‘ ) -fair with respect to  ๐‘‘  and  ๐ท ๐‘‹ }

Realizable learnability is defined similarly, where the adversary is constrained to picking distributions which have 0 error with respect to some ( ๐›ผ , ๐›พ ) -fair classifier in ๐ป . We show that property generalization holds for the PACF model.

Theorem 12.3 (Agnostic โ†’ Realizable (PACF Setting)).

Let ( ๐’Ÿ , ๐‘‹ , ๐ป ) be any class that is realizably ( ๐›ผ , ๐›พ ) -PACF learnable with sample complexity ๐‘› โข ( ๐œ€ , ๐œ€ ๐›ผ , ๐œ€ ๐›พ , ๐›ฟ ) . Then ( ๐’Ÿ , ๐‘‹ , ๐ป ) is agnostically ( ๐›ผ , ๐›พ ) -fair-PAC learnable in only

๐‘š ๐‘ˆ โข ( ๐œ€ , ๐œ€ ๐›ผ , ๐œ€ ๐›พ , ๐›ฟ ) โ‰ค ๐‘› โข ( ๐œ€ / 2 , ๐œ€ ๐›ผ , ๐œ€ ๐›พ , ๐›ฟ / 2 )

unlabeled samples, and

๐‘š ๐ฟ โข ( ๐œ€ , ๐œ€ ๐›ผ , ๐œ€ ๐›พ , ๐›ฟ ) โ‰ค ๐‘‚ โข ( log โก ( ฮ  ๐ป โข ( ๐‘› โข ( ๐œ€ / 2 , ๐œ€ ๐›ผ , ๐œ€ ๐›พ , ๐›ฟ / 2 ) ) ) + log โก ( 1 / ๐›ฟ ) ๐œ€ 2 )

labeled samples.

Proof 12.4.

The key observation is that the definition of fairness depends only on the classifier โ„Ž and the marginal distribution ๐ท ๐‘‹ . Let โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ be the hypothesis achieving the minimum error over ๐ป ๐ท ๐‘‹ , ๐›ผ , ๐›พ . By the above observation, with probability 1 โˆ’ ๐›ฟ the hypothesis set ๐ถ โข ( ๐‘† ๐‘ˆ ) returned by LearningToCover contains an ( ๐›ผ + ๐œ€ ๐›ผ , ๐›พ + ๐œ€ ๐›พ ) -fair โ„Ž satisfying:

๐”ผ ๐‘ฅ โˆผ ๐ท ๐‘‹ โข [ | โ„Ž โข ( ๐‘ฅ ) โˆ’ โ„Ž ๐‘‚ โข ๐‘ƒ โข ๐‘‡ โข ( ๐‘ฅ ) | ] โ‰ค ๐œ€ / 2 .

Since โ„“ 1 error is a metric (and therefore satisfies the triangle inequality), we can use our argument for ๐‘ -pseudometric loss functions from Theorem 6.6 to argue that choosing the lowest empirical risk ( ๐›ผ + ๐œ€ ๐›ผ , ๐›พ + ๐œ€ ๐›พ ) -fair classifier in ๐ถ โข ( ๐‘† ๐‘ˆ ) with respect to a sufficiently large labeled sample ๐‘† ๐ฟ gives the desired learner.

With care, this result can be extended to a broader range of loss functions as well as to other finitely-satisfiable properties covered in this work.

13Notions of Coverability

In this section we discuss the connection between non-uniform covers and several previous notions of coverability used in various learning applications. For simplicity, weโ€™ll restrict our attention to covering with respect to standard classification distance, that is given a distribution ๐ท and hypotheses โ„Ž and โ„Ž โ€ฒ over some instance space ๐‘‹ :

๐‘‘ ๐ท โข ( โ„Ž , โ„Ž โ€ฒ )

Pr ๐‘ฅ โˆผ ๐ท โก [ โ„Ž โข ( ๐‘ฅ ) โ‰  โ„Ž โ€ฒ โข ( ๐‘ฅ ) ] .

To start, letโ€™s recall the basic notion of an ๐œ€ -cover specified to this measure for simplicity.

Definition 13.1 ( ๐œ€ -cover).

Let ๐‘‹ be an instance space, ๐‘Œ a label space, and let ๐ฟ ๐‘‹ , ๐‘Œ denote the set of all labelings ๐‘ : ๐‘‹ โ†’ ๐‘Œ . A set ๐ถ โŠ‚ ๐ฟ ๐‘‹ , ๐‘Œ is said to form an ๐œ€ -cover for ( ๐ท , ๐‘‹ , ๐ป ) if for every hypothesis โ„Ž โˆˆ ๐ป , there exists ๐‘ โˆˆ ๐ถ such that

๐‘‘ ๐ท โข ( ๐‘ , โ„Ž ) โ‰ค ๐œ€ .

๐ถ is called proper if ๐ถ โŠ‚ ๐ป .

Finite ๐œ€ -covers are exceedingly useful in learning theory. As discussed in Section 3, a common strategy in the literature is to use unlabeled samples to construct an ๐œ€ -cover with high probability [8, 35, 2, 10]. This results in a distribution over potential covers we call a uniform ( ๐œ€ , ๐›ฟ ) -cover.

Definition 13.2 (Uniform ( ๐œ€ , ๐›ฟ ) -cover).

Let ๐‘‹ be an instance space, ๐‘Œ a label space, and let ๐ฟ ๐‘‹ , ๐‘Œ denote the set of all labelings ๐‘ : ๐‘‹ โ†’ ๐‘Œ . A distribution ๐ท ๐ถ over the power set ๐‘ƒ โข ( ๐ฟ ๐‘‹ , ๐‘Œ ) is said to form a uniform ( ๐œ€ , ๐›ฟ ) -cover for ( ๐ท , ๐‘‹ , ๐ป ) if:

Pr ๐ถ โˆผ ๐ท ๐ถ โก [ ๐ถ โข is an  ๐œ€ -cover for โข ( ๐ท , ๐‘‹ , ๐ป ) ] โ‰ฅ 1 โˆ’ ๐›ฟ .

๐ท ๐ถ is called proper if its support lies entirely in ๐ป .

In this work, we introduce a weaker non-uniform variant of this notion where each โ„Ž has an individual guarantee of being covered by the distribution, but it is not necessarily the case that a sample will cover all โ„Ž โˆˆ ๐ป simultaneously.

Definition 13.3 (Non-Uniform ( ๐œ€ , ๐›ฟ ) -cover).

Let ๐‘‹ be an instance space, ๐‘Œ a label space, and let ๐ฟ ๐‘‹ , ๐‘Œ denote the set of all labelings ๐‘ : ๐‘‹ โ†’ ๐‘Œ . A distribution ๐ท ๐ถ over the power set ๐‘ƒ โข ( ๐ฟ ๐‘‹ , ๐‘Œ ) is said to form a non-uniform ( ๐œ€ , ๐›ฟ ) -cover for ( ๐ท , ๐‘‹ , ๐ป ) if for every fixed hypothesis โ„Ž โˆˆ ๐ป ,

Pr ๐ถ โˆผ ๐ท ๐ถ โก [ ๐ถ โข is an  ๐œ€ -cover for โข ( ๐ท , ๐‘‹ , { โ„Ž } ) ] โ‰ฅ 1 โˆ’ ๐›ฟ .

๐ท ๐ถ is called proper if its support lies entirely in ๐ป .

In the context of learning, we are usually interested not just in the existence of these covers, but in the more challenging problem of constructing them from a small number of unlabeled samples. In other words, given a class ( ๐’Ÿ , ๐‘‹ , ๐ป ) , weโ€™d like to know how many unlabeled samples from an adversarially chosen distribution ๐ท โˆˆ ๐’Ÿ are necessary to build a uniform (or non-uniform) ( ๐œ€ , ๐›ฟ ) -cover for ( ๐ท , ๐‘‹ , ๐ป ) . In Section 6.3, we saw that the ability to construct a non-uniform ( ๐œ€ , ๐›ฟ ) -cover from ๐‘‚ โข ( log โก ( 1 / ๐›ฟ ) ๐œ€ ) samples was crucial to give a semi-private learner with optimal public sample complexity. This improved over recent work of Alon, Bassily, and Moran (ABM) [2], who showed that it is possible to build a uniform ( ๐œ€ , ๐›ฟ ) -cover in ๐‘‚ โข ( log โก ( 1 / ๐œ€ ) + log โก ( 1 / ๐›ฟ ) ๐œ€ ) samples.

It is interesting to ask whether non-uniformity is really necessary here, or whether ABMโ€™s analysis is simply sub-optimal. Weโ€™ll show that the former is true, at least in the proper distribution-family setting: the log โก ( 1 / ๐œ€ ) gap between these models is necessary and uniform covers cannot be used to build optimal semi-private learners.

Theorem 13.4 (Separation of Uniform and Non-Uniform Covers).

There exists an instance space ๐‘‹ , hypothesis class ๐ป , and family of distributions ๐’Ÿ such that for any sufficiently small ๐œ€

0 , the following statements holds:

1.

Any algorithm which returns a finite proper uniform ( ๐œ€ , 1 / 3 ) -cover for ( ๐’Ÿ , ๐‘‹ , ๐ป ) requires at least ฮฉ โข ( 1 / ๐œ€ โ‹… log โก ( 1 / ๐œ€ ) ) samples.

2.

There exists an algorithm which returns a finite proper non-uniform ( ๐œ€ , ๐›ฟ ) -cover for ( ๐’Ÿ , ๐‘‹ , ๐ป ) in ๐‘‚ โข ( log โก ( 1 / ๐›ฟ ) / ๐œ€ ) samples.

Proof 13.5.

Let the instance space ๐‘‹

โ„• and ๐ป be the class of indicators along with the all 0 โ€™s function, that is ๐ป

{ โ„Ž ๐‘– : ๐‘– โˆˆ โ„• } โˆช { โ„Ž 0 } where โ„Ž ๐‘– โข ( ๐‘ฅ )

๐Ÿ™ โข { ๐‘ฅ

๐‘– } and โ„Ž 0 is 0 everywhere. We consider the family of distribution ๐’Ÿ

{ ๐’Ÿ ๐‘› , ๐‘˜ } ๐‘› , ๐‘˜

0 given by ๐‘˜ -sets of [ ๐‘› ] where

๐’Ÿ ๐‘› , ๐‘˜

{ unif โข ( ๐‘‡ ) : ๐‘‡ โŠ‚ [ ๐‘› ] โข ๐‘Ž๐‘›๐‘‘ โข | ๐‘‡ |

๐‘˜ } ,

where unif โข ( ๐‘‡ ) is a uniform distribution over ๐‘‡ .

We start with the first claim, that building a bounded uniform ( ๐œ€ , 1 / 2 ) -cover needs at least ฮฉ โข ( 1 / ๐œ€ โข log โก ( 1 / ๐œ€ ) ) samples. More formally, for any error parameter ๐œ€ > 0 and size bound ๐‘š

๐‘š โข ( ๐œ€ ) โˆˆ โ„• , let ๐‘˜

โŒŠ 1 / ( 2 โข ๐œ– ) โŒ‹ . We will show that for any algorithm ๐’œ on ๐‘˜ โข log โก ( ๐‘˜ ) samples that outputs at most ๐‘š hypotheses, ๐’œ must fail to output an ๐œ€ -cover with probability at least 1 / 2 .

Let ๐‘› โ‰ซ ๐‘š , ๐‘˜ be some natural number to be fixed later and consider the family of distributions ๐’Ÿ ๐‘› , ๐‘˜ . By Yaoโ€™s minimax principle, it is sufficient to show that there exists a distribution over the elements in ๐’Ÿ ๐‘› , ๐‘˜ such that any deterministic algorithm over ๐‘˜ โข log โก ( ๐‘˜ ) samples outputting a set of (at most) ๐‘š hypotheses fails to give a proper ๐œ€ -cover with probability 1 / 2 . We claim that taking the uniform distribution over ๐’Ÿ ๐‘› , ๐‘˜ suffices. To formalize this, it is useful to observe the following claim.

Claim 2.

Any subset of hypotheses ๐ถ โŠ‚ ๐ป of size ๐‘š can be a proper ๐œ€ -cover for ๐ป under at most ( ๐‘š ๐‘˜ ) distributions in ๐’Ÿ ๐‘› , ๐‘˜ .

Letโ€™s prove the result under this assumption. The key observation is that by standard lower bounds on the coupon collector problem, a sample ๐‘† of ๐‘˜ โข log โก ( ๐‘˜ ) points from any unif โข ( ๐‘‡ ) โˆˆ ๐’Ÿ ๐‘› , ๐‘˜ will not include unif โข ( ๐‘‡ ) โ€™s entire support with probability at least 1 / 2 . With this in mind, assume that the input sample ๐‘† contains only ๐‘ ๐‘ข๐‘๐‘ โข ( ๐‘† )

๐‘— < ๐‘˜

๐‘ ๐‘ข๐‘๐‘ โข ( unif โข ( ๐‘‡ ) ) elements. As a result, there are ( ๐‘› โˆ’ ๐‘— ๐‘˜ โˆ’ ๐‘— ) consistent distributions with ๐‘† , and by Claim 2, ๐’œ โข ( ๐‘† ) is a proper ๐œ€ -cover for at most ( ๐‘š ๐‘˜ ) of them. Since ๐‘† is equally likely to have been sampled from any of these distributions, the probability that ๐’œ โข ( ๐‘† ) is a proper ๐œ€ -cover is at most:

Pr โก [ ๐’œ โข  fails given  ๐‘ ๐‘ข๐‘๐‘ โข ( ๐‘† )

๐‘— < ๐‘˜ ] โ‰ฅ ( ๐‘› โˆ’ ๐‘— ๐‘˜ โˆ’ ๐‘— ) โˆ’ ( ๐‘š ๐‘˜ ) ( ๐‘› โˆ’ ๐‘— ๐‘˜ โˆ’ ๐‘— ) .

Taking ๐‘› sufficiently larger than ๐‘š and ๐‘˜ , we can make this probability as close to 1 as desired for any 0 < ๐‘— < ๐‘˜ . Finally, since samples of this form occur with probability at least 1 / 2 , the algorithm fails with probability at least 1 / 3 as desired. It is left to prove Claim 2.

Proof 13.6 (Proof of Claim 2).

Notice that for any distribution unif โข ( ๐‘‡ ) โˆˆ ๐’Ÿ ๐‘› , ๐‘˜ , any ๐‘– โˆˆ ๐‘‡ and any ๐‘— โ‰  ๐‘– , ๐‘‘ unif โข ( ๐‘‡ ) โข ( โ„Ž ๐‘– , โ„Ž ๐‘— ) > 2 โข ๐œ€ . Let ๐ถ be any proper ๐œ€ -cover of ๐ป under distribution unif โข ( ๐‘‡ ) . Then, by the above argument, it must contain { โ„Ž ๐‘– : ๐‘– โˆˆ ๐‘‡ } . Since | ๐‘‡ |

๐‘˜ , ๐ถ can be a proper ๐œ€ -cover of ๐ป under at most ( | ๐ถ | ๐‘˜ ) distributions in ๐’Ÿ ๐‘› , ๐‘˜ .

We now move to proving that a proper non-uniform ( ๐œ€ , ๐›ฟ ) -cover can be built in only ๐‘‚ โข ( log โก ( 1 / ๐›ฟ ) / ๐œ€ ) samples. This follows from the fact that for any ๐‘› โ‰ฅ ๐‘˜ and distribution unif โข ( ๐‘‡ ) โˆˆ ๐’Ÿ ๐‘› , ๐‘˜ , each ๐‘– โˆˆ ๐‘‡ is in the random sample ๐‘† with probability 1 โˆ’ ๐›ฟ . Since each โ„Ž ๐‘— for ๐‘— โˆ‰ ๐‘‡ is covered by โ„Ž 0 , outputting { โ„Ž ๐‘– : ๐‘– โˆˆ ๐‘† } โˆช { โ„Ž 0 } generates a proper non-uniform ( ๐œ€ , ๐›ฟ ) -cover.

The construction in Theorem 13.4 can easily be modified to give a class with the same gap which is not privately learnable (say by embedding a single copy of a threshold over [ 0 , 1 ] ). Since any such class requires at least ฮฉ โข ( 1 ๐œ€ ) public samples to semi-privately learn by Theorem 6.25,21 Theorem 13.4 then provides a separation between using uniform and non-uniform covers in semi-private learning: the former provably requires an extra log factor, while the latter matches the lower bound exactly. Unfortunately, our proof of this result only holds in the proper setting, as Claim 2 fails when improper hypotheses are allowed. We conjecture that this is not an inherent barrier: the separation should continue to hold in the improper case, albeit with some different analysis.

We have now seen a weak separation between uniform and non-uniform covers, but one might reasonably wonder whether a much stronger separation is possible. In particular, all previous constructions of uniform covers use uniform convergence, but there exist simple examples of learnable classes in the distribution-family model that fail this property: do such classes provide an example of objects which are non-uniformly coverable but not uniformly coverable? Surprisingly, the answer is no! It turns out that an algorithm for non-uniform covering can always be used to construct a uniform covering without too much overhead. Moreover, weโ€™ll see that the log โก ( 1 / ๐œ€ ) gap is tight when ( ๐‘‹ , ๐ป ) has finite VC dimension.

To prove this, it will actually be useful to make a brief aside and introduce another closely related notion of covering called fractional covers. These objects are essentially a form of non-uniform covering which output a single hypothesis instead of a set of them.

Definition 13.7 (Fractional cover).

Let ๐‘‹ be an instance space, ๐‘Œ a label space, and let ๐ฟ ๐‘‹ , ๐‘Œ denote the set of all labelings ๐‘ : ๐‘‹ โ†’ ๐‘Œ . A distribution ๐ท ๐ถ over ๐ฟ ๐‘‹ , ๐‘Œ is said to form a fractional ( ๐œ€ , ๐‘ ) -cover for a hypothesis class ๐ป for ( ๐ท , ๐‘‹ , ๐ป ) if for any fixed โ„Ž โˆˆ ๐ป , a sample from ๐ท ๐ถ covers โ„Ž with probability ๐‘ :

Pr ๐‘ โˆผ ๐ท ๐ถ โก [ ๐‘‘ โข ( ๐‘ , โ„Ž ) โ‰ค ๐œ€ ] โ‰ฅ ๐‘ .

Fractional covers are closely connected to non-uniform covers. In fact, one can easily move between the two by sampling or subsampling.

Proposition 13.8 (Non-uniform cover โ‡” Fractional cover).

Let ( ๐ท , ๐‘‹ , ๐ป ) be any class, ๐ถ ๐‘“๐‘Ÿ๐‘Ž๐‘ a fractional ( ๐œ€ , ๐‘ ) -cover, and ๐ถ n-u a non-uniform ( ๐œ€ , 1 / 2 ) -cover. Then the following hold:

1.

Drawing log 1 / ( 1 โˆ’ ๐‘ ) โก ( 1 / ๐›ฟ ) samples from ๐ถ ๐‘“๐‘Ÿ๐‘Ž๐‘ gives a non-uniform ( ๐œ€ , ๐›ฟ ) -cover.

2.

Choosing a random hypothesis from ๐ถ n-u gives a fractional ( ๐œ€ , 1 / 2 โข | ๐ถ | ) -cover.

Proof 13.9.

Both statements are essentially immediate from definition. For any fixed โ„Ž โˆˆ ๐ป , if we draw ๐‘€ samples from ๐ถ ๐‘“๐‘Ÿ๐‘Ž๐‘ , the probability we fail to cover โ„Ž is ( 1 โˆ’ ๐‘ ) ๐‘€ , so setting ๐‘€

log 1 / ( 1 โˆ’ ๐‘ ) โก ( 1 / ๐›ฟ ) gives the desired non-uniform cover. On the other hand, for any fixed โ„Ž โˆˆ ๐ป , a sample from ๐ถ โˆผ ๐ถ n-u contains ๐‘

๐œ€ -close to โ„Ž with probability 1 / 2 . Outputting a uniformly random element of ๐ถ then gives an element within ๐œ€ of โ„Ž with probability 1 / 2 โข | ๐ถ | as desired.

It will also be useful to note a classical relation between covers and fractional covers.

Lemma 13.10.

If there exists a fractional ( ๐œ€ , ๐‘ ) -cover for ( ๐ท , ๐‘‹ , ๐ป ) , then there exists a 2 โข ๐œ€ -cover of size 1 / ๐‘ .

Proof 13.11.

This follows from classical packing-covering duality. The existence of a fractional ( ๐œ€ , ๐‘ ) -cover implies there cannot exist a 2 โข ๐œ€ -packing of size greater than 1 / ๐‘ (that is, a set of more than 1 / ๐‘ hypotheses in ๐ป that are pairwise 2 โข ๐œ€ -separated with respect to ๐ท ). By packing-covering duality, this implies the existence of a 2 โข ๐œ€ -cover of size โŒˆ 1 / ๐‘ โŒ‰ + 1 .

With this in hand, letโ€™s show that uniform covers can be constructed for any realizably learnable class, regardless of whether we have uniform convergence.

Theorem 13.12 (Realizable learning โ†’ Uniform cover).

Let ( ๐’Ÿ , ๐‘‹ , ๐ป ) be realizably PAC-learnable with sample complexity ๐‘› โข ( ๐œ€ , ๐›ฟ ) . Then it is possible to construct a uniform ( ๐œ€ , ๐›ฟ ) -cover for ( ๐’Ÿ , ๐‘‹ , ๐ป ) in ๐‘› โข ( ๐œ€ / 2 , ๐›ฟ โ€ฒ ) samples where ๐›ฟ โ€ฒ

๐‘‚ โข ( ๐›ฟ ฮ  ๐ป โข ( ๐‘› โข ( ๐œ€ / 2 , 1 / 2 ) ) ) .

Proof 13.13.

Weโ€™ll start by proving a slightly more general fact. If for every ๐ท โˆˆ ๐’Ÿ , ( ๐ท , ๐‘‹ , ๐ป ) has a proper ( ๐œ€ / 2 ) -cover ๐ถ ๐ท of size at most ๐ถ

๐ถ โข ( ๐œ€ / 2 ) , then it is possible to construct a uniform ( ๐œ€ , ๐›ฟ ) -cover in ๐‘› โข ( ๐œ€ / 2 , ๐›ฟ / ๐ถ ) samples. This is essentially immediate from Lemma 5.2, which states that running LearningToCover over a sample of size ๐‘› โข ( ๐œ€ / 2 , ๐›ฟ / ๐ถ ) gives a non-uniform ( ๐œ€ / 2 , ๐›ฟ / ๐ถ ) -cover. Union bounding over ๐ถ ๐ท then gives that a sample from the non-uniform cover ( ๐œ€ / 2 ) -covers ๐ถ ๐ท with probability at least 1 โˆ’ ๐›ฟ . Since ๐ถ ๐ท is itself an ( ๐œ€ / 2 ) -cover, this implies that the entire class ๐ป

๐œ€ -covered by the sample with probability at least 1 โˆ’ ๐›ฟ as desired.

It remains to show that for every ๐ท โˆˆ ๐’Ÿ , ( ๐ท , ๐‘‹ , ๐ป ) has a proper ( ๐œ€ / 2 ) -cover of size ๐‘‚ โข ( ฮ  ๐ป โข ( ๐‘› โข ( ๐œ€ / 2 , 1 / 2 ) ) ) . This follows from combining Proposition 13.8 and Lemma 13.10. In particular, Lemma 5.2 implies that running LearningToCover over a sample of size ๐‘› โข ( ๐œ€ / 2 , 1 / 2 ) produces a non-uniform ( ๐œ€ / 2 , 1 / 2 ) -cover of size at most ฮ  ๐ป โข ( ๐‘› โข ( ๐œ€ / 2 , 1 / 2 ) ) . Proposition 13.8 states that subsampling from this cover gives a fractional ( ๐œ€ / 2 , 1 / ( 2 โข ฮ  ๐ป โข ( ๐‘› โข ( ๐œ€ / 2 , 1 / 2 ) ) ) ) -cover, which in turn implies the existence of a ( ๐œ€ / 2 ) -cover of size ๐‘‚ โข ( ฮ  ๐ป โข ( ๐‘› โข ( ๐œ€ / 2 , 1 / 2 ) ) ) as desired. We note that this last argument is similar to an observation made in Benedek and Itaiโ€™s [17] seminal work on the distribution-dependent model.

When ( ๐‘‹ , ๐ป ) has finite VC-dimension ๐‘‘ , note that Theorem 13.12 exactly matches the lower bound exhibited in Theorem 13.4 as the required number of samples for a uniform ( ๐œ€ , ๐›ฟ ) -cover becomes:

๐‘› โข ( ๐œ€ / 2 , ๐›ฟ โ€ฒ ) โ‰ค ๐‘‚ โข ( ๐‘‘ โข log โก ( 1 / ๐œ€ ) + log โก ( 1 / ๐›ฟ ) ๐œ€ ) .

This also matches the bound given by ABM [2] using uniform convergence.

Acknowledgements

The authors would like to thank Shay Moran, Russell Impagliazzo, Omar Montasser, and Avrim Blum for enlightening discussions. We also thank anonymous referees for constructive feedback, and especially for pointing out the notion of probabilistic representations and that prior work discussed in Section 3 falls under the general framework of our reduction.

References [1] โ†‘ Nir Ailon, Anup Bhattacharya and Ragesh Jaiswalโ€œApproximate correlation clustering using same-cluster queriesโ€In Latin American Symposium on Theoretical Informatics, LATIN โ€™18, 2018, pp. 14โ€“27SpringerDOI: 10.1007/978-3-319-77404-6ห™2 [2] โ†‘ Noga Alon, Raef Bassily and Shay Moranโ€œLimits of Private Learning with Access to Public Dataโ€In Proceedings of the 33rd International Conference on Neural Information Processing Systems, NeurIPS โ€™19NY, USA: Curran Associates Inc., 2019DOI: 10.5555/3454287.3455215 [3] โ†‘ Noga Alon, Amos Beimel, Shay Moran and Uri Stemmerโ€œClosure properties for private classification and online predictionโ€In Conference on Learning Theory, COLT โ€™20, 2020, pp. 119โ€“152PMLRURL: https://proceedings.mlr.press/v125/alon20a.html [4] โ†‘ Noga Alon et al.โ€œAdversarial Laws of Large Numbers and Optimal Regret in Online Classificationโ€In Proceedings of the 53rd Annual ACM SIGACT Symposium on Theory of Computing, STOC 2021Virtual, Italy: Association for Computing Machinery, 2021, pp. 447โ€“455DOI: 10.1145/3406325.3451041 [5] โ†‘ Noga Alon, Steve Hanneke, Ron Holzman and Shay Moranโ€œA Theory of PAC Learnability of Partial Concept Classesโ€In 2021 IEEE 62nd Annual Symposium on Foundations of Computer Science, FOCS โ€™21CA, USA: IEEE Computer Society, 2021, pp. 658โ€“671DOI: 10.1109/FOCS52979.2021.00070 [6] โ†‘ Noga Alon, Roi Livni, Maryanthe Malliaris and Shay Moranโ€œPrivate PAC learning implies finite Littlestone dimensionโ€In Proceedings of the 51st Annual ACM SIGACT Symposium on Theory of Computing, STOC โ€™19, 2019, pp. 852โ€“860DOI: 10.1145/3313276.3316312 [7] โ†‘ Anish Athalye, Logan Engstrom, Andrew Ilyas and Kevin Kwokโ€œSynthesizing Robust Adversarial Examplesโ€In Proceedings of the 35th International Conference on Machine Learning 80, ICML โ€™18PMLR, 2018, pp. 284โ€“293URL: https://proceedings.mlr.press/v80/athalye18b.html [8] โ†‘ Maria-Florina Balcan and Avrim Blumโ€œA discriminative model for semi-supervised learningโ€In Journal of the ACM (JACM) 57.3ACM New York, NY, USA, 2010, pp. 1โ€“46DOI: 10.1145/1706591.1706599 [9] โ†‘ Peter L Bartlett, Philip M Long and Robert C Williamsonโ€œFat-shattering and the learnability of real-valued functionsโ€In Journal of computer and system sciences 52.3Elsevier, 1996, pp. 434โ€“452DOI: 10.1006/jcss.1996.0033 [10] โ†‘ Raef Bassily et al.โ€œPrivate Query Release Assisted by Public Dataโ€In Proceedings of the 37th International Conference on Machine Learning 119, ICML โ€™20PMLR, 2020, pp. 695โ€“703URL: https://proceedings.mlr.press/v119/bassily20a.html [11] โ†‘ Raef Bassily, Shay Moran and Anupama Nandiโ€œLearning from Mixtures of Private and Public Populationsโ€In Advances in Neural Information Processing Systems 33, NeurIPS โ€™20Curran Associates, Inc., 2020, pp. 2947โ€“2957URL: https://proceedings.neurips.cc/paper_files/paper/2020/file/1ee942c6b182d0f041a2312947385b23-Paper.pdf [12] โ†‘ Amos Beimel, Kobbi Nissim and Uri Stemmerโ€œCharacterizing the sample complexity of private learnersโ€In Proceedings of the 4th conference on Innovations in Theoretical Computer Science, ITCS โ€™13, 2013, pp. 97โ€“110DOI: 10.1145/2422436.2422450 [13] โ†‘ Amos Beimel, Kobbi Nissim and Uri Stemmerโ€œLearning privately with labeled and unlabeled examplesโ€In Algorithmica 83Springer, 2021, pp. 177โ€“215DOI: 10.1007/s00453-020-00753-z [14] โ†‘ Amos Beimel, Kobbi Nissim and Uri Stemmerโ€œPrivate Learning and Sanitization: Pure vs. Approximate Differential Privacyโ€In Theory of Computing 12.1Theory of Computing Exchange, 2016, pp. 1โ€“61DOI: 10.4086/toc.2016.v012a001 [15] โ†‘ Shai Ben-David, Nicolรฒ Cesa-Bianchi, David Haussler and Philip Longโ€œCharacterizations of Learnability for Classes of { 0 , โ€ฆ , ๐‘› } -Valued Functionsโ€In Journal of Computer and System Sciences 50.1, 1995, pp. 74โ€“86DOI: 10.1006/jcss.1995.1008 [16] โ†‘ Shai Ben-David, Dรกvid Pรกl and Shai Shalev-Shwartzโ€œAgnostic Online Learning.โ€In The 22nd Annual Conference on Learning Theory 3, COLT โ€™09, 2009, pp. 1URL: https://www.cs.huji.ac.il/~shais/papers/BendavidPalShalevtech09.pdf [17] โ†‘ Gyora M Benedek and Alon Itaiโ€œLearnability with respect to fixed distributionsโ€In Theoretical Computer Science 86.2Elsevier, 1991, pp. 377โ€“389DOI: 10.1016/0304-3975(91)90026-X [18] โ†‘ Avrim Blum and Tom Mitchellโ€œCombining Labeled and Unlabeled Data with Co-Trainingโ€In Proceedings of the Eleventh Annual Conference on Computational Learning Theory, COLTโ€™ 98Wisconsin, USA: Association for Computing Machinery, 1998, pp. 92โ€“100DOI: 10.1145/279943.279962 [19] โ†‘ Anselm Blumer, Andrzej Ehrenfeucht, David Haussler and Manfred K Warmuthโ€œLearnability and the Vapnik-Chervonenkis dimensionโ€In Journal of the ACM (JACM) 36.4ACM, 1989, pp. 929โ€“965DOI: 10.1145/76359.76371 [20] โ†‘ Olivier Bousquet and Andrรฉ Elisseeffโ€œStability and generalizationโ€In The Journal of Machine Learning Research 2JMLR. org, 2002, pp. 499โ€“526URL: https://www.jmlr.org/papers/volume2/bousquet02a/bousquet02a.pdf [21] โ†‘ Nicolรฒ Cesa-Bianchi et al.โ€œSample-efficient strategies for learning in the presence of noiseโ€In Journal of the ACM (JACM) 46.5ACM New York, NY, USA, 1999, pp. 684โ€“719DOI: 10.1145/324133.324221 [22] โ†‘ Kamalika Chaudhuri and Daniel Hsuโ€œSample complexity bounds for differentially private learningโ€In Proceedings of the 24th Annual Conference on Learning Theory, COLT โ€™11, 2011, pp. 155โ€“186JMLR WorkshopConference ProceedingsURL: https://proceedings.mlr.press/v19/chaudhuri11a.html [23] โ†‘ Koby Crammer, Michael Kearns and Jennifer Wortmanโ€œLearning from Multiple Sourcesโ€In Journal of Machine Learning Research 9.57, 2008, pp. 1757โ€“1774URL: http://jmlr.org/papers/v9/crammer08a.html [24] โ†‘ Yuval Dagan and Vitaly Feldmanโ€œPAC learning with stable and private predictionsโ€In Conference on Learning Theory, COLT โ€™20, 2020, pp. 1389โ€“1410PMLRURL: https://proceedings.mlr.press/v125/dagan20a.html [25] โ†‘ Amit Daniely, Sivan Sabato, Shai Ben-David and Shai Shalev-Shwartzโ€œMulticlass learnability and the ERM principleโ€In Journal of Machine Learning Research 16, 2015, pp. 2377โ€“2404URL: https://www.jmlr.org/papers/volume16/daniely15a [26] โ†‘ Amit Daniely and Shai Shalev-Shwartzโ€œOptimal learners for multiclass problemsโ€In Proceedings of The 27th Conference on Learning Theory 35, COLT โ€™14Barcelona, Spain: PMLR, 2014, pp. 287โ€“316URL: https://proceedings.mlr.press/v35/daniely14b.html [27] โ†‘ Sanjoy Dasguptaโ€œCoarse sample complexity bounds for active learningโ€In Advances in Neural Information Processing Systems 18, NeurIPS โ€™05MIT Press, 2005URL: https://proceedings.neurips.cc/paper_files/paper/2005/file/6e82873a32b95af115de1c414a1849cb-Paper.pdf [28] โ†‘ Sanjoy Dasgupta, Michael Littman and David McAllesterโ€œPAC Generalization Bounds for Co-trainingโ€In Advances in Neural Information Processing Systems 14, NeurIPS โ€™02MIT Press, 2002URL: https://proceedings.neurips.cc/paper/2001/file/4c144c47ecba6f8318128703ca9e2601-Paper.pdf [29] โ†‘ Ofir David, Shay Moran and Amir Yehudayoffโ€œSupervised learning through the lens of compressionโ€In Advances in Neural Information Processing Systems 29, NeurIPS โ€™16, 2016, pp. 2784โ€“2792URL: https://papers.nips.cc/paper_files/paper/2016/hash/59f51fd6937412b7e56ded1ea2470c25-Abstract.html [30] โ†‘ Shai Ben David, Tyler Lu, Teresa Luu and David Palโ€œImpossibility Theorems for Domain Adaptationโ€In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics 9, AISTATS โ€™10Sardinia, Italy: PMLR, 2010, pp. 129โ€“136URL: https://proceedings.mlr.press/v9/david10a.html [31] โ†‘ Richard M. Dudley, Sanjeev R. Kulkarni, Thomas Richardson and Ofer Zeitouniโ€œA metric entropy bound is not sufficient for learnabilityโ€In IEEE Transactions on Information Theory 40.3, 1994, pp. 883โ€“885DOI: 10.1109/18.335898 [32] โ†‘ Cynthia Dwork and Vitaly Feldmanโ€œPrivacy-preserving predictionโ€In Conference On Learning Theory, COLT โ€™18, 2018, pp. 1693โ€“1702PMLRURL: https://proceedings.mlr.press/v75/dwork18a [33] โ†‘ Vitaly Feldman, Parikshit Gopalan, Subhash Khot and Ashok Kumar Ponnuswamiโ€œOn agnostic learning of parities, monomials, and halfspacesโ€In SIAM Journal on Computing 39.2SIAM, 2009, pp. 606โ€“645DOI: 10.1137/070684914 [34] โ†‘ Steve Hannekeโ€œProper PAC learning VC dimension boundsโ€, Theoretical Computer Science Stack Exchange, 2018URL: https://cstheory.stackexchange.com/q/44252 [35] โ†‘ Steve Hanneke and Liu Yangโ€œMinimax Analysis of Active Learningโ€In Journal of Machine Learning Research 16.109, 2015, pp. 3487โ€“3602URL: http://jmlr.org/papers/v16/hanneke15a.html [36] โ†‘ David Hausslerโ€œDecision theoretic generalizations of the PAC model for neural net and other learning applicationsโ€In Information and computation 100.1Elsevier, 1992, pp. 78โ€“150DOI: 10.1016/0890-5401(92)90010-D [37] โ†‘ Daniel Hsu and Sivan Sabatoโ€œLoss Minimization and Parameter Estimation with Heavy Tailsโ€In Journal of Machine Learning Research 17.18, 2016, pp. 1โ€“40URL: http://jmlr.org/papers/v17/14-273.html [38] โ†‘ Thorsten Joachimsโ€œTransductive Inference for Text Classification Using Support Vector Machinesโ€In Proceedings of the Sixteenth International Conference on Machine Learning, ICML โ€™99CA, USA: Morgan Kaufmann Publishers Inc., 1999, pp. 200โ€“209DOI: 10.5555/645528.657646 [39] โ†‘ Shiva Prasad Kasiviswanathan et al.โ€œWhat can we learn privately?โ€In SIAM Journal on Computing 40.3SIAM, 2011, pp. 793โ€“826DOI: 10.1137/090756090 [40] โ†‘ Michael Kearnsโ€œEfficient Noise-Tolerant Learning from Statistical Queriesโ€In J. ACM 45.6New York, NY, USA: Association for Computing Machinery, 1998, pp. 983โ€“1006DOI: 10.1145/293347.293351 [41] โ†‘ Michael Kearns and Ming Liโ€œLearning in the presence of malicious errorsโ€In SIAM Journal on Computing 22.4SIAM, 1993, pp. 807โ€“837DOI: 10.1137/0222052 [42] โ†‘ Michael J. Kearns and Robert E. Schapireโ€œEfficient distribution-free learning of probabilistic conceptsโ€In Journal of Computer and System Sciences 48.3, 1994, pp. 464โ€“497DOI: https://doi.org/10.1016/S0022-0000(05)80062-5 [43] โ†‘ Sanjeev R. Kulkarni and Mathukumalli Vidyasagarโ€œLearning decision rules for pattern classification under a family of probability measuresโ€In IEEE Transactions on Information Theory 43.1, 1997, pp. 154โ€“166DOI: 10.1109/18.567668 [44] โ†‘ Tosca Lechner and Shai Ben-Davidโ€œImpossibility of Characterizing Distribution Learningโ€“a simple solution to a long-standing problemโ€In Arxiv, 2023URL: https://arxiv.org/abs/2304.08712 [45] โ†‘ Yi Li, Philip M Long and Aravind Srinivasanโ€œImproved bounds on the sample complexity of learningโ€In Journal of Computer and System Sciences 62.3Elsevier, 2001, pp. 516โ€“527DOI: 10.1006/jcss.2000.1741 [46] โ†‘ Nick Littlestone and Manfred Warmuthโ€œThe Weighted Majority Algorithmโ€In Information and Computation 108.2, 1994, pp. 212โ€“261DOI: https://doi.org/10.1006/inco.1994.1009 [47] โ†‘ Philip M Longโ€œOn agnostic learning with { 0,*, 1 } -valued and real-valued hypothesesโ€In International Conference on Computational Learning Theory, COLT โ€™01, 2001, pp. 289โ€“302SpringerURL: https://link.springer.com/chapter/10.1007/3-540-44581-1_19 [48] โ†‘ Andreas Maurer and Massimiliano Pontilโ€œConcentration inequalities under sub-Gaussian and sub-exponential conditionsโ€In Advances in Neural Information Processing Systems 34, NeurIPS โ€™11Curran Associates, Inc., 2021, pp. 7588โ€“7597URL: https://proceedings.neurips.cc/paper_files/paper/2021/file/3e33b970f21d2fc65096871ea0d2c6e4-Paper.pdf [49] โ†‘ Frank McSherry and Kunal Talwarโ€œMechanism design via differential privacyโ€In 48th Annual IEEE Symposium on Foundations of Computer Science, FOCS โ€™07, 2007, pp. 94โ€“103IEEEDOI: 10.1109/FOCS.2007.66 [50] โ†‘ Omar Montasser, Steve Hanneke and Nathan Srebroโ€œAdversarially Robust Learning with Unknown Perturbation Setsโ€In Proceedings of Thirty Fourth Conference on Learning Theory 134, COLT โ€™21PMLR, 2021, pp. 3452โ€“3482URL: https://proceedings.mlr.press/v134/montasser21a.html [51] โ†‘ Omar Montasser, Steve Hanneke and Nathan Srebroโ€œReducing Adversarially Robust Learning to Non-Robust PAC Learningโ€In Proceedings of the 34th International Conference on Neural Information Processing Systems, NeurIPS โ€™20BC, Canada: Curran Associates Inc., 2020URL: https://dl.acm.org/doi/10.5555/3495724.3496950 [52] โ†‘ Omar Montasser, Steve Hanneke and Nathan Srebroโ€œVC classes are adversarially robustly learnable, but only improperlyโ€In Conference on Learning Theory, COLT โ€™19, 2019, pp. 2512โ€“2530PMLRURL: https://proceedings.mlr.press/v99/montasser19a/montasser19a.pdf [53] โ†‘ Vaishnavh Nagarajan and J.Zico Kolterโ€œUniform convergence may be unable to explain generalization in deep learningโ€In Advances in Neural Information Processing Systems 32, NeurIPS โ€™11Curran Associates, Inc., 2019URL: https://proceedings.neurips.cc/paper_files/paper/2019/file/05e97c207235d63ceb1db43c60db7bbb-Paper.pdf [54] โ†‘ Balas K. Natarajanโ€œOn Learning Sets and Functionsโ€In Machine Learning 4.1USA: Kluwer Academic Publishers, 1989, pp. 67โ€“97DOI: 10.1007/BF00114804 [55] โ†‘ Shai Shalev-Shwartz, Ohad Shamir, Nathan Srebro and Karthik Sridharanโ€œLearnability, stability and uniform convergenceโ€In Journal of Machine Learning Research 11.90, 2010, pp. 2635โ€“2670URL: http://jmlr.org/papers/v11/shalev-shwartz10a.html [56] โ†‘ Hidetoshi Shimodairaโ€œImproving predictive inference under covariate shift by weighting the log-likelihood functionโ€In Journal of Statistical Planning and Inference 90.2, 2000, pp. 227โ€“244DOI: https://doi.org/10.1016/S0378-3758(00)00115-4 [57] โ†‘ Hans Ulrich Simonโ€œA characterization of strong learnability in the statistical query modelโ€In Annual Symposium on Theoretical Aspects of Computer Science, STACS โ€™07โ€™, 2007, pp. 393โ€“404SpringerDOI: 10.1007/978-3-540-70918-3ห™34 [58] โ†‘ Leslie G Valiantโ€œA theory of the learnableโ€In Proceedings of the sixteenth annual ACM symposium on Theory of computing, STOC โ€™84, 1984, pp. 436โ€“445ACMDOI: 10.1145/1968.1972 [59] โ†‘ Vladimir Vapnik and Alexey Chervonenkisโ€œOn the uniform convergence of relative frequencies of events to their probabilities.โ€In Theory of Probability and Its Applications, 1971DOI: 10.1137/1116025 [60] โ†‘ Vladimir Vapnik and Alexey Chervonenkisโ€œTheory of pattern recognitionโ€Moscow: Nauka Publishing House, 1974URL: https://cml.rhul.ac.uk/resources/PatternRecognition/patternreclowres.pdf [61] โ†‘ Mathuicumalli Vidyasagar, Srinivasan Balaji and Barbara Hammerโ€œClosure properties of uniform convergence of empirical means and PAC learnability under a family of probability measuresโ€In Systems and Control Letters 42.2, 2001, pp. 151โ€“157DOI: https://doi.org/10.1016/S0167-6911(00)00086-4 [62] โ†‘ Michael M Wolfโ€œMathematical foundations of supervised learningโ€Munich: mediaTUM, 2023URL: https://mediatum.ub.tum.de/doc/1723378/1723378.pdf [63] โ†‘ Gal Yona and Guy Rothblumโ€œProbably Approximately Metric-Fair Learningโ€In Proceedings of the 35th International Conference on Machine Learning 80, ICML โ€™18PMLR, 2018, pp. 5680โ€“5688URL: https://proceedings.mlr.press/v80/yona18a.html [64] โ†‘ Chiyuan Zhang et al.โ€œUnderstanding Deep Learning (Still) Requires Rethinking Generalizationโ€In Commun. ACM 64.3New York, NY, USA: Association for Computing Machinery, 2021, pp. 107โ€“115DOI: 10.1145/3446776 [65] โ†‘ Xiaojin Zhuโ€œSemi-supervised Learningโ€In Encyclopedia of Machine Learning and Data MiningSpringer US, 2017, pp. 1142โ€“1147DOI: 10.1007/978-0-387-30164-8ห™749 Generated by L A T E xml Instructions for reporting errors

We are continuing to improve HTML versions of papers, and your feedback helps enhance accessibility and mobile support. To report errors in the HTML that will help us improve conversion and rendering, choose any of the methods listed below:

Click the "Report Issue" button. Open a report feedback form via keyboard, use "Ctrl + ?". Make a text selection and click the "Report Issue for Selection" button near your cursor. You can use Alt+Y to toggle on and Alt+Shift+Y to toggle off accessible reporting links at each section.

Our team has already identified the following issues. We appreciate your time reviewing and reporting rendering errors we may not have found yet. Your efforts will help us improve the HTML versions for all readers, because disability should not be a barrier to accessing research. Thank you for your continued support in championing open access for all.

Have a free development cycle? Help support accessibility at arXiv! Our collaborators at LaTeXML maintain a list of packages that need conversion, and welcome developer contributions.

Report Issue Report Issue for Selection

Xet Storage Details

Size:
199 kB
ยท
Xet hash:
e0c3f43bc88d9bdd81065278202defdaa1887e7e011d1e15458561c0885d3dc1

Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.