This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

\useunder

\ul

Improving group robustness under noisy
labels using predictive uncertainty

Dongpin Oh , Dae Lee11footnotemark: 1 & Jeunghyun Byun
Deargen Inc.
Seoul, South Korea
{dongpin.oh,daelee,jhbyun}@deargen.me
&Bonggun Shin
Deargen USA Inc.
Atlanta, GA
{bonggun.shin}@deargen.me
These authors contributed equally
Abstract

The standard empirical risk minimization (ERM) can underperform on certain minority groups (i.e., waterbirds in lands or landbirds in water) due to the spurious correlation between the input and its label. Several studies have improved the worst-group accuracy by focusing on the high-loss samples. The hypothesis behind this is that such high-loss samples are spurious-cue-free (SCF) samples. However, these approaches can be problematic since the high-loss samples may also be samples with noisy labels in the real-world scenarios. To resolve this issue, we utilize the predictive uncertainty of a model to improve the worst-group accuracy under noisy labels. To motivate this, we theoretically show that the high-uncertainty samples are the SCF samples in the binary classification problem. This theoretical result implies that the predictive uncertainty is an adequate indicator to identify SCF samples in a noisy label setting. Motivated from this, we propose a novel ENtropy based Debiasing (END) framework that prevents models from learning the spurious cues while being robust to the noisy labels. In the END framework, we first train the identification model to obtain the SCF samples from a training set using its predictive uncertainty. Then, another model is trained on the dataset augmented with an oversampled SCF set. The experimental results show that our END framework outperforms other strong baselines on several real-world benchmarks that consider both the noisy labels and the spurious-cues.

1 Introduction

The standard Empirical Risk Minimization (ERM) has shown a high error on specific groups of data although it achieves the low test error on the in-distribution datasets. One of the reasons accounting for such degradation is the presence of spurious-cues. The spurious cue refers to the feature which is highly correlated with labels on certain training groups—thus, easy to learn—but not correlated with other groups in the test set (Nagarajan et al., 2020; Wiles et al., 2022). This spurious-cue is problematic especially occurs when the model cannot classify the minority samples although the model can correctly classify the majority of the training samples using the spurious cue. In practice, deep neural networks tend to fit easy-to-learn simple statistical correlations like the spurious-cues (Geirhos et al., 2020). This problem arises in the real-world scenarios due to various factors such as an observation bias and environmental factors (Beery et al., 2018; Wiles et al., 2022). For instance, an object detection model can predict an identical object differently simply because of the differences in the background(Ribeiro et al., 2016; Dixon et al., 2018; Xiao et al., 2020).

In nutshell, there is a low accuracy problem caused by the spurious-cues being present in a certain group of data. In that sense, importance weighting (IW) is one of the classical techniques to resolve this problem. Recently, several deep learning methods related to IW (Sagawa et al., 2019; 2020; Liu et al., 2021; Nam et al., 2020) have shown a remarkable empirical success. The main idea of those IW-related methods is to train a model with using data oversampled with hard (high-loss) samples. The assumption behind such approaches is that the high-loss samples are free from the spurious cues because these shortcut features generally reside mostly in the low-loss samples Geirhos et al. (2020). For instance, Just-Train-Twice (JTT) trains a model using an oversampled training set containing the error set generated by the identification model.

On the other hand, noisy labels are another factor of performance degradation in the real-world scenario. Noisy labels commonly occur in massive-scale human annotation data, biology and chemistry data with inevitable observation noise (Lloyd et al., 2004; Ladbury & Arold, 2012; Zhang et al., 2016). In practice, the proportions of incorrectly labeled samples in the real-world human-annotated image datasets can be up to 40% (Wei et al., 2021). Moreover, the presence of noisy labels can lead to the failure of the high-loss-based IW approaches, since a large value of the loss indicates not only that the sample may belong to a minority group but also that the label may be noisy (Ghosh et al., 2017). In practice, we observed that even a relatively small noise ratio (10%) can impair the high-loss-based methods on the benchmarks with spurious-cues, such as Waterbirds and CelebA. This is because the high loss-based approaches tend to focus on the noisy samples without focusing on the minority group with spurious cues.

Our observation motivates the principal goal of this paper: how can we better select only spurious-cue-free (SCF) samples while excluding the noisy samples? As an answer to this question, we propose the predictive uncertainty-based sampling as an oversampling criterion, which outperforms the error-set-based sampling. The predictive uncertainty has been used to discover the minority or unseen samples (Liang et al., 2017; Van Amersfoort et al., 2020). We utilize such uncertainty to detect the SCF samples. In practice, we train the identification model via the noise-robust loss and the Bayesian neural network framework to obtain reliable uncertainty for the minority group samples. By doing so, the proposed identification model is capable of properly identifying the SCF sample while preventing the noisy labels from being focused on. After training the identification model, similar to JTT, the debiased model is trained with the SCF set oversampled dataset. Our novel framework, ENtropy-based Debiasing (END), shows an impressive worst-group accuracy on several benchmarks with various degrees of symmetric label noise. Furthermore, as a theoretical motivation, we demonstrate that the predictive uncertainty (entropy) is a proper indicator for identifying the SCF set regardless of the existence of the noisy labels in the simple binary classification problem setting.

To summarize, our key contributions are three folds:

  1. 1.

    We propose a novel predictive uncertainty-based oversampling method that effectively selects the SCF samples while minimizing the selection of noisy samples.

  2. 2.

    We rigorously prove that the predictive uncertainty is an appropriate indicator for identifying a SCF set in the presence of the noisy labels, which well supports the proposed method.

  3. 3.

    We propose additional model considerations for real-world applications in both classification and regression tasks. The overall framework shows superior worst-group accuracy compared to recent strong baselines in various benchmarks.

2 Related works

Noisy label robustness: small loss samples

In this paper, we focus on two types of the noisy label robustness studies: (1) a sample re-weighting based approach and (2) a robust loss functions based approach. First, the sample re-weighting methods assign sample weights during model training to achieve the robustness against the noisy label (Han et al., 2018; Ren et al., 2018; Wei et al., 2020; Yao et al., 2021). Alternatively, the robust loss function based approaches design the loss function which implicitly focuses on the clean label (Reed et al., 2015; Zhang & Sabuncu, 2018; Thulasidasan et al., 2019; Ma et al., 2020). The common premise of the sample re-weighting and robust loss function methods are that the low-loss samples are likely to be the clean samples. For instance, Co-teaching uses two models which select the clean sample for each model by choosing samples of small losses (Han et al., 2018). Similarly, (Zhang & Sabuncu, 2018) design the generalized cross entropy loss function to have less emphasis on the samples of large loss than the vanilla cross entropy.

Group robustness: large loss samples

The model with the group robustness should yield a low test error regardless of the group specific information of samples (i.e., groups by background images). This group robustness can be improved if the model does not focus on the spurious cues (i.e., the background). The common assumption of prior works on the group robustness is that the large loss samples are spurious-cue-free. Sagawa et al. (2019); Zhang et al. (2020) propose the Distributionally Robust Optimization (DRO) methods which directly minimize the worst-group loss via group information of the training datasets given a priori. On the other hand, the group information-free approaches (Namkoong & Duchi (2017); Arjovsky et al. (2019); Oren et al. (2019)) have been proposed due to the non-negligible cost of group information. These approaches aim at achieving the group robustness by training the model via an implicit worst-case loss (e.g., maximum loss around the current loss). Nam et al. (2020) trains the debiased model by focusing on the samples with a large-loss with respect to the biased model. Notably, Just-Train-Twice (JTT) by Liu et al. (2021) is simple but effective framework. In JTT, an identification model is trained to build an error set, which consists of the wrongly predicted samples by the identification model. Then, the final model is trained after adding the oversampled error set into the original training dataset.

3 Preliminary: worst group performance and spurious cue

Refer to caption
Figure 1: The example case of the minority groups and its attribute on the cow-camel classification problem.

We consider a supervised learning task with inputs 𝐱(i)𝒳\mathbf{x}^{(i)}\in\mathcal{X}, corrupted labels y(i)𝒴y^{(i)}\in\mathcal{Y}, and true labels z(i)𝒴z^{(i)}\in\mathcal{Y}. We let a latent attribute of corresponding 𝐱(i)\mathbf{x}^{(i)} as a(i)𝒜a^{(i)}\in\mathcal{A} (e.g., different backgrounds in Figure 1). We assume each triplet (𝐱(i),y(i),z(i))(\mathbf{x}^{(i)},y^{(i)},z^{(i)}) belongs to a corresponding group, g(i)𝒢g^{(i)}\in\mathcal{G} (e.g., group by a background attribute and its true class (z(i),a(i)z^{(i)},a^{(i)}) in Figure 1). We also denote a training dataset as D={(𝐱^(i),y^(i))}i=1ND=\{(\hat{\mathbf{x}}^{(i)},\hat{y}^{(i)})\}_{i=1}^{N} where each pair (𝐱^(i),y^(i))(\hat{\mathbf{x}}^{(i)},\hat{y}^{(i)}) is sampled from a data distribution DD^{*} on 𝒳×𝒴\mathcal{X}\times\mathcal{Y}. Importantly, an attribute a(i)a^{(i)} can be spuriously correlated with a label on a certain dataset DD. For instance, in Figure 1, the label (cow/camel) can be highly correlated with the background feature (green pasture/desert), implying a false causal relationship. In this case, the background feature is a spurious cue.

Ideally, we aim to minimize the worst-group risk of a model fθ:𝒳cf_{\theta}:\mathcal{X}\rightarrow\mathbb{R}^{c} (cc is the number of class), parameterized by θ\theta, with an unknown data distribution DD^{*} and true labels zz as follows:

θ=argminθ[maxg𝒢RDg(θ)],RDg(θ)=𝔼(𝐱,z)Dg[L(fθ(𝐱),z)]\begin{gathered}\theta^{*}=\operatorname*{arg\,min}_{\theta}[\max_{g\in\mathcal{G}}R_{D^{*}_{g}}(\theta)],\qquad R_{D^{*}_{g}}(\theta)=\mathbb{E}_{(\mathbf{x},z)\sim D^{*}_{g}}[L(f_{\theta}(\mathbf{x}),z)]\end{gathered} (1)

where DgD^{*}_{g} is the data distribution for the group gg and L:c×𝒴L:\mathbb{R}^{c}\times\mathcal{Y}\rightarrow\mathbb{R} is a loss function. To achieve the above goal of improving the worst group accuracy, the prediction of the model fθf_{\theta} should not depend on the spurious cues. We instantiate this with the cow or camel classification dataset (Beery et al., 2018) which includes the background features as the spurious cue. When this relationship is abused, a model can easily classify the majority groups while failing to classify the minority groups such as a cow in the desert. Thus, the model has a poor accuracy on the minority group, leading to a high worst-group error.

To improve the worst-group accuracy, it is ideal to directly optimize the model via Eq 1. However, in the real-world, we assume that only the training samples DD are given, with no information about groups gg, attributes aa, and true labels zz during training. Therefore, ERM with the corrupted label is the alternative solution for the parameter θ\theta:

θD=argminθR^D(θ),R^D(θ)=1Ni=1NL(fθ(𝐱^(i)),y^(i))\begin{gathered}\theta_{D}^{*}=\operatorname*{arg\,min}_{\theta}\hat{R}_{D}(\theta),\qquad\hat{R}_{D}(\theta)=\tfrac{1}{N}\textstyle\sum_{i=1}^{N}L(f_{\theta}(\hat{\mathbf{x}}^{(i)}),\hat{y}^{(i)})\end{gathered} (2)

Our goal in this paper is to resolve the problem caused by the unavoidable alternative, ERM. The problem lies in the poor accuracy on a specific group due to the model’s reliance on the spurious cues (Sagawa et al., 2019). To address the problem of the poor worst-group accuracy, the common assumption in the literature has been that the model training should focus more on the particular samples that are not classifiable using the spurious cues (e.g., oversampling the cow on the desert samples) (Sagawa et al., 2019; Xu et al., 2020; Liu et al., 2021). In this paper, we call such samples the Spurious-Cue-Free (SCF) samples.

Given that we lack information to determine which samples are SCF, the remaining question is “how can we identify which samples belong to a SCF set?”. The previous studies (Liu et al., 2021) hypothesize that the samples with large loss values correspond to the SCF set (Sec 2). These approaches, however, have limitations because large loss values can be attributed to both the SCF and the noisy samples. Going a step further than other approaches, our primary strategy for identifying a SCF set is obtaining samples with a high predictive uncertainty. By doing so, our approach allows a more careful selection of SCF samples while excluding noisy ones as much as possible. As theoretical support, we rigorously show in the following section that the predictive uncertainty is a proper indicator to identify a SCF set in the presence of the noisy labels.

4 group robustness under noisy labels via uncertainty

In this section, we primarily show that the predictive uncertainty is a proper indicator to identify the SCF samples under the noisy label environment. Not only that, we can verify that utilizing the loss values as an oversampling metric can fail to distinguish SCF samples from the noisy samples. To rigorously prove this, we theoretically analyze the binary classification problem including both the spurious cues and the noise.

Problem setup: data distribution and model hypothesis

Consider a dd-dimensional input 𝐱=(x1,,xd)\mathbf{x}=(x_{1},\dots,x_{d}) and its features and labels are binary: xi,y,z{1,1}x_{i},y,z\in\{-1,1\}. A data generation procedure for triplets (𝐱,y,z)(\mathbf{x},y,z) is defined as following: first, zz is uniformly sampled over {1,1}\{-1,1\}. Then, if the true label is positive (z=1z=1), 𝐱𝐩\mathbf{x}\sim\mathcal{B}_{\mathbf{p}} where 𝐩\mathcal{B}_{\mathbf{p}} is a distribution of independent Bernoulli random samples and 𝐩=(p1,,pd)[0,1]d\mathbf{p}=(p_{1},\dots,p_{d})\in[0,1]^{d}. Here, pip_{i} represents the probability that the ii-th feature has a value of 1. On the contrary, if z=1z=-1, 𝐱\mathbf{x} is sampled from a different distribution: 𝐱𝐩\mathbf{x}\sim\mathcal{B}_{\mathbf{p^{\prime}}} where 𝐩=(p1,,pd)[0,1]d\mathbf{p}^{\prime}=(p^{\prime}_{1},\dots,p^{\prime}_{d})\in[0,1]^{d}. Furthermore, with the probability η\eta, there is a label noise: y=zy=-z. Otherwise, y=zy=z. Finally, we consider a linear model with parameters β=(β0,βd)\mathbf{\beta}=(\beta_{0},\dots\beta_{d}), which are optimized via risk minimization over the joint distribution of the features and noisy labels (𝐱,y\mathbf{x},y). This problem setup is inspired by the problem definition of Nagarajan et al. (2020) and Sagawa et al. (2020), representing the spurious features.

Next, the concept of the spurious cue in the defined classification task is demonstrated using the cow and camel images shown in Figure  1. Let’s assume that the jj’th feature represents a background attribute (xj=ax_{j}=a), which is either a green pasture (xj=1x_{j}=1) or a desert (xj=1x_{j}=-1). Suppose 98% of cows have a green pasture background (thus, pj=0.98p_{j}=0.98) while only 5% of camels have the green pasture background (pj=0.05p^{\prime}_{j}=0.05). In this case, the majority of the data could likely be classified only using the xjx_{j} feature (e.g., only utilizing βj\beta_{j}). However, abusing this spurious feature could hinder a model from accurately classifying the minority groups such as cows in a desert. Note that in practice, there can be many spurious features that correspond to a given latent attribute (aa).

To quantify the spuriousness, we define the Spurious-Cue Score (SCS) function.

Definition 1 (Spurious cue score function)

we define the spurious cue score function Ψ𝐩,𝐩:d\Psi_{\mathbf{p},\mathbf{p}^{\prime}}:\mathbb{R}^{d}\to\mathbb{R} with any function s(,)s(\cdot,\cdot) which satisfies followings:

Ψ𝐩,𝐩(𝐱)=i=1ds(pi,pi)xi,{s(pi,pi)>0,If pi>pis(pi,pi)0,If pipi\Psi_{\mathbf{p},\mathbf{p}^{\prime}}(\mathbf{x})=\sum_{i=1}^{d}s(p_{i},p^{\prime}_{i})x_{i},\quad\begin{cases}s(p_{i},p^{\prime}_{i})>0,\quad\text{If }p_{i}>p^{\prime}_{i}\\ s(p_{i},p^{\prime}_{i})\leq 0,\quad\text{If }p_{i}\leq p^{\prime}_{i}\end{cases} (3)

Intuitively, if the SCS function value is low, the model will struggle to correctly predict the label of a sample 𝐱\mathbf{x} by relying solely on features having a high correlation with the label. A simple example of s(,)s(\cdot,\cdot) can be s(pj,pj)=pjpjs(p_{j},p^{\prime}_{j})=p_{j}-p^{\prime}_{j}. In the cow/camel classification problem, the cow in the desert sample has the lower Ψ𝐩,𝐩\Psi_{\mathbf{p},\mathbf{p}^{\prime}} value due to the term s(pj,pj)xj=(0.980.05)(1)=0.93s(p_{j},p^{\prime}_{j})x_{j}=(0.98-0.05)(-1)=-0.93. In contrast, the majority of cow has 0.930.93. For the negative class samples (camel), Ψ𝐩,𝐩\Psi_{\mathbf{p}^{\prime},\mathbf{p}} can be used instead. Importantly, the SCS function is only determined by the true labels (zz) and the input features (𝐱\mathbf{x}) but not the noisy labels (yy).

With Definition 1, the following result formalizes our goal: to find the SCF samples (having a low SCS function values) by obtaining samples with the high predictive uncertainty.

Theorem 1

Given any 0<ϵ<1/20<\epsilon<1/2, for a sufficiently small δ\delta, a sufficiently large dd and any 𝐩,𝐩[ϵ,1ϵ]d\mathbf{p},\mathbf{p}^{\prime}\in[\epsilon,1-\epsilon]^{d} with |pipi|ϵ|p_{i}-p_{i}^{\prime}|\geq\epsilon (1id1\leq i\leq d), the following holds for any 0η<1/20\leq\eta<1/2:

𝐱,y,z[R(z,𝐱)ϵ|F(Hβ(𝐱))1δ]1ϵ,R(𝐱,z)=𝟏z=1F𝐱𝐩(Ψ𝐩,𝐩(𝐱))+𝟏z=1F𝐱𝐩(Ψ𝐩,𝐩(𝐱))\begin{gathered}\mathbb{P}_{\mathbf{x},y,z}[\enskip R(z,\mathbf{x})\leq\epsilon\enskip|\enskip F(H_{\beta^{*}}(\mathbf{x}))\geq 1-\delta\enskip]\geq 1-\epsilon,\\ R(\mathbf{x},z)=\mathbf{1}_{z=1}F_{\mathbf{x}\sim\mathcal{B}_{\mathbf{p}}}(\Psi_{\mathbf{p},\mathbf{p}^{\prime}}(\mathbf{x}))+\mathbf{1}_{z=-1}F_{\mathbf{x}\sim\mathcal{B}_{\mathbf{p}^{\prime}}}(\Psi_{\mathbf{p}^{\prime},\mathbf{p}}(\mathbf{x}))\end{gathered} (4)

where β\beta^{*} is the risk minimization solution of the linear regression on the distribution on (𝐱,y\mathbf{x},y) and HβH_{\beta^{*}} is the predictive entropy of the model with its parameter β\beta^{*}. FF, F𝐱𝐩F_{\mathbf{x}\sim\mathcal{B}_{\mathbf{p}}} and F𝐱𝐩F_{\mathbf{x}\sim\mathcal{B}_{\mathbf{p}^{\prime}}} are the cumulative distribution functions with respect to the data distribution, 𝐩\mathcal{B}_{\mathbf{p}} and 𝐩\mathcal{B}_{\mathbf{p}^{\prime}} respectively.

Theorem 1 states that the highly uncertain samples (F(Hβ(𝐱))1δF(H_{\beta^{*}}(\mathbf{x}))\geq 1-\delta) are more likely to have a low SCS function values among samples belonging to the same class (R(𝐱,z)ϵR(\mathbf{x},z)\leq\epsilon). Thus, this statement implies that utilizing the uncertainty (predictive entropy) can be useful in identifying the SCF samples. Additionally, it can be observed that the selectivity of the predictive entropy (HβH_{\beta^{*}}) is independent of the presence of the label noise (whether yz=1yz=-1). Particularly, it’s worth noting that the probability of the samples with a high predictive uncertainty to have a label noise is not greater than η\eta, the original label noise ratio. On the contrary, the probability of the label noise to be present in the samples with high loss exceeds η\eta (Theorem 2 in Appendix A.6). We empirically demonstrate that the proposed framework with the predictive uncertainty outperforms the existing loss value-based ones in the real-world benchmarks with the noisy labels (Sec 6). The formal form of this theorem and its proof are in Appendix A.

Entropy based debiasing for neural networks trained via ERM

We theoretically show that utilizing the predictive uncertainty can allow us to obtain the SCF samples under the ideal condition. However, there are two obstacles in applying this to the real-world scenario. Firstly, an overparameterized neural network trained via ERM (with a finite number of samples) can perfectly memorize the label noises (Zhang et al., 2017). This memorization could cause samples with the noisy labels to have a high entropy during training. For instance, as the model reduces the loss of the noisy samples once after it has fitted to the noise-free samples, the predictive entropy of those noisy samples increases (Xia et al., 2020). Secondly, since the neural network models generally tend to be overconfident (Guo et al., 2017; Hein et al., 2019), the proposed framework could potentially be unreliable in entropy-based acquisition of the SCF samples.

Refer to caption
Figure 2: 2-D classification results of identification models on synthetic data. The blue/red represent classes. The deeper the background, the more confident the prediction. The dots and translucent stars represent training and test data, respectively. Here, true SCF samples are in dotted circles (cannot be classified via spurious feature). Two classification results ((a) and (b)) shows that Model-A (proposed) well identifies the SCF set (large overlaps between yellow circles in (b) and true SCF samples) while Model-B (without regularization) fails due to its overconfident uncertainty for the true SCF samples.

To summarize, utilizing the predictive uncertainty with the neural networks necessitates two requirements for a neural network: (1: robustness to label noise) the model should not memorize samples with noisy labels; (2: reliable uncertainty) the model should not be overconfident when identifying the SCF samples. As an illustrative example, Figure 2 presents how well two different models identify the SCF samples. Model-A (proposed) is successful because it satisfies these two requirements. In contrast, model-B fails due to the overconfident prediction (low predictive uncertainty) for the true SCF samples. In the following section, we describe the proposed framework along with the training process that is designed to satisfy these two requirements.

5 Entropy Based Debiasing

Overview

The proposed ENtropy-based Debiasing (END) framework focuses on training the samples with high predictive uncertainty. The END framework, in particular, is made up of two models with identical architectures. First, the identification model uses the predictive uncertainty to identify the SCF samples. Second, the debiased model is trained on a newly constructed training data set by oversampling the SCF set. As a result, END achieves group and noise label robustness. In addition, our framework can also be extended to regression problems, whereas other baselines cannot.

5.1 Identification model

Robustness to the label noise and reliable uncertainty are two major requirements for using the predictive uncertainty with neural networks. To achieve these goals, we design the proposed identification model with the loss function robust to noisy labels and the overconfidence regularizer. In addition, we employ a Bayesian Neural Network (BNN) to obtain reliable uncertainty. Additionally, we explain how to modify the proposed identification model to fit a regression task.

The noise-robust loss function and the overconfidence regularizer

For the loss function, we use the mean absolute error (MAE) loss instead of the typical cross-entropy loss because the noise-label robustness of MAE has been well demonstrated in the literature (Ghosh et al., 2017; Zhang & Sabuncu, 2018). The MAE loss LMAEL_{MAE} is defined as the following:

LMAE(fθ(𝐱^),𝐲^)=𝐲^σ(fθ(𝐱^))1=22σi(fθ(𝐱^))\begin{gathered}L_{MAE}(f_{\theta}(\hat{\mathbf{x}}),\hat{\mathbf{y}})=\|\hat{\mathbf{y}}-\sigma(f_{\theta}(\hat{\mathbf{x}}))\|_{1}=2-2\sigma_{i^{*}}(f_{\theta}(\hat{\mathbf{x}}))\end{gathered} (5)

where ii^{*} is the index at which 𝐲^i=1\hat{\mathbf{y}}_{i^{*}}=1 in the one-hot encoded 𝐲^\hat{\mathbf{y}} and σi()\sigma_{i}(\cdot) is the ii-th value of the softmax function.

Although a model with the MAE loss alone could generally be noise-robust, it may occasionally be overconfident in predicting the SCF samples, meaning that the model produces low uncertainties in its predictions of those samples. As a result, the framework decides to exclude them from the SCF set. This problem is visually demonstrated in Model-B of Figure 2, which is trained with the MAE alone. To resolve this, we employ the confidence regularization. Importantly, the role of this regularization is to prevent overconfident prediction for the SCF samples (Liang et al., 2018; Müller et al., 2019; Utama et al., 2020). Specifically, this confidence regularization (Pereyra et al., 2017) penalizes the predictive entropy. The regularizer RentR_{ent} is defined as follows:

Rent(fθ(𝐱^))=i=1cσi(fθ(𝐱^))log(σi(fθ(𝐱^)))\begin{gathered}R_{ent}(f_{\theta}(\hat{\mathbf{x}}))=\textstyle\sum_{i=1}^{c}\sigma_{i}(f_{\theta}(\hat{\mathbf{x}}))\log(\sigma_{i}(f_{\theta}(\hat{\mathbf{x}})))\end{gathered} (6)

In practice, we empirically show that combining MAE with the confidence regularization yields a better set of SCF samples (Model-A in Figure 2), which enhances the worst-group accuracy (Sec 6.1 and  6.2). In addition, the ablation study (Appendix C.1) shows that the contribution of their combination is significant on the classification benchmarks.

Bayesian neural network

We chose BNN as a network architecture to ensure that the identification model’s uncertainty is reliable. This identification model is trained using the widely used stochastic gradient Markov-chain Monte Carlo sampling algorithm, Stochastic Gradient Langevin Dynamics (SGLD) (Welling & Teh, 2011). The SGLD updates the parameter θt\theta_{t} with the batch {𝐱^(i),y^(i)}i=1n\{\hat{\mathbf{x}}^{(i)},\hat{y}^{(i)}\}^{n}_{i=1} at step tt via the following equation:

θt+1θt[ϵt2(θlogp(θt)Nni=1nθlogp(y^(i)|θt,𝐱^(i)))+ρt]\begin{gathered}\theta_{t+1}\leftarrow\theta_{t}-\Bigl{[}-\frac{\epsilon_{t}}{2}(\nabla_{\theta}\log p(\theta_{t})-\frac{N}{n}\sum^{n}_{i=1}\nabla_{\theta}\log p(\hat{y}^{(i)}|\theta_{t},\hat{\mathbf{x}}^{(i)}))+\rho_{t}\Bigr{]}\end{gathered} (7)

where ϵt\epsilon_{t} is the step size and ρt𝒩(0,ϵt)\rho_{t}\sim\mathcal{N}(0,\epsilon_{t}). The negative log-likelihood term (logp(y^(i)|θt,𝐱^(i)))(-\log p(\hat{y}^{(i)}|\theta_{t},\hat{\mathbf{x}}^{(i)})) can be interpreted as a loss function. The prior term (logp(θt))(\log p(\theta_{t})) is equivalent to the L2 regularization if we use the Gaussian prior over the θ\theta. During the parameter updates, we obtain the snapshots of the parameters at every KK steps. The final predictive value of the identification model is the empirical mean of the predictions: f~(𝐱)=1Mj=1Mfθ(j)(𝐱)\tilde{f}(\mathbf{x})=\frac{1}{M}\sum^{M}_{j=1}f_{\theta^{(j)}}(\mathbf{x}) where MM is the number of parameter snapshots, θ(j)\theta^{(j)} is the parameters at the jj’th snapshot.

Extension to regression

Another benefit of the END framework is that it can be easily extended to fit a regression task with minor modifications. Since we utilize the BNN, the entropy can be calculated over a gaussian distribution with a predictive mean and a variance of SGLD weight samples 𝒩(f~(𝐱),Varj[fθ(j)(𝐱)])\mathcal{N}(\tilde{f}(\mathbf{x}),\text{Var}_{j}[f_{\theta^{(j)}}(\mathbf{x})]) (Kendall & Gal, 2017). Instead of the classification-MAE loss (Eq 5), the regression version of the identification model uses the common regression-MAE loss (L(𝐱,y)=|f(𝐱)y|L(\mathbf{x},y)=|f(\mathbf{x})-y|) similar to the classification task. Another change in the regression version is that the confidence regularization is no longer used because it is not defined in the regression task. The regression version of END also improves worst-group performance in the regression task as shown in Sec 6.3.

5.2 Debiased model

Once the identification model is trained, we build a new training dataset by oversampling the SCF samples to train the debiased model. First, we assume that the predictive entropy of the identification model follows the Gamma distribution because it is a common assumption that the variance (uncertainty) of the Gaussian distribution follows the Gamma distribution (Bernardo & Smith, 2009). Then, we obtain the SCF samples based on the p-value cut-off in a fitted Gamma distribution. Formally, the SCF set (DSCFkD^{k}_{SCF}) obtained by the identification model is as follows:

DSCFk=D^τD^τk times,D^τ={(𝐱^(i),y^(i))|Φ(H(σ(f~(𝐱^(i)));α,β)>1τ}\begin{gathered}D^{k}_{SCF}=\underbrace{\hat{D}_{\tau}\cup\dots\cup\hat{D}_{\tau}}_{k\text{ times}},\quad\hat{D}_{\tau}=\{(\hat{\mathbf{x}}^{(i)},\hat{y}^{(i)})|\Phi(H(\sigma(\tilde{f}(\hat{\mathbf{x}}^{(i)}));\alpha^{*},\beta^{*})>1-\tau\}\vspace{-8pt}\end{gathered} (8)

where H()H(\cdot) is entropy; Φ\Phi is the CDF of the Gamma distribution; τ\tau is the p-value threshold111Note the similarity between the definition of D^τ\hat{D}_{\tau} and the condition F(Hβ(𝐱))1δF(H_{\beta^{*}}(\mathbf{x}))\geq 1-\delta in Theorem 1.; kk is the hyperparameter to represent the degree of oversampling. The parameters of the gamma distribution, α\alpha^{*} and β\beta^{*}, are fitted via the moment estimation method (Hansen, 1982).

Finally, the debiased model is trained via ERM with the new dataset D^DSCFk\hat{D}\cup D^{k}_{SCF} after acquisition of the SCF set via eq 8. Note that training the debiased model follows a conventional ERM procedure: it does not use the confidence regularization or the MAE loss. The final prediction of the END framework is the predictive value of the trained debiased model.

6 Experiments

6.1 Synthetic dataset experiment

Refer to caption
Figure 3: 2-D classification results on the synthetic data. The blue/red represents the classes. The dots and translucent stars represents training and test data respectively. The minority (SCF) groups are in the dotted circles. The ideal decision boundary is the vertical purple dotted lines. The decision boundary of END is the closest to the ideal boundary.

We begin with the 2D-classification synthetic dataset experiment to qualitatively substantiate the group and noise label robustness of the END framework. The dataset has two characteristics: it has (1) the spurious features and (2) the noisy labels. Initially, we describe the two features of the dataset. One feature is the spurious feature which is easy to learn but exploiting this feature cannot classify the test set (Figure 3). The other one is the invariant feature. This feature is hard to learn because we manually scale down this feature value. Only a few training samples’ labels solely rely on this feature. The ideal model can correctly classify both the train and the test sets only when using the invariant features. Therefore, the ideal decision boundary will be the vertical line in Figure 3. Secondly, we assign random labels to the training samples with the probability of 20% to evaluate the model’s robustness to the noisy labels. The details of the dataset and the neural network model are in Appendix B.

The experimental results show that END correctly classifies both the majority and the minority groups (Figure 3(c)). We posit that the outperformance is due to the well-identified SCF sample (Model-A in Figure 2). On the other hand, ERM is insufficient to classify the minority group (samples in the dotted circles), although it perfectly fits the majority group samples(Figure 3(a)). This poor performance of ERM is consistent with the empirical studies on the real-world datasets like Waterbirds and CelebA (Liu et al., 2021). Notably, JTT has learned a wrong decision boundary in favor of the minority group while completely overlooking the majority group. This is because JTT focuses too much on the noisy labels.

6.2 Group Robustness Benchmark Dataset

Table 1: Average accuracy (ACC) and worst-group accuracy (WG Acc) evaluated on the Waterbirds and CelebA dataset, with varying noise levels. END consistently outperforms other baselines, especially for the noisy datasets.
Noise level Clean 10% noise 20% noise 30% noise
Waterbirds WG Acc Acc WG Acc Acc WG Acc Acc WG Acc Acc
ERM 0.687 (0.01) 0.969 (0.00) 0.648 (0.03) 0.945 (0.01) 0.649 (0.05) 0.913 (0.01) 0.629 (0.06) 0.893 (0.03)
ERM (GCE) 0.674 (0.01) 0.968 (0.00) 0.665 (0.03) 0.945 (0.00) 0.651 (0.04) 0.902 (0.00) 0.660 (0.07) 0.885 (0.03)
LfF 0.710 (0.03) 0.947 (0.02) 0.710 (0.00) 0.914 (0.03) 0.726 (0.03) 0.858 (0.04) 0.660 (0.04) 0.899 (0.02)
JTT 0.846 (0.03) 0.865 (0.02) 0.565 (0.08) 0.670 (0.04) 0.060 (0.03) 0.135 (0.01) 0.027 (0.01) 0.107 (0.00)
SSA 0.887 (0.01) 0.918 (0.00) 0.872 (0.01) 0.885 (0.02) \ul0.803 (0.00) 0.825 (0.02) \ul0.747 (0.02) 0.773 (0.02)
END 0.828 (0.01) 0.934 (0.01) \ul0.842 (0.01) 0.914 (0.01) 0.832 (0.01) 0.887 (0.02) 0.818 (0.01) 0.854 (0.01)
CelebA WG Acc Acc WG Acc Acc WG Acc Acc WG Acc Acc
ERM 0.487 (0.03) 0.952 (0.00) 0.477 (0.01) 0.927 (0.01) 0.480 (0.02) 0.891 (0.01) 0.485 (0.01) 0.858 (0.03)
ERM (GCE) 0.502 (0.03) 0.956 (0.00) 0.524 (0.01) 0.950 (0.00) 0.522 (0.02) 0.941 (0.00) 0.526 (0.04) 0.920 (0.01)
LfF 0.788 (0.03) 0.871 (0.04) 0.080 (0.06) 0.217 (0.01) 0.027 (0.02) 0.089 (0.02) 0.052 (0.06) 0.236 (0.25)
JTT \ul0.822 (0.02) 0.915 (0.01) \ul0.748 (0.02) 0.810 (0.01) 0.245 (0.36) 0.357 (0.29) 0.151 (0.16) 0.258 (0.12)
SSA 0.899 (0.00) 0.906 (0.01) 0.735 (0.01) 0.811 (0.00) \ul0.674 (0.01) 0.767 (0.00) \ul0.632 (0.02) 0.729 (0.01)
END \ul0.826 (0.02) 0.889 (0.00) 0.797 (0.01) 0.893 (0.01) 0.811 (0.02) 0.901 (0.01) 0.778 (0.03) 0.892 (0.01)
Refer to caption
Figure 4: 2-D PCA projection of the latent feature of the identification model (first row) and its variant (w/o LMAEL_{MAE} and RentR_{ent}) (second row). Each dot represents the training sample. (first column) The green dots are noisy labels, and the blue dots are clean labels; (second column) The yellow dots are the obtained SCF samples; (third column) Each color represents corresponding groups.

In this subsection, we evaluate the END framework on the two benchmark image datasets: the CelebA and Waterbirds, which have the spurious-cues (Wah et al., 2011; Liu et al., 2015). To evaluate both the group and the noise label robustness, we add simple symmetric label noises (uniformly flip the label) to the datasets, as shown in Table 1.

In this experiment, we use two kinds of baselines: the group-robust and the noise-robust baselines. The group-robust baselines include JTT (Liu et al., 2021), ERM (Zhang & Sabuncu, 2018), Learning-from-Failure (LfF) (Nam et al., 2020), and recently proposed Spread-Spurious-Attribute (SSA) (Nam et al., 2022). There is one noise-robust baseline, ERM with GCE (Generalized Cross Entropy) (Zhang & Sabuncu, 2018). We use the identical model architecture, ResNet50 He et al. (2016), for all baselines and END. For JTT, LfF, and ERM, we follow the identical experimental setup as Liu et al. (2021) and Nam et al. (2022) presented. Details of the experimental setup are in Appendix B.

The results from Table 1 substantiate that, unlike the other baselines, END achieves both the group and the noisy-label robustness on the noisy label setting. The primary reason for this is that END employs the predictive entropy, which is unaffected by the label noise. Specifically, we observe that the worst-group accuracy of the END framework consistently outperforms the other models in the noisy cases (Table 1; >> 10% noise). Not only that, END also shows the competitive worst-group accuracy on the noise free case. On the other hand, as the noise rate increases, the performances of the group-robust baselines are harmed much more severely. We interpreted that the reason for the degradation of these baselines is that they focus on the large loss samples, which are likely to be the noisy labels, as discussed in Sec 4. The noise label robust loss, GCE, improves the ERM with noisy labels, but its group robustness is insufficient. In addition, in the ablation study in Appendix C.1 shows that (1) utilizing entropy is the major contribution of the group and the noisy label robustness, as Theorem 1 stated, and (2) cooperation between the noisy-robust loss and the overconfident regularizer has an important role in its performance.

Additionally, we qualitatively show that our identification model can identify the SCF samples while being robust to the noisy labels. Concretely, we visualize the 2-D projection of the latent features (before the last linear layer) of the identification model on the Waterbirds with 30% label noise. This experiment has two implications. Firstly, the SCF set identified by our framework corresponds with the minority group (the true SCF samples). Specifically, the first row of Figure 4 (END) shows that both the minority group (red and green dots in the third column) and the SCF set are mainly located around the middle of the images. Quantitatively, up to 30% of the SCF set consists of the minority group, which is higher than the actual proportion (5%). Secondly, the contamination of the SCF set with the noisy labeled samples is significantly mitigated by our framework. The first row of Figure 4 (END) shows that (1) the noise labels are almost identically distributed over the space, and (2) the noise labels do not severely overlap with the SCF set. This result substantiates that the proposed identification model (Sec 5.1) effectively identifies the SCF samples while including less noisy samples. In contrast, the baseline identification model (the second row) shows an overlap between the noisy samples and the SCF set. This is in line with our claim: to utilize the uncertainty with the neural networks, the model should not memorize the noise labels and should have reliable uncertainty.

6.3 Real-world Regression Dataset

In this subsection, we conduct experiments on the regression datasets with non-synthetic label noise to demonstrate the followings: (1) the END framework can be extended to a regression problem; (2) END achieve the group robustness under the non-artificial label noise. Particularly, we evaluate models on two drug-target affinity (DTA) benchmarks, Davis (Davis et al., 2011) and KIBA (Tang et al., 2014), see Appendix B for the details. Inputs of these datasets are drug molecules and protein sequences. The target value is a physical experiment-derived affinity values between the drug and the protein. We use the DeepDTA architecture (Öztürk et al., 2018) as the base architecture. See Appendix B for the details. Similar to the classification benchmarks, the DTA benchmarks have two characteristics. Firstly, the dataset has the spurious correlation: the DTA model typically relies on a single modality (e.g., predict its affinity by only leveraging the drug molecule, not considering the interaction with the target protein, or vice versa), which is inconsistent with the physicochemical laws (Özçelik et al., 2021; Yang et al., 2022). To get the worst group information of each benchmark, we group data samples by their distinct drug molecules and target proteins, respectively. Secondly, the target values are naturally noisy due to the different environments of data acquisition (Davis et al., 2011; Tang et al., 2014; 2018). Thus, it can be seen as the non-synthetic label noise.

Table 2: Evaluation results on the DTA datasets.
Davis MSE WG MSE
ERM 0.268 (0.01) \ul0.860 (0.14)
Hard 0.299 (0.02) 1.196 (0.24)
END 0.262 (0.01) 0.785 (0.07)
KIBA MSE WG MSE
ERM 0.217 (0.04) 8.915 (0.33)
Hard 0.203 (0.02) \ul8.748 (0.39)
END 0.207 (0.05) 8.358 (0.53)

Since Lff and JTT cannot be extended to the regression problems, we propose an alternative baseline, “hard.” Akin to JTT and END, the hard algorithm picks up the top-KK largest loss samples after the first phase of training. Next, another model is trained with the oversampled training dataset with those hard samples. Table 2 shows that END outperforms others in terms of worst-group MSE metrics. We posit that it is the well-identified SCF set via the proposed uncertainty-based approach that contributes to this improvement. On the other hand, hard shows no improvement over ERM due to the oversampled noise labels.

7 Discussion

In this study, we present a new approach that can significantly improve the group robustness under the label noise scenario. We theoretically show that the predictive uncertainty is the proper criterion for identifying the SCF samples. Upon this foundation, we propose the END framework consisting of two procedures. (1) Obtaining the SCF set via predictive uncertainty of the noise-robust model with reliable uncertainty; (2) Training the debiased model on the oversampled training set with the selected SCF samples. In practice, we empirically demonstrate that END achieves both the group and the noisy label robustness.

For future works, we discuss several potential areas of improvement. Firstly, the END framework adopts simple approaches (the MAE loss and the SGLD) for the identification model. Thus, the future works can employ a more advanced approach for the identification model which (1) obtains the reliable uncertainty and (2) prevents the memorization of the noisy label. Secondly, we only consider the total predictive uncertainty of the model in this study. However, the predictive uncertainty can be decomposed into two different types of uncertainty: aleatoric (uncertainty arising from the data noise) and epistemic (uncertainty arising from the model parameters) Kendall & Gal (2017). If we disregard the decomposed aleatoric uncertainty, we believe that it could improve the END framework.

References

  • Arjovsky et al. (2019) Martin Arjovsky, Léon Bottou, Ishaan Gulrajani, and David Lopez-Paz. Invariant risk minimization. arXiv preprint arXiv:1907.02893, 2019.
  • Beery et al. (2018) Sara Beery, Grant Van Horn, and Pietro Perona. Recognition in terra incognita. In Proceedings of the European conference on computer vision (ECCV), pp.  456–473, 2018.
  • Bentkus (2005) Vidmantas Bentkus. A lyapunov-type bound in rd. Theory of Probability & Its Applications, 49(2):311–323, 2005.
  • Bernardo & Smith (2009) José M Bernardo and Adrian FM Smith. Bayesian theory, volume 405. John Wiley & Sons, 2009.
  • Davis et al. (2011) Mindy I Davis, Jeremy P Hunt, Sanna Herrgard, Pietro Ciceri, Lisa M Wodicka, Gabriel Pallares, Michael Hocker, Daniel K Treiber, and Patrick P Zarrinkar. Comprehensive analysis of kinase inhibitor selectivity. Nature biotechnology, 29(11):1046–1051, 2011.
  • Dixon et al. (2018) Lucas Dixon, John Li, Jeffrey Sorensen, Nithum Thain, and Lucy Vasserman. Measuring and mitigating unintended bias in text classification. In Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society, pp.  67–73, 2018.
  • Geirhos et al. (2020) Robert Geirhos, Jörn-Henrik Jacobsen, Claudio Michaelis, Richard Zemel, Wieland Brendel, Matthias Bethge, and Felix A Wichmann. Shortcut learning in deep neural networks. Nature Machine Intelligence, 2(11):665–673, 2020.
  • Ghosh et al. (2017) Aritra Ghosh, Himanshu Kumar, and P Shanti Sastry. Robust loss functions under label noise for deep neural networks. In Proceedings of the AAAI conference on artificial intelligence, volume 31, 2017.
  • Guo et al. (2017) Chuan Guo, Geoff Pleiss, Yu Sun, and Kilian Q Weinberger. On calibration of modern neural networks. In International conference on machine learning, pp. 1321–1330. PMLR, 2017.
  • Han et al. (2018) Bo Han, Quanming Yao, Xingrui Yu, Gang Niu, Miao Xu, Weihua Hu, Ivor Tsang, and Masashi Sugiyama. Co-teaching: Robust training of deep neural networks with extremely noisy labels. Advances in neural information processing systems, 31, 2018.
  • Hansen (1982) Lars Peter Hansen. Large sample properties of generalized method of moments estimators. Econometrica: Journal of the econometric society, pp. 1029–1054, 1982.
  • He et al. (2016) Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp.  770–778, 2016.
  • Hein et al. (2019) Matthias Hein, Maksym Andriushchenko, and Julian Bitterwolf. Why relu networks yield high-confidence predictions far away from the training data and how to mitigate the problem. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  41–50, 2019.
  • Juškevičius & Kurauskas (2019) Tomas Juškevičius and Valentas Kurauskas. On littlewood-offord theory for arbitrary distributions. arXiv preprint arXiv:1912.08770, 2019.
  • Kendall & Gal (2017) Alex Kendall and Yarin Gal. What uncertainties do we need in bayesian deep learning for computer vision? Advances in neural information processing systems, 30, 2017.
  • Krishnapur (2016) MANJUNATH Krishnapur. Anti-concentration inequalities. Lecture notes, 2016.
  • Ladbury & Arold (2012) John E Ladbury and Stefan T Arold. Noise in cellular signaling pathways: causes and effects. Trends in biochemical sciences, 37(5):173–178, 2012.
  • Liang et al. (2017) Shiyu Liang, Yixuan Li, and Rayadurgam Srikant. Enhancing the reliability of out-of-distribution image detection in neural networks. arXiv preprint arXiv:1706.02690, 2017.
  • Liang et al. (2018) Shiyu Liang, Yixuan Li, and R Srikant. Enhancing the reliability of out-of-distribution image detection in neural networks. In International Conference on Learning Representations, 2018.
  • Liu et al. (2021) Evan Z Liu, Behzad Haghgoo, Annie S Chen, Aditi Raghunathan, Pang Wei Koh, Shiori Sagawa, Percy Liang, and Chelsea Finn. Just train twice: Improving group robustness without training group information. In International Conference on Machine Learning, pp. 6781–6792. PMLR, 2021.
  • Liu et al. (2015) Ziwei Liu, Ping Luo, Xiaogang Wang, and Xiaoou Tang. Deep learning face attributes in the wild. In Proceedings of the IEEE international conference on computer vision, pp.  3730–3738, 2015.
  • Lloyd et al. (2004) Ricardo V Lloyd, Lori A Erickson, Mary B Casey, King Y Lam, Christine M Lohse, Sylvia L Asa, John KC Chan, Ronald A DeLellis, H Ruben Harach, Kennichi Kakudo, et al. Observer variation in the diagnosis of follicular variant of papillary thyroid carcinoma. The American journal of surgical pathology, 28(10):1336–1340, 2004.
  • Ma et al. (2020) Xingjun Ma, Hanxun Huang, Yisen Wang, Simone Romano, Sarah Erfani, and James Bailey. Normalized loss functions for deep learning with noisy labels. In International conference on machine learning, pp. 6543–6553. PMLR, 2020.
  • Müller et al. (2019) Rafael Müller, Simon Kornblith, and Geoffrey E Hinton. When does label smoothing help? Advances in neural information processing systems, 32, 2019.
  • Nagarajan et al. (2020) Vaishnavh Nagarajan, Anders Andreassen, and Behnam Neyshabur. Understanding the failure modes of out-of-distribution generalization. In International Conference on Learning Representations, 2020.
  • Nam et al. (2020) Junhyun Nam, Hyuntak Cha, Sungsoo Ahn, Jaeho Lee, and Jinwoo Shin. Learning from failure: De-biasing classifier from biased classifier. Advances in Neural Information Processing Systems, 33:20673–20684, 2020.
  • Nam et al. (2022) Junhyun Nam, Jaehyung Kim, Jaeho Lee, and Jinwoo Shin. Spread spurious attribute: Improving worst-group accuracy with spurious attribute estimation. In International Conference on Learning Representations, 2022.
  • Namkoong & Duchi (2017) Hongseok Namkoong and John C Duchi. Variance-based regularization with convex objectives. Advances in neural information processing systems, 30, 2017.
  • Oren et al. (2019) Yonatan Oren, Shiori Sagawa, Tatsunori B Hashimoto, and Percy Liang. Distributionally robust language modeling. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp.  4227–4237, 2019.
  • Özçelik et al. (2021) Rıza Özçelik, Alperen Bağ, Berk Atıl, Arzucan Özgür, and Elif Özkırımlı. Debiaseddta: Model debiasing to boost drug-target affinity prediction. arXiv preprint arXiv:2107.05556, 2021.
  • Öztürk et al. (2018) Hakime Öztürk, Arzucan Özgür, and Elif Ozkirimli. Deepdta: deep drug–target binding affinity prediction. Bioinformatics, 34(17):i821–i829, 2018.
  • Pereyra et al. (2017) Gabriel Pereyra, George Tucker, Jan Chorowski, Lukasz Kaiser, and Geoffrey Hinton. Regularizing neural networks by penalizing confident output distributions. In ICLR, 2017.
  • Reed et al. (2015) Scott E Reed, Honglak Lee, Dragomir Anguelov, Christian Szegedy, Dumitru Erhan, and Andrew Rabinovich. Training deep neural networks on noisy labels with bootstrapping. In ICLR (Workshop), 2015.
  • Ren et al. (2018) Mengye Ren, Wenyuan Zeng, Bin Yang, and Raquel Urtasun. Learning to reweight examples for robust deep learning. In International conference on machine learning, pp. 4334–4343. PMLR, 2018.
  • Ribeiro et al. (2016) Marco Tulio Ribeiro, Sameer Singh, and Carlos Guestrin. ” why should i trust you?” explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, pp.  1135–1144, 2016.
  • Sagawa et al. (2019) Shiori Sagawa, Pang Wei Koh, Tatsunori B Hashimoto, and Percy Liang. Distributionally robust neural networks. In International Conference on Learning Representations, 2019.
  • Sagawa et al. (2020) Shiori Sagawa, Aditi Raghunathan, Pang Wei Koh, and Percy Liang. An investigation of why overparameterization exacerbates spurious correlations. In International Conference on Machine Learning, pp. 8346–8356. PMLR, 2020.
  • Tang et al. (2014) Jing Tang, Agnieszka Szwajda, Sushil Shakyawar, Tao Xu, Petteri Hintsanen, Krister Wennerberg, and Tero Aittokallio. Making sense of large-scale kinase inhibitor bioactivity data sets: a comparative and integrative analysis. Journal of Chemical Information and Modeling, 54(3):735–743, 2014.
  • Tang et al. (2018) Jing Tang, Balaguru Ravikumar, Zaid Alam, Anni Rebane, Markus Vähä-Koskela, Gopal Peddinti, Arjan J van Adrichem, Janica Wakkinen, Alok Jaiswal, Ella Karjalainen, et al. Drug target commons: a community effort to build a consensus knowledge base for drug-target interactions. Cell chemical biology, 25(2):224–229, 2018.
  • Thulasidasan et al. (2019) Sunil Thulasidasan, Tanmoy Bhattacharya, Jeff Bilmes, Gopinath Chennupati, and Jamal Mohd-Yusof. Combating label noise in deep learning using abstention. In International Conference on Machine Learning, pp. 6234–6243. PMLR, 2019.
  • Utama et al. (2020) Prasetya Ajie Utama, Nafise Sadat Moosavi, and Iryna Gurevych. Mind the trade-off: Debiasing nlu models without degrading the in-distribution performance. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp.  8717–8729, 2020.
  • Van Amersfoort et al. (2020) Joost Van Amersfoort, Lewis Smith, Yee Whye Teh, and Yarin Gal. Uncertainty estimation using a single deep deterministic neural network. In International conference on machine learning, pp. 9690–9700. PMLR, 2020.
  • Wah et al. (2011) Catherine Wah, Steve Branson, Peter Welinder, Pietro Perona, and Serge Belongie. The caltech-ucsd birds-200-2011 dataset. 2011.
  • Wei et al. (2020) Hongxin Wei, Lei Feng, Xiangyu Chen, and Bo An. Combating noisy labels by agreement: A joint training method with co-regularization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  13726–13735, 2020.
  • Wei et al. (2021) Jiaheng Wei, Zhaowei Zhu, Hao Cheng, Tongliang Liu, Gang Niu, and Yang Liu. Learning with noisy labels revisited: A study using real-world human annotations. In International Conference on Learning Representations, 2021.
  • Welling & Teh (2011) Max Welling and Yee W Teh. Bayesian learning via stochastic gradient langevin dynamics. In Proceedings of the 28th international conference on machine learning (ICML-11), pp.  681–688. Citeseer, 2011.
  • Wiles et al. (2022) Olivia Wiles, Sven Gowal, Florian Stimberg, Sylvestre Alvise-Rebuffi, Ira Ktena, Taylan Cemgil, et al. A fine-grained analysis on distribution shift. In International Conference on Learning Representations, 2022.
  • Xia et al. (2020) Xiaobo Xia, Tongliang Liu, Bo Han, Chen Gong, Nannan Wang, Zongyuan Ge, and Yi Chang. Robust early-learning: Hindering the memorization of noisy labels. In International conference on learning representations, 2020.
  • Xiao et al. (2020) Kai Xiao, Logan Engstrom, Andrew Ilyas, and Aleksander Madry. Noise or signal: The role of image backgrounds in object recognition. arXiv preprint arXiv:2006.09994, 2020.
  • Xu et al. (2020) Da Xu, Yuting Ye, and Chuanwei Ruan. Understanding the role of importance weighting for deep learning. In International Conference on Learning Representations, 2020.
  • Yang et al. (2022) Xixi Yang, Zhangming Niu, Yuansheng Liu, Bosheng Song, Weiqiang Luins, Li Zeng, and Xiangxiang Zeng. Modality-dta: Multimodality fusion strategy for drug-target affinity prediction. IEEE/ACM Transactions on Computational Biology and Bioinformatics, 2022.
  • Yao et al. (2021) Yazhou Yao, Zeren Sun, Chuanyi Zhang, Fumin Shen, Qi Wu, Jian Zhang, and Zhenmin Tang. Jo-src: A contrastive approach for combating noisy labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  5192–5201, 2021.
  • Zhang et al. (2017) Chiyuan Zhang, Samy Bengio, Moritz Hardt, Benjamin Recht, and Oriol Vinyals. Understanding deep learning requires rethinking generalization. iclr (2016), 2017.
  • Zhang et al. (2016) Jing Zhang, Xindong Wu, and Victor S Sheng. Learning from crowdsourced labeled data: a survey. Artificial Intelligence Review, 46(4):543–576, 2016.
  • Zhang et al. (2020) Jingzhao Zhang, Aditya Krishna Menon, Andreas Veit, Srinadh Bhojanapalli, Sanjiv Kumar, and Suvrit Sra. Coping with label shift via distributionally robust optimisation. In International Conference on Learning Representations, 2020.
  • Zhang & Sabuncu (2018) Zhilu Zhang and Mert Sabuncu. Generalized cross entropy loss for training deep neural networks with noisy labels. Advances in neural information processing systems, 31, 2018.

Appendix

Appendix A Formal theorems and proofs

In this section, we formally state and prove our main theorem and an additional theorem.

  • In A.1, we briefly summarize our conventions on notations.

  • In A.2, we formalize the concept of “risk minimization” on our toy dataset, and provide the solution.

  • In A.3, we list some basic definitions and lemmas that will be used in our theorem statements and proofs.

  • In A.4, we re-state our main theorem formally.

  • In A.5, we prove our main theorem.

  • In A.6, we state the additional theorem.

  • In A.7, we prove the additional theorem.

A.1 Notations

Vectors are written in bold, while (one-dimensional) numbers are not. Random variables are always written in upper case, while plain values are written in lower case except for constants. For example, 𝐗\mathbf{X} is a random vector, while ww is a plain real number. Note that this is a slightly different from the notations in the main paper, where random variables are written in lower case.

X𝒟X\sim\mathcal{D} reads as “XX follows the distribution 𝒟\mathcal{D}” or “XX is sampled from 𝒟\mathcal{D}”. \mathbb{P} stands for the probability, and 𝔼\mathbb{E} stands for the expectation.

A.2 Risk Minimization on the toy dataset

Definition 1 (Risk-minimizing linear solution of a distribution).

Let nn\in\mathbb{N}, and let 𝒟\mathcal{D} be a distribution on n+1\mathbb{R}^{n+1}. We call 𝛃=(β0,β1,,βn)n+1\bm{\beta}^{*}=(\beta_{0},\beta_{1},\cdots,\beta_{n})\in\mathbb{R}^{n+1} the risk-minimizing linear solution of 𝒟\mathcal{D} if it is the solution of the minimization problem

𝜷=argminβ0,β1,,βn𝔼(X1,,Xn,Y)𝒟(β0+i=1nβiXiY)2\bm{\beta}^{*}=\operatorname*{arg\,min}_{\beta_{0},\beta_{1},\cdots,\beta_{n}}\mathbb{E}_{(X_{1},\cdots,X_{n},Y)\sim\mathcal{D}}(\beta_{0}+\sum_{i=1}^{n}\beta_{i}X_{i}-Y)^{2} (9)
Definition 2 (𝐩\mathcal{B}_{\mathbf{p}}).

Let 𝐩=(p1,,pn)[0,1]n\mathbf{p}=(p_{1},\cdots,p_{n})\in[0,1]^{n}. We write 𝐩\mathcal{B}_{\mathbf{p}} for the joint distribution of (X1,,Xn)(X_{1},\cdots,X_{n}) for independent random variables Xi{1,1}X_{i}\in\left\{-1,1\right\} where Xi=1X_{i}=1 with probability pip_{i}. When n=1n=1, we simply write p1=(p1)\mathcal{B}_{p_{1}}=\mathcal{B}_{(p_{1})}

Definition 3 (The toy dataset).

Let 𝐩=(p1,,pn),𝐩=(p1,,pn)[0,1]n\mathbf{p}=(p_{1},\cdots,p_{n}),\mathbf{p}^{\prime}=(p_{1}^{\prime},\cdots,p_{n}^{\prime})\in[0,1]^{n} and 0η10\leq\eta\leq 1. We define a distribution 𝒟=𝒟𝐩,𝐩η\mathcal{D}=\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta} on n+2\mathbb{R}^{n+2} by its sampling procedure:

  1. 1.

    Sample Z0.5Z\sim\mathcal{B}_{0.5} uniformly.

  2. 2.

    If Z=1Z=1, sample 𝐗=(X1,Xn)𝐩\mathbf{X}=(X_{1}\cdots,X_{n})\sim\mathcal{B}_{\mathbf{p}}. Otherwise if Z=1Z=-1, sample 𝐗=(X1,Xn)𝐩\mathbf{X}=(X_{1}\cdots,X_{n})\sim\mathcal{B}_{\mathbf{p}^{\prime}}.

  3. 3.

    With probability 1η1-\eta, let Y=ZY=Z. Otherwise, let Y=ZY=-Z.

  4. 4.

    Output (X1,,Xn,Y,Z)(X_{1},\cdots,X_{n},Y,Z)

Note that the mathematical objects used here can be interpreted as follows:

  • XiX_{i}: ii-th feature of the sample.

  • ZZ: The true label of the sample, either positive(11) or negative(1-1).

  • pip_{i}: The probability that the ii-th feature has value 11 for a positive sample.

  • pip_{i}^{\prime}: The probability that the ii-th feature has value 11 for a negative sample.

  • η\eta: The probability of “label noise”.

  • YY: The post-noise label of the sample

Proposition 1.

Given 𝐩=(p1,,pn),𝐩=(p1,,pn)[0,1]n\mathbf{p}=(p_{1},\cdots,p_{n}),\mathbf{p}^{\prime}=(p_{1}^{\prime},\cdots,p_{n}^{\prime})\in[0,1]^{n} and 0η10\leq\eta\leq 1, let 𝛃=(β0,β1,,βn)\bm{\beta}^{*}=(\beta_{0},\beta_{1},\cdots,\beta_{n}) be the solution of 𝒟𝐩,𝐩η\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta}. Then we have, for some k=k(𝐩,𝐩)>0k=k(\mathbf{p},\mathbf{p}^{\prime})>0,

βi=k(12η)pipi112(2pi1)212(2pi1)2(i=1,,n)\beta_{i}=k(1-2\eta)\frac{p_{i}-p_{i}^{\prime}}{1-\frac{1}{2}(2p_{i}-1)^{2}-\frac{1}{2}(2p_{i}^{\prime}-1)^{2}}\quad\text{($i=1,\cdots,n$)} (10)

and

β0+i=1n(pi+pi1)βi=0\beta_{0}+\sum_{i=1}^{n}(p_{i}+p_{i}^{\prime}-1)\beta_{i}=0 (11)
Proof.

All the expectations and probabilities that follow are with respect to (𝐗,Y,Z)𝒟𝐩,𝐩η(\mathbf{X},Y,Z)\sim\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta}. Let

𝜷=argmin𝜷L(𝜷)=argmin𝜷𝔼(β0+i=1nβiXiY)2\bm{\beta}^{*}=\operatorname*{arg\,min}_{\bm{\beta}}L(\bm{\beta})=\operatorname*{arg\,min}_{\bm{\beta}}\mathbb{E}(\beta_{0}+\sum_{i=1}^{n}\beta_{i}X_{i}-Y)^{2}

Using

  • 𝔼[Xi]=12𝔼[Xi|Z=1]+12𝔼[Xi|Z=1]=pi+pi1\mathbb{E}[X_{i}]=\frac{1}{2}\mathbb{E}[X_{i}|Z=1]+\frac{1}{2}\mathbb{E}[X_{i}|Z=-1]=p_{i}+p_{i}^{\prime}-1

  • 𝔼[Xi|Y=1]=(1η)𝔼[Xi|Y=1,Z=1]+η𝔼[Xi|Y=1,Z=1]=(1η)(pi(1pi))+η(pi(1pi))=2pi(1η)+2piη1\mathbb{E}[X_{i}|Y=1]=(1-\eta)\mathbb{E}[X_{i}|Y=1,Z=1]+\eta\mathbb{E}[X_{i}|Y=1,Z=-1]=(1-\eta)(p_{i}-(1-p_{i}))+\eta(p_{i}^{\prime}-(1-p_{i}^{\prime}))=2p_{i}(1-\eta)+2p_{i}^{\prime}\eta-1

  • 𝔼[Xi|Y=1]=(1η)𝔼[Xi|Y=1,Z=1]+η𝔼[Xi|Y=1,Z=1]=(1η)(pi(1pi))+η(pi(1pi))=2piη+2pi(1η)1\mathbb{E}[X_{i}|Y=-1]=(1-\eta)\mathbb{E}[X_{i}|Y=-1,Z=-1]+\eta\mathbb{E}[X_{i}|Y=-1,Z=1]=(1-\eta)(p_{i}^{\prime}-(1-p_{i}^{\prime}))+\eta(p_{i}-(1-p_{i}))=2p_{i}\eta+2p_{i}^{\prime}(1-\eta)-1

  • For iji\neq j, 𝔼[XiXj]=12(𝔼[XiXj|Z=1]+𝔼[XiXj|Z=1])=12(𝔼[Xi|Z=1]𝔼[Xj|Z=1]+𝔼[Xi|Z=1]𝔼[Xj|Z=1])=12((2pi1)(2pj1)+(2pi1)(2pj1))\mathbb{E}[X_{i}X_{j}]=\frac{1}{2}(\mathbb{E}[X_{i}X_{j}|Z=1]+\mathbb{E}[X_{i}X_{j}|Z=-1])=\frac{1}{2}(\mathbb{E}[X_{i}|Z=1]\mathbb{E}[X_{j}|Z=1]+\mathbb{E}[X_{i}|Z=-1]\mathbb{E}[X_{j}|Z=-1])=\frac{1}{2}((2p_{i}-1)(2p_{j}-1)+(2p_{i}^{\prime}-1)(2p_{j}^{\prime}-1))

  • [Y=1]=[Y=1]=1/2\mathbb{P}[Y=1]=\mathbb{P}[Y=-1]=1/2, 𝔼[Y]=0\mathbb{E}[Y]=0

  • 𝔼[Xi2]=1\mathbb{E}[X_{i}^{2}]=1

, we get

0\displaystyle 0 =12β0L(𝜷)|𝜷=𝜷\displaystyle=\frac{1}{2}\frac{\partial}{\partial\beta_{0}}L(\bm{\beta})\Bigr{|}_{\bm{\beta}=\bm{\beta}^{*}}
=𝔼(β0+i=1nβiXiY)=β0+i=1nβi𝔼[Xi]\displaystyle=\mathbb{E}(\beta_{0}+\sum_{i=1}^{n}\beta_{i}X_{i}-Y)=\beta_{0}+\sum_{i=1}^{n}\beta_{i}\mathbb{E}[X_{i}]
=β0+i=1n(pi+pi1)βi\displaystyle=\beta_{0}+\sum_{i=1}^{n}(p_{i}+p_{i}^{\prime}-1)\beta_{i}

,

β0+i=1n(pi+pi1)βi=0\beta_{0}+\sum_{i=1}^{n}(p_{i}+p_{i}^{\prime}-1)\beta_{i}=0 (12)

(which proves equation 11) and

0\displaystyle 0 =12βiL(𝜷)|𝜷=𝜷\displaystyle=\frac{1}{2}\frac{\partial}{\partial\beta_{i}}L(\bm{\beta})\Bigr{|}_{\bm{\beta}=\bm{\beta}^{*}}
=𝔼Xi(β0+j=1nβjXjY)\displaystyle=\mathbb{E}X_{i}(\beta_{0}+\sum_{j=1}^{n}\beta_{j}X_{j}-Y)
=β0𝔼Xi+βi𝔼Xi2+jiβj𝔼[XiXj]𝔼[XiY]\displaystyle=\beta_{0}\mathbb{E}X_{i}+\beta_{i}\mathbb{E}X_{i}^{2}+\sum_{j\neq i}\beta_{j}\mathbb{E}[X_{i}X_{j}]-\mathbb{E}[X_{i}Y]
=β0𝔼Xi+βi𝔼Xi2+jiβj𝔼[XiXj]12𝔼[Xi|Y=1]+12𝔼[Xi|Y=1]\displaystyle=\beta_{0}\mathbb{E}X_{i}+\beta_{i}\mathbb{E}X_{i}^{2}+\sum_{j\neq i}\beta_{j}\mathbb{E}[X_{i}X_{j}]-\frac{1}{2}\mathbb{E}[X_{i}|Y=1]+\frac{1}{2}\mathbb{E}[X_{i}|Y=-1]
=β0(pi+pi1)+βi+12jiβj((2pi1)(2pj1)+(2pi1)(2pj1))\displaystyle=\beta_{0}(p_{i}+p_{i}^{\prime}-1)+\beta_{i}+\frac{1}{2}\sum_{j\neq i}\beta_{j}((2p_{i}-1)(2p_{j}-1)+(2p_{i}^{\prime}-1)(2p_{j}^{\prime}-1))
(12η)(pipi)\displaystyle-(1-2\eta)(p_{i}-p_{i}^{\prime})

,

β0(pi+pi1)+βi+12jiβj((2pi1)(2pj1)+(2pi1)(2pj1))=(12η)(pipi)\beta_{0}(p_{i}+p_{i}^{\prime}-1)+\beta_{i}+\frac{1}{2}\sum_{j\neq i}\beta_{j}((2p_{i}-1)(2p_{j}-1)+(2p_{i}^{\prime}-1)(2p_{j}^{\prime}-1))=(1-2\eta)(p_{i}-p_{i}^{\prime}) (13)

Letting w=i=1n(2pi1)βiw=\sum_{i=1}^{n}(2p_{i}-1)\beta_{i} and w=i=1n(2pi1)βiw^{\prime}=\sum_{i=1}^{n}(2p_{i}^{\prime}-1)\beta_{i}, we get

β0=w+w2\beta_{0}=-\frac{w+w^{\prime}}{2} (14)

from equation 12 and

β0(pi+pi1)+(pi12)w+(pi12)w+(112(2pi1)212(2pi1)2)βi=(12η)(pipi)\beta_{0}(p_{i}+p_{i}^{\prime}-1)+(p_{i}-\frac{1}{2})w+(p_{i}^{\prime}-\frac{1}{2})w^{\prime}+(1-\frac{1}{2}(2p_{i}-1)^{2}-\frac{1}{2}(2p_{i}^{\prime}-1)^{2})\beta_{i}=(1-2\eta)(p_{i}-p_{i}^{\prime}) (15)

from equation 13. Plugging equation 14 into equation 15, we get

ww2piww2pi+(112(2pi1)212(2pi1)2)βi=(12η)(pipi)\frac{w-w^{\prime}}{2}p_{i}-\frac{w-w^{\prime}}{2}p_{i}^{\prime}+(1-\frac{1}{2}(2p_{i}-1)^{2}-\frac{1}{2}(2p_{i}^{\prime}-1)^{2})\beta_{i}=(1-2\eta)(p_{i}-p_{i}^{\prime})

hence

βi=kpipi112(2pi1)212(2pi1)2\beta_{i}=k^{\prime}\frac{p_{i}-p_{i}^{\prime}}{1-\frac{1}{2}(2p_{i}-1)^{2}-\frac{1}{2}(2p_{i}^{\prime}-1)^{2}} (16)

, where

k=12ηww2.k^{\prime}=1-2\eta-\frac{w-w^{\prime}}{2}. (17)

Plugging equation 16 back into equation 17 using the definitions of ww and ww^{\prime}, we get

k=12η1+i=1n(pipi)2112(2pi1)212(2pi1)2k^{\prime}=\frac{1-2\eta}{1+\sum_{i=1}^{n}\frac{(p_{i}-p_{i}^{\prime})^{2}}{1-\frac{1}{2}(2p_{i}-1)^{2}-\frac{1}{2}(2p_{i}^{\prime}-1)^{2}}}

, so we conclude

βi=k(12η)pipi112(2pi1)212(2pi1)2\beta_{i}=k(1-2\eta)\frac{p_{i}-p_{i}^{\prime}}{1-\frac{1}{2}(2p_{i}-1)^{2}-\frac{1}{2}(2p_{i}^{\prime}-1)^{2}}

where

k=k(𝐩,𝐩)=11+i=1n(pipi)2112(2pi1)212(2pi1)2>0.k=k(\mathbf{p},\mathbf{p}^{\prime})=\frac{1}{1+\sum_{i=1}^{n}\frac{(p_{i}-p_{i}^{\prime})^{2}}{1-\frac{1}{2}(2p_{i}-1)^{2}-\frac{1}{2}(2p_{i}^{\prime}-1)^{2}}}>0.

A.3 Basic lemmas

Definition 4 (cumulative distribution function FF).

Let 𝒟\mathcal{D} be a distribution, and let ff be a real function on the domain of 𝒟\mathcal{D}. Then for yy\in\mathbb{R}, we define

Ff(𝒟)(y)=𝐗𝒟[f(𝐗)y]F_{f(\mathcal{D})}(y)=\mathbb{P}_{\mathbf{X}\in\mathcal{D}}[f(\mathbf{X})\leq y]

We treat “F(f(𝐗))F(f(\mathbf{X}))” as a shorthand of Ff(𝒟)(f(𝐗))F_{f(\mathcal{D})}(f(\mathbf{X})) when the definition of 𝒟\mathcal{D} is clear from the context.

Definition 5 (anti-concentration of a discrete random variable).

Let WW be a discrete random variable. We write

AC(W)=maxw[W=w]AC(W)=\max_{w}\mathbb{P}[W=w]

We make use of the following estimates of [F(W)a)]\mathbb{P}[F(W)\leq a)] and [F(W)1a]\mathbb{P}[F(W)\geq 1-a] for discrete random variables WW throughout our proof:

Lemma 1.

For any discrete random variable WW and any a[0,1]a\in[0,1], we have

aAC(W)<[F(W)a]aa-AC(W)<\mathbb{P}[F(W)\leq a]\leq a (18)

and

aAC(W)<[F(W)1a]aa-AC(W)<\mathbb{P}[F(W)\geq 1-a]\leq a (19)
Proof.

It is enough to prove equation 18. equation 19 can be shown from equation 18 by letting W=WW=-W. Let

w=max{w|F(w)a}w^{*}=\max\left\{w\in\mathbb{R}|F(w)\leq a\right\}

We have

[F(W)a]=[Ww]=F(w)\mathbb{P}[F(W)\leq a]=\mathbb{P}[W\leq w^{*}]=F(w^{*})

Since

F(w)aF(w^{*})\leq a

by the definition of ww^{*}, we have the inequality on the right hand side. To show the inequality on the left hand side, suppose F(w)aAC(W)F(w^{*})\leq a-AC(W). Let ww^{**} be the smallest real number in the range of WW bigger than ww^{*}. By the definition of ww^{*}, we have F(w)>aF(w^{**})>a. Then,

AC(W)[W=w]=[Ww][Ww]=F(w)F(w)>a(aAC(W))=AC(W)AC(W)\geq\mathbb{P}[W=w^{**}]=\mathbb{P}[W\leq w^{**}]-\mathbb{P}[W\leq w^{*}]=F(w^{**})-F(w^{*})>a-(a-AC(W))=AC(W)

, which is a contradiction.

Upper bounds on AC(W)AC(W) are called anti-concentration inequalities in the literature (Krishnapur (2016)). One such inequality is Littlewood-Offord inequality, and a version of it that encompasses our case was proved in Juškevičius & Kurauskas (2019):

Lemma 2.

Let W=i=1nβiXiW=\sum_{i=1}^{n}\beta_{i}X_{i}, where βi0\beta_{i}\neq 0 and 𝐗=(X1,,Xn)𝐩\mathbf{X}=(X_{1},\cdots,X_{n})\sim\mathcal{B}_{\mathbf{p}} for 𝐩[ϵ,1ϵ]n\mathbf{p}\in[\epsilon,1-\epsilon]^{n}. Then we have

AC(W)CnAC(W)\leq\frac{C}{\sqrt{n}}

for some C=C(ϵ)>0C=C(\epsilon)>0.

Proof.

This is a direct consequence of Corollary 2 in Juškevičius & Kurauskas (2019). ∎

An application of Lemma 1 is the following.

Lemma 3.

Let WW be a discrete random variable. Let α=[W<0]\alpha=\mathbb{P}[W<0] and γ=AC(W)\gamma=AC(W). For any δ>0\delta>0, we have

[F(|W|)δ and F(W)>δ]2α+γ\mathbb{P}[F(\left|W\right|)\leq\delta\text{ and }F(W)>\delta]\leq 2\alpha+\gamma

and

[F(W)δ and F(|W|)>δ]α\mathbb{P}[F(W)\leq\delta\text{ and }F(\left|W\right|)>\delta]\leq\alpha

In particular, for any event EE, for any δ>2γ\delta>2\gamma, we have

[E|F(|W|)δ][E|F(W)δ]11+(2α+γ)/(δγ)αδγ\mathbb{P}[E\enskip|\enskip F(\left|W\right|)\leq\delta]\geq\mathbb{P}[E\enskip|\enskip F(W)\leq\delta]\cdot\frac{1}{1+(2\alpha+\gamma)/(\delta-\gamma)}-\frac{\alpha}{\delta-\gamma}
Proof.

We have

[F(|W|)δ and F(W)>δ]\displaystyle\mathbb{P}[F(\left|W\right|)\leq\delta\text{ and }F(W)>\delta] [δ<F(W)δ+α]+[W<0]\displaystyle\leq\mathbb{P}[\delta<F(W)\leq\delta+\alpha]+\mathbb{P}[W<0]
+[W0 and F(|W|)δ and F(W)>δ+α]\displaystyle+\mathbb{P}[W\geq 0\text{ and }F(\left|W\right|)\leq\delta\text{ and }F(W)>\delta+\alpha]
2α+γ+[W0 and F(|W|)δ and F(W)>δ+α]\displaystyle\leq 2\alpha+\gamma+\mathbb{P}[W\geq 0\text{ and }F(\left|W\right|)\leq\delta\text{ and }F(W)>\delta+\alpha]
(by Lemma 1)
=2α+γ\displaystyle=2\alpha+\gamma

, where the last equality is valid because if W0W\geq 0, then F(|W|)δF(\left|W\right|)\leq\delta and F(|W|)>δ+αF(\left|W\right|)>\delta+\alpha cannot hold simultaneously since if they do, δ+α<F(W)=[WW]=[W|W|][W<0]+[|W|<|W|]=α+F(|W|)δ+α\delta+\alpha<F(W)=\mathbb{P}[W^{\prime}\leq W]=\mathbb{P}[W^{\prime}\leq\left|W\right|]\leq\mathbb{P}[W^{\prime}<0]+\mathbb{P}[\left|W^{\prime}\right|<\left|W\right|]=\alpha+F(\left|W\right|)\leq\delta+\alpha, which is a contradiction.

Similarly, we have

[F(W)δ and F(|W|)>δ]\displaystyle\mathbb{P}[F(W)\leq\delta\text{ and }F(\left|W\right|)>\delta] [W<0]+[W0 and F(W)δ and F(|W|)>δ]\displaystyle\leq\mathbb{P}[W<0]+\mathbb{P}[W\geq 0\text{ and }F(W)\leq\delta\text{ and }F(\left|W\right|)>\delta]
=α+[W0 and F(W)δ and F(|W|)>δ]\displaystyle=\alpha+\mathbb{P}[W\geq 0\text{ and }F(W)\leq\delta\text{ and }F(\left|W\right|)>\delta]
=α\displaystyle=\alpha

We can show the remaining statement as follows:

[E|F(|W|)δ]\displaystyle\mathbb{P}[E\enskip|\enskip F(\left|W\right|)\leq\delta] =[E and F(|W|)δ][F(|W|)δ]\displaystyle=\frac{\mathbb{P}[E\text{ and }F(\left|W\right|)\leq\delta]}{\mathbb{P}[F(\left|W\right|)\leq\delta]}
[E and F(W)δ][F(W)δ and F(|W|)>δ][F(|W|)δ]\displaystyle\geq\frac{\mathbb{P}[E\text{ and }F(W)\leq\delta]-\mathbb{P}[F(W)\leq\delta\text{ and }F(\left|W\right|)>\delta]}{\mathbb{P}[F(\left|W\right|)\leq\delta]}
[E and F(W)δ][F(|W|)δ]αδγ\displaystyle\geq\frac{\mathbb{P}[E\text{ and }F(W)\leq\delta]}{\mathbb{P}[F(\left|W\right|)\leq\delta]}-\frac{\alpha}{\delta-\gamma}
[E and F(W)δ][F(W)δ]+[F(|W|)δ and F(W)>δ]αδγ\displaystyle\geq\frac{\mathbb{P}[E\text{ and }F(W)\leq\delta]}{\mathbb{P}[F(W)\leq\delta]+\mathbb{P}[F(\left|W\right|)\leq\delta\text{ and }F(W)>\delta]}-\frac{\alpha}{\delta-\gamma}
[E and F(W)δ][F(W)δ]+2α+γαδγ\displaystyle\geq\frac{\mathbb{P}[E\text{ and }F(W)\leq\delta]}{\mathbb{P}[F(W)\leq\delta]+2\alpha+\gamma}-\frac{\alpha}{\delta-\gamma}
=[E and F(W)δ][F(W)δ](1+(2α+γ)/[F(W)δ])αδγ\displaystyle=\frac{\mathbb{P}[E\text{ and }F(W)\leq\delta]}{\mathbb{P}[F(W)\leq\delta](1+(2\alpha+\gamma)/\mathbb{P}[F(W)\leq\delta])}-\frac{\alpha}{\delta-\gamma}
[E|F(W)δ]11+(2α+γ)/(δγ)αδγ\displaystyle\geq\mathbb{P}[E\enskip|\enskip F(W)\leq\delta]\cdot\frac{1}{1+(2\alpha+\gamma)/(\delta-\gamma)}-\frac{\alpha}{\delta-\gamma}

Also, we need a quantitative version of multi-dimensional central limit theorem:

Lemma 4 (Bentkus (2005)).

Let 𝐕id\mathbf{V}_{i}\in\mathbb{R}^{d} (i=1,,ni=1,\cdots,n) be independent random variables with 𝔼[𝐕i]=0\mathbb{E}[\mathbf{V}_{i}]=0. Let

𝐒=i=1n𝐕i.\mathbf{S}=\sum_{i=1}^{n}\mathbf{V}_{i}.

If the covariance matrix Σ=Σ(𝐒)\Sigma=\Sigma(\mathbf{S}) is invertible, for 𝐙N(0,Σ)\mathbf{Z}\sim N(0,\Sigma), we have

|[𝐒U][𝐙U]|Cγ for any convex U2\left|\mathbb{P}[\mathbf{S}\in U]-\mathbb{P}[\mathbf{Z}\in U]\right|\leq C\gamma\text{ for any convex }U\subseteq\mathbb{R}^{2}

, where CC is an absolute constant and

γ=i=1n𝔼[Σ1𝐕i23].\gamma=\sum_{i=1}^{n}\mathbb{E}[\left\|\Sigma^{-1}\mathbf{V}_{i}\right\|_{2}^{3}].

A.4 The Entropy theorem

Definition 6 (entropy function).

For 𝛃=(β0,,βn)n+1\bm{\beta}=(\beta_{0},\cdots,\beta_{n})\in\mathbb{R}^{n+1}, we define the entropy function H𝛃:nH_{\bm{\beta}}:\mathbb{R}^{n}\to\mathbb{R} as follows:

H𝜷(x1,,xn)=H(exp(β0+i=1nβixi)1+exp(β0+i=1nβixi)),where H(p)=plogp(1p)log(1p)H_{\bm{\beta}}(x_{1},\cdots,x_{n})=H(\frac{\exp(\beta_{0}+\sum_{i=1}^{n}\beta_{i}x_{i})}{1+\exp(\beta_{0}+\sum_{i=1}^{n}\beta_{i}x_{i})}),\quad\text{where }H(p)=-p\log p-(1-p)\log(1-p)

Note that H𝛃(x1,,xn)H_{\bm{\beta}}(x_{1},\cdots,x_{n}) can be expressed as f(|β0+i=1nβixi|)f(\left|\beta_{0}+\sum_{i=1}^{n}\beta_{i}x_{i}\right|) for some monotonically decreasing function ff.

Definition 7 (sign-respecting function).

Let A+={(p,p)[0,1]×[0,1]:p>p}A_{+}=\left\{(p,p^{\prime})\in[0,1]\times[0,1]:p>p^{\prime}\right\} and A={(p,p)[0,1]×[0,1]:p<p}A_{-}=\left\{(p,p^{\prime})\in[0,1]\times[0,1]:p<p^{\prime}\right\}. A function Λ:([0,1]×[0,1]){(x,x)|x[0,1]}=A+A\Lambda:([0,1]\times[0,1])\setminus\left\{(x,x)|x\in[0,1]\right\}=A_{+}\cup A_{-}\to\mathbb{R} will be called sign-respecting if it is continuous on each of A+A_{+} and AA_{-}, positive on A+A_{+} and negative on AA_{-}. For example, f(p,p)=pp|pp|f(p,p^{\prime})=\frac{p-p^{\prime}}{\left|p-p^{\prime}\right|} is sign-respecting.

Definition 8 (spurious cue score function).

Given a sign-respecting function Λ\Lambda, under the distribution defined in Definition 3, we define the spurious cue score function Ψ𝐩,𝐩Λ:n\Psi^{\Lambda}_{\mathbf{p},\mathbf{p}^{\prime}}:\mathbb{R}^{n}\to\mathbb{R} as follows:

Ψ𝐩,𝐩Λ(𝐗)=i=1nΛ(pi,pi)Xi\Psi^{\Lambda}_{\mathbf{p},\mathbf{p}^{\prime}}(\mathbf{X})=\sum_{i=1}^{n}\Lambda(p_{i},p_{i}^{\prime})X_{i}

Roughly speaking, Ψ𝐩,𝐩Λ\Psi^{\Lambda}_{\mathbf{p},\mathbf{p}^{\prime}} measures how easy a given positive sample 𝐗\mathbf{X} is in terms of the number of label-compatible features. Here, a feature XiX_{i} of a positive sample is label-compatible if either Xi=1X_{i}=1 and pi>pip_{i}>p_{i}^{\prime} (i.e. positive samples have higher probability of having Xi=1X_{i}=1) or Xi=1X_{i}=-1 and pi>pip_{i}^{\prime}>p_{i} (i.e. negative samples have higher probability of having Xi=1X_{i}=1). Here, the “number” is weighted in terms of Λ(pi,pi)\Lambda(p_{i},p_{i}^{\prime}).

If we want to measure a similar thing for a negative sample, we can use Ψ𝐩,𝐩\Psi_{\mathbf{p}^{\prime},\mathbf{p}} instead.

Theorem 1 (formal).

Let Λ\Lambda be a sign-respecting function. Given 0<ϵ<1/20<\epsilon<1/2, there exists δ0=δ0(Λ,ϵ)>0\delta_{0}=\delta_{0}(\Lambda,\epsilon)>0 such that for any 0<δδ00<\delta\leq\delta_{0}, for sufficiently large nn (nNn\geq N for some N=N(Λ,ϵ,δ)N=N(\Lambda,\epsilon,\delta)), the following holds:

For any combination of

  • 0η<1/20\leq\eta<1/2

  • 𝐩=(p1,,pn),𝐩=(p1,,pn)[ϵ,1ϵ]n\mathbf{p}=(p_{1},\cdots,p_{n}),\mathbf{p}^{\prime}=(p_{1}^{\prime},\cdots,p_{n}^{\prime})\in[\epsilon,1-\epsilon]^{n} with |pipi|ϵ\left|p_{i}-p_{i}^{\prime}\right|\geq\epsilon (1in1\leq i\leq n)

, when 𝛃=(β0,β1,βn)\bm{\beta}^{*}=(\beta_{0},\beta_{1}\cdots,\beta_{n}) is the risk-minimizing linear solution of (𝐗,Y)(\mathbf{X},Y) for (𝐗,Y,Z)𝒟𝐩,𝐩η(\mathbf{X},Y,Z)\sim\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta}, we have

𝐗,Y,Z𝒟𝐩,𝐩η[R𝐩,𝐩Λ(𝐗,Z)ϵ|F(H𝜷(𝐗))1δ]1ϵ.\mathbb{P}_{\mathbf{X},Y,Z\sim\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta}}[\enskip R^{\Lambda}_{\mathbf{p},\mathbf{p}^{\prime}}(\mathbf{X},Z)\leq\epsilon\enskip|\enskip F(H_{\bm{\beta}^{*}}(\mathbf{X}))\geq 1-\delta\enskip]\geq 1-\epsilon. (20)

, where R𝐩,𝐩ΛR_{\mathbf{p},\mathbf{p}^{\prime}}^{\Lambda} is the spurious cue score rank function defined as

R𝐩,𝐩Λ(𝐗,Z)=𝟏Z=1FΨ𝐩,𝐩(𝐩)(Ψ𝐩,𝐩Λ(𝐗))+𝟏Z=1FΨ𝐩,𝐩Λ(𝐩)(Ψ𝐩,𝐩Λ(𝐗))R_{\mathbf{p},\mathbf{p}^{\prime}}^{\Lambda}(\mathbf{X},Z)=\mathbf{1}_{Z=1}F_{\Psi_{\mathbf{p},\mathbf{p}^{\prime}}(\mathcal{B}_{\mathbf{p}})}(\Psi_{\mathbf{p},\mathbf{p}^{\prime}}^{\Lambda}(\mathbf{X}))+\mathbf{1}_{Z=-1}F_{\Psi_{\mathbf{p}^{\prime},\mathbf{p}}^{\Lambda}(\mathcal{B}_{\mathbf{p}^{\prime}})}(\Psi_{\mathbf{p}^{\prime},\mathbf{p}}^{\Lambda}(\mathbf{X}))

That is, R𝐩,𝐩Λ(𝐗,Z)R_{\mathbf{p},\mathbf{p}^{\prime}}^{\Lambda}(\mathbf{X},Z) is the rank of the spurious cue score of 𝐗\mathbf{X} within its ground truth label ZZ.

This means intuitively that under our dataset generation process (as described by Definition 3) with some minor conditions (Those for 𝐩\mathbf{p} and 𝐩\mathbf{p}^{\prime}), relatively high (top-δ\delta) entropy w.r.t. 𝛃\bm{\beta}^{*} implies relatively small (bottom-ϵ\epsilon within the ground truth label ZZ) spurious cue score with high probability (probability at least 1ϵ1-\epsilon).

A.5 Proof of Theorem 1

A.5.1 A reduction

We reduce Theorem 1 to the following variant:

Lemma 5.

Let Λ\Lambda be a sign-respecting function. Given 0<ϵ<1/20<\epsilon<1/2, there exists δ0=δ0(Λ,ϵ)>0\delta_{0}=\delta_{0}(\Lambda,\epsilon)>0 such that for any 0<δminδ00<\delta_{min}\leq\delta_{0}, for sufficiently large nn (nNn\geq N for some N=N(Λ,ϵ,δmin)N=N(\Lambda,\epsilon,\delta_{min})), the following holds:

For any combination of

  • 0η<1/20\leq\eta<1/2

  • 𝐩=(p1,,pn),𝐩=(p1,,pn)[ϵ,1ϵ]n\mathbf{p}=(p_{1},\cdots,p_{n}),\mathbf{p}^{\prime}=(p_{1}^{\prime},\cdots,p_{n}^{\prime})\in[\epsilon,1-\epsilon]^{n} with |pipi|ϵ\left|p_{i}-p_{i}^{\prime}\right|\geq\epsilon (1in1\leq i\leq n)

  • δ[δmin,δ0]\delta\in[\delta_{min},\delta_{0}]

, when 𝛃=(β0,β1,βn)\bm{\beta}^{*}=(\beta_{0},\beta_{1}\cdots,\beta_{n}) is the risk-minimizing linear solution of (𝐗,Y)(\mathbf{X},Y) for (𝐗,Y,Z)𝒟𝐩,𝐩η(\mathbf{X},Y,Z)\sim\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta},

𝐗𝐩[F(Ψ𝐩,𝐩Λ(𝐗))ϵ|F(H𝜷(𝐗))1δ]1ϵ.\mathbb{P}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}[\enskip F(\Psi^{\Lambda}_{\mathbf{p},\mathbf{p}^{\prime}}(\mathbf{X}))\leq\epsilon\enskip|\enskip F(H_{\bm{\beta}^{*}}(\mathbf{X}))\geq 1-\delta\enskip]\geq 1-\epsilon. (21)

and

𝐗𝐩[F(Ψ𝐩,𝐩Λ(𝐗))ϵ|F(H𝜷(𝐗))1δ]1ϵ.\mathbb{P}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}^{\prime}}}[\enskip F(\Psi^{\Lambda}_{\mathbf{p}^{\prime},\mathbf{p}}(\mathbf{X}))\leq\epsilon\enskip|\enskip F(H_{\bm{\beta}^{*}}(\mathbf{X}))\geq 1-\delta\enskip]\geq 1-\epsilon. (22)

where H𝛃H_{\bm{\beta}^{*}} is the entropy function (Definition 6), Ψ𝐩,𝐩Λ\Psi^{\Lambda}_{\mathbf{p},\mathbf{p}^{\prime}} and Ψ𝐩,𝐩Λ\Psi^{\Lambda}_{\mathbf{p}^{\prime},\mathbf{p}} are the spurious cue score functions (Definition 8) and FF are the cumulative distribution functions(Definition 4) related to them with respect to B𝐩B_{\mathbf{p}} or B𝐩B_{\mathbf{p}^{\prime}}.

The key differences from Theorem 1 are (1) the introduction of δmin\delta_{min} and (2) that equation 20 has been separated into equation 21 and equation 22.

Before presenting the proof of Lemma 5, we will show how to derive Theorem 1 from Lemma 5.

proof of Theorem 1 assuming Lemma 5.

Let Λ\Lambda be a sign-respecting function, and let 0<ϵ<1/20<\epsilon<1/2. Let

δ0=δ0Lemma5(Λ,ϵ2)\delta_{0}^{*}=\delta_{0}^{\hyperref@@ii[main_thm]{Lemma\ref{main_thm}}}(\Lambda,\frac{\epsilon}{2})

and take

δ0=δ03\delta_{0}=\frac{\delta_{0}^{*}}{3}

. Given δ\delta with

0<δ<δ00<\delta<\delta_{0}

, let

δmin=δϵ2\delta_{min}=\frac{\delta\epsilon}{2}

and take

N=max(NLemma5(Λ,ϵ,δmin),(6Cδ0)2,(4Cδ)2)N=\max(N^{\hyperref@@ii[main_thm]{Lemma\ref{main_thm}}}(\Lambda,\epsilon^{\prime},\delta_{min}),(\frac{6C}{\delta_{0}^{*}})^{2},(\frac{4C}{\delta})^{2})

, where C=C(ϵ)C=C(\epsilon) is from Lemma 2. To show that NN satisfies the theorem statement, let nNn\geq N.

Since δminδ0\delta_{min}\leq\delta_{0}^{*}, by Lemma 5, we have

δminδδ0,𝐗𝐩[F(Ψ𝐩,𝐩Λ(𝐗))ϵ2|F(H𝜷(𝐗))1δ]1ϵ2\forall\delta_{min}\leq\delta^{\prime}\leq\delta_{0}^{*},\quad\mathbb{P}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}[\enskip F(\Psi^{\Lambda}_{\mathbf{p},\mathbf{p}^{\prime}}(\mathbf{X}))\leq\frac{\epsilon}{2}\enskip|\enskip F(H_{\bm{\beta}^{*}}(\mathbf{X}))\geq 1-\delta^{\prime}\enskip]\geq 1-\frac{\epsilon}{2} (23)

and

δminδδ0,𝐗𝐩[F(Ψ𝐩,𝐩Λ(𝐗))ϵ2|F(H𝜷(𝐗))1δ]1ϵ2.\forall\delta_{min}\leq\delta^{\prime}\leq\delta_{0}^{*},\quad\mathbb{P}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}^{\prime}}}[\enskip F(\Psi^{\Lambda}_{\mathbf{p}^{\prime},\mathbf{p}}(\mathbf{X}))\leq\frac{\epsilon}{2}\enskip|\enskip F(H_{\bm{\beta}^{*}}(\mathbf{X}))\geq 1-\delta^{\prime}\enskip]\geq 1-\frac{\epsilon}{2}. (24)

Our goal is to show that

Goal: 𝐗,Y,Z𝒟𝐩,𝐩η[R𝐩,𝐩Λ(𝐗,Z))ϵ|F(H𝜷(𝐗))1δ]1ϵ\text{Goal: }\mathbb{P}_{\mathbf{X},Y,Z\sim\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta}}[\enskip R^{\Lambda}_{\mathbf{p},\mathbf{p}^{\prime}}(\mathbf{X},Z))\leq\epsilon\enskip|\enskip F(H_{\bm{\beta}^{*}}(\mathbf{X}))\geq 1-\delta\enskip]\geq 1-\epsilon (25)

, where

R𝐩,𝐩Λ(𝐗,Z)=𝟏Z=1FΨ𝐩,𝐩(𝐩)(Ψ𝐩,𝐩Λ(𝐗))+𝟏Z=1FΨ𝐩,𝐩Λ(𝐩)(Ψ𝐩,𝐩Λ(𝐗)).R_{\mathbf{p},\mathbf{p}^{\prime}}^{\Lambda}(\mathbf{X},Z)=\mathbf{1}_{Z=1}F_{\Psi_{\mathbf{p},\mathbf{p}^{\prime}}(\mathcal{B}_{\mathbf{p}})}(\Psi_{\mathbf{p},\mathbf{p}^{\prime}}^{\Lambda}(\mathbf{X}))+\mathbf{1}_{Z=-1}F_{\Psi_{\mathbf{p}^{\prime},\mathbf{p}}^{\Lambda}(\mathcal{B}_{\mathbf{p}^{\prime}})}(\Psi_{\mathbf{p}^{\prime},\mathbf{p}}^{\Lambda}(\mathbf{X})).

To simplify the equations that follow, let us make the following definitions:

  • a=min{a|FH𝜷(𝒟𝐩,𝐩η)(a)1δ)}a^{*}=\min\left\{a\in\mathbb{R}|F_{H_{\bm{\beta}^{*}}(\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta})}(a)\geq 1-\delta)\right\}

  • p=𝐗,Y,Z𝒟𝐩,𝐩η[Z=1|H𝜷(𝐗)a]p=\mathbb{P}_{\mathbf{X},Y,Z\sim\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta}}[Z=1|H_{\bm{\beta}^{*}}(\mathbf{X})\geq a^{*}]

  • δ1=1FH𝜷(𝐩)(a)\delta_{1}=1-F_{H_{\bm{\beta}^{*}}(\mathcal{B}_{\mathbf{p}})}(a^{*})

  • δ2=1FH𝜷(𝐩)(a)\delta_{2}=1-F_{H_{\bm{\beta}^{*}}(\mathcal{B}_{\mathbf{p}^{\prime}})}(a^{*})

We have the following (in)equalities:

  1. 1.
    𝐗𝐩[F(H𝜷(𝐗))1δ1]=2p𝐗,Y,Z𝒟𝐩,𝐩η[F(H𝜷(𝐗))1δ]\mathbb{P}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}[F(H_{\bm{\beta}^{*}}(\mathbf{X}))\geq 1-\delta_{1}]=2p\mathbb{P}_{\mathbf{X},Y,Z\sim\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta}}[F(H_{\bm{\beta}^{*}}(\mathbf{X}))\geq 1-\delta] (26)

    and

    𝐗𝐩[F(H𝜷(𝐗))1δ2]=2(1p)𝐗,Y,Z𝒟𝐩,𝐩η[F(H𝜷(𝐗))1δ]\mathbb{P}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}^{\prime}}}[F(H_{\bm{\beta}^{*}}(\mathbf{X}))\geq 1-\delta_{2}]=2(1-p)\mathbb{P}_{\mathbf{X},Y,Z\sim\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta}}[F(H_{\bm{\beta}^{*}}(\mathbf{X}))\geq 1-\delta] (27)
    Proof.

    We have

    𝐗𝐩[\displaystyle\mathbb{P}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}[ F(H𝜷(𝐗))1δ1]=𝐗𝐩[H𝜷(𝐗)a]\displaystyle F(H_{\bm{\beta}^{*}}(\mathbf{X}))\geq 1-\delta_{1}]=\mathbb{P}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}[H_{\bm{\beta}^{*}}(\mathbf{X})\geq a^{*}]
    =𝐗,Y,Z𝒟𝐩,𝐩η[H𝜷(𝐗)a|Z=1]\displaystyle=\mathbb{P}_{\mathbf{X},Y,Z\sim\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta}}[H_{\bm{\beta}^{*}}(\mathbf{X})\geq a^{*}|Z=1]
    =𝐗,Y,Z𝒟𝐩,𝐩η[Z=1|H𝜷(𝐗)a]𝐗,Y,Z𝒟𝐩,𝐩η[H𝜷(𝐗)a]𝐗,Y,Z𝒟𝐩,𝐩η[Z=1]\displaystyle=\frac{\mathbb{P}_{\mathbf{X},Y,Z\sim\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta}}[Z=1|H_{\bm{\beta}^{*}}(\mathbf{X})\geq a^{*}]\cdot\mathbb{P}_{\mathbf{X},Y,Z\sim\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta}}[H_{\bm{\beta}^{*}}(\mathbf{X})\geq a^{*}]}{\mathbb{P}_{\mathbf{X},Y,Z\sim\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta}}[Z=1]}
    =2p𝐗,Y,Z𝒟𝐩,𝐩η[H𝜷(𝐗)a]\displaystyle=2p\mathbb{P}_{\mathbf{X},Y,Z\sim\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta}}[H_{\bm{\beta}^{*}}(\mathbf{X})\geq a^{*}]
    =2p𝐗,Y,Z𝒟𝐩,𝐩η[F(H𝜷(𝐗))1δ]\displaystyle=2p\mathbb{P}_{\mathbf{X},Y,Z\sim\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta}}[F(H_{\bm{\beta}^{*}}(\mathbf{X}))\geq 1-\delta]

    , which proves equation 26. equation 27 can be proved similarly. ∎

  2. 2.
    δ1,δ2δ0\delta_{1},\delta_{2}\leq\delta_{0}^{*} (28)
    Proof.

    We have

    δ1AC𝐗𝐩(H𝜷(𝐗))\displaystyle\delta_{1}-AC_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}(H_{\bm{\beta}^{*}}(\mathbf{X})) 𝐗𝐩[F(H𝜷(𝐗))1δ1] (By Lemma 1)\displaystyle\leq\mathbb{P}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}[F(H_{\bm{\beta}^{*}}(\mathbf{X}))\geq 1-\delta_{1}]\text{ (By \hyperref@@ii[F_estimate]{Lemma \ref{F_estimate}})}
    =2p𝐗,Y,Z𝒟𝐩,𝐩η[F(H𝜷(𝐗))1δ] (By equation 26)\displaystyle=2p\mathbb{P}_{\mathbf{X},Y,Z\sim\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta}}[F(H_{\bm{\beta}^{*}}(\mathbf{X}))\geq 1-\delta]\text{ (By equation~{}\ref{delta_one_equality})}
    2pδ (By Lemma 1)\displaystyle\leq 2p\delta\text{ (By \hyperref@@ii[F_estimate]{Lemma \ref{F_estimate}})}
    2δ2δ023δ0.\displaystyle\leq 2\delta\leq 2\delta_{0}\leq\frac{2}{3}\delta_{0}^{*}.

    Since

    AC𝐗𝐩(H𝜷(𝐗))\displaystyle AC_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}(H_{\bm{\beta}^{*}}(\mathbf{X})) =AC𝐗𝐩(|i=1n𝜷iXi|)\displaystyle=AC_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}(\left|\sum_{i=1}^{n}\bm{\beta}^{*}_{i}X_{i}\right|)
    2AC𝐗𝐩(i=1n𝜷iXi)\displaystyle\leq 2AC_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}(\sum_{i=1}^{n}\bm{\beta}^{*}_{i}X_{i})
    2Cn (By Lemma 2)\displaystyle\leq\frac{2C}{\sqrt{n}}\text{ (By \hyperref@@ii[anti-concentration_lemma]{Lemma \ref{anti-concentration_lemma}})}
    δ03 (Since n(6Cδ0)2)\displaystyle\leq\frac{\delta_{0}^{*}}{3}\text{ (Since $n\geq(\frac{6C}{\delta_{0}^{*}})^{2}$)}

    , this proves δ1δ0\delta_{1}\leq\delta_{0}^{*}. δ2δ0\delta_{2}\leq\delta_{0}^{*} can be proved similarly. ∎

  3. 3.
    δ1pδ,δ2(1p)δ\delta_{1}\geq p\delta,\quad\delta_{2}\geq(1-p)\delta (29)
    Proof.

    We have

    δ1\displaystyle\delta_{1} 𝐗𝐩[F(H𝜷(𝐗))1δ1] (By Lemma 1)\displaystyle\geq\mathbb{P}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}[F(H_{\bm{\beta}^{*}}(\mathbf{X}))\geq 1-\delta_{1}]\text{ (By \hyperref@@ii[F_estimate]{Lemma \ref{F_estimate}})}
    =2p𝐗,Y,Z𝒟𝐩,𝐩η[F(H𝜷(𝐗))1δ] (By equation 26)\displaystyle=2p\mathbb{P}_{\mathbf{X},Y,Z\sim\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta}}[F(H_{\bm{\beta}^{*}}(\mathbf{X}))\geq 1-\delta]\text{ (By equation~{}\ref{delta_one_equality})}
    2p(δAC𝐗,Y,Z𝒟𝐩,𝐩η(H𝜷(𝐗))) (By Lemma 1)\displaystyle\geq 2p(\delta-AC_{\mathbf{X},Y,Z\sim\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta}}(H_{\bm{\beta}^{*}}(\mathbf{X})))\text{ (By \hyperref@@ii[F_estimate]{Lemma \ref{F_estimate}})}
    2p(δ12(AC𝐗𝐩(H𝜷(𝐗))+AC𝐗𝐩(H𝜷(𝐗))))\displaystyle\geq 2p(\delta-\frac{1}{2}(AC_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}(H_{\bm{\beta}^{*}}(\mathbf{X}))+AC_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}^{\prime}}}(H_{\bm{\beta}^{*}}(\mathbf{X}))))
    =2p(δ12(AC𝐗𝐩(|i=1n𝜷iXi|)+AC𝐗𝐩(|i=1n𝜷iXi|)))\displaystyle=2p(\delta-\frac{1}{2}(AC_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}(\left|\sum_{i=1}^{n}\bm{\beta}^{*}_{i}X_{i}\right|)+AC_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}^{\prime}}}(\left|\sum_{i=1}^{n}\bm{\beta}^{*}_{i}X_{i}\right|)))
    2p(δ(AC𝐗𝐩(i=1n𝜷iXi)+AC𝐗𝐩(i=1n𝜷iXi)))\displaystyle\geq 2p(\delta-(AC_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}(\sum_{i=1}^{n}\bm{\beta}^{*}_{i}X_{i})+AC_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}^{\prime}}}(\sum_{i=1}^{n}\bm{\beta}^{*}_{i}X_{i})))
    2p(δ2Cn) (By applying Lemma 2 twice)\displaystyle\geq 2p(\delta-\frac{2C}{\sqrt{n}})\text{ (By applying \hyperref@@ii[anti-concentration_lemma]{Lemma \ref{anti-concentration_lemma}} twice)}
    pδ (Since n(4Cδ)2)\displaystyle\geq p\delta\text{ (Since $n\geq(\frac{4C}{\delta})^{2}$)}

    , which proves δ1pδ\delta_{1}\geq p\delta. δ2(1p)δ\delta_{2}\geq(1-p)\delta can be proved similarly. ∎

Using these inequalities, we can show equation 25 as follows:

\displaystyle\mathbb{P} [R𝐩,𝐩Λ(𝐗,Z))𝐗,Y,Z𝒟𝐩,𝐩ηϵ|F(H𝜷(𝐗))1δ]{}_{\mathbf{X},Y,Z\sim\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta}}[\enskip R^{\Lambda}_{\mathbf{p},\mathbf{p}^{\prime}}(\mathbf{X},Z))\leq\epsilon\enskip|\enskip F(H_{\bm{\beta}^{*}}(\mathbf{X}))\geq 1-\delta\enskip]
𝐗,Y,Z𝒟𝐩,𝐩η[R𝐩,𝐩Λ(𝐗,Z))ϵ2|F(H𝜷(𝐗))1δ]\displaystyle\geq\mathbb{P}_{\mathbf{X},Y,Z\sim\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta}}[\enskip R^{\Lambda}_{\mathbf{p},\mathbf{p}^{\prime}}(\mathbf{X},Z))\leq\frac{\epsilon}{2}\enskip|\enskip F(H_{\bm{\beta}^{*}}(\mathbf{X}))\geq 1-\delta\enskip]
=𝐗,Y,Z𝒟𝐩,𝐩η[R𝐩,𝐩Λ(𝐗,Z))ϵ2|H𝜷(𝐗)a] (By definition of a)\displaystyle=\mathbb{P}_{\mathbf{X},Y,Z\sim\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta}}[\enskip R^{\Lambda}_{\mathbf{p},\mathbf{p}^{\prime}}(\mathbf{X},Z))\leq\frac{\epsilon}{2}\enskip|\enskip H_{\bm{\beta}^{*}}(\mathbf{X})\geq a^{*}\enskip]\text{ (By definition of $a^{*}$)}
=p𝐗,Y,Z𝒟𝐩,𝐩η[FΨ𝐩,𝐩(𝐩)(Ψ𝐩,𝐩(𝐗))ϵ2|H𝜷(𝐗)a,Z=1]\displaystyle=p\cdot\mathbb{P}_{\mathbf{X},Y,Z\sim\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta}}[\enskip F_{\Psi_{\mathbf{p},\mathbf{p}^{\prime}}(\mathcal{B}_{\mathbf{p}})}(\Psi_{\mathbf{p},\mathbf{p}^{\prime}}(\mathbf{X}))\leq\frac{\epsilon}{2}\enskip|\enskip H_{\bm{\beta}^{*}}(\mathbf{X})\geq a^{*},Z=1\enskip]
+(1p)𝐗,Y,Z𝒟𝐩,𝐩η[FΨ𝐩,𝐩(𝐩)(Ψ𝐩,𝐩(𝐗))ϵ2|H𝜷(𝐗)a,Z=1]\displaystyle+(1-p)\cdot\mathbb{P}_{\mathbf{X},Y,Z\sim\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta}}[\enskip F_{\Psi_{\mathbf{p}^{\prime},\mathbf{p}}(\mathcal{B}_{\mathbf{p}^{\prime}})}(\Psi_{\mathbf{p}^{\prime},\mathbf{p}}(\mathbf{X}))\leq\frac{\epsilon}{2}\enskip|\enskip H_{\bm{\beta}^{*}}(\mathbf{X})\geq a^{*},Z=-1\enskip]
=p𝐗𝐩[FΨ𝐩,𝐩(𝐩)(Ψ𝐩,𝐩(𝐗))ϵ2|H𝜷(𝐗)a]\displaystyle=p\cdot\mathbb{P}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}[\enskip F_{\Psi_{\mathbf{p},\mathbf{p}^{\prime}}(\mathcal{B}_{\mathbf{p}})}(\Psi_{\mathbf{p},\mathbf{p}^{\prime}}(\mathbf{X}))\leq\frac{\epsilon}{2}\enskip|\enskip H_{\bm{\beta}^{*}}(\mathbf{X})\geq a^{*}\enskip]
+(1p)𝐗𝐩[FΨ𝐩,𝐩(𝐩)(Ψ𝐩,𝐩(𝐗))ϵ2|H𝜷(𝐗)a]\displaystyle+(1-p)\cdot\mathbb{P}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}^{\prime}}}[\enskip F_{\Psi_{\mathbf{p}^{\prime},\mathbf{p}}(\mathcal{B}_{\mathbf{p}^{\prime}})}(\Psi_{\mathbf{p}^{\prime},\mathbf{p}}(\mathbf{X}))\leq\frac{\epsilon}{2}\enskip|\enskip H_{\bm{\beta}^{*}}(\mathbf{X})\geq a^{*}\enskip]
=p𝐗𝐩[FΨ𝐩,𝐩(𝐩)(Ψ𝐩,𝐩(𝐗))ϵ2|F(H𝜷(𝐗))1δ1]\displaystyle=p\cdot\mathbb{P}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}[\enskip F_{\Psi_{\mathbf{p},\mathbf{p}^{\prime}}(\mathcal{B}_{\mathbf{p}})}(\Psi_{\mathbf{p},\mathbf{p}^{\prime}}(\mathbf{X}))\leq\frac{\epsilon}{2}\enskip|\enskip F(H_{\bm{\beta}^{*}}(\mathbf{X}))\geq 1-\delta_{1}\enskip]
+(1p)𝐗𝐩[FΨ𝐩,𝐩(𝐩)(Ψ𝐩,𝐩(𝐗))ϵ2|F(H𝜷(𝐗))1δ2]\displaystyle+(1-p)\cdot\mathbb{P}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}^{\prime}}}[\enskip F_{\Psi_{\mathbf{p}^{\prime},\mathbf{p}}(\mathcal{B}_{\mathbf{p}^{\prime}})}(\Psi_{\mathbf{p}^{\prime},\mathbf{p}}(\mathbf{X}))\leq\frac{\epsilon}{2}\enskip|\enskip F(H_{\bm{\beta}^{*}}(\mathbf{X}))\geq 1-\delta_{2}\enskip]
(1ϵ2)(p𝟏δ1δmin+(1p)𝟏δ2δmin) (By equation 23, equation 24 and equation 28)\displaystyle\geq(1-\frac{\epsilon}{2})(p\mathbf{1}_{\delta_{1}\geq\delta_{min}}+(1-p)\mathbf{1}_{\delta_{2}\geq\delta_{min}})\text{ (By equation~{}\ref{goal1_reminder}, equation~{}\ref{goal2_reminder} and equation~{}\ref{delta_onetwo_upper_bound})}
(1ϵ2)(p𝟏pδmin/δ+(1p)𝟏1pδmin/δ) (By equation 29)\displaystyle\geq(1-\frac{\epsilon}{2})(p\mathbf{1}_{p\geq\delta_{min}/\delta}+(1-p)\mathbf{1}_{1-p\geq\delta_{min}/\delta})\text{ (By equation~{}\ref{delta_onetwo_lower_bound})}
(1ϵ2)(1δmin/δ) (Case analysis based on the magnitude of p)\displaystyle\geq(1-\frac{\epsilon}{2})(1-\delta_{min}/\delta)\text{ (Case analysis based on the magnitude of $p$)}
=(1ϵ2)2\displaystyle=(1-\frac{\epsilon}{2})^{2}
1ϵ\displaystyle\geq 1-\epsilon

A.5.2 Proof of Lemma 5

Now, let us prove Lemma 5. The proof relies on the calculation results of Proposition 1 and the following lemma.

Lemma 6.

Given 0<ϵ<1/20<\epsilon<1/2 and M1M\geq 1, there exists δ0=δ0(ϵ,M)>0\delta_{0}=\delta_{0}(\epsilon,M)>0 such that for any 0<δminδ00<\delta_{min}\leq\delta_{0}, for sufficiently large nn (i.e. nNn\geq N for some N=N(ϵ,M,δmin)N=N(\epsilon,M,\delta_{\min})), the following holds:

For any combination of

  • 𝐩=(p1,,pn)[ϵ,1ϵ]n\mathbf{p}=(p_{1},\cdots,p_{n})\in[\epsilon,1-\epsilon]^{n}

  • a1,,an,b1,,bn[1M,M]a_{1},\cdots,a_{n},b_{1},\cdots,b_{n}\in[\frac{1}{M},M]

  • δ[δmin,δ0]\delta\in[\delta_{min},\delta_{0}]

, we have

𝐗𝐩[F(i=1naiXi)ϵ|F(i=1nbiXi)δ]1ϵ\mathbb{P}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}[F(\sum_{i=1}^{n}a_{i}X_{i})\leq\epsilon\enskip|\enskip F(\sum_{i=1}^{n}b_{i}X_{i})\leq\delta]\geq 1-\epsilon (30)

Before presenting the proof of this lemma, we will show how to conclude Lemma 5 from it.

Let 0<ϵ<1/20<\epsilon<1/2, M1M\geq 1. It is enough to find δ0\delta_{0} that satisfies the theorem statement where only equation 21 is considered but not equation 22. This is because since then a symmetric argument that accounts for equation 22 can be made and we can take the minimum of the both δ0\delta_{0}.

Let

β(a,b)=ab112(2a1)212(2b1)2.\beta(a,b)=\frac{a-b}{1-\frac{1}{2}(2a-1)^{2}-\frac{1}{2}(2b-1)^{2}}.

Take M=M(Λ,ϵ)M=M(\Lambda,\epsilon) such that

1Mmin(a,b)Iϵ|β(a,b)|,min(a,b)Iϵ|Λ(a,b)|andmax(a,b)Iϵ|β(a,b)|,max(a,b)Iϵ|Λ(a,b)|M\frac{1}{M}\leq\min_{(a,b)\in I_{\epsilon}}\left|\beta(a,b)\right|,\min_{(a,b)\in I_{\epsilon}}\left|\Lambda(a,b)\right|\quad\text{and}\quad\max_{(a,b)\in I_{\epsilon}}\left|\beta(a,b)\right|,\max_{(a,b)\in I_{\epsilon}}\left|\Lambda(a,b)\right|\leq M

, where Iϵ={(a,b)[ϵ,1ϵ]×[ϵ,1ϵ]:|ab|ϵ}I_{\epsilon}=\left\{(a,b)\in[\epsilon,1-\epsilon]\times[\epsilon,1-\epsilon]:\left|a-b\right|\geq\epsilon\right\}.

Let δ0=δ0Lemma6(ϵ/2,M)\delta_{0}=\delta_{0}^{\hyperref@@ii[different_linear_sums_lemma]{Lemma\ref{different_linear_sums_lemma}}}(\epsilon/2,M), where δ0Lemma6(ϵ,M)\delta_{0}^{\hyperref@@ii[different_linear_sums_lemma]{Lemma\ref{different_linear_sums_lemma}}}(\epsilon,M) stands for the δ0\delta_{0} found by using Lemma 6. Suppose 0<δminδ00<\delta_{min}\leq\delta_{0}. Suppose δ[δmin,δ0]\delta\in[\delta_{min},\delta_{0}], 𝐩\mathbf{p}, 𝐩\mathbf{p}^{\prime} and η\eta have been chosen, and 𝜷=(β0,,βn)\bm{\beta}^{*}=(\beta_{0},\cdots,\beta_{n}) has been determined accordingly. Then our goal is to show that

Goal: 𝐗𝐩[F(Ψ𝐩,𝐩Λ(𝐗))ϵ|F(H𝜷(𝐗))1δ]1ϵ\text{Goal: }\mathbb{P}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}[\enskip F(\Psi^{\Lambda}_{\mathbf{p},\mathbf{p}^{\prime}}(\mathbf{X}))\leq\epsilon\enskip|\enskip F(H_{\bm{\beta}^{*}}(\mathbf{X}))\geq 1-\delta\enskip]\geq 1-\epsilon

for sufficiently large nn. Letting

Xi=pipi|pipi|Xi(i=1,,n)X_{i}^{\prime}=\frac{p_{i}-p_{i}^{\prime}}{\left|p_{i}-p_{i}^{\prime}\right|}X_{i}\quad(i=1,\cdots,n)

,

α=𝐗𝐩[β0+i=1nβiXi<0]\alpha=\mathbb{P}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}[\beta_{0}+\sum_{i=1}^{n}\beta_{i}X_{i}<0]

and

γ=AC(β0+i=1nβiXi)\gamma=AC(\beta_{0}+\sum_{i=1}^{n}\beta_{i}X_{i})

, we get

\displaystyle\mathbb{P} [F(Ψ𝐩,𝐩Λ(𝐗))ϵ|F(H𝜷(𝐗))1δ]𝐗𝐩{}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}[\enskip F(\Psi^{\Lambda}_{\mathbf{p},\mathbf{p}^{\prime}}(\mathbf{X}))\leq\epsilon\enskip|\enskip F(H_{\bm{\beta}^{*}}(\mathbf{X}))\geq 1-\delta\enskip]
=𝐗𝐩[F(Ψ𝐩,𝐩Λ(𝐗))ϵ|F(|β0+i=1nβiXi|)δ]\displaystyle=\mathbb{P}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}[\enskip F(\Psi^{\Lambda}_{\mathbf{p},\mathbf{p}^{\prime}}(\mathbf{X}))\leq\epsilon\enskip|\enskip F(\left|\beta_{0}+\sum_{i=1}^{n}\beta_{i}X_{i}\right|)\leq\delta\enskip]
αδγ+11+(2α+γ)/(δγ)𝐗𝐩[F(Ψ𝐩,𝐩Λ(𝐗))ϵ|F(β0+i=1nβiXi)δ]\displaystyle\geq-\frac{\alpha}{\delta-\gamma}+\frac{1}{1+(2\alpha+\gamma)/(\delta-\gamma)}\cdot\mathbb{P}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}[\enskip F(\Psi^{\Lambda}_{\mathbf{p},\mathbf{p}^{\prime}}(\mathbf{X}))\leq\epsilon\enskip|\enskip F(\beta_{0}+\sum_{i=1}^{n}\beta_{i}X_{i})\leq\delta\enskip]
(by Lemma3)\displaystyle\quad(\text{by }\hyperref@@ii[almost_positive_lemma]{Lemma\ref{almost_positive_lemma}})
=αδγ+11+(2α+γ)/(δγ)𝐗𝐩[F(Ψ𝐩,𝐩Λ(𝐗))ϵ|F(i=1nβiXi)δ]\displaystyle=-\frac{\alpha}{\delta-\gamma}+\frac{1}{1+(2\alpha+\gamma)/(\delta-\gamma)}\cdot\mathbb{P}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}[\enskip F(\Psi^{\Lambda}_{\mathbf{p},\mathbf{p}^{\prime}}(\mathbf{X}))\leq\epsilon\enskip|\enskip F(\sum_{i=1}^{n}\beta_{i}X_{i})\leq\delta\enskip]
=αδγ+11+(2α+γ)/(δγ)𝐗𝐩[\displaystyle=-\frac{\alpha}{\delta-\gamma}+\frac{1}{1+(2\alpha+\gamma)/(\delta-\gamma)}\cdot\mathbb{P}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}[
F(i=1npipi|pipi|Λ(pi,pi)Xi)ϵ|F(i=1npipi|pipi|β(pi,pi)Xi)δ](by equation10)\displaystyle\enskip F(\sum_{i=1}^{n}\frac{p_{i}-p_{i}^{\prime}}{\left|p_{i}-p_{i}^{\prime}\right|}\Lambda(p_{i},p_{i}^{\prime})X_{i}^{\prime})\leq\epsilon\enskip|\enskip F(\sum_{i=1}^{n}\frac{p_{i}-p_{i}^{\prime}}{\left|p_{i}-p_{i}^{\prime}\right|}\beta(p_{i},p_{i}^{\prime})X_{i}^{\prime})\leq\delta\enskip]\quad(\text{by }equation~{}\ref{main-beta_formula})
αδγ+11+(2α+γ)/(δγ)(1ϵ/2)\displaystyle\geq-\frac{\alpha}{\delta-\gamma}+\frac{1}{1+(2\alpha+\gamma)/(\delta-\gamma)}\cdot(1-\epsilon/2)
(by Lemma 6, provided that nNLemma6(ϵ/2,M,δmin)n\geq N^{\hyperref@@ii[different_linear_sums_lemma]{Lemma\ref{different_linear_sums_lemma}}}(\epsilon/2,M,\delta_{min}))
αδminγ+11+(2α+γ)/(δminγ)(1ϵ/2)\displaystyle\geq-\frac{\alpha}{\delta_{min}-\gamma}+\frac{1}{1+(2\alpha+\gamma)/(\delta_{min}-\gamma)}\cdot(1-\epsilon/2)

In the second last inequality, we could apply Lemma6 provided that nNLemma6(ϵ/2,M,δmin)n\geq N^{\hyperref@@ii[different_linear_sums_lemma]{Lemma\ref{different_linear_sums_lemma}}}(\epsilon/2,M,\delta_{min}) because

  1. 1.

    δ[δmin,δ0Lemma6(ϵ/2,M)]\delta\in[\delta_{min},\delta_{0}^{\hyperref@@ii[different_linear_sums_lemma]{Lemma\ref{different_linear_sums_lemma}}}(\epsilon/2,M)]

  2. 2.

    pipi|pipi|Λ(pi,pi)>0\frac{p_{i}-p_{i}^{\prime}}{\left|p_{i}-p_{i}^{\prime}\right|}\Lambda(p_{i},p_{i}^{\prime})>0 (since Λ\Lambda is sign-respecting) and 1M|Λ(pi,pi)|M\frac{1}{M}\leq\left|\Lambda(p_{i},p_{i}^{\prime})\right|\leq M

  3. 3.

    pipi|pipi|β(pi,pi)>0\frac{p_{i}-p_{i}^{\prime}}{\left|p_{i}-p_{i}^{\prime}\right|}\beta(p_{i},p_{i}^{\prime})>0 and 1M|β(pi,pi)|M\frac{1}{M}\leq\left|\beta(p_{i},p_{i}^{\prime})\right|\leq M

  4. 4.

    𝐗=(X1,,Xn)𝐩′′\mathbf{X}^{\prime}=(X_{1}^{\prime},\cdots,X_{n}^{\prime})\sim\mathcal{B}_{\mathbf{p}^{\prime\prime}}, where pi′′={pi(pi>pi)1pi(pi<pi)[ϵ,1ϵ]p_{i}^{\prime\prime}=\begin{cases}p_{i}\quad(p_{i}>p_{i}^{\prime})\\ 1-p_{i}\quad(p_{i}<p_{i}^{\prime})\end{cases}\in[\epsilon,1-\epsilon] (i=1,,ni=1,\cdots,n).

Now, it is enough to show that the last quantity is 1ϵ\geq 1-\epsilon when nNn\geq N^{\prime} for some N=N(ϵ,δmin)N^{\prime}=N^{\prime}(\epsilon,\delta_{min}). (Then, N=max(N,NLemma6(ϵ/2,M,δmin))N=\max(N^{\prime},N^{\hyperref@@ii[different_linear_sums_lemma]{Lemma\ref{different_linear_sums_lemma}}}(\epsilon/2,M,\delta_{min})) will satisfy the theorem statement) Since γ0\gamma\to 0 as nn\to\infty in the convergence rate that depends only on ϵ\epsilon, it remains to show that the same is true for α\alpha. Using equation 11 and equation 10, we get

𝔼(β0+i=1nβiXi)=β0+i=1n(2pi1)βi=i=1n(pipi)βi=k(12η)i=1n|pipi||β(pi,pi)|k(12η)ϵM(ϵ)n\displaystyle\begin{split}\mathbb{E}(\beta_{0}+\sum_{i=1}^{n}\beta_{i}X_{i})&=\beta_{0}+\sum_{i=1}^{n}(2p_{i}-1)\beta_{i}=\sum_{i=1}^{n}(p_{i}-p_{i}^{\prime})\beta_{i}\\ &=k(1-2\eta)\sum_{i=1}^{n}\left|p_{i}-p_{i}^{\prime}\right|\left|\beta(p_{i},p_{i}^{\prime})\right|\\ &\geq k(1-2\eta)\frac{\epsilon}{M(\epsilon)}n\end{split} (31)

Therefore by Hoeffding’s concentration inequality, since βi[k(12η)M(ϵ),k(12η)M(ϵ)]=[ai,bi]\beta_{i}\in[-k(1-2\eta)M(\epsilon),k(1-2\eta)M(\epsilon)]=[a_{i},b_{i}] (i=1,,ni=1,\cdots,n), we get

α=[β0+i=1nβiXi<0][i=1nβiXi𝔼[i=1nβiXi]<k(12η)ϵM(ϵ)n]exp(2(k(12η)ϵM(ϵ)n)2i=1n(biai)2)=exp(ϵ22M(ϵ)4n)\displaystyle\begin{split}\alpha&=\mathbb{P}[\beta_{0}+\sum_{i=1}^{n}\beta_{i}X_{i}<0]\\ &\leq\mathbb{P}[\sum_{i=1}^{n}\beta_{i}X_{i}-\mathbb{E}[\sum_{i=1}^{n}\beta_{i}X_{i}]<-k(1-2\eta)\frac{\epsilon}{M(\epsilon)}n]\\ &\leq\exp(-\frac{2(k(1-2\eta)\frac{\epsilon}{M(\epsilon)}n)^{2}}{\sum_{i=1}^{n}(b_{i}-a_{i})^{2}})\\ &=\exp(-\frac{\epsilon^{2}}{2M(\epsilon)^{4}}n)\end{split} (32)

, finishing the proof.

A.5.3 Proof of Lemma 6

Now we will prove Lemma 6. We will make a bootstrapping argument composed of two stages: (1) proving the theorem imposing a certain restriction on (p1,,pn,a1,,an,b1,,bn)(p_{1},\cdots,p_{n},a_{1},\cdots,a_{n},b_{1},\cdots,b_{n}) (2) proving the general case using the “restricted theorem”.

First, let us prove the restricted theorem. The restriction will be clarified later on. Let 0<ϵ<1/20<\epsilon<1/2 and M1M\geq 1. Define

α=Φ1(ϵ/2)\alpha=\Phi^{-1}(\epsilon/2)

and

δ0=Φ(M2(αΦ1(1ϵ/2)))\delta_{0}=\Phi(M^{2}(\alpha-\Phi^{-1}(1-\epsilon/2)))

, where Φ\Phi is the cumulative density function of the standard normal distribution. We argue that δ0\delta_{0} satisfies the theorem statement. To show that, take arbitrary δmin(0,δ0]\delta_{min}\in(0,\delta_{0}], δ[δmin,δ0]\delta\in[\delta_{min},\delta_{0}], 𝐩=(p1,,pn)[ϵ,1ϵ]n\mathbf{p}=(p_{1},\cdots,p_{n})\in[\epsilon,1-\epsilon]^{n} and a1,,an,b1,,bn[1M,M]a_{1},\cdots,a_{n},b_{1},\cdots,b_{n}\in[\frac{1}{M},M]. We have to show that

Goal: 𝐗𝐩[F(i=1naiXi)ϵ|F(i=1nbiXi)δ]1ϵ\text{Goal: }\mathbb{P}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}[F(\sum_{i=1}^{n}a_{i}X_{i})\leq\epsilon\enskip|\enskip F(\sum_{i=1}^{n}b_{i}X_{i})\leq\delta]\geq 1-\epsilon (33)

Define

𝐕i=(ai(Xiμ(pi))/i=1nai2σ(pi)2bi(Xiμ(pi))/i=1nbi2σ(pi)2),(i=1,,n)\mathbf{V}_{i}=\begin{pmatrix}a_{i}(X_{i}-\mu(p_{i}))/\sqrt{\sum_{i=1}^{n}a_{i}^{2}\sigma(p_{i})^{2}}\\ b_{i}(X_{i}-\mu(p_{i}))/\sqrt{\sum_{i=1}^{n}b_{i}^{2}\sigma(p_{i})^{2}}\end{pmatrix},\quad(i=1,\cdots,n) (34)

and

𝐒=(S1S2)=i=1n𝐯i\mathbf{S}=\begin{pmatrix}S_{1}\\ S_{2}\end{pmatrix}=\sum_{i=1}^{n}\mathbf{v}_{i}

, where μ(pi)=𝔼[Xi]\mu(p_{i})=\mathbb{E}[X_{i}] and σ(pi)=𝔼[(Xi𝔼Xi)2]\sigma(p_{i})=\mathbb{E}[(X_{i}-\mathbb{E}X_{i})^{2}]. Since 𝐯i\mathbf{v}_{i} are independent and 𝔼[𝐯i]=(00)\mathbb{E}[\mathbf{v}_{i}]=\begin{pmatrix}0\\ 0\end{pmatrix}, we can apply Lemma 4, provided that the covariance matrix Σ=Cov[𝐒]\Sigma=Cov[\mathbf{S}] is invertible. Indeed, we may assume that Σ\Sigma is invertible because Σ\Sigma has the form

Σ=(1κκ1),κ=i=1naibiσ(pi)2i=1nai2σ(pi)2i=1nbi2σ(pi)2\Sigma=\begin{pmatrix}1&\kappa\\ \kappa&1\end{pmatrix},\enskip\kappa=\frac{\sum_{i=1}^{n}a_{i}b_{i}\sigma(p_{i})^{2}}{\sqrt{\sum_{i=1}^{n}a_{i}^{2}\sigma(p_{i})^{2}}\sqrt{\sum_{i=1}^{n}b_{i}^{2}\sigma(p_{i})^{2}}} (35)

and the conclusion of the theorem is trivial when κ=1\kappa=1 (Since then (a1,,an)(b1,,bn)(a_{1},\cdots,a_{n})\sim(b_{1},\cdots,b_{n})). The inverse is:

Σ1=11κ2(1κκ1)\Sigma^{-1}=\frac{1}{1-\kappa^{2}}\begin{pmatrix}1&\kappa\\ \kappa&1\end{pmatrix} (36)

Therefore, applying Lemma 4, we get, for 𝐙N(0,Σ)\mathbf{Z}\sim N(0,\Sigma),

|[𝐒U][𝐙U]|Cγ for any convex U2\left|\mathbb{P}[\mathbf{S}\in U]-\mathbb{P}[\mathbf{Z}\in U]\right|\leq C\gamma\text{ for any convex }U\subseteq\mathbb{R}^{2} (37)

, where CC is an absolute constant and

γ=i=1n𝔼[Σ1𝐕i23].\gamma=\sum_{i=1}^{n}\mathbb{E}[\left\|\Sigma^{-1}\mathbf{V}_{i}\right\|_{2}^{3}]. (38)

κ\kappa has the following lower bound:

κ=i=1naibiσ(pi)2i=1naibi(aibi)σ(pi)2i=1naibi(biai)σ(pi)21maxi=1naibimaxi=1nbiaiM2\kappa=\frac{\sum_{i=1}^{n}a_{i}b_{i}\sigma(p_{i})^{2}}{\sqrt{\sum_{i=1}^{n}a_{i}b_{i}(\frac{a_{i}}{b_{i}})\sigma(p_{i})^{2}}\sqrt{\sum_{i=1}^{n}a_{i}b_{i}(\frac{b_{i}}{a_{i}})\sigma(p_{i})^{2}}}\geq\frac{1}{\sqrt{\max_{i=1}^{n}\frac{a_{i}}{b_{i}}\max_{i=1}^{n}\frac{b_{i}}{a_{i}}}}\geq M^{-2} (39)

On the other hand, an upper bound comes from our “restriction”.

bootstrapping restriction: κ0.99\text{bootstrapping restriction: }\kappa\leq 0.99 (40)

Note that the case where κ>0.99\kappa>0.99 should be, at least intuitvely, easier to prove, since the two events in equation 33 become more correlated in that case. We assume this upper bound temporarily to make sure γ\gamma is bounded from above properly. From equation 36 and equation 40, it can be easily deduced that

γCn\gamma\leq\frac{C^{\prime}}{\sqrt{n}}

for some C=C(ϵ,M)C^{\prime}=C^{\prime}(\epsilon,M).

Now, we’re going to prove equation 33 in light of equation 37. To do so, note that the probability density function of ZZ can be written as:

fZ(z1,z2)\displaystyle f_{Z}(z_{1},z_{2}) =12πdetΣexp(12(z1z2)TΣ1(z1z2))\displaystyle=\frac{1}{2\pi\det\Sigma}\exp(-\frac{1}{2}\begin{pmatrix}z_{1}\\ z_{2}\end{pmatrix}^{T}\Sigma^{-1}\begin{pmatrix}z_{1}\\ z_{2}\end{pmatrix})
=12π(1κ2)exp(12(1κ2)(z122κz1z2+z22))\displaystyle=\frac{1}{2\pi(1-\kappa^{2})}\exp(-\frac{1}{2(1-\kappa^{2})}(z_{1}^{2}-2\kappa z_{1}z_{2}+z_{2}^{2}))

Also, note that from this we can calculate:

αfZ(z1,z2)𝑑z1fZ(z1,z2)𝑑z1=Φ(ακz21κ2).\frac{\int_{-\infty}^{\alpha}f_{Z}(z_{1},z_{2})dz_{1}}{\int_{-\infty}^{\infty}f_{Z}(z_{1},z_{2})dz_{1}}=\Phi(\frac{\alpha-\kappa z_{2}}{\sqrt{1-\kappa^{2}}}). (41)

For sufficiently large nn, we have

𝐗𝐩[\displaystyle\mathbb{P}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}[ F(i=1naiXi)ϵ|F(i=1nbiXi)δ]\displaystyle F(\sum_{i=1}^{n}a_{i}X_{i})\leq\epsilon\enskip|\enskip F(\sum_{i=1}^{n}b_{i}X_{i})\leq\delta]
=𝐗𝐩[F(S1)ϵ|F(S2)δ]\displaystyle=\mathbb{P}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}[F(S_{1})\leq\epsilon\enskip|\enskip F(S_{2})\leq\delta]
=𝐗[F(S1)ϵ and F(S2)δ]/δ (by Lemma 1)\displaystyle=\mathbb{P}_{\mathbf{X}}[F(S_{1})\leq\epsilon\text{ and }F(S_{2})\leq\delta]\enskip/\enskip\delta\text{ (by \hyperref@@ii[F_estimate]{Lemma \ref{F_estimate}})}
=𝐗[𝐗[S1S1]ϵ and 𝐗[S2S2]δ]/δ\displaystyle=\mathbb{P}_{\mathbf{X}}[\mathbb{P}_{\mathbf{X}^{\prime}}[S_{1}^{\prime}\leq S_{1}]\leq\epsilon\text{ and }\mathbb{P}_{\mathbf{X}^{\prime}}[S_{2}^{\prime}\leq S_{2}]\leq\delta]\enskip/\enskip\delta
𝐗[Z[Z1S1]ϵCγ and Z[Z2S2]δCγ]/δ (by equation 37)\displaystyle\geq\mathbb{P}_{\mathbf{X}}[\mathbb{P}_{Z^{\prime}}[Z_{1}^{\prime}\leq S_{1}]\leq\epsilon-C\gamma\text{ and }\mathbb{P}_{Z^{\prime}}[Z_{2}^{\prime}\leq S_{2}]\leq\delta-C\gamma]\enskip/\enskip\delta\text{ (by equation~{}\ref{main-lemma2_normal_estimation})}
=𝐗[S1Φ1(ϵCγ) and S2Φ1(δCγ)]/δ\displaystyle=\mathbb{P}_{\mathbf{X}}[S_{1}\leq\Phi^{-1}(\epsilon-C\gamma)\text{ and }S_{2}\leq\Phi^{-1}(\delta-C\gamma)]\enskip/\enskip\delta
𝐗[S1α and S2Φ1(δCγ)]/δ\displaystyle\geq\mathbb{P}_{\mathbf{X}}[S_{1}\leq\alpha\text{ and }S_{2}\leq\Phi^{-1}(\delta-C\gamma)]\enskip/\enskip\delta
Cγ/δ+Z[Z1α and Z2Φ1(δCγ)]/δ (by equation 37)\displaystyle\geq-C\gamma/\delta+\mathbb{P}_{Z}[Z_{1}\leq\alpha\text{ and }Z_{2}\leq\Phi^{-1}(\delta-C\gamma)]\enskip/\enskip\delta\text{ (by equation~{}\ref{main-lemma2_normal_estimation})}
=Cγ/δ+(1Cγ/δ)Φ1(δCγ)(αfZ(z1,z2)𝑑z1)𝑑z2Φ1(δCγ)(fZ(z1,z2)𝑑z1)𝑑z2\displaystyle=-C\gamma/\delta+(1-C\gamma/\delta)\cdot\frac{\int_{-\infty}^{\Phi^{-1}(\delta-C\gamma)}(\int_{-\infty}^{\alpha}f_{Z}(z_{1},z_{2})dz_{1})dz_{2}}{\int_{-\infty}^{\Phi^{-1}(\delta-C\gamma)}(\int_{-\infty}^{\infty}f_{Z}(z_{1},z_{2})dz_{1})dz_{2}}
Cγ/δ+(1Cγ/δ)infz2Φ1(δCγ)αfZ(z1,z2)𝑑z1fZ(z1,z2)𝑑z1\displaystyle\geq-C\gamma/\delta+(1-C\gamma/\delta)\cdot\inf_{z_{2}\leq\Phi^{-1}(\delta-C\gamma)}\frac{\int_{-\infty}^{\alpha}f_{Z}(z_{1},z_{2})dz_{1}}{\int_{-\infty}^{\infty}f_{Z}(z_{1},z_{2})dz_{1}}
=Cγ/δ+(1Cγ/δ)infz2Φ1(δCγ)Φ(ακz21κ2) (by equation 41)\displaystyle=-C\gamma/\delta+(1-C\gamma/\delta)\cdot\inf_{z_{2}\leq\Phi^{-1}(\delta-C\gamma)}\Phi(\frac{\alpha-\kappa z_{2}}{\sqrt{1-\kappa^{2}}})\text{ (by equation~{}\ref{main-lemma2_normal_conditional_cdf})}
=Cγ/δ+(1Cγ/δ)Φ(ακΦ1(δCγ)1κ2)\displaystyle=-C\gamma/\delta+(1-C\gamma/\delta)\cdot\Phi(\frac{\alpha-\kappa\Phi^{-1}(\delta-C\gamma)}{\sqrt{1-\kappa^{2}}})

Now, let us further estimate the last term:

Φ(ακΦ1(δCγ)1κ2)\displaystyle\Phi(\frac{\alpha-\kappa\Phi^{-1}(\delta-C\gamma)}{\sqrt{1-\kappa^{2}}}) Φ(ακΦ1(δ0)1κ2)\displaystyle\geq\Phi(\frac{\alpha-\kappa\Phi^{-1}(\delta_{0})}{\sqrt{1-\kappa^{2}}})
=Φ(ακM2(αΦ1(1ϵ/2))1κ2)\displaystyle=\Phi(\frac{\alpha-\kappa M^{2}(\alpha-\Phi^{-1}(1-\epsilon/2))}{\sqrt{1-\kappa^{2}}})
Φ(α(αΦ1(1ϵ/2))1κ2)\displaystyle\geq\Phi(\frac{\alpha-(\alpha-\Phi^{-1}(1-\epsilon/2))}{\sqrt{1-\kappa^{2}}})
(αΦ(1ϵ/2) is negative + equation39)\displaystyle\quad(\alpha-\Phi(1-\epsilon/2)\text{ is negative + }equation~{}\ref{main-lemma2_kappa_lower_bound})
1ϵ/2\displaystyle\geq 1-\epsilon/2

Therefore continuing the previous chain of inequalities,

𝐗𝐩[\displaystyle\mathbb{P}_{\mathbf{X}\sim\mathcal{B}_{\mathbf{p}}}[ F(i=1naiXi)ϵ|F(i=1nbiXi)δ]\displaystyle F(\sum_{i=1}^{n}a_{i}X_{i})\leq\epsilon\enskip|\enskip F(\sum_{i=1}^{n}b_{i}X_{i})\leq\delta]
Cγ/δ+(1Cγ/δ)(1ϵ/2)\displaystyle\geq-C\gamma/\delta+(1-C\gamma/\delta)\cdot(1-\epsilon/2)
Cδ1C/n+(1Cδ1C/n)(1ϵ/2)\displaystyle\geq-C\delta^{-1}C^{\prime}/\sqrt{n}+(1-C\delta^{-1}C^{\prime}/\sqrt{n})\cdot(1-\epsilon/2)
Cδmin1C/n+(1Cδmin1C/n)(1ϵ/2)\displaystyle\geq-C\delta_{min}^{-1}C^{\prime}/\sqrt{n}+(1-C\delta_{min}^{-1}C^{\prime}/\sqrt{n})\cdot(1-\epsilon/2)
1ϵ\displaystyle\geq 1-\epsilon

, when nNn\geq N for some NN that depends only on ϵ\epsilon, MM and δmin\delta_{min}. This finishes the proof with the restriction of equation 40.

Turning to the general case, let 0<ϵ<1/20<\epsilon<1/2 and M1M\geq 1. We take

δ0=min(δ0bootstrap(ϵ,M),δ0bootstrap(ϵ/2,2M)).\delta_{0}=\min(\delta_{0}^{\text{bootstrap}}(\epsilon,M),\delta_{0}^{\text{bootstrap}}(\epsilon/2,2M)).

We claim that δ0\delta_{0} satisfies the theorem statement. Let δmin(0,δ0]\delta_{min}\in(0,\delta_{0}], δ[δmin,δ0]\delta\in[\delta_{min},\delta_{0}], 𝐩=(p1,,pn)[ϵ,1ϵ]n\mathbf{p}=(p_{1},\cdots,p_{n})\in[\epsilon,1-\epsilon]^{n} and 1Ma1,,anM\frac{1}{M}\leq a_{1},\cdots,a_{n}\leq M. Take

N=max(Nbootstrap(ϵ,M,δmin),Nbootstrap(ϵ/2,2M,δmin)).N=\max(N^{\text{bootstrap}}(\epsilon,M,\delta_{min}),N^{\text{bootstrap}}(\epsilon/2,2M,\delta_{min})).

We may assume that

κ=i=1naibiσ(pi)2i=1nai2σ(pi)2i=1nbi2σ(pi)2>0.99\kappa=\frac{\sum_{i=1}^{n}a_{i}b_{i}\sigma(p_{i})^{2}}{\sqrt{\sum_{i=1}^{n}a_{i}^{2}\sigma(p_{i})^{2}}\sqrt{\sum_{i=1}^{n}b_{i}^{2}\sigma(p_{i})^{2}}}>0.99

, since otherwise, we can rely on δ[δmin,δ0bootstrap(ϵ,M)]\delta\in[\delta_{min},\delta_{0}^{\text{bootstrap}}(\epsilon,M)] to conclude.

We will construct 12Ma1,,an2M\frac{1}{2M}\leq a_{1}^{\prime},\cdots,a_{n}^{\prime}\leq 2M and 12Ma1′′,,an′′2M\frac{1}{2M}\leq a_{1}^{\prime\prime},\cdots,a_{n}^{\prime\prime}\leq 2M such that

κ=i=1naibiσ(pi)2i=1nai2σ(pi)2i=1nbi2σ(pi)20.99\kappa^{\prime}=\frac{\sum_{i=1}^{n}a_{i}^{\prime}b_{i}\sigma(p_{i})^{2}}{\sqrt{\sum_{i=1}^{n}a_{i}^{\prime 2}\sigma(p_{i})^{2}}\sqrt{\sum_{i=1}^{n}b_{i}^{2}\sigma(p_{i})^{2}}}\leq 0.99 (42)

,

κ′′=i=1nai′′biσ(pi)2i=1nai′′2σ(pi)2i=1nbi2σ(pi)20.99\kappa^{\prime\prime}=\frac{\sum_{i=1}^{n}a_{i}^{\prime\prime}b_{i}\sigma(p_{i})^{2}}{\sqrt{\sum_{i=1}^{n}a_{i}^{\prime\prime 2}\sigma(p_{i})^{2}}\sqrt{\sum_{i=1}^{n}b_{i}^{2}\sigma(p_{i})^{2}}}\leq 0.99 (43)

and

3ai=ai+ai′′(1in)3a_{i}=a_{i}^{\prime}+a_{i}^{\prime\prime}\quad(1\leq i\leq n) (44)

. Then since δ[δmin,δ0bootstrap(ϵ/2,2M)]\delta\in[\delta_{min},\delta_{0}^{\text{bootstrap}}(\epsilon/2,2M)], we get

p=[F(i=1naiXi)ϵ/2|F(i=1nbiXi)δ]1ϵ/2p^{\prime}=\mathbb{P}[F(\sum_{i=1}^{n}a_{i}^{\prime}X_{i})\leq\epsilon/2\enskip|\enskip F(\sum_{i=1}^{n}b_{i}X_{i})\leq\delta]\geq 1-\epsilon/2

and

p′′=[F(i=1nai′′Xi)ϵ/2|F(i=1nbiXi)δ]1ϵ/2p^{\prime\prime}=\mathbb{P}[F(\sum_{i=1}^{n}a_{i}^{\prime\prime}X_{i})\leq\epsilon/2\enskip|\enskip F(\sum_{i=1}^{n}b_{i}X_{i})\leq\delta]\geq 1-\epsilon/2

for nNn\geq N. Then we can finish the proof as follows:

[\displaystyle\mathbb{P}[ F(i=1naiXi)ϵ|F(i=1nbiXi)δ]\displaystyle F(\sum_{i=1}^{n}a_{i}X_{i})\leq\epsilon\enskip|\enskip F(\sum_{i=1}^{n}b_{i}X_{i})\leq\delta]
=[F(i=1n3aiXi)ϵ|F(i=1nbiXi)δ]\displaystyle=\mathbb{P}[F(\sum_{i=1}^{n}3a_{i}X_{i})\leq\epsilon\enskip|\enskip F(\sum_{i=1}^{n}b_{i}X_{i})\leq\delta]
=[F(i=1n(ai+ai′′)Xi)ϵ|F(i=1nbiXi)δ]\displaystyle=\mathbb{P}[F(\sum_{i=1}^{n}(a_{i}^{\prime}+a_{i}^{\prime\prime})X_{i})\leq\epsilon\enskip|\enskip F(\sum_{i=1}^{n}b_{i}X_{i})\leq\delta]
[F(i=1naiXi)+F(i=1nai′′Xi)ϵ|F(i=1nbiXi)δ]\displaystyle\geq\mathbb{P}[F(\sum_{i=1}^{n}a_{i}^{\prime}X_{i})+F(\sum_{i=1}^{n}a_{i}^{\prime\prime}X_{i})\leq\epsilon\enskip|\enskip F(\sum_{i=1}^{n}b_{i}X_{i})\leq\delta]
[F(i=1naiXi)ϵ/2 and F(i=1nai′′Xi)ϵ/2|F(i=1nbiXi)δ]\displaystyle\geq\mathbb{P}[F(\sum_{i=1}^{n}a_{i}^{\prime}X_{i})\leq\epsilon/2\text{ and }F(\sum_{i=1}^{n}a_{i}^{\prime\prime}X_{i})\leq\epsilon/2\enskip|\enskip F(\sum_{i=1}^{n}b_{i}X_{i})\leq\delta]
p+p′′1\displaystyle\geq p^{\prime}+p^{\prime\prime}-1
(1ϵ/2)+(1ϵ/2)1=1ϵ\displaystyle\geq(1-\epsilon/2)+(1-\epsilon/2)-1=1-\epsilon

In the middle, we relied on the inequality F(Y1+Y2)F(Y1)+F(Y2)F(Y_{1}+Y_{2})\leq F(Y_{1})+F(Y_{2}) where Y1=i=1naiXiY_{1}=\sum_{i=1}^{n}a_{i}^{\prime}X_{i} and Y2=i=1nai′′XiY_{2}=\sum_{i=1}^{n}a_{i}^{\prime\prime}X_{i}, which can be easily checked.

Now it’s enough to construct ai,ai′′a_{i}^{\prime},a_{i}^{\prime\prime} that satisfy equation 42, equation 43 and equation 44. If we find S{1,,n}S\subseteq\left\{1,\cdots,n\right\} such that

iSaibiσ(pi)12i=1naibiσ(pi)2\sum_{i\in S}a_{i}b_{i}\sigma(p_{i})\simeq\frac{1}{2}\sum_{i=1}^{n}a_{i}b_{i}\sigma(p_{i})^{2} (45)

and

iSai2σ(pi)12i=1nai2σ(pi)2\sum_{i\in S}a_{i}^{2}\sigma(p_{i})\simeq\frac{1}{2}\sum_{i=1}^{n}a_{i}^{2}\sigma(p_{i})^{2} (46)

, we can let

ai={2ai(iS)ai(iSc) andai′′={ai(iS)2ai(iSc)a_{i}^{\prime}=\begin{cases}2a_{i}\quad(i\in S)\\ a_{i}\quad(i\in S^{c})\end{cases}\text{ and}\quad a_{i}^{\prime\prime}=\begin{cases}a_{i}\quad(i\in S)\\ 2a_{i}\quad(i\in S^{c})\end{cases}

, which leads to:

κ\displaystyle\kappa^{\prime} =i=1naibiσ(pi)2i=1nai2σ(pi)2i=1nbi2σ(pi)2\displaystyle=\frac{\sum_{i=1}^{n}a_{i}^{\prime}b_{i}\sigma(p_{i})^{2}}{\sqrt{\sum_{i=1}^{n}a_{i}^{\prime 2}\sigma(p_{i})^{2}}\sqrt{\sum_{i=1}^{n}b_{i}^{2}\sigma(p_{i})^{2}}}
=2iSaibiσ(pi)2+iScaibiσ(pi)24iSai2σ(pi)2+iScai2σ(pi)2i=1nbi2σ(pi)2\displaystyle=\frac{2\sum_{i\in S}a_{i}b_{i}\sigma(p_{i})^{2}+\sum_{i\in S^{c}}a_{i}b_{i}\sigma(p_{i})^{2}}{\sqrt{4\sum_{i\in S}a_{i}^{2}\sigma(p_{i})^{2}+\sum_{i\in S^{c}}a_{i}^{\prime 2}\sigma(p_{i})^{2}}\sqrt{\sum_{i=1}^{n}b_{i}^{2}\sigma(p_{i})^{2}}}
i=1n(212+12)aibiσ(pi)2(412+12)i=1nai2σ(pi)2i=1nbi2σ(pi)2\displaystyle\sim\frac{\sum_{i=1}^{n}(2\cdot\frac{1}{2}+\frac{1}{2})a_{i}b_{i}\sigma(p_{i})^{2}}{\sqrt{(4\cdot\frac{1}{2}+\frac{1}{2})\sum_{i=1}^{n}a_{i}^{2}\sigma(p_{i})^{2}}\sqrt{\sum_{i=1}^{n}b_{i}^{2}\sigma(p_{i})^{2}}}
=1.52.5κ1.52.5<0.99\displaystyle=\frac{1.5}{\sqrt{2.5}}\kappa\leq\frac{1.5}{\sqrt{2.5}}<0.99

SS that satisfies equation 45 and equation 46 can found, for example, by using a probabilistic argument. (Define a random set and show that it satisfies the desired properties with high probability using Hoeffding’s concentration inequality)

A.6 The error theorem

Definition 9.

A function Err:×{1,1}Err:\mathbb{R}\times\left\{-1,1\right\}\to\mathbb{R} will be called a good error function if it satisfies the following properties:

  1. 1.

    For any (y^1,y1),(y^2,y2)×{1,1}(\hat{y}_{1},y_{1}),(\hat{y}_{2},y_{2})\in\mathbb{R}\times\left\{-1,1\right\},

    y^1y1>0,y^2y20Err(y^1,y1)<Err(y^2,y2)\hat{y}_{1}y_{1}>0,\hat{y}_{2}y_{2}\leq 0\Rightarrow Err(\hat{y}_{1},y_{1})<Err(\hat{y}_{2},y_{2})
  2. 2.

    For any y^1,y^2\hat{y}_{1},\hat{y}_{2}\in\mathbb{R} and y{1,1}y\in\left\{-1,1\right\}

    y^1y^2Err(y^1,y)Err(y^2,y)\hat{y}_{1}\neq\hat{y}_{2}\Rightarrow Err(\hat{y}_{1},y)\neq Err(\hat{y}_{2},y)

That is,

  1. 1.

    Predictions that have the same sign with the target label always have lower error values than those that do not.

  2. 2.

    Different predictions on inputs that have the same target label result in distinguishable error values.

Theorem 2.

Let ErrErr be a good error function. Given 0<ϵ<1/20<\epsilon<1/2 and 0η<1/20\leq\eta<1/2, there exists C=C(ϵ,η)>0C=C(\epsilon,\eta)>0 such that for any 0<θ10<\theta\leq 1, for sufficiently large nn, the following holds:

For any combination of 𝐩=(p1,,pn)\mathbf{p}=(p_{1},\cdots,p_{n}) and 𝐩=(p1,,pn)\mathbf{p}^{\prime}=(p_{1}^{\prime},\cdots,p_{n}^{\prime}) in [ϵ,1ϵ]n[\epsilon,1-\epsilon]^{n} with |pipi|ϵ\left|p_{i}-p_{i}^{\prime}\right|\geq\epsilon (1in1\leq i\leq n), when 𝛃=(β0,β1,βn)\bm{\beta}^{*}=(\beta_{0},\beta_{1}\cdots,\beta_{n}) is the risk-minimizing linear solution of (𝐗,Y)(\mathbf{X},Y) for (𝐗,Y,Z)𝒟𝐩,𝐩η(\mathbf{X},Y,Z)\sim\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta}, we have

𝐗,Y,Z𝒟𝐩,𝐩η[YZ=1|F(Err(Y^,Y))1θ]min(1,ηθ)Cn\mathbb{P}_{\mathbf{X},Y,Z\sim\mathcal{D}_{\mathbf{p},\mathbf{p}^{\prime}}^{\eta}}[\enspace YZ=-1\enspace|\enspace F(Err(\hat{Y},Y))\geq 1-\theta\enspace]\geq\min(1,\frac{\eta}{\theta})-\frac{C}{\sqrt{n}}

, where Y^=β0+i=1nβiXi\hat{Y}=\beta_{0}+\sum_{i=1}^{n}\beta_{i}X_{i}.

A.7 Proof of Theorem 2

In proving Theorem 2, we tackle the different ranges of θ\theta (1) θ>η\theta>\eta, (2) θ<η\theta<\eta (3) θ=η\theta=\eta separately. In all cases, we need the following facts:

  • When

    ξ=[Y^Z0]\xi=\mathbb{P}[\hat{Y}Z\leq 0]

    , we have

    ξexp(C1n)\xi\leq\exp(-C_{1}n) (47)

    for some C1=C1(ϵ)C_{1}=C_{1}(\epsilon).

    Proof.

    Use Hoeffding’s inequality in conjunction with equation 11 and equation 10 in the same manner as in equation 31 and equation 32. ∎

  • When

    η0=[Y^Y0]\eta_{0}=\mathbb{P}[\hat{Y}Y\leq 0]

    , we have

    |η0η|ξ\left|\eta_{0}-\eta\right|\leq\xi (48)
    Proof.
    |η0η|\displaystyle\left|\eta_{0}-\eta\right| =|[Y^Y0][YZ=1]|\displaystyle=\left|\mathbb{P}[\hat{Y}Y\leq 0]-\mathbb{P}[YZ=-1]\right|
    [(Y^Y0YZ=1)(Y^Y>0YZ=1)] (symmetric difference)\displaystyle\leq\mathbb{P}[(\hat{Y}Y\leq 0\land YZ=1)\lor(\hat{Y}Y>0\land YZ=-1)]\text{ (symmetric difference)}
    [Y^Z0] (logical implication)\displaystyle\leq\mathbb{P}[\hat{Y}Z\leq 0]\text{ (logical implication)}
    =ξ\displaystyle=\xi

  • When

    γ=AC(Err(Y^,Y)) (See Lemma 1)\gamma=AC(Err(\hat{Y},Y))\text{ (See \hyperref@@ii[F_estimate]{Lemma \ref{F_estimate}})}

    , we have

    γC2n\gamma\leq\frac{C_{2}}{\sqrt{n}} (49)

    for some C2=C2(ϵ)C_{2}=C_{2}(\epsilon).

    Proof.
    γ\displaystyle\gamma =maxx[Err(Y^,Y)=x]\displaystyle=\max_{x\in\mathbb{R}}\mathbb{P}[Err(\hat{Y},Y)=x]
    maxx([Err(Y^,0)=x]+[Err(Y^,1)=x])\displaystyle\leq\max_{x\in\mathbb{R}}(\mathbb{P}[Err(\hat{Y},0)=x]+\mathbb{P}[Err(\hat{Y},1)=x])
    AC(Err(Y^,0))+AC(Err(Y^,1))\displaystyle\leq AC(Err(\hat{Y},0))+AC(Err(\hat{Y},1))
    AC(Y^)+AC(Y^) (Since Err is “good”)\displaystyle\leq AC(\hat{Y})+AC(\hat{Y})\text{ (Since $Err$ is ``good")}
    =2AC(i=1nβiXi)\displaystyle=2AC(\sum_{i=1}^{n}\beta_{i}X_{i})
    C2n (equation 10Lemma 2)\displaystyle\leq\frac{C_{2}}{\sqrt{n}}\text{ (equation~{}\ref{main-beta_formula}, \hyperref@@ii[anti-concentration_lemma]{Lemma \ref{anti-concentration_lemma}})}

A.7.1 The case θ>η\theta>\eta

Assume θ>η\theta>\eta. From equation 48, equation 47 and equation 49, we see that

θη0+γ\theta\geq\eta_{0}+\gamma (50)

when nn is sufficiently large.

First, we argue that

Claim: Y^Y0F(Err(Y^,Y))1θ.\text{Claim: }\hat{Y}Y\leq 0\Rightarrow F(Err(\hat{Y},Y))\geq 1-\theta. (51)

Suppose otherwise. Then it happens with nonzero probability that Y^Y0\hat{Y}Y\leq 0 and F(Err(Y^,Y))<1θF(Err(\hat{Y},Y))<1-\theta. That is,

(y^1,y1),y^1y10F(Err(y1^,y1))<1θ.\exists(\hat{y}_{1},y_{1}),\hat{y}_{1}y_{1}\leq 0\land F(Err(\hat{y_{1}},y_{1}))<1-\theta. (52)

On the other hand, we have

[\displaystyle\mathbb{P}[ F(Err(Y^,Y))1θY^Y>0]\displaystyle F(Err(\hat{Y},Y))\geq 1-\theta\land\hat{Y}Y>0]
[F(Err(Y^,Y))1θ]η0\displaystyle\geq\mathbb{P}[F(Err(\hat{Y},Y))\geq 1-\theta]-\eta_{0}
>θγη0 (by Lemma 1)\displaystyle>\theta-\gamma-\eta_{0}\text{ (by \hyperref@@ii[F_estimate]{Lemma \ref{F_estimate}})}
0 (by equation 50)\displaystyle\geq 0\text{ (by equation~{}\ref{error_thm_case1_ineq})}

Since the event F(Err(Y^,Y))1θY^Y>0F(Err(\hat{Y},Y))\geq 1-\theta\land\hat{Y}Y>0 occurs with a nonzero probability, we have

(y^2,y2),y^2y2>0F(Err(y2^,y2))1θ.\exists(\hat{y}_{2},y_{2}),\hat{y}_{2}y_{2}>0\land F(Err(\hat{y_{2}},y_{2}))\geq 1-\theta. (53)

From equation 52 and equation 53, since y^1y10\hat{y}_{1}y_{1}\leq 0, y^2y2>0\hat{y}_{2}y_{2}>0 but Err(y^1,y1)<Err(y^2,y2)Err(\hat{y}_{1},y_{1})<Err(\hat{y}_{2},y_{2}), we find a contradiction with the definition of “good error function”.

Now, we have

[\displaystyle\mathbb{P}[ YZ=1|F(Err(Y^,Y))1θ]\displaystyle YZ=-1|F(Err(\hat{Y},Y))\geq 1-\theta]
=[YZ=1F(Err(Y^,Y))1θ]][F(Err(Y^,Y))1θ]]\displaystyle=\frac{\mathbb{P}[YZ=-1\land F(Err(\hat{Y},Y))\geq 1-\theta]]}{\mathbb{P}[F(Err(\hat{Y},Y))\geq 1-\theta]]}
[YZ=1Y^Y0]][F(Err(Y^,Y))1θ]] (by equation 51)\displaystyle\geq\frac{\mathbb{P}[YZ=-1\land\hat{Y}Y\leq 0]]}{\mathbb{P}[F(Err(\hat{Y},Y))\geq 1-\theta]]}\text{ (by equation~{}\ref{error_thm_case1_logical_claim})}
[YZ=1Y^Y0]]θ (by Lemma 1)\displaystyle\geq\frac{\mathbb{P}[YZ=-1\land\hat{Y}Y\leq 0]]}{\theta}\text{ (by \hyperref@@ii[F_estimate]{Lemma \ref{F_estimate}})}
=[YZ=1Y^Z0]]θ\displaystyle=\frac{\mathbb{P}[YZ=-1\land\hat{Y}Z\geq 0]]}{\theta}
[YZ=1][Y^Z<0]]θηξθ\displaystyle\geq\frac{\mathbb{P}[YZ=-1]-\mathbb{P}[\hat{Y}Z<0]]}{\theta}\geq\frac{\eta-\xi}{\theta}
ηθ1θexp(C1n) (by equation 47)\displaystyle\geq\frac{\eta}{\theta}-\frac{1}{\theta}\exp(-C_{1}n)\text{ (by equation~{}\ref{error_thm_fact1})}

A.7.2 The case θ<η\theta<\eta

Assume θ<η\theta<\eta. From equation 48 and equation 47, we see that

θ<η0\theta<\eta_{0} (54)

when nn is sufficiently large.

First, we argue that

Claim: F(Err(Y^,Y))1θY^Y0\text{Claim: }F(Err(\hat{Y},Y))\geq 1-\theta\Rightarrow\hat{Y}Y\leq 0 (55)

Suppose otherwise. Then it happens with nonzero probability that F(Err(Y^,Y))1θF(Err(\hat{Y},Y))\geq 1-\theta and Y^Y>0\hat{Y}Y>0. That is,

(y^1,y1),y^1y1>0F(Err(y1^,y1))1θ.\exists(\hat{y}_{1},y_{1}),\hat{y}_{1}y_{1}>0\land F(Err(\hat{y_{1}},y_{1}))\geq 1-\theta. (56)

On the other hand, we have

[\displaystyle\mathbb{P}[ F(Err(Y^,Y))<1θY^Y0]\displaystyle F(Err(\hat{Y},Y))<1-\theta\land\hat{Y}Y\leq 0]
η0[F(Err(Y^,Y))1θ]\displaystyle\geq\eta_{0}-\mathbb{P}[F(Err(\hat{Y},Y))\geq 1-\theta]
η0θ (by Lemma 1)\displaystyle\geq\eta_{0}-\theta\text{ (by \hyperref@@ii[F_estimate]{Lemma \ref{F_estimate}})}
>0 (by equation 54)\displaystyle>0\text{ (by equation~{}\ref{error_thm_case2_ineq})}

Since the event F(Err(Y^,Y))<1θY^Y0F(Err(\hat{Y},Y))<1-\theta\land\hat{Y}Y\leq 0 happens with a nonzero probability, we have

(y^2,y2),y^2y20F(Err(y2^,y2))<1θ.\exists(\hat{y}_{2},y_{2}),\hat{y}_{2}y_{2}\leq 0\land F(Err(\hat{y_{2}},y_{2}))<1-\theta. (57)

From equation 56 and equation 57, since y^1y1>0\hat{y}_{1}y_{1}>0, y^2y20\hat{y}_{2}y_{2}\leq 0 but Err(y2^,y2)<Err(y1^,y1)Err(\hat{y_{2}},y_{2})<Err(\hat{y_{1}},y_{1}), we find a contradiction with the definition of “good error function”.

Now, we have

[\displaystyle\mathbb{P}[ YZ=1|F(Err(Y^,Y))1θ]\displaystyle YZ=-1|F(Err(\hat{Y},Y))\geq 1-\theta]
=[YZ=1F(Err(Y^,Y))1θ]][F(Err(Y^,Y))1θ]]\displaystyle=\frac{\mathbb{P}[YZ=-1\land F(Err(\hat{Y},Y))\geq 1-\theta]]}{\mathbb{P}[F(Err(\hat{Y},Y))\geq 1-\theta]]}
[Y^Z>0F(Err(Y^,Y))1θ]][F(Err(Y^,Y))1θ]]\displaystyle\geq\frac{\mathbb{P}[\hat{Y}Z>0\land F(Err(\hat{Y},Y))\geq 1-\theta]]}{\mathbb{P}[F(Err(\hat{Y},Y))\geq 1-\theta]]}
(equation 55, Y^Y0Y^Z>0YZ0YZ=1\hat{Y}Y\leq 0\land\hat{Y}Z>0\Rightarrow YZ\leq 0\Rightarrow YZ=-1)
1ξ[F(Err(Y^,Y))1θ]]\displaystyle\geq 1-\frac{\xi}{\mathbb{P}[F(Err(\hat{Y},Y))\geq 1-\theta]]}
1ξθγ (by Lemma 1)\displaystyle\geq 1-\frac{\xi}{\theta-\gamma}\text{ (by \hyperref@@ii[F_estimate]{Lemma \ref{F_estimate}})}
11θ/2exp(C1n) (by equation 47, equation 49)\displaystyle\geq 1-\frac{1}{\theta/2}\exp(-C_{1}n)\text{ (by equation~{}\ref{error_thm_fact1}, equation~{}\ref{error_thm_fact3})}

, when nn is sufficiently large.

A.7.3 The case θ=η\theta=\eta

The remaining case is when θ=η\theta=\eta. In this case, we rely on an approximation to the case θ<η0\theta<\eta_{0} (Note that the proof for the previous case θ<η\theta<\eta relied on that θ<η0\theta<\eta_{0} when nn is sufficiently large.). Let

η1=min(η,(1exp(n))η0).\eta_{1}=\min(\eta,(1-\exp(-n))\eta_{0}).

We get

[\displaystyle\mathbb{P}[ YZ=1|F(Err(Y^,Y))1η]\displaystyle YZ=-1|F(Err(\hat{Y},Y))\geq 1-\eta]
[YZ=1|F(Err(Y^,Y))1η1][F(Err(Y^,Y))1η1][F(Err(Y^,Y))1η]\displaystyle\geq\mathbb{P}[YZ=-1|F(Err(\hat{Y},Y))\geq 1-\eta_{1}]\cdot\frac{\mathbb{P}[F(Err(\hat{Y},Y))\geq 1-\eta_{1}]}{\mathbb{P}[F(Err(\hat{Y},Y))\geq 1-\eta]}
(12η1exp(C1n))[F(Err(Y^,Y))1η1][F(Err(Y^,Y))1η]\displaystyle\geq(1-\frac{2}{\eta_{1}}\exp(-C_{1}n))\cdot\frac{\mathbb{P}[F(Err(\hat{Y},Y))\geq 1-\eta_{1}]}{\mathbb{P}[F(Err(\hat{Y},Y))\geq 1-\eta]}
(Applying the lower bound from the previous case)
(12η1exp(C1n))η1γη (Lemma 1)\displaystyle\geq(1-\frac{2}{\eta_{1}}\exp(-C_{1}n))\cdot\frac{\eta_{1}-\gamma}{\eta}\text{ (\hyperref@@ii[F_estimate]{Lemma \ref{F_estimate}})}
(14ηexp(C1n))(1exp(n))η0γη\displaystyle\geq(1-\frac{4}{\eta}\exp(-C_{1}n))\cdot\frac{(1-\exp(-n))\eta_{0}-\gamma}{\eta}
(14ηexp(C1n))(1exp(n))(ηexp(C1n))C2nη\displaystyle\geq(1-\frac{4}{\eta}\exp(-C_{1}n))\cdot\frac{(1-\exp(-n))(\eta-\exp(-C_{1}n))-\frac{C_{2}}{\sqrt{n}}}{\eta}
(equation 48, equation 47, equation 49)
12C2/ηn\displaystyle\geq 1-\frac{2C_{2}/\eta}{\sqrt{n}}

, when nn is sufficiently large.

Appendix B Experimental details

B.1 Synthetic dataset

Dataset details

The 2-D synthetic classification dataset consists of the majority groups (can be classified with the spurious features) and the minority groups (cannot be classified with the spurious features). The training set consists of 1,000 majority group samples and five minority group samples. Specifically, the majority group sample is sampled from the multivariate Gaussian distributions 𝒩([5.0,5.0],1.3𝐈)\mathcal{N}([5.0,5.0],1.3\mathbf{I}) and 𝒩([5.0,5.0],1.3𝐈)\mathcal{N}([-5.0,-5.0],1.3\mathbf{I}) for positive and negative classes, respectively. On the contrary, the minority group sample is sampled from the 𝒩([5.0,5.0],1.3𝐈)\mathcal{N}([5.0,-5.0],1.3\mathbf{I}) and 𝒩([5.0,5.0],1.3𝐈)\mathcal{N}([-5.0,5.0],1.3\mathbf{I}). Notably, before we feed the training samples into the model, the second feature (invariant feature) is manually scaled down 1/10 times. This downscaling leads to relatively insignificant gradients for the invariant features. Thus, the invariant feature is hard to learn.

Model details

We use the simplest neural network architecture for this experiment: the fully connected neural network with a single hidden layer with ReLU activation. The number of the hidden neuron is 50,000. For the training, we use SGD optimization for all evaluated models. For the ERM, we use 1e-4 as the learning rate; 1e-5 as the weight decay. For the END’s identification model, we use 5e-4 as the learning rates; 1e-5 as the weight decay; 1e-3 as the confidence regularization; 1e-2 as the p-value threshold. The SGLD weight sample is saved every 3,000 iterations. For the JTT and END’s debiased model, we use 1e-5 as the learning rate; 5e-4 as the weight decay for both the identification model and final model. The hyperparameters for JTT and END’s debiased model are determined based on the argument from Sagawa et al. (2019); Liu et al. (2021): the importance weighting approaches should use relatively lower learning rates and higher weight decays. Every evaluated method is trained in 100,000 iterations with batch-size 32, excepts for the identification model of END and JTT: 30,000 iterations. Additionally, we add the additional datasets (error set for JTT, SCF set for END) 30 times.

B.2 Group robustness benchmark dataset

Dataset details

We use the two image classification datasets, the CelebA and Waterbirds. The two datasets have been used to evalulate the group robustness (worst-group accuracy). Each dataset has the group information (attributes, labels). And the attributes is the spurious-cue. For instance, the target class of the CelebA dataset is hair color, but the model can predict the class by abusing the spurious cue ”gender”. This abuse leads to the degradation of accuracy for a certain group. If the model has group-robustness, this degradation is insignificant.

Here, we elucidate the input features, the group information, and the target classes of the dataset:

  • Waterbirds (Wah et al., 2011): The input data is the image of the birds and their backgrounds. The target classes are either ”waterbirds” or ”landbirds”. Notably, the dataset has the group information (background images, labels): ”waterbirds with (water/land) background” or ”landbirds (water/land) background”. In the training dataset, most of the waterbird images have the water background, and vice-versa. Here, only 5% of the training dataset has a contradictory background (e.g., landbird on the water background). We consider 4 groups: [waterbird, water background], [landbird, water background], [waterbird, land background], and [landbird, land background].

  • CelebA (Liu et al., 2015) The input data is the image of the celebrities. Similar to the Sagawa et al. (2019) and Liu et al. (2021), the target class is the hair color, ”blond” or ”not blond”. Here, the spurious attributes are the ”male” or ”female.” The label and spurious attributes are spuriously correlated like in the Waterbird case (e.g., most of the ”male” group has the ”not blond” label). We consider 4 groups: [blond, female], [not blond, female], [blond, male], and [not blond, male].

Moreover, we add symmetric noise to the two datasets. In practice, we randomly change the target labels of each sample with probabilities of [0, 10%, 20%, 30%].

Model details

In our experiments on the benchmark datasets, we follow the experimental setup of Liu et al. (2021) for LfF, ERM, and JTT. Specifically, we use the same model architecture for all evaluated methods: ResNet50 (He et al., 2016), pre-trained with the ImageNet. For the baseline approaches’ hyperparameters, we follow the prior studies (Nam et al., 2020; Liu et al., 2021). Moreover, similar to the Liu et al. (2021), the model selection is based on the validation worst-group accuracy.

For the Waterbirds dataset, all evaluated models are trained with up to 300 epochs and batch-size 64, except for the identification model of END and JTT, 50 epochs. The SGD optimizers with the 0.9 momentum are used, except for the LfF. The model selection is based on the worst-group accuracy on the validation dataset for all methods. The ERM uses 1e-3 as the learning rate, 1e-4 as the L2 regularization; The JTT and END’s debiased model uses 1e-5 as the learning rates, 1.0 as the L2 regularization, and adding the additional dataset for JTT (error set) 50 times and for END (SCF set) 100 times; The LfF uses the Adam optimizer, with 1e-4 learning rates and 1e-4 L2 regularization. The qq of the LfF is 1e-3; END uses 1e-3 as the learning rate, the 1e-4 as the L2 regularization, and 3e-1 as the confidence regularization for the identification model. We save the SGLD weight samples every 5 epochs.

For the CelebA dataset, the models are trained with up to 50 epochs and batch-size 64, except for the identification of END and JTT, 5 and 1 epochs respectively. Similar to the Waterbirds, the SGD optimizers with the 0.9 momentum are used, except for the LfF. The model selection is based on the worst-group accuracy on the validation dataset for all methods. The ERM uses 1e-4 as the learning rate, 1e-4 as the L2 regularization; The JTT uses 1e-5 as the learning rates, 1e-1 as the L2 regularization, and adding the additional dataset (error set of JTT, SCF set of END) 50 times; The LfF uses the Adam optimizer, with 1e-4 learning rates and 1e-4 L2 regularization. The qq of the LfF is 1e-3; END uses 5e-2 as the learning rate, the 1e-4 as the L2 regularization, and 1e-3 as the confidence regularization for the identification model. We save the SGLD weight samples every 1 epochs.

B.3 Drug-target affinity regression dataset

Refer to caption
Figure 5: The schema of the DTA task with DeepDTA architecture.
DTA task

The Drug-Target Affinity (DTA) regression task is a significant task for early-stage drug discovery. We use two well-known benchmark datasets: Davis (Davis et al., 2011) and KIBA (Tang et al., 2014). The inputs of the datasets are the Simplified Molecular Input Line-entry System (SMILES) sequence—the sequence of the drug molecule—and the amino acid sequence—the sequence of the target protein. Thus, similar to other sequence-based deep learning tasks, the input data are the one-hot encoded sequences. The target value is real-valued drug-target affinity. The dataset has 5 different folds, which is suggested in (Öztürk et al., 2018). Thus, the results in Table 2 are average and standard deviation over 5 trials.

  • Davis dataset consists of clinically relevant kinase inhibitor ligands and their affinity values–dissociation constant KdK_{d}. As Öztürk et al. (2018) does, we rescale the target affinity value: logKd1e6-\tfrac{\log K_{d}}{1e^{6}}. The Davis dataset consists of 68 drugs and 442 target protein sequences, a total of 30,056 affinity values. The additional detail of the Davis dataset is that the target affinity value is 5 if the affinity is lower than 5 (Davis et al., 2011). This could be a noisy label that does not represent the true affinity value.

  • KIBA dataset includes kinase protein sequences and SMILES molecule sequences, similar to the Davis. The KIBA dataset uses its own affinity score, the KIBA score. The dataset has 2,111 compounds and 229 proteins, a total of 118,254 affinity values.

DeepDTA model

We use the well-known simple, but effective DeepDTA architecture Öztürk et al. (2018). The DeepDTA consists of the one-dimensional convolution layers to encode the sequences and the fully-connected layers to output the affinity value from the concatenated latent features. Figure 5 illustrates the DeepDTA architecture.

Details

For the all evaluated model, we use 1e-3 as the learning rate; 1e-4 as the weight decay; 0.1 as the dropout probability; 256 as the batch size, except for the 1e-1 learning rate for the identification model. We train the models with 200 epochs. For the identification model, we train the model with 200 and 100 epochs for the Davis and KIBA, respectively. We save the SGLD samples every 20 and 10 epochs for the Davis and Kiba, respectively. The baseline ”hard” pick up the top-500 high-loss samples for the oversampling.

Appendix C Additional experiments

C.1 Ablation study

In this subsection, we conduct the experiments on the CelebA and Waterbirds with 30% label noise to verify contributions of END’s components (Table 3). Specifically, we train the models via the END framework with its variants: training the identification model without the MAE loss or confidence regularization, or both.

In this experiment, we have several implications. Firstly, utilizing uncertainty is a major factor in improving the worst-case group accuracy compared to exploiting the error (loss). In particular, compared to JTT, every END variant improves the worst group accuracy by at least 0.4. This supports one of our main claims: utilizing uncertainty is a proper approach to improving the group robustness under the noisy label scenarios. Secondly, the cooperation between the noisy-label robust loss and the overconfidence regularization significantly contributes to the group robustness in the presence of the label noise. In practice, the combination of both shows the outstanding worst group accuracy with noisy labels, as shown in Table 2 and  3. In contrast, without the overconfidence regularization (END w/o reg), only utilizing the noisy robust loss (MAE) does not show outstanding improvement compared to the proposed method (END). We interpret this degradation is due to the overconfident uncertainty for the minority group, as Model-B in Figure 2 shown. In addition, without the noisy label robust loss (END w/o reg and MAE), the worst group accuracy is degraded since the identification model memorizes the noisy labels, as we stated in the last two paragraphs of Sec . Hence, we conclude that the cooperation between the noisy robust MAE loss and the overconfidence regularization is the key component in the classification tasks, as we argued.

Table 3: Worst-group accuracy (WG ACC) and average accuracy (ACC) of END with different loss functions and the regularization.
Waterbirds 30% noise CelebA 30% noise
AVG ACC WG ACC AVG ACC WG ACC
JTT 0.012 (0.00) 0.106 (0.01) 0.151 (0.16) 0.258 (0.12)
END (w/o both) 0.787 (0.04) 0.716 (0.04) 0.698 (0.03) 0.611 (0.06)
END (w/o reg) 0.888 (0.02) 0.570 (0.02) 0.901 (0.01) 0.757 (0.06)
END (w/o MAE) 0.932 (0.01) 0.459 (0.08) 0.725 (0.04) 0.618 (0.03)
END 0.854 (0.01) 0.818 (0.03) 0.892 (0.01) 0.778 (0.03)
Refer to caption
Figure 6: 2-D classification results of the identification model on synthetic data. (a) The prediction results. The colors represents the classes. The dots and translucent stars represents training and test data respectively; (b) the histogram of the predictive entropy and the corresponding Gamma distribution; (c) the detected SCF samples.