| { |
| "File Number": "1094", |
| "Title": "When Does Group Invariant Learning Survive Spurious Correlations?", |
| "Limitation": "Considering the limitations of previous group-IL methods, we propose a new method SCILL to satisfy the two criteria. Furthermore, we theoretically prove that SCILL has the ability to learn a spurious-free predictor. Finally, we conduct extensive experiments on both synthetic and real data to evaluate the proposed new method. Experimental results show that SCILL significantly outperforms existing SOTA group-IL methods, owing to its ability to satisfy the two criteria. The empirical studies validate our theoretical findings.", |
| "Reviewer Comment": "Reviewer_2: Strengths\nThe motivation is reasonable and problem formulation is well constructed with related works.\nTheir criterion analysis is well explained, and theoretically supported.\nProposed model can be applied to most of invariant learning methods.\nThe experiment results show the effectiveness of proposed method.\nThe paper is well-written.\nWeaknesses\nTheir methods are based reference model and they use ERM models. It might be give ambiguous information of spurious correlations.\nQuestions:\nIt would be nice to modify the caption of each figure of Figure 1, instead of reference numbers.\nIs\nh\nin Section 4.1 (e.g. Criterion 4.1) different with\nh\nin Line 131, which indicates a classifier?\nHow about group inference time compared with baseline?\nReference model like ERM might be give ambiguous information of spurious correlations. How does it affected?\nTypo\nLine 72: Classification -> classification\nLine 207: )) -> )\nLimitations:\nThey addressed the limitations in Discussions Section.\nEthics Flag: No\nSoundness: 3 good\nPresentation: 4 excellent\nContribution: 3 good\n\nReviewer_3: Strengths\nTwo criteria that they designed are suitable for describing the sufficiency of the group invariant learning methods.\nThe manuscripts are well-written and easy to follow.\nThe analysis for the limitations of the existing works are reasonable and the proposed model is properly designed with this point.\nWeaknesses\nCompared to the analysis until section 4, the proposed model in section 5 is quite simple in terms of the technical novelty.\nIn terms of dataset, their experiments are quite limited.\nIn terms of baselines, it would be good to compare with Group DRO or its variants.\nThe detailed experimental analysis are limited since they just provided the quantitative results.\nQuestions:\nI think that CelebA or CivilComments datasets are suitable for real-world cases. Is there any reasons that this paper does not conduct on these dataset?\nGroup DRO or its variants should be compared. If not necessary, please provide the reasons.\nThere is a hyper-parameter, \\lambda, in Equation (3). Is the proposed model sensitive to this value? It would be interesting if there are analysis of the interaction between the value of \\lambda and two criteria the authors define.\nLimitations:\nThey addressed the limitation and its further direction adequately.\nEthics Flag: Yes\nEthics Review Area: I don’t know\nSoundness: 2 fair\nPresentation: 3 good\nContribution: 2 fair\n\nReviewer_4: This paper studies an important question regarding the effectiveness of group invariant learning at the presence of spurious features. The proposed method is compared in the experiment with a good variety of methods and datasets and has demonstrated potential to improve current methods.\nHowever, there are some issues in the technical results and presentations that prevent the readers from fully following the results. For example, In Proposition 4.5, line 217, it is not quite clear why\nP\n(\nY\n=\nj\n|\nB\ni\n)\n=\np\ni\n, where the right side depends on\ni\nalone. In line 215,\nr\nis not defined although its meaning can be vaguely guessed. Similarly, in line 260, the\nG\nis not defined. Another common issue is the usage of conditional probability/distributions given group\ng\n, which seems to be really referring to conditional on the event\nG\n=\ng\n, where\nG\nis the random mapping from the sample\n(\nX\n,\nY\n)\nto the deterministic group label\nG\nin which\ng\nis an element.\nWhat is more important than the notation issues is that the main results of this paper seem to center around the construction of the random mapping\nG\nas a function of\nY\n,\nX\ni\nn\nv\n, and\nX\ns\np\nto achieve the independent criterions for group invariant learning to work, but the current presentation of results does not completely bridge the gap between the criterion and the final conclusion and nail the requirement of\nG\nin terms of the spurious features. Since the probability model framework plays a central role in this paper, it would be helpful to make sure the arguments are rigorous and seamless.\nThe paper would also benefit from an ablation study to check the importance of the two proposed criteria as well as varying arguments that affect the results including the grouping\nG\nand the weights in the objective\nw\n.\nQuestions:\nline 123: please clarify definition of\nD\n′\nline 186: please give the definition of\nB\n, even if appeared in the appendix\nPlease clarify in the main text the relation between\nX\nand\nX\ni\nn\nv\nand\nX\ns\np\n. e.g. whether\nX\nis generated as a mapping from\n(\nX\ni\nn\nv\n,\nX\ns\np\n)\nHow strong is the condition in Criterion 4.3 and how would the results be affected if the criterion is violated?\nLimitations:\nYes\nEthics Flag: No\nSoundness: 2 fair\nPresentation: 2 fair\nContribution: 3 good", |
| "abstractText": "By inferring latent groups in the training data, recent works introduce invariant learning to the case where environment annotations are unavailable. Typically, learning group invariance under a majority/minority split is empirically shown to be effective in improving out-of-distribution generalization on many datasets. However, theoretical guarantee for these methods on learning invariant mechanisms is lacking. In this paper, we reveal the insufficiency of existing group invariant learning methods in preventing classifiers from depending on spurious correlations in the training set. Specifically, we propose two criteria on judging such sufficiency. Theoretically and empirically, we show that existing methods can violate both criteria and thus fail in generalizing to spurious correlation shifts. Motivated by this, we design a new group invariant learning method, which constructs groups with statistical independence tests, and reweights samples by group label proportion to meet the criteria. Experiments on both synthetic and real data demonstrate that the new method significantly outperforms existing group invariant learning methods in generalizing to spurious correlation shifts1.", |
| "1 Introduction": "In many real-world applications, machine learning models inevitably encounter data that are rarely presented in the training environment, i.e. being out-of-distribution (OOD). For example, data collected under new weather [36], locations [6], or light conditions [9] in vision tasks. However, machine learning models often fail in generalizing to OOD data, which blocks their deployment to critical applications [12; 38]. The dependence on spurious correlations that are prone to change across environments has been recognized as a major cause of such failure [4; 12; 37]. For example, it has been shown that models trained on MNLI [39] usually classify sentence pairs with high word overlap as the label ‘entailment’, regardless of their semantics [25]. On a new dataset where such relation no longer holds, the performance drops over 25% [25; 7].\nA notable line of research on improving the robustness of models to distribution shifts is learning features with invariant conditioned label distribution across training environments [28; 4; 15], which has been termed invariant learning (IL). These methods are based on the assumption that the causal mechanism keeps invariant across environments [28], while the spurious correlation varies. By\n∗Work done during Yimeng Chen’s internship at AIR, Tsinghua University. †Corresponding author. 1Code is availiable at https://github.com/Beastlyprime/group-invariant-learning.\n36th Conference on Neural Information Processing Systems (NeurIPS 2022).\npenalizing the variance of model prediction across environments, models are then encouraged to capture the causal mechanism instead of spurious correlations.\nRecently, invariant learning has been introduced to the scenario where environment labels are unknown [33; 8], which we term the group invariant learning (group-IL). These methods utilize prior knowledge of spurious correlations to split the training data into groups. For example, Teney et al. [33] cluster training samples with their predefined spurious features. A more generic method, EIIL [8], splits training data into the majority/minority sets on which the spurious feature conditioned label distribution varies maximally. Similar to a priori environments, these groups are supposed to encode variations of spurious information, while holding the causal mechanism.\nThough some performance improvements have been gained on several datasets, much uncertainty still exists on the effectiveness of group-IL methods. In particular, when these methods can effectively address spurious correlations remain a question. Though some theoretical analysis on the success and failure cases of IL with known environments has been proposed [2; 31; 21; 3], they are not sufficient for group-IL. First, inferred groups may not meet the assumptions on environments in existing theoretical analysis, thus their conclusions cannot generalize to group-IL. For example, in each environment, the spurious feature is assumed to have a Gaussian type distribution in [31]. However, the inferred group may not satisfy that condition, for example each group may only contain one unique value of spurious feature. Second, as environments are known and defined with causal structures, exiting analysis on the effect of environment on IL focus mostly on their number [31; 3] but less on their property or validity. However, the later are important for group-IL. Therefore, we need specific theoretical analysis under the setting of group-IL.\nIn this study, we discuss necessary conditions for group-IL to survive spurious correlations. For this purpose, we first formalize the setting of group-IL and clarify the necessary assumptions required for group-IL, which underlie our theoretical analysis. We then propose two criteria for group-IL, namely falsity exposure criterion and label balance criterion. They are respectively for judging whether spurious correlations are sufficiently exposed through their variation across groups and whether group invariance can reach spurious-free conditions. Based on that, we discuss the success and failure cases of existing methods. In some synthetic benchmarks (e.g. colored-MNIST in [4; 8]), the majority/minority groups meet the two criteria. However, in a case when the spurious feature is multivariate, the majority/minority split violates both criteria according to our theoretical analysis and observations on real datasets. As a result, existing group-IL methods are insufficient for solving spurious correlations.\nTo fix these problems, we propose a new group-IL method guided by the two criteria. Specifically, this method contains the following two steps to meet the two criteria. First, groups are defined by stratifying the prediction of a reference predictor which encodes spurious correlations. The strata is constructed with statistical tests, such that the spurious prediction is independent of the label on each group. Second, the label proportion of each inferred group is balanced by attaching weights to each instance within the group. Models are then trained with invariant learning objectives on the defined groups. We term this method Spurious-Correlation-Strata Invariant Learning with Label-balancing, abbreviated as SCILL. We further show that SCILL is provably sufficient in reaching spurious-free with ideal reference models.\nTo demonstrate the effectiveness of our proposed strategy, we conduct experiments on both synthetic and real data benchmarks on spurious correlations shifts in image classification and natural language inference (NLI). Specifically, we adopt two different invariant learning objectives, IRM (IRMv1) [4] and REx (V-REx) [15], to show the consistency of SCILL. To show the availability of SCILL, we also experiment with PGI [1] and cMMD [16; 1], which are feature invariance targets used with EIIL [8] in [1]. The experimental results show that SCILL with all the four invariance objectives consistently outperforms the existing state-of-the-art method EIIL in generalizing to spurious correlation shifts. Ablation study further shows the effectiveness of each component in SCILL.\nOur main contributions can be summarized as follows.\n• We propose two criteria for group-IL and analyze the insufficiency of existing methods. • Guided by the two criteria, we propose a new practical group-IL method which is provably\nsufficient in solving spurious correlations. • Extensive experiments on both synthetic and real-data benchmarks show that SCILL signifi-\ncantly outperforms existing methods on image classification and NLI tasks.", |
| "2 Related works": "Combating spurious correlations. A typical kind of distribution-shift is caused by the shift of spurious correlations [38], which are correlations between meaningless features (e.g. hospital tokens on a lung scan) and labels (e.g. has pneumonia or not) in the training set. The existence of spurious correlations in popular benchmarks have been revealed by many works [13; 29; 25; 32]. For example, predictive models with only incomplete semantic inputs or syntactic statistics can achieve high accuracy in NLI benchmarks [13; 29]. Geirhos et al. [11] point out that deep neural networks are prone to take easy-to-fit spurious correlations, i.e. shortcut strategies, in solving problems. As a result, resolving model’s dependence on spurious correlation is important for their robustness to distribution shifts.\nInvariant learning. Many works on domain generalization focus on capturing invariancies across training environments [12]. Recently a new kind of strategy which has made significant impacts is to learn features that permit an invariant predictor across environments [28; 4; 15; 37], termed invariant learning in this paper. Such strategy is grounded upon the theory of causality [27], where Structural Equation Models [28] or causal graphs [37] are used to describe assumptions on the data generation process. The feature-conditioned label distribution invariance is then induced by the invariance of causal mechanisms in different environments. Recent theoretical works on IL study their failure cases in the domain generalization task where environments are known a priori [2; 31; 21; 3]. In this study, we focus on the setting of group-IL.\nGroup invariant learning. In recent works, invariant learning is extended to the scenario without a priori environment labels, but with knowledge on spurious correlations in the training data [33; 8]. Such knowledge is proven to be necessary in this setting [17]. They are utilized to split the training data into groups, which are supposed to encode variations of spurious information so that they can be avoided by learning the invariance. For example, Teney et al. [33] cluster samples according to their question types. Liu et al. [19] construct groups with varying spurious correlations, based on the spurious features uncovered with feature selection. A more generic method proposed in EIIL [8] assume the access to a reference predictor which encodes spurious correlations, and exploit the outputs of the reference predictor to split training data into two groups, namely the majority and the minority. This strategy is shown effective, sometimes even outperform an oracle method using true environment labels in proving the OOD generalization [8] and also systematic generalization [1]. These empirical observations show the potential of group-IL methods, while also reveal the difference between the inferred groups and prior environments. As will be discussed in Section 3, group-IL methods need different assumptions on the causal structure from those for IL. Thus existing theoretical conclusions for IL can not generalize to group-IL. Therefore, we need specific theoretical analysis under the setting of group-IL. In a recent work, Lin et al. [17] derived the sufficient and necessary assumptions for their proposed algorithm. However, in this paper we focus on group criteria for general group-IL methods.", |
| "3 Formalization of group invariant learning": "In this section, we introduce the problem setting and formalization of the scheme of group-IL.\nConsider the task of learning a classifier f : X → Y , which maps a value x ∈ X of the input variable X to a value y ∈ Y of the target variable Y . For example, map an image of horse on grass to the label ‘horse’. We denote xinv as the essential features of an instance x of the input variable X which define its class label (e.g. the shape of the horse), while xsp as features of x should not inform the label of x (e.g. the grass background). Xinv, Xsp denote the corresponding random variables. The target is then to learn a spurious-free predictor whose predictions only depend on the feature xinv, thus is supposed to have invariant performance on any dataset.\nWe first introduce the setting of IL. IL methods suppose that the training data D are collected under multiple environments E , i.e. D = {De}e∈E . De = {xei , yei } ne\ni=1 contains data i.i.d. sampled from the probability distribution Pe(X ×Y). Pe(Y |Xinv) is invariant among different e, while Pe(Y |Xsp) varies. Such an idea is based on the invariance of causal mechanisms across environments [28; 4], with assumptions on the causal structure of the data generating process. Figure 1 demonstrates four kinds of different assumptions in existing works on invariant learning. In these causal structures, environment\nis treated as a random variable E take values in Eall, which satisfies E ⊂ Eall. Pe(·) := P(·|E = e). Xinv and Xsp are latent feature variables generating the observation X , i.e. X = r(Xinv, Xsp). Here r is generally assumed as a bijective function so that the latent features can be recovered from the observations [31; 2; 37]. Align with the formalization in the former paragraph where Xsp, Xinv are assumed recognizable from X , in this paper we adopt the same assumption. In all the four kinds of causal graphs, Pe(Y |Xinv) := P(Y |Xinv, E = e) keeps invariant under different e ∈ Eall, while Pe(Y |Xsp), Pe(Xsp), Pe(Xinv), and Pe(Xinv|Y ) can vary across different e ∈ Eall. Suppose the predictor f can be decomposed into f = c ◦ Φ, where Φ : X → H denotes a feature encoder which maps the input into a representation space H, c : H → Y is a classifier. The target of invariant learning is then to search for a Φ which satisfies the following constraint:\nP(Y |Φ(X), E = e) = P(Y |Φ(X), E = e′),∀e, e′ ∈ E . (EIC)\nIt is termed as Environment Invariance Constraint (EIC). Note that the EIC stated in [8] is a weaker form of this EIC. The constraint is incorporated into the training target via a penalty term. In a generic form, the learning target of invariant learning methods can be written as follows:\nmin f ∑ e∈E λeRe(f) + λ · penalty({Se(f)}e∈E) (1)\nwhere Re(f) stands for the expected loss of f on the environment e, weighted by a scalar λe. Se(f) stands for some statistics of f on e, and the penalty is on the variation of Se(f) to measure the deviation degree of EIC. In IRM [4], the penalty is the summation of Se(f) = ∥∇wRe(w ◦ f)∥2, where w is a constant scalar multiplier of 1.0 for each output dimension. In V-REx [15], Se(f) = Re(f), and the penalty is the variance of Se(f) on different environments. In CLOvE [37], the penalty is defined as the summation of calibration errors of the model on each environment.\nGroup invariant learning methods release the dependency of invariant learning on predefined environments by splitting the training data into groups. Suppose D is sampled from the distribution P(X ×Y). Intuitively, those groups are expected to hold the invariant mechanism P(Y |Xinv), while informing the variation of Xsp. As a result, it is meaningless to divide groups according to Xinv. In group-IL methods [33; 19; 8], group inference algorithms are designed to utilize knowledge on Xsp or the correlation between Xsp and Y . Formally, denote the inferred groups as G := {g1, g2, . . . , gm}. Define G : X × Y → I as the function which maps a sample to its group identity, i.e. G(x, y) = i if and only if (x, y) ∈ gi. G is then the set of events {G = i}, i ∈ I. We have G is σ(Xsp, Y )-measurable, i.e. it is a function of Xsp, as in [33; 19], or both Xsp and Y [8].\nIn the following sections, we ground our analysis on group-IL with the causal structures in Figure 1 (a) and (b), while without additional assumptions on the causal models. Our choice of causal structures is based on the following two observations. First, in the causal graph (d), Xsp is a backdoor variable [27] between Xinv and Y and confounded with Xinv by an unobserved variable. As a result, whether the invariant mechanism holds on each group is indeterminate without additional assumptions on the mechanisms between Xsp and Xinv . Second, as shown by Ahuja et al. [2], invariance itself can not deal with spurious feature for causal structure (c). Additional knowledge or penalty, e.g. information bottleneck, is needed together with group-IL. Therefore, we investigate group-IL under the causal structures depicted by graphs (a) and (b).", |
| "4 Two group criteria": "With the above formulation, we are ready to theoretically analyze the ability of existing group-IL methods in learning spurious-free predictors. For this purpose, in this section we derive two necessary conditions for group-IL in surviving spurious correlations, i.e. falsity exposure, and label balance. Both conditions can be used as criteria to judge the sufficiency of group-IL methods. We then show that existing methods can violate the two criteria, thus become insufficient for learning a spurious-free predictor.", |
| "4.1 Falsity exposure criterion": "As groups are supposed to expose variance of spurious features so that they can be avoided by invariant learning, a natural idea is to take into account the sufficiency of such exposure on inferred groups. Ideally, if groups are split according to Xsp, any variance of P(Y |Xsp) is then fully exposed. However, such split is only practical when the value of Xsp is accessible and sparse. On the contrary, we consider the condition when groups provide insufficient exposure. Intuitively, if some spurious correlation keeps invariant across groups, its variation is then not exposed, thus group invariant predictor may still depend on such correlation. Formally, this can be written as the following criterion. Criterion 4.1 (Falsity Exposure). For any σ(Xsp)-measurable function h that satisfies ∀g, g′ ∈ G, P(Y |h(Xsp), g) = P(Y |h(Xsp), g′), it must satisfies P(Y |h(Xsp)) = P(Y ).\nIntuitively, if the falsity exposure criterion is not satisfied, h(Xsp) will be an invariant feature across environments, with predictive ability on Y . The predictor depending on both Xinv and h(Xsp) can satisfy EIC but fails to be free of spurious features. The following theorem formalizes the significance of the falsity exposure criterion. Theorem 4.2. Suppose the falsity exposure criterion is violated, i.e. ∃h which satisfies P(Y |h(Xsp), g) = P(Y |h(Xsp), g′) ̸= P(Y ),∀g, g′∈ G. Then the optimal solution of group-IL is f(X)=P[Y |Xinv, h(Xsp)], which fails to generalize when P(Y |Xsp) shifts.", |
| "4.2 Label balance criterion": "Even if we have sufficient falsity exposure, would the constraint in group-IL, i.e. EIC, guarantee the model to be free of spurious correlations? We study this problem by analyzing the relation between EIC and the constraint for a predictor to be spurious-free. In our assumed causal structures, Xinv ⊥ Xsp|Y . Thus, a spurious-free predictor, which only depends on Xinv, satisfies f(X) ⊥ Xsp|Y . In fact, it can be proved that this is a sufficient condition for a predictor f(X) to be invariant to the intervention [27] on Xsp (See the appendix). As a result, we term f(X) ⊥ Xsp|Y as the spurious-free constraint, i.e.\nP(f(X)|Xsp = b, Y ) = P(f(X)|Xsp = b′, Y ),∀b, b′ ∈ B, (SFC)\nwhere B is the image set of Xsp. The following is a necessary condition for EIC to induce the above constraint, which we term as the label balance criterion. It states that the label proportion in different groups should be the same. Criterion 4.3 (Label Balance). For any g, g′ ∈ G and y, y′ ∈ Y with non-zero P(Y = y|g),P(Y = y′|g),P(Y = y|g′) and P(Y = y′|g′), the following equation holds.\nP(Y = y|g)/P(Y = y′|g) = P(Y = y|g′)/P(Y = y′|g′) (2)\nFormally, the following theorem shows the significance of this criterion on the effectiveness of group-IL. Theorem 4.4. With a set of groups G inferred by (Xsp, Y ), i.e. G ⊂ σ(Xsp, Y ), if the label balance criterion is violated, functions satisfying EIC can not satisfy SFC.", |
| "4.3 Analysis of existing group invariant learning methods": "Now we analyze whether the groups in existing group-IL methods meet the two criteria. Directly, randomly grouped clusters of Xsp as in [33] do not guarantee to meet either criteria, and clusters of\nP(Y |Xsp) [19] do not meet the label balance criterion. In the following discussions, we focus on the majority/minority groups inferred by the EI algorithm in EIIL [8]. We find that in some synthetic datasets where the spurious feature only has two distinct values, the majority/minority groups satisfy the two criteria. However, observation on real datasets and theoretical results on the case when the spurious feature is multivariate show that they can violate both criteria.\nOn some synthetic datasets constructed in existing works [1; 8], the majority/minority groups satisfy these two criteria. For example, on both colored-MNIST [4] and coloured-MNIST [1], Y has a uniform distribution, and the spurious correlation has the same ratio for any spurious features, e.g. P(Y = 0|color = green) = P(Y = 1|color = red) on colored-MNIST. It can be proved that in this case, the majority/minority groups satisfy both criteria (See the appendix).\nHowever, it no longer holds in general cases. Empirically, we observe that on MNLI, the label distributions of the two groups inferred by EI are significantly different. Specifically, the ratio of the counts of label 0 and label 1 in the two groups are 2.87 and 0.17 respectively. The following proposition theoretically provides a case where the majority/minority split satisfies the label balance but breaks the falsity exposure, and invariant learning objectives fail to find the spurious-free classifier.\nProposition 4.5. Suppose we have (X,Y ) ∼ P(X,Y ). Y takes value in {0, 1}. X is formed with spurious feature variable Xsp = (B0, B1), and invariant feature variable S, i.e. X = r(B0, B1, S), for some bijective function r. B0 and B1 are both binary variables, which take values in {b00, b10} and {b01, b11} respectively. B0, B1 and S are conditionally independent given Y . Suppose P(Y = j|Bi = bji ) = pi, ∀i, j ∈ {0, 1}, and p0 > p1. Then we have 1) the majority/minority groups emar, emin violate the falsity exposure criterion. 2) the optimal classifier under invariant learning objectives depends on B1.\nIn this case, we suppose the spurious feature can be decomposed into two variables that are conditionally independent with each other given the label. Such case can realize when the dataset contains multiple kinds of spurious features. For example, in the image classification task, the background pattern and the color of an object can be independent but both correlate with the label spuriously.", |
| "5 SCILL: a new method": "According to the above discussion in Section 4, existing group-IL methods may fail to meet the two criteria, leading to insufficient training loss for solving spurious correlations. Faced with this challenge, we propose a new group-IL method, to satisfy the two criteria. For the generality, we only assume the access to a reference model, as in EIIL, instead of spurious features. Our new method includes two steps. In the group inference step, we define groups as spurious correlation strata constructed with the reference model, for the falsity exposure criterion. While in the training objective step, we reweight each sample with group label proportion to meet the label balance criterion. To highlight the two parts, our method is named as Spurious-Correlation-strata Invariant Learning with Label-balance, abbreviated as SCILL. Theoretically, we prove that this method is provably sufficient with ideal reference models, i.e. its optimal solution is spurious-free.", |
| "5.1 Group inference with statistical split": "We first introduce the group inference step in SCILL. As in EIIL, we assume a reference classifier fr is available, which is expected to predict only depending on Xsp. fr can be a model trained with empirical risk minimization (ERM) [8] or a model designed to capture spurious correlations [20].\nWe propose to construct groups G by stratify the outputs of fr, with the target that Y ⊥ fr(X)|g, ∀g ∈ G. The motivation for introducing spurious correlation strata comes from the deduction of the falsity exposure criterion, i.e. when groups meet the requirements that Xsp ⊥ Y |g, the falsity exposure criterion is satisfied. That is because with the above condition, we have for any function h, h(Xsp) ⊥ Y |g, as a result P(Y |h(Xsp), g) = P(Y |g). If P(Y |h(Xsp), g) = P(Y |h(Xsp), g′), ∀g, g′ ∈ G, we have P(Y |g) = P(Y |g′) = P(Y ). Thus P(Y |h(Xsp), g) = P(Y ). As a result, the falsity exposure criterion is satisfied. As different output values of the reference model fr inform the difference in P(Y |Xsp), Y ⊥ fr(X)|g approximates P(Y |Xsp) ⊥ Y |g, which is equivalent to Xsp ⊥ Y |g.\nWe propose to construct such groups through the statistical-split algorithm, inspired by the algorithm proposed in [14] for propensity score estimation. Specifically, we split current groups into subgroups according to the hypothesis-test statistics. For example, for the binary classification case, a sample set B is first divided into two subsets L0, L1 according to the labels of samples. Then the two-sample t-statistics tB of log[fr(x)0/fr(x)1] are computed on the two sets. If tB exceeds a fixed threshold thr, B is then split into two subsets according to the median of fr(x)0 on B. In this way, we enhance fr(X) ⊥ Y in each group. Note that thr is a hyperparameter of this algorithm. Empirical robustness analysis on thr is conducted in our experiments. More details about the algorithm and the robustness study are provided in the appendix.", |
| "5.2 Training with reweighted loss": "Now we introduce the second step. To guarantee the label balance criterion, we reweight each sample with group label proportion in the training loss. Correspondingly, the objective of invariant learning becomes the following form:\nL(f) := ∑ g∈G R̃g(f) + λ · penalty({Sg(f)}g∈G) (3)\nwhere R̃g(f) = E[wg(Y )Lg(f(X), Y )], ωg(y) := P(Y = y)/P(Y = y|g) for nonzero P(Y = y|g). The weight function is defined to balance the label distribution between groups, i.e. P(Y |g) = P(Y |g′),∀g, g′, for achieving the label balance criterion. With the above two steps, SCILL is then able to meet the two group criteria. Now we further investigate the theoretical capability of SCILL in solving spurious correlations. The following theorem shows that with a purely spurious reference model, SCILL can find spurious-free predictors. Theorem 5.1. If G satisfies f∗r (X) ⊥ Y |g, ∀g ∈ G, where f∗r : X → Y is spurious-only, i.e. σ(Xsp)-measurable, and minimizes the prediction loss Lrce = E[ ∑ y P(Y = y|X) log fr(X)y], the optimal model minimizing the objective (3) satisfies SFC.", |
| "6 Experiments": "In this section, we first conduct experiments to show that SCILL outperforms existing group-IL methods in generalizing to spurious correlation shifts. Then we empirically analyze whether the experimental improvements are consistent with the theoretical findings.", |
| "6.1 Experimental settings": "Now we describe our experimental settings, including datasets, models, and some training details. More details are provided in the appendix.\nDatasets We conduct experiments on both synthetic and real-world datasets. The synthetic dataset, Patched-Colored MNIST (PC-MNIST), is constructed as a realization of the conditions in the Proposition 4.5 to verify the proposed criteria. It is derived from MNIST, by assigning two conditionally independent spurious features given label, namely the color and patch bias to each image. The design of the patch bias is inspired by [5]. MNLI-HANS is a benchmark widely used in many previous works on combating spurious correlations, such as [7; 34]. In our experiments, we follow the practice to utilize MNLI [39] as the training data and HANS [25] as the test data.\nBaselines and configurations In our experiments, we compare SCILL with two baselines, i.e. ERM and EIIL [8]. ERM represents the method with the traditional empirical risk minimization (ERM) approach. EIIL is a state-of-the-art group-IL method, where groups are inferred by searching an assignment to make the reference model maximally violates the invariant learning principle. We experiment with four different invariance penalties: IRM [4], REx (V-REx) [15], cMMD [16; 1] and PGI [1]. Note that cMMD and PGI target to learn group invariant predictions conditioning on the label, different from EIC. See Appendix for more details of the four penalties.\nThe training configurations are presented as follows. For PC-MNIST, we adopt the classifier proposed in [4] for Colored MNIST, which is a MLP with two hidden layers of 390 neurons. The reference\nmodel is a MLP with the same structure trained with ERM on the training set, following the setting in EIIL on Colored MNIST. While for MNLI, we use a BERT-based classifier with the standard setup for sentence pair classification [10]. The reference model is the same as the biased classifier propose in [34], which is trained on top of some hand-crafted syntactic features. For each task, all implementations of SCILL and EIIL adopt the same model configurations and pretrained reference models. Since models are tested with OOD data, it is important to specify the model selection strategy, as has been revealed by Gulrajani and Lopez-Paz [12] for the case of domain generalization. In our experiments, we report results with 3 different model selection strategies, including ID, Oracle, and TEV. ID refers to the strategy based on model performance on the in-distribution validation set as used in [34]. Oracle refers to the selection based on data from the test data distribution, as used in [8; 12]. While TEV is a new strategy adapted from the training-domain validation method in [12] to the inferred groups, which alleviates the dependence on the test data as ID. Details can be found in the appendix.", |
| "6.2 Experimental results": "Now we demonstrate our experimental results, including performance comparison and detailed analysis. More empirical results including the robustness analysis on the hyperparameter in SCILL can be found in the appendix.", |
| "6.2.1 Performance comparison": "Table 1 and 2 show the experimental results on PC-MNIST and MNLI-HANS, respectively. The main observation is that all implementations of SCILL consistently outperform the counterpart of EIIL across all model selection strategies, in terms of the performance on OOD data. Comparing different model selection strategies, Oracle performs the best for both EIIL and SCILL. However, SCILL with the TEV strategy has the ability to outperform EIIL with Oracle, demonstrating the superiority of our new objective. Additionally, SCILL also gains improvements against some debiasing methods utilizing the same reference model (See the appendix).", |
| "6.2.2 Ablation study and verification of the two criteria": "Our main theoretical results in Section 4, i.e. Theorem 4.2 and 4.4, reveal that the two group criteria are necessary conditions for group-IL to survive spurious correlations. Now we discuss the empirical verification of the significance of the two criteria.\nFalsity exposure criterion. To show the significance of the falsity exposure criterion, we compare the performance of methods under the case when the label balance criterion is satisfied. On PCMNIST, both SCILL and EIIL groups satisfy the label balance criterion3, while EIIL groups provably violate the falsity exposure, according to Proposition 4.5. The significant improvement of SCILL over EIIL on Table 1 then shows the importance of the falsity exposure criterion. To exclude the effect of the noise in the reference model in the group inference, we further implement SCILL with the ground-truth spurious predictor, obtaining SCILLgt in Table 1. The groups then satisfy the falsity exposure criterion. We construct the ground truth majority/minority split and experiment with IL methods, obtaining results in the row maj./min. in Table 1. The significant performance drop of maj./min. compared with SCILLgt verifies the importance of falsity exposure for group-IL. On MNLI, the label in EIIL groups is unbalanced (See the appendix). Therefore we attach the instance reweight step as in SCILL to EIIL, obtaining EIILlb which satisfies the label balance criterion. As shown in Table 2, EIILlb fails to achieve improved performance, which verifies the necessity of falsity exposure.\nLabel balance criterion. To verify the necessity of the label balance criterion, we investigate the cases when the falsity exposure is satisfied. As the SCILLgt on PC-MNIST satisfies the falsity exposure, we construct such cases by disturbing the label balancing weights in SCILL. We multiply the estimated label proportion of class 0 by 0.5 to get the unbalanced SCILLub. As shown in Table 1, under Oracle selection, the test accuracy of SCILLub drops approximately 15% compared with SCILLgt, thus verifying the impact of label balance. More results can be found in the appendix.\n3The ratio of label 0 and label 1 on group 0 and group 1 in EIIL is 1:1.03 and 1:1.04, respectively.\nWe further show the importance of the instance reweight step in SCILL, which is designed following the label balance criterion. For this, we remove the instance reweight step in SCILL, obtaining SCILLuw. The experimental results in Table 1 and 2 show that SCILLuw performs worse than SCILL, demonstrating the importance of the instance reweight step in SCILL.", |
| "6.2.3 Robusteness analysis": "As shown in Section 5.1, the statistical-split algorithm contains a hyper-parameter thr. So We study the robustness of SCILL w.r.t. thr by experiments on PC-MNIST with thr set as 5, 10, 15, 20 in SCILL-IRM. From the results shown in Table 3, the models are robust with different thr = 10, 15, 20, though the model with thr = 5 is worse than others. More results can be found in the Appendix.", |
| "7 Discussions": "The main limitation of the paper is our assumptions on the causal structure. In fact, our conclusions can be generalized to more complex structures, e.g. those in [35] (See the appendix). The most central assumption is the conditional independence between Xinv and Xsp given Y , which is adopted in many existing works on solving spurious correlations [35; 7; 40; 30]. It would be an important direction to find causal structures on which the assumption is not satisfied while group invariant learning can still be effective.\nIn this paper, the algorithm SCIIL is proposed as a possible but not necessarily optimal solution to meet the two criteria for group-IL. As this paper focuses on analyzing group invariant learning, comparing SCILL with other algorithms besides group-IL is beyond the scope of this paper. It is noteworthy that the objective function of SCIIL exhibits similarities to those utilized in two recent methods. [24; 30]. Both methods incorporate a risk term reweighted by estimations of spurious correlations and a feature invariance penalty. However, these methods are limited to scenarios when spurious features can be explicitly defined [30], and are also discrete as assumed in [24]. Additionally, their feature invariance penalty differs from those employed in IL. Compared to some other methods that utilize a reference model [7; 22; 34; 26; 18; 40], the first term of the SCIIL algorithm is in similar form as it also involves reweighting samples based on the outputs of the reference model. However, the IL penalty in SCIIL serves as an additional regularization term. An extended discussion is provided in the appendix.\nBesides the setting of group invariant learning, the two criteria may also bring benefits to the study of domain generalization. Furthermore, our empirical results show SCILL can achieve good performance with other kinds of invariances besides that in invariant learning, e.g. the invariance of the distribution of model outputs conditioned on the label. Discussing the effect of the two group criteria on other kinds of group invariance is a potential research direction.", |
| "8 Conclusion": "This paper is concerned with when group invariant learning (group-IL) can survive spurious correlations. We first formulate the setting of group-IL and necessary assumptions. Then we theoretically analyze the necessary conditions for group-IL in learning spurious-free predictors, and obtain two group criteria, i.e. falsity exposure and label balance. Considering the limitations of previous group-IL methods, we propose a new method SCILL to satisfy the two criteria. Furthermore, we theoretically prove that SCILL has the ability to learn a spurious-free predictor. Finally, we conduct extensive experiments on both synthetic and real data to evaluate the proposed new method. Experimental results show that SCILL significantly outperforms existing SOTA group-IL methods, owing to its ability to satisfy the two criteria. The empirical studies validate our theoretical findings.", |
| "9 Acknowledgement": "This work was supported by National Key R&D Program of China No. 2021YFF1201600, Vanke Special Fund for Public Health and Health Discipline Development, Tsinghua University (NO.20221080053), and Beijing Academy of Artificial Intelligence (BAAI). The authors would like to thank Keyue Qiu and Yuan Li for providing useful feedback on the draft.", |
| "Reviewer Summary": "Reviewer_2: The paper proposed the group invariant learning (group-IL) which splits the training data into groups based on spurious correlations. First, they first theoretically analyzed and clarified necessary conditions of group-IL: falsity exposure criterion and label balance criterion. They also discuss that many benchmark datasets violate these conditions. Therefore, they propose a novel methods guided by these conditions. They conduct experiments on image and natural language, and show the effectiveness of their model.\n\nReviewer_3: They claimed that the existing group invariant learning methods are insufficient to prevent classifiers from depending on spurious correlations in the training set. In detail, they proposed two criteria on judging such sufficient characteristics. And then, they showed the limitations of the existing works in terms of the criteria. From this point, they designed a new group invariant learning method, i.e. constructing group and reweighting to meet the criteria. The experimental results support the superiority of the proposed method.\n\nReviewer_4: This paper proposes and evaluates conditions for group invariant learning methods in preventing classifiers from depending on spurious correlations. Motivated by the proposed conditions, the authors develop a new group learning approach with a two-step procedure involving constructing group structures to achieve these conditions and re-weighting samples in model training according to group information." |
| } |