This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

ABC: Auxiliary Balanced Classifier for Class-imbalanced Semi-supervised Learning

Hyuck Lee    Seungjae Shin    Heeyoung Kim
Department of Industrial and Systems Engineering, KAIST
{dlgur0921, tmdwo0910, heeyoungkim}@kaist.ac.kr
Abstract

Existing semi-supervised learning (SSL) algorithms typically assume class-balanced datasets, although the class distributions of many real-world datasets are imbalanced. In general, classifiers trained on a class-imbalanced dataset are biased toward the majority classes. This issue becomes more problematic for SSL algorithms because they utilize the biased prediction of unlabeled data for training. However, traditional class-imbalanced learning techniques, which are designed for labeled data, cannot be readily combined with SSL algorithms. We propose a scalable class-imbalanced SSL algorithm that can effectively use unlabeled data, while mitigating class imbalance by introducing an auxiliary balanced classifier (ABC) of a single layer, which is attached to a representation layer of an existing SSL algorithm. The ABC is trained with a class-balanced loss of a minibatch, while using high-quality representations learned from all data points in the minibatch using the backbone SSL algorithm to avoid overfitting and information loss. Moreover, we use consistency regularization, a recent SSL technique for utilizing unlabeled data in a modified way, to train the ABC to be balanced among the classes by selecting unlabeled data with the same probability for each class. The proposed algorithm achieves state-of-the-art performance in various class-imbalanced SSL experiments using four benchmark datasets.

1 Introduction

Recently, numerous deep neural network (DNN)-based semi-supervised learning (SSL) algorithms have been proposed to improve the performance of DNNs by utilizing unlabeled data when only a small amount of labeled data is available. These algorithms have shown effective performance in various tasks. However, most existing SSL algorithms assume class-balanced datasets, whereas the class distributions of many real-world datasets are imbalanced. It is well known that classifiers trained on class-imbalanced data tend to be biased toward the majority classes. This issue can be more problematic for SSL algorithms that use predicted labels of unlabeled data for their training, because the labels predicted by the algorithm trained on class-imbalanced data become even more severely imbalanced [18]. For example, Figure 1 (b) presents biased predictions of ReMixMatch [3], a recent SSL algorithm, trained on CIFAR-1010-LT, which is a class-imbalanced dataset with the amount of Class 0 being 100 times more than that of Class 9, as depicted in Figure 1 (a). Although there are various class-imbalanced learning techniques, they are usually designed for labeled data, and thus cannot be simply combined with SSL algorithms under class-imbalanced SSL (CISSL) scenarios. Recently, a few CISSL algorithms have been proposed, but the CISSL problem is still underexplored.

Refer to caption Refer to caption Refer to caption
 (a) Class-imbalanced training set      (b) ReMixMatch       (c) Proposed algorithm
Figure 1: Predictions on a class-balanced test set using ReMixMatch (b) and the proposed algorithm (c) trained on a class-imbalanced training set (a).

We propose a new CISSL algorithm that can effectively use unlabeled data, while mitigating class imbalance by using an existing DNN-based SSL algorithm [3, 29] as the backbone and introducing an auxiliary balanced classifier (ABC) of a single layer. The ABC is attached to a representation layer immediately preceding the classification layer of the backbone, based on the argument that a classification algorithm (i.e., backbone) can learn high-quality representations even if its classifier is biased toward the majority classes [17]. The ABC is trained to be balanced across all classes by using a mask that rebalances the class distribution, similar to re-sampling in previous SSL studies [2, 7, 13, 16]. Specifically, the mask stochastically regenerates a class-balanced subset of a minibatch on which the ABC is trained. The ABC is trained simultaneously with the backbone, so that the ABC can use high-quality representations learned from all data points in the minibatch using the backbone. In this way, the ABC can overcome the limitations of the previous resampling techniques, overfitting on minority-class data or loss of information on majority-class data [6, 27].

Moreover, to place decision boundaries in low-density regions by utilizing unlabeled data, we use consistency regularization, a recent SSL technique, which enforces the classification outputs of two augmented or perturbed versions of the same unlabeled example to remain unchanged. In particular, we encourage the ABC to be balanced across classes when using consistency regularization by selecting unlabeled examples with the same probability for each class using a mask. Figure 1 (c) illustrates that compared to the results of ReMixMatch in Figure 1 (b), the class distribution of the predicted labels becomes more balanced using the proposed algorithm trained on the same dataset. Our experimental results under various scenarios demonstrate that the proposed algorithm achieves state-of-the-art performance. Through qualitative analysis and an ablation study, we further investigate the contribution of each component of the proposed algorithm. The code for the proposed algorithm is available at https://github.com/LeeHyuck/ABC.

2 Related Work

Semi-supervised learning (SSL) Recently, several SSL techniques that utilize unlabeled data have been proposed. Entropy minimization [12] encourages the classifier outputs to have low entropy for unlabeled data, as in pseudo-labels [22]. Mixup regularization [4, 32] makes the decision boundaries farther away from the data clusters by encouraging the prediction for an interpolation of two inputs to be the same as the interpolation of the prediction for each input. Consistency regularization [26, 24, 30] encourages a classifier to produce similar predictions for perturbed versions of the same unlabeled input. To create perturbed unlabeled inputs, various data augmentation techniques have been used. For example, FixMatch [29] and ReMixMatch [3] used strong augmentation methods such as Cutout [10] and RandomAugment [8]. FixMatch and ReMixMatch are used as the backbone of the proposed algorithm; they are described in Section 3.2.

Class-imbalanced learning (CIL) As a popular approach for CIL, re-sampling techniques [16, 7, 2, 13] balance the number of training samples for each class in the training set. As another popular approach, re-weighting techniques [23, 14, 33] re-weight the loss for each class by a factor inversely proportional to the number of data points belonging to that class. Although these approaches are simple, they have some drawbacks. For example, oversampling from minority classes can cause overfitting, whereas undersampling from majority classes can cause information loss [6]. In the case of re-weighting, gradients can be calculated to be abnormally large when the class imbalance is severe, resulting in unstable training [6, 1]. Many attempts have been made to alleviate these problems, such as effective re-weighting [9] and meta-learning-based re-weighting [28, 15]. New forms of losses have also been proposed [6, 27]. In [36, 19], knowledge is transferred from the data of majority classes to the data of minority classes. These CIL algorithms were designed for labeled data and require label information; thus, they are not applicable to unlabeled data. In [17], it was found that biased classification is mainly due to the classification layer and that a classification algorithm can learn meaningful representations even from a class-imbalanced training set. Based on this finding, we design the ABC to use high-quality representations learned from class-imbalanced data utilizing FixMatch [29] and ReMixMatch [3].

Class-imbalanced semi-supervised learning (CISSL) There have been few studies on CISSL. In [35], it was found that more accurate decision boundaries can be obtained in class-imbalanced settings through self-supervised learning and semi-supervised learning. DARP [18] refines biased pseudo-labels by solving a convex optimization problem. CReST [34], a recent self-training technique, mitigates class imbalance by using pseudo-labeled unlabeled data points classified as minority classes with a higher probability than those classified as majority classes.

3 Methodology

3.1 Problem setting

Suppose that we have a labeled dataset 𝒳={(xn,yn):n(1,,N)}\mathcal{X}=\left\{\left(x_{n},y_{n}\right):n\in\left(1,...,N\right)\right\}, where xndx_{n}\in\mathbb{R}^{d} is the nnth labeled data point and yn{1,,L}y_{n}\in\left\{1,\ldots,L\right\} is the corresponding label. We also have an unlabeled dataset 𝒰={(um):m(1,,M)}\mathcal{U}=\left\{\left(u_{m}\right):m\in\left(1,\ldots,M\right)\right\}, where umdu_{m}\in\mathbb{R}^{d} is the mmth unlabeled data point. We express the ratio of the amount of labeled data as β=NM+N\beta=\frac{N}{M+N}. Generally, β<0.5\beta<0.5, because label acquisition is costly and laborious. We denote the number of labeled data points of class ll as NlN_{l}, i.e., l=1LNl=N\sum_{l=1}^{L}N_{l}=N, and assume that the LL classes are sorted according to cardinality in descending order, i.e., N1N2NLN_{1}\geq N_{2}\geq\cdots\geq N_{L}. We denote the ratio of the class imbalance as γ=N1NL\gamma=\frac{N_{1}}{N_{L}}. Under class-imbalanced scenarios, γ1\gamma\gg 1. Following previous CIL studies, we define the half of the classes containing a large amount of data as the majority classes, and the other half of the classes, containing a small amount of data, as the minority classes. Following [34], we assume that 𝒳\mathcal{X} and 𝒰\mathcal{U} share the same class distribution, i.e., the labeled and unlabeled datasets are class-imbalanced to the same extent. From 𝒳\mathcal{X} and 𝒰\mathcal{U}, we generate minibatches 𝒳={(xb,yb):b(1,,B)}𝒳\mathcal{MB}_{\mathcal{X}}=\left\{\left(x_{b},y_{b}\right):b\in\left(1,...,B\right)\right\}\subset\mathcal{X} and 𝒰={(ub):b(1,,B)}𝒰\mathcal{MB}_{\mathcal{U}}=\left\{\left(u_{b}\right):b\in\left(1,\ldots,B\right)\right\}\subset\mathcal{U} for each iteration of training, where BB is the minibatch size. Using these minibatches for training, we aim to learn a model f:d{1,L}f:\mathbb{R}^{d}\rightarrow\left\{1,...L\right\} that performs effectively on a class-balanced test set.

3.2 Backbone SSL algorithm

We attach the ABC to the backbone’s representation layer, so that it can utilize the high-quality representations learned by the backbone. We use FixMatch [29] or ReMixMatch [3] as the backbone, as these two have achieved state-of-the-art SSL performance. FixMatch uses the classification loss calculated from the weakly augmented labeled data point α(xb)\alpha\left(x_{b}\right) generated by flipping and cropping the image, and the consistency regularization loss calculated from the weakly augmented unlabeled data point α(ub)\alpha\left(u_{b}\right) and strongly augmented unlabeled data point 𝒜(ub)\mathcal{A}\left(u_{b}\right) generated by Cutout [10] and RandomAugment [8]. ReMixMatch predicts the class label of the weakly augmented unlabeled data point α(ub)\alpha\left(u_{b}\right) using distribution alignment and sharpening, and assigns the predicted label to the strongly augmented unlabeled data point 𝒜(ub)\mathcal{A}\left(u_{b}\right). These strongly augmented unlabeled data point 𝒜(ub)\mathcal{A}\left(u_{b}\right) and strongly augmented labeled data point 𝒜(xb)\mathcal{A}\left(x_{b}\right) are used to conduct mixup regularization. ReMixMatch also conducts consistency regularization in a manner similar to FixMatch and self-supervised learning using the rotation of the image [11, 39]. FixMatch and ReMixMatch have greatly improved the SSL performance by learning high-quality representations using strong data augmentation. However, these algorithms can be significantly biased toward the majority classes in class-imbalanced settings.

Using FixMatch and ReMixMatch as the backbone of the proposed algorithm, we ensure that the ABC enjoys high-quality representations learned by the backbone, while replacing the backbone’s biased classifier. To train the ABC, we reuse the weakly augmented data and strongly augmented data used by the backbone to decrease the computational cost. Although we use FixMatch and ReMixMatch as the backbone in this study, the ABC can also be combined with other DNN-based SSL algorithms, as long as they use weakly augmented data and strongly augmented data.

3.3 ABC for class-imbalanced Semi-supervised learning

To train the ABC to be balanced, we first generate 0/10/1 mask M(xb)M\left(x_{b}\right) for each labeled data point xbx_{b} using a Bernoulli distribution ()\mathcal{B}\left(\cdot\right) with the parameter set to be inversely proportional to the number of data points of each class. This setting makes ()\mathcal{B}\left(\cdot\right) generate mask 1 with high probability for the data points in the minority classes, but with low probability for those in the majority classes. Then, the classification loss is multiplied by the generated mask, so that the ABC can be trained with a balanced classification loss. Multiplying the classification loss by the 0/10/1 mask can be interpreted as oversampling of the data points in the minority classes, whereas it can be interpreted as undersampling of those in the majority classes. In representation learning, oversampling and undersampling techniques have shown overfitting and information loss problems, respectively. In contrast, the ABC can overcome these problems because it uses the representations learned by the backbone, which is trained on all data points in the minibatch. The use of the 0/10/1 mask to construct the balanced loss, instead of directly creating a balanced subset, allows the backbone and the ABC to be trained from the same minibatches. Therefore, the representations of minibatches calculated for training the backbone can be used again for training the ABC. Consequently, the proposed algorithm only requires a slightly increased time cost compared to training the backbone alone. This is confirmed in Section 4.3. The overall procedure of balanced training with 0/10/1 mask for the ABC attached to a representation layer of the backbone is presented in Figure 2. The classification loss for the ABC, LclsL_{cls}, with 0/10/1 mask M()M\left(\cdot\right) is expressed as

Lcls=1Bb=1BM(xb)𝐇(ps(y|α(xb)),pb),\displaystyle L_{cls}=\frac{1}{B}\sum_{b=1}^{B}M\left(x_{b}\right)\mathbf{H}\left(p_{s}\left(y|\alpha\left(x_{b}\right)\right),p_{b}\right), (1)
M(xb)=(NLNyb),\displaystyle M\left(x_{b}\right)=\mathcal{B}\left(\frac{N_{L}}{N_{y_{b}}}\right), (2)

where 𝐇\mathbf{H} is the standard cross-entropy loss, α(xb)\alpha\left(x_{b}\right) is an augmented labeled data point, ps(y|α(xb))p_{s}\left(y|\alpha\left(x_{b}\right)\right) is the predicted class distribution using the ABC for α(xb)\alpha\left(x_{b}\right), and pbp_{b} is the one-hot label for xbx_{b}.

Refer to caption
Figure 2: Overall procedure for balanced training of the ABC with a 0/10/1 mask

3.4 Consistency regularization for ABC

To increase the margin between the decision boundary and the data points using unlabeled data, we conduct consistency regularization for the ABC, similar to the way in FixMatch. Specifically, we first obtain the predicted class distribution ps(y|α(ub))p_{s}\left(y|\alpha\left(u_{b}\right)\right) for a weakly augmented unlabeled data point α(ub)\alpha\left(u_{b}\right) using the ABC and use it as a soft pseudo-label qbq_{b}. Then, for two strongly augmented unlabeled data points 𝒜1(ub)\mathcal{A}_{1}\left(u_{b}\right) and 𝒜2(ub)\mathcal{A}_{2}\left(u_{b}\right), we train the ABC to produce their predicted class distributions, ps(y|𝒜1(ub))p_{s}\left(y|\mathcal{A}_{1}\left(u_{b}\right)\right) and ps(y|𝒜2(ub))p_{s}\left(y|\mathcal{A}_{2}\left(u_{b}\right)\right), to be close to qbq_{b}.

In class-imbalanced settings, because most unlabeled data points belong to majority classes, most weakly augmented unlabeled data points can be predicted as the majority classes. Then, consistency regularization would be conducted with a higher frequency for the majority classes, which can cause a classifier to be biased toward the majority classes. To prevent this issue, we conduct consistency regularization in a modified manner that is suitable for class-imbalance problems. Specifically, whereas FixMatch minimizes entropy by converting the predicted class distribution for a weakly augmented data point into a one-hot pseudo-label, we directly use the predicted class distribution as a soft pseudo-label. We do not pursue entropy minimization for the ABC because it can accelerate biased classification toward certain classes. Moreover, we once again generate 0/10/1 mask M()M\left(\cdot\right) for each unlabeled data point ubu_{b} based on a soft pseudo label qbq_{b}, and multiply the consistency regularization loss for ubu_{b} by the generated mask, so that the ABC can be trained with a class-balanced consistency regularization loss. Note that existing resampling techniques are not applicable to unlabeled data, because they require a label for each data point. In contrast, we make it possible to resample unlabeled data by using the soft pseudo-label and the 0/10/1 mask. The consistency regularization loss, LconL_{con}, with 0/10/1 mask M()M\left(\cdot\right) is expressed as

Lcon=1Bb=1Bk=12M(ub)𝐈(max(qb)τ)𝐇(ps(y|𝒜k(ub)),qb),\displaystyle L_{con}=\frac{1}{B}\sum_{b=1}^{B}\sum_{k=1}^{2}M\left(u_{b}\right)\mathbf{I}\left(\max\left(q_{b}\right)\geq\tau\right)\mathbf{H}\left(p_{s}\left(y|\mathcal{A}_{k}\left(u_{b}\right)\right),q_{b}\right), (3)
M(ub)=(NLNq^b),\displaystyle M\left(u_{b}\right)=\mathcal{B}\left(\frac{N_{L}}{N_{\widehat{q}_{b}}}\right), (4)

where 𝐈\mathbf{I} is the indicator function, max(qb)\max\left(q_{b}\right) is the highest predicted assignment probability for any class, representing the confidence of prediction, and τ\tau is the confidence threshold. To avoid the unwanted effects of inaccurate soft pseudo-labels during consistency regularization, we only use the weakly augmented unlabeled data points whose confidence is higher than the threshold τ\tau, similar to that in FixMatch. To take full advantage of few unlabeled data points with prediction confidence values that are higher than the confidence threshold τ\tau in the early stage of training, we gradually decrease the parameter of the Bernoulli distribution ()\mathcal{B}\left(\cdot\right) for ubu_{b} from 1 to NL/Nq^b{N_{L}}/{N_{\widehat{q}_{b}}}, where q^b\widehat{q}_{b} is the one-hot pseudo-label obtained from qbq_{b}. Following previous studies [3, 24, 29, 4], we do not backpropagate gradients for pseudo-label prediction. The overall procedure for consistency regularization for the ABC is shown in Appendix A.

3.5 End-to-end training

Unlike a recent CIL trend to finetune a classifier in a balanced manner after representation learning is completed (i.e., decoupled learning of representations and a classifier) [17, 27], we obtain a balanced classifier by training the proposed algorithm end-to-end. We train the proposed algorithm with the sum of losses from Sections 3.3 and 3.4, and the loss for the backbone, LbackL_{back}. The total loss function LtotalL_{total} is expressed as

Ltotal=Lcls+Lcon+Lback.\displaystyle L_{total}=L_{cls}+L_{con}+L_{back}. (5)

Whereas we use the sum of the losses for the backbone and ABC for training the proposed algorithm, we predict the class labels of new data points using only the ABC. In our experiments in Sections 4.4 and 4.5, we show that the proposed algorithm trained end-to-end produces better performance than competing algorithms with decoupled learning of representations and a classifier, and we analyze possible reasons. We present the pseudo code of the proposed algorithm in Appendix B.

4 Experiments

4.1 Experimental setup

We created class-imbalanced versions of CIFAR-10, CIFAR-100 [21], and SVHN [25] datasets to conduct experiments under various ratios of class imbalance γ\gamma and various ratios of the amount of labeled data β\beta. For class-imbalance types, we first consider long-tailed (LT) imbalance in which the number of data points exponentially decreases from the largest to the smallest class, i.e., Nk=N1×γk1L1N_{k}=N_{1}\times\gamma^{-\frac{k-1}{L-1}}, where γ=N1NL\gamma=\frac{N_{1}}{N_{L}}. We also consider step imbalance [5] in which the whole majority classes have the same amount of data and the whole minority classes also have the same amount of data. Two types of class imbalance for the considered datasets are illustrated in Appendix C. For the main setting, we set γ=100\gamma=100, N1=1000N_{1}=1000, and β=20%\beta=20\% for CIFAR-1010 and SVHN, and γ=20\gamma=20, N1=200N_{1}=200 and β=40%\beta=40\% for CIFAR-100100. Similar to [18], we set γ\gamma of CIFAR-100 to be relatively small because CIFAR-100100 has only 500 training data points for each class. To evaluate the performance of the proposed algorithm on large-scale datasets, we also conducted experiments on 7.5M data points of 256 by 256 images from the LSUN dataset [37].

We compared the performance of the proposed algorithm with that of various baseline algorithms. Specifically, we considered the following baseline algorithms:

  • Deep CNN (vanilla algorithm): This is trained on only labeled data with the cross-entropy loss.

  • BALMS [27] (CIL algorithm): This state-of-the-art CIL algorithm does not use unlabeled data.

  • VAT [24], ReMixMatch [3], and FixMatch [29] (SSL algorithms): These are state-of-the-art SSL algorithms, but do not consider class imbalance.

  • FixMatch+CReST+PDA and ReMixMatch+CReST+PDA (CISSL algorithms): CReST+PDA [34] mitigates class imbalance by using unlabeled data points classified as the minority classes with a higher probability than those classified as the majority classes.

  • ReMixMatch+DARP and FixMatch+DARP (CISSL algorithms): These algorithms use DARP [18] to refine the pseudo labels obtained from ReMixMatch or FixMatch.

  • ReMixMatch+DARP+cRT and FixMatch+DARP+cRT (CISSL algorithms): Compared to ReMixMatch+DARP and FixMatch+DARP, these algorithms finetune the classifier using cRT [17].

For the structure of the deep CNN used in the proposed and baseline algorithms, we used Wide ResNet-2828-22 [38]. We trained the proposed algorithm for 250,000250,000 iterations with a batch size of 64. The confidence threshold τ\tau was set to 0.950.95 based on experiments with various values of τ\tau in Appendix D. We used the Adam optimizer [20] with a learning rate of 0.0020.002, and used Cutout [10] and RandomAugment [8] for strong data augmentation, following [18]. Similar to [3, 4], we evaluated the performance of the proposed algorithm using an exponential moving average of the parameters over iterations with a decay rate of 0.9990.999, instead of scheduling the learning rate. In Tables 1-5, we used the overall accuracy and the accuracy only for minority classes as performance measures. We repeated the experiments five times under the main setting, and three times under the step imbalance and other settings of β\beta and γ\gamma. We report the average and standard deviation of the performance measures over repeated experiments. For the vanilla algorithm, FixMatch+DARP+cRT, and ReMixMatch+DARP+cRT, which suffered from overfitting, we measured performance every 500 iterations and recorded the best performance. Further details of the experimental setup are described in Appendix E.

4.2 Experimental results

The performance of the competing algorithms under the main setting are summarized in Table 1. We can observe that the proposed algorithm achieved the highest overall performance, with improved performance for minority classes. Interestingly, VAT, an SSL algorithm, showed similar performance to the vanilla algorithm, and worse performance than BALMS, a CIL algorithm. Similarly, FixMatch and ReMixMatch, which do not consider class imbalance, showed poor performance for minority classes. Although BALMS mitigated class imbalance, it produced poor overall performance, as it did not use unlabeled data for training. This demonstrates the importance of using unlabeled data for training, even in the class-imbalanced setting. FixMatch+CReST+PDA and ReMixMatch+CReST+PDA mitigated class imbalance by using unlabeled data points classified as the minority classes with a higher probability, but produced lower performance than the proposed algorithm. This may be because even if all unlabeled data points classified as minority classes are additionally used for training, their amount is still less than that of the data in majority classes, while the proposed algorithm uses class-balanced minibatches by generating the 0/10/1 mask. Fixmatch+DARP and ReMixMatch+DARP slightly mitigated class imbalance by refining biased pseudo-labels, but resulted in lower performance than the proposed algorithm. This may be because even perfect pseudo labels cannot change the underlying class-imbalanced distribution of the training data. By additionally using a rebalancing technique cRT, FixMatch(ReMixMatch)+DARP+cRT performed better than FixMatch(ReMixMatch)+DARP. However, FixMatch(ReMixMatch)+DARP+cRT still performed worse than FixMatch(ReMixMatch)+ABC, although it also uses high-quality representations learned by FixMatch(ReMixMatch) and techniques for mitigating class imbalance. The superior performance of FixMatch(ReMixMatch)+ABC over FixMatch(ReMixMatch)+DARP+cRT is probably because FixMatch(ReMixMatch)+ABC was trained end-to-end, and the ABC was also trained using unlabeled data. We discuss this in more detail in Sections 4.4 and 4.5. Overall, the algorithms combined with ReMixMatch performed better than the algorithms combined with FixMatch. In addition to the overall accuracy and minority-class-accuracy, we also compared the performance of the competing algorithms in terms of the geometric mean (G-mean) of class-wise accuracy under the main setting in Appendix F.

Table 1: Overall accuracy/minority-class-accuracy under the main setting
CIFAR-1010-LT SVHN-LT CIFAR-100100-LT
Algorithm γ=100\gamma=100, β=20%\beta=20\% γ=100\gamma=100, β=20%\beta=20\% γ=20\gamma=20, β=40%\beta=40\%
Vanilla 55.355.3±1.30\pm 1.30 / 33.933.9±1.88\pm 1.88 77.077.0±0.67\pm 0.67 / 63.363.3±1.25\pm 1.25 40.140.1±1.15\pm 1.15 / 25.225.2±0.95\pm 0.95
VAT [24] 55.355.3±0.88\pm 0.88 / 28.228.2±1.55\pm 1.55 81.381.3±0.47\pm 0.47 / 68.268.2±0.88\pm 0.88 40.440.4±0.34\pm 0.34 / 24.824.8±0.38\pm 0.38
BALMS [27] 70.770.7±0.59\pm 0.59 / 69.869.8±1.03\pm 1.03 87.687.6±0.53\pm 0.53 / 85.085.0±0.67\pm 0.67 50.250.2±0.54\pm 0.54 / 42.942.9±1.03\pm 1.03
FixMatch [29] 72.372.3±0.33\pm 0.33 / 53.853.8±0.63\pm 0.63 88.088.0±0.30\pm 0.30 / 79.479.4±0.54\pm 0.54 51.051.0±0.20\pm 0.20 / 32.832.8±0.41\pm 0.41
w/ CReST+PDA [34] 76.676.6±0.46\pm 0.46 / 61.461.4±0.85\pm 0.85 89.189.1±0.69\pm 0.69 / 81.781.7±1.18\pm 1.18 51.651.6±0.29\pm 0.29 / 36.436.4±0.46\pm 0.46
w/ DARP [18] 73.773.7±0.98\pm 0.98 / 57.057.0±2.12\pm 2.12 88.688.6±0.19\pm 0.19 / 80.580.5±0.54\pm 0.54 51.451.4±0.37\pm 0.37 / 33.933.9±0.77\pm 0.77
w/ DARP+cRT [18] 78.178.1±0.89\pm 0.89 / 66.666.6±1.55\pm 1.55 89.989.9±0.44\pm 0.44 / 83.583.5±0.61\pm 0.61 54.754.7±0.46\pm 0.46 / 41.241.2±0.42\pm 0.42
w/ ABC 81.1±0.82\pm 0.82 / 72.0±1.77\pm 1.77 92.0±0.38\pm 0.38 / 87.9±0.73\pm 0.73 56.3±0.19\pm 0.19 / 43.4±0.42\pm 0.42
ReMixMatch [3] 73.773.7±0.39\pm 0.39 / 55.955.9±0.87\pm 0.87 89.889.8±0.42\pm 0.42 / 82.882.8±0.68\pm 0.68 54.054.0±0.29\pm 0.29 / 37.137.1±0.37\pm 0.37
w/ CReST+PDA [34] 75.775.7±0.34\pm 0.34 / 59.659.6±0.76\pm 0.76 90.990.9±0.20\pm 0.20 / 85.285.2±0.39\pm 0.39 54.654.6±0.48\pm 0.48 / 38.138.1±0.69\pm 0.69
w/ DARP [18] 74.474.4±0.41\pm 0.41 / 56.956.9±0.67\pm 0.67 90.290.2±0.22\pm 0.22 / 83.583.5±0.40\pm 0.40 54.554.5±0.33\pm 0.33 / 37.737.7±0.58\pm 0.58
w/ DARP+cRT [18] 78.578.5±0.61\pm 0.61 / 66.466.4±1.68\pm 1.68 92.192.1±0.48\pm 0.48 / 87.687.6±0.75\pm 0.75 55.155.1±0.45\pm 0.45 / 43.643.6±0.58\pm 0.58
w/ ABC 82.4±0.45\pm 0.45 / 75.7±1.18\pm 1.18 93.9±0.16\pm 0.16 / 92.5±0.4\pm 0.4 57.6±0.26\pm 0.26 / 46.7±0.50\pm 0.50

To evaluate the performance of the proposed algorithm in various settings, we conducted experiments using ReMixMatch, FixMatch, and the CISSL algorithms considered in Table 1, while changing the ratio of class imbalance γ\gamma and the ratio of the amount of labeled data β\beta. The results for CIFAR-1010 are presented in Table 2, and the results for SVHN and CIFAR-100100 are presented in Appendix G. In Table 2, we can observe that the proposed algorithm achieved the highest overall accuracy with greatly improved performance for minority classes for all settings. Because FixMatch+DARP+cRT and ReMixMatch+DARP+cRT do not use unlabeled data for classifier tuning, the difference in performance between FixMatch(ReMixMatch)+DARP+cRT and the proposed algorithm increased as the ratio of the amount of labeled data β\beta decreased and as the ratio of class imbalance γ\gamma increased. In addition, the difference in performance between FixMatch(ReMixMatch)+CReST+PDA and the proposed algorithm tended to increase as the ratio of class imbalance γ\gamma increased, because the difference between the number of labeled data points belonging to the majority classes and the number of unlabeled data points classified as the minority classes increases with γ\gamma.

Table 2: Overall accuracy/minority-class accuracy for CIFAR-1010 under various settings
CIFAR-1010-LT
Algorithm γ=100\gamma=100, β=10%\beta=10\% γ=100\gamma=100, β=30%\beta=30\% γ=50\gamma=50, β=20%\beta=20\% γ=150\gamma=150, β=20%\beta=20\%
FixMatch [29] 70.070.0±0.59\pm 0.59 / 48.948.9±1.04\pm 1.04 74.974.9±0.63\pm 0.63 / 58.258.2±1.28\pm 1.28 81.281.2±0.07\pm 0.07 / 70.770.7±0.36\pm 0.36 68.568.5±0.60\pm 0.60 / 45.845.8±1.15\pm 1.15
w/ CReST+PDA [34] 73.973.9±0.40\pm 0.40 / 58.958.9±1.14\pm 1.14 77.677.6±0.73\pm 0.73 / 64.064.0±1.39\pm 1.39 83.383.3±0.10\pm 0.10 / 75.775.7±0.39\pm 0.39 70.070.0±0.82\pm 0.82 / 49.449.4±1.52\pm 1.52
w/ DARP+cRT [18] 74.674.6±0.98\pm 0.98 / 59.259.2±2.12\pm 2.12 79.079.0±0.25\pm 0.25 / 67.767.7±0.95\pm 0.95 83.683.6±0.42\pm 0.42 / 77.177.1±1.19\pm 1.19 73.273.2±0.85\pm 0.85 / 57.157.1±1.13\pm 1.13
w/ ABC 77.2±1.60\pm 1.60 / 65.7±2.85\pm 2.85 81.5±0.29\pm 0.29 / 72.9±0.96\pm 0.96 85.2±0.51\pm 0.51 / 80.2±0.64\pm 0.64 77.1±0.46\pm 0.46 / 64.4±0.92\pm 0.92
ReMixMatch [3] 71.571.5±0.51\pm 0.51 / 52.252.2±1.08\pm 1.08 75.875.8±0.10\pm 0.10 / 59.459.4±0.17\pm 0.17 81.581.5±0.17\pm 0.17 / 70.770.7±0.32\pm 0.32 69.969.9±0.23\pm 0.23 / 48.448.4±0.60\pm 0.60
w/ CReST+PDA [34] 73.873.8±0.32\pm 0.32 / 56.656.6±0.43\pm 0.43 78.678.6±0.73\pm 0.73 / 64.864.8±1.49\pm 1.49 83.983.9±0.26\pm 0.26 / 75.475.4±0.52\pm 0.52 71.371.3±0.77\pm 0.77 / 50.850.8±1.59\pm 1.59
w/ DARP+cRT [18] 75.975.9±1.20\pm 1.20 / 62.162.1±3.10\pm 3.10 81.081.0±0.16\pm 0.16 / 70.770.7±0.72\pm 0.72 84.584.5±0.80\pm 0.80 / 77.877.8±1.67\pm 1.67 73.973.9±0.59\pm 0.59 / 57.457.4±1.45\pm 1.45
w/ ABC 79.8±0.36\pm 0.36 / 70.8±0.92\pm 0.92 84.3±1.03\pm 1.03 / 80.6±0.97\pm 0.97 87.5±0.31\pm 0.31 / 84.6±1.19\pm 1.19 80.6±0.66\pm 0.66 / 72.1±1.51\pm 1.51

We also conducted experiments under a step-imbalance setting, where the class imbalance was more noticeable. This setting assumes a more severely imbalanced class distribution than the LT imbalance settings, because half of the classes have very scarce data. The experimental results for CIFAR-1010 are presented in Table 3, and the results for SVHN and CIFAR-100100 are presented in Appendix H. In Table 3, we can see that the proposed algorithm achieved the best performance, and the performance margin is greater than that of the LT imbalance settings. ReMixMatch+CReST+PDA showed relatively low performance compared to the other algorithms.

Table 3: Overall accuracy/minority-class accuracy on CIFAR-1010 under a step imbalance setting
CIFAR-1010-Step,   γ=100\gamma=100,   β=20%\beta=20\%
Algorithm w/ - w/ CReST+PDA [34] w/ DARP+cRT [18] w/ ABC
FixMatch [29] 54.054.0±0.84\pm 0.84 / 11.811.8±1.71\pm 1.71 71.171.1±0.78\pm 0.78 / 48.248.2±2.26\pm 2.26 69.869.8±1.51\pm 1.51 / 45.145.1±2.70\pm 2.70 75.9±0.49\pm 0.49 / 57.0±1.07\pm 1.07
ReMixMatch [3] 60.860.8±0.10\pm 0.10 / 25.125.1±1.28\pm 1.28 64.664.6±0.97\pm 0.97 / 33.533.5±2.05\pm 2.05 72.372.3±1.77\pm 1.77 / 50.650.6±3.53\pm 3.53 76.4±1.70\pm 1.70 / 65.7±1.30\pm 1.30

To evaluate the performance of the proposed algorithm on a large-scale dataset, we also conducted experiments on the LSUN dataset [37], which is naturally a long-tailed dataset. Among the algorithms considered in Tables 2 and 3, those combined with CReST were excluded for comparison, because CReST requires loading of the whole unlabeled data in the repeated process of updating pseudo-labels, which is not possible for the large-scale LSUN dataset. Instead, we additionally considered FixMatch+cRT and ReMixMatch+cRT for comparison. The experimental results are presented in Table 4. The proposed algorithm showed better performance than the other baseline algorithms. DARP resulted in degradation of the performance, possibly because the scale of the LSUN dataset is very large. Specifically, DARP solves a convex optimization with all unlabeled data points to refine the pseudo labels. As the scale of the unlabeled dataset increases, this optimization problem becomes more difficult to solve and, consequently, the pseudo-labels could be refined inaccurately. Unlike the results for other datasets, the algorithms combined with FixMatch performed better than the algorithms combined with ReMixMatch.

Table 4: Overall accuracy/minority-class accuracy for the large-scale LSUN dataset
LSUN,    γ=100\gamma=100,   β=20%\beta=20\%
Algorithm w/ - w/ cRT [17] w/ DARP [18] w/ DARP+cRT [18]        w/ ABC
FixMatch [29] 73.173.1 / 55.355.3 77.077.0 / 71.571.5 71.071.0 / 51.851.8 75.875.8 / 69.569.5 78.9 / 75.5
ReMixMatch [3] 69.469.4 / 49.149.1 75.475.4 / 69.569.5 65.665.6 / 44.144.1 72.172.1 / 67.567.5 76.9 / 69.5

4.3 Complexity of the proposed algorithm

The proposed algorithm requires additional parameters for the ABC, but the number of the additional parameters is negligible compared to the number of parameters of the backbone. For example, the ABC additionally required only 0.09%0.09\% and 0.87%0.87\% of the number of backbone parameters for CIFAR-1010 with 1010 classes and CIFAR-100100 with 100100 classes, respectively. Moreover, because the ABC shares the representation layer of the backbone, it does not significantly increase the memory usage and training time. Furthermore, we could train the proposed algorithm on the large-scale LSUN dataset without a significant increase in computation cost, because the entire training procedure could be carried out using minibatches of data. In contrast, the algorithms combined with DARP required convex optimization for all pseudo-labels, which significantly increased the computation cost as the number of classes or the amount of data increased. Similarly, it required significant time to train the algorithms combined with CReST, because CReST requires iterative re-training with a labeled set expanded by adding unlabeled data points with pseudo-labels. We present the floating point operations per second (FLOPS) for each algorithm using Nvidia Tesla-V100 in Appendix I.

4.4 Qualitative analysis of high-quality representations and balanced classification

The ABC can use high-quality representations learned by the backbone when performing balanced classification. To verify this, in Figure 3, we present t-distributed stochastic neighbor embedding (t-SNE) [31] of the representations of the CIFAR-1010 test set learned by the ABC (without SSL backbone), FixMatch+ABC, and ReMixMatch+ABC on CIFAR-1010-LT under the main setting. Different colors indicate different classes. As expected, “ABC (without SSL backbone)" failed to learn class-separable representations because sufficient data were not used for training while using the 0/10/1 mask. In contrast, by training the backbone (FixMatch or ReMixMatch) together with the ABC, the proposed algorithm could use the entire data and learn high-quality representations. In this example, ReMixMatch produced more separable representations than FixMatch, which shows that the choice of the backbone affects the performance of the proposed algorithm, as expected.

Refer to caption          Refer to caption          Refer to caption
 (a)ABC (without SSL backbone)           (b)FixMatch+ABC          (c) ReMixMatch+ABC
Figure 3: t-SNE of the proposed algorithm and the ABC (without SSL backbone)

The proposed algorithm can also mitigate class imbalance by using the ABC. To verify this, we compare the confusion matrices of the predictions on the test set of CIFAR-1010 using ReMixMatch, ReMixMatch+DARP+cRT, and ReMiMatch+ABC trained on CIFAR-1010 under the main setting in Figure 4. In the confusion matrices, the value in the iith row and the jjth column represents the ratio of the amount of data belonging to the iith class to the amount of data predicted as the jjth class. Each cell has a darker red color when the ratio is larger. We can see that ReMixMatch often misclassified data points in the minority classes (e.g., classes 88 and 99 into classes 0 and 11). This may be because ReMixMatch does not consider class imbalance, and thus biased pseudo-labels were used for training. ReMixMatch+DARP+cRT produced a more balanced class-distribution compared to ReMixMatch by additionally using DARP+cRT. However, a significant number of data points in the minority classes were still misclassified as majority classes. In contrast, ReMixMatch+ABC classified the test data points in the minority classes with higher accuracy, and produced a significantly more balanced class distribution than ReMixMatch+DARP+cRT, as shown in Figure 4 (c). As both ReMixMatch+DARP+cRT and ReMixMatch+ABC use ReMixMatch to learn representations, the performance gap between these two algorithms results from the different characteristics of the ABC versus DARP+cRT as follows. First, DARP+cRT does not use unlabeled data for training its classifier after representations learning is completed, whereas the ABC uses unlabeled data with unbiased pseudo-labels for its training. Second, whereas DARP+cRT decouples the learning of representations and training of a classifier, the ABC is trained end-to-end interactively with representations learned by the backbone. We also present the confusion matrices of the predictions on the test set of CIFAR-1010 using FixMatch, FixMatch+DARP+cRT, and FixMatch+ABC as well as the confusion matrices of the pseudo-labels on the same dataset using ReMixMatch, ReMixMatch+DARP+cRT, ReMixMatch+ABC, FixMatch, FixMatch+DARP+cRT, and FixMatch+ABC in Appendix J. Moreover, we compare the ABC and the classifier of DARP+cRT in more detail using the validation loss plots in Appendix K.

Refer to caption Refer to caption Refer to caption Refer to caption
 (a)ReMixMatch (b)ReMixMatch+DARP+cRT (c) ReMixMatch+ABC
Figure 4: Confusion matrices of the predictions on the test set of CIFAR-1010

4.5 Ablation study

We conducted an ablation study on CIFAR-1010-LT in the main setting to investigate the effect of each element of the proposed algorithm. The results for ReMixMatch+ABC are presented in Table 5, where each row indicates the proposed algorithm with the described conditions in that row. The results are summarized as follows. 1) If we did not gradually decrease the parameter of the Bernoulli distribution ()\mathcal{B}\left(\cdot\right) when conducting consistency regularization, then an overbalance problem occurred because of unlabeled data misclassified as minority classes. 2) Without consistency regularization for the ABC, the decision boundary did not clearly separate each class. 3) Without using the 0/10/1 mask for LclsL_{cls} and LconL_{con}, the ABC was trained to be biased toward the majority classes. 4) Without confidence threshold τ\tau for consistency regularization, training became unstable and, consequently, the ABC was trained to be biased toward certain classes. 5) Similarly, if hard pseudo-labels, instead of soft pseudo-labels, were used for consistency regularization, then the ABC was biased toward certain classes. 6) If the ABC was solely used without the backbone, the performance decreased because the ABC could not use high-quality representations learned by the backbone. 7) When we used a re-weighting technique [13] instead of a mask for the ABC, training became unstable because of abnormally large gradients calculated for training on the data of the minority classes. 8) The decoupled training of the backbone and ABC resulted in decreased classification performance, as was also analyzed in Section 4.4. Similarly, we present the results of the ablation study for FixMatch+ABC in Appendix L.

Table 5: Ablation study for ReMixMatch+ABC on CIFAR-1010-LT, γ=100\gamma=100, β=20%\beta=20\%
Ablation study Overall Minority
ReMixMatch+ABC (proposed algorithm) 82.4\mathbf{82.4} 75.7\mathbf{75.7}
Without gradually decreasing the parameter of ()\mathcal{B}\left(\cdot\right) for consistency regularization 81.881.8 74.674.6
Without consistency regularization for the ABC 79.479.4 66.966.9
Without using the 0/1 mask for the consistency regularization loss LconL_{con} 79.079.0 69.269.2
Without using the 0/1 mask for the classification loss LclsL_{cls} 74.474.4 57.857.8
Without using the confidence threshold τ\tau for consistency regularization 74.374.3 75.475.4
Using hard pseudo labels for consistency regularization 70.270.2 75.175.1
Without training backbone (ABC without SSL backbone) 68.768.7 56.256.2
Training the ABC with a re-weighting technique 81.281.2 74.174.1
Decoupled training of the backbone and ABC 79.579.5 72.372.3

5 Conclusion

We introduced the ABC, which is attached to a state-of-the-art SSL algorithm, for CISSL. The ABC can utilize high-quality representations learned by the backbone, while being trained to make class-balanced predictions. The ABC also utilizes unlabeled data by conducting consistency regularization in a modified way for class-imbalance problems. The experimental results obtained under various settings demonstrate that the proposed algorithm outperforms the baseline algorithms. We also conducted a qualitative analysis and an ablation study to verify the contribution of each element of the proposed algorithm. The proposed algorithm assumes that the labeled and unlabeled data are class-imbalanced to the same extent. In the future, we plan to release this assumption by adopting a module for estimating class distribution. Deep learning algorithms can be applied to many societal problems. However, if the training data are imbalanced, the algorithms could be trained to make socially biased decisions in favor of the majority groups. The proposed algorithm can contribute to solving these issues. However, there is also a potential risk that the proposed algorithm could be used as a tool to identify minorities and discriminate against them. It should be ensured that the proposed method cannot be used for any purpose that may have negative social impacts.

6 Acknowledgments

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (2018R1C1B6004511, 2020R1A4A10187747).

References

  • An et al., [2021] An, J., Ying, L., and Zhu, Y. (2021). Why resampling outperforms reweighting for correcting sampling bias with stochastic gradients. In International Conference on Learning Representations.
  • Barandela et al., [2003] Barandela, R., Rangel, E., Sánchez, J. S., and Ferri, F. J. (2003). Restricted decontamination for the imbalanced training sample problem. In Iberoamerican congress on pattern recognition, pages 424–431. Springer.
  • [3] Berthelot, D., Carlini, N., Cubuk, E. D., Kurakin, A., Sohn, K., Zhang, H., and Raffel, C. (2019a). Remixmatch: Semi-supervised learning with distribution matching and augmentation anchoring. In International Conference on Learning Representations.
  • [4] Berthelot, D., Carlini, N., Goodfellow, I., Papernot, N., Oliver, A., and Raffel, C. A. (2019b). Mixmatch: A holistic approach to semi-supervised learning. In Wallach, H., Larochelle, H., Beygelzimer, A., d'Alché-Buc, F., Fox, E., and Garnett, R., editors, Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc.
  • Buda et al., [2018] Buda, M., Maki, A., and Mazurowski, M. A. (2018). A systematic study of the class imbalance problem in convolutional neural networks. Neural Networks, 106:249–259.
  • Cao et al., [2019] Cao, K., Wei, C., Gaidon, A., Arechiga, N., and Ma, T. (2019). Learning imbalanced datasets with label-distribution-aware margin loss. In Wallach, H., Larochelle, H., Beygelzimer, A., d'Alché-Buc, F., Fox, E., and Garnett, R., editors, Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc.
  • Chawla et al., [2002] Chawla, N. V., Bowyer, K. W., Hall, L. O., and Kegelmeyer, W. P. (2002). Smote: synthetic minority over-sampling technique. Journal of artificial intelligence research, 16:321–357.
  • Cubuk et al., [2020] Cubuk, E. D., Zoph, B., Shlens, J., and Le, Q. V. (2020). Randaugment: Practical automated data augmentation with a reduced search space. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, pages 702–703.
  • Cui et al., [2019] Cui, Y., Jia, M., Lin, T.-Y., Song, Y., and Belongie, S. (2019). Class-balanced loss based on effective number of samples. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9268–9277.
  • DeVries and Taylor, [2017] DeVries, T. and Taylor, G. W. (2017). Improved regularization of convolutional neural networks with cutout. arXiv preprint arXiv:1708.04552.
  • Gidaris et al., [2018] Gidaris, S., Singh, P., and Komodakis, N. (2018). Unsupervised representation learning by predicting image rotations. In International Conference on Learning Representations.
  • Grandvalet and Bengio, [2005] Grandvalet, Y. and Bengio, Y. (2005). Semi-supervised learning by entropy minimization. In Saul, L., Weiss, Y., and Bottou, L., editors, Advances in Neural Information Processing Systems, volume 17. MIT Press.
  • He and Garcia, [2009] He, H. and Garcia, E. A. (2009). Learning from imbalanced data. IEEE Transactions on knowledge and data engineering, 21(9):1263–1284.
  • Huang et al., [2016] Huang, C., Li, Y., Loy, C. C., and Tang, X. (2016). Learning deep representation for imbalanced classification. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 5375–5384.
  • Jamal et al., [2020] Jamal, M. A., Brown, M., Yang, M.-H., Wang, L., and Gong, B. (2020). Rethinking class-balanced methods for long-tailed visual recognition from a domain adaptation perspective. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 7610–7619.
  • JAPKOWICZ, [2000] JAPKOWICZ, N. (2000). The class imbalance problem: Significance and strategies. In Proc. 2000 International Conference on Artificial Intelligence, volume 1, pages 111–117.
  • Kang et al., [2020] Kang, B., Xie, S., Rohrbach, M., Yan, Z., Gordo, A., Feng, J., and Kalantidis, Y. (2020). Decoupling representation and classifier for long-tailed recognition. In International Conference on Learning Representations.
  • [18] Kim, J., Hur, Y., Park, S., Yang, E., Hwang, S. J., and Shin, J. (2020a). Distribution aligning refinery of pseudo-label for imbalanced semi-supervised learning. In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M. F., and Lin, H., editors, Advances in Neural Information Processing Systems, volume 33, pages 14567–14579. Curran Associates, Inc.
  • [19] Kim, J., Jeong, J., and Shin, J. (2020b). M2m: Imbalanced classification via major-to-minor translation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 13896–13905.
  • Kingma and Ba, [2015] Kingma, D. P. and Ba, J. (2015). Adam: A method for stochastic optimization. In ICLR, 2015.
  • Krizhevsky, [2009] Krizhevsky, A. (2009). Learning multiple layers of features from tiny images. Technical report, Department of Computer Science, University of Toronto.
  • Lee et al., [2013] Lee, D.-H. et al. (2013). Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks. In Workshop on challenges in representation learning, ICML, volume 3.
  • Mikolov et al., [2013] Mikolov, T., Sutskever, I., Chen, K., Corrado, G. S., and Dean, J. (2013). Distributed representations of words and phrases and their compositionality. In Burges, C. J. C., Bottou, L., Welling, M., Ghahramani, Z., and Weinberger, K. Q., editors, Advances in Neural Information Processing Systems, volume 26. Curran Associates, Inc.
  • Miyato et al., [2018] Miyato, T., Maeda, S.-i., Koyama, M., and Ishii, S. (2018). Virtual adversarial training: a regularization method for supervised and semi-supervised learning. IEEE transactions on pattern analysis and machine intelligence, 41(8):1979–1993.
  • Netzer et al., [2011] Netzer, Y., Wang, T., Coates, A., Bissacco, A., Wu, B., and Ng, A. Y. (2011). Reading digits in natural images with unsupervised feature learning. In NIPS Workshop, 2011.
  • Park et al., [2018] Park, S., Park, J., Shin, S.-J., and Moon, I.-C. (2018). Adversarial dropout for supervised and semi-supervised learning. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 32.
  • Ren et al., [2020] Ren, J., Yu, C., sheng, s., Ma, X., Zhao, H., Yi, S., and Li, h. (2020). Balanced meta-softmax for long-tailed visual recognition. In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M. F., and Lin, H., editors, Advances in Neural Information Processing Systems, volume 33, pages 4175–4186. Curran Associates, Inc.
  • Ren et al., [2018] Ren, M., Zeng, W., Yang, B., and Urtasun, R. (2018). Learning to reweight examples for robust deep learning. In International Conference on Machine Learning, pages 4334–4343. PMLR.
  • Sohn et al., [2020] Sohn, K., Berthelot, D., Carlini, N., Zhang, Z., Zhang, H., Raffel, C. A., Cubuk, E. D., Kurakin, A., and Li, C.-L. (2020). Fixmatch: Simplifying semi-supervised learning with consistency and confidence. Advances in Neural Information Processing Systems, 33.
  • Tarvainen and Valpola, [2017] Tarvainen, A. and Valpola, H. (2017). Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In Guyon, I., Luxburg, U. V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., and Garnett, R., editors, Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc.
  • Van der Maaten and Hinton, [2008] Van der Maaten, L. and Hinton, G. (2008). Visualizing data using t-sne. Journal of machine learning research, 9(11).
  • Verma et al., [2019] Verma, V., Lamb, A., Kannala, J., Bengio, Y., and Lopez-Paz, D. (2019). Interpolation consistency training for semi-supervised learning. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI-19, pages 3635–3641. International Joint Conferences on Artificial Intelligence Organization.
  • Wang et al., [2017] Wang, Y.-X., Ramanan, D., and Hebert, M. (2017). Learning to model the tail. In Guyon, I., Luxburg, U. V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., and Garnett, R., editors, Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc.
  • Wei et al., [2021] Wei, C., Sohn, K., Mellina, C., Yuille, A., and Yang, F. (2021). Crest: A class-rebalancing self-training framework for imbalanced semi-supervised learning. arXiv preprint arXiv:2102.09559.
  • Yang and Xu, [2020] Yang, Y. and Xu, Z. (2020). Rethinking the value of labels for improving class-imbalanced learning. In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M. F., and Lin, H., editors, Advances in Neural Information Processing Systems, volume 33, pages 19290–19301. Curran Associates, Inc.
  • Yin et al., [2018] Yin, X., Yu, X., Sohn, K., Liu, X., and Chandraker, M. (2018). Feature transfer learning for deep face recognition with under-represented data. arXiv e-prints, pages arXiv–1803.
  • Yu et al., [2015] Yu, F., Zhang, Y., Song, S., Seff, A., and Xiao, J. (2015). Lsun: Construction of a large-scale image dataset using deep learning with humans in the loop. arXiv preprint arXiv:1506.03365.
  • Zagoruyko and Komodakis, [2016] Zagoruyko, S. and Komodakis, N. (2016). Wide residual networks. arXiv preprint arXiv:1605.07146.
  • Zhai et al., [2019] Zhai, X., Oliver, A., Kolesnikov, A., and Beyer, L. (2019). S4l: Self-supervised semi-supervised learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 1476–1485.

Appendix A Overall procedure of consistency regularization for ABC

Figure 5 illustrates the overall procedure of consistency regularization for the ABC. Detailed procedure is described in Section 3.4 of the main paper.

Refer to caption
Figure 5: Overall procedure of consistency regularization for the ABC

Appendix B Pseudo code of the proposed algorithm

The pseudo code of the proposed algorithm is presented in Algorithm 1. The for loop (lines 2 14) can be run in parallel. The classification loss LclsL_{cls} and consistency regularization loss LconL_{con} are expressed in detail in Sections 3.3 and 3.4 of the main paper.

Algorithm 1 Pseudo code of the proposed algorithm

Input: 𝒳={(xb,yb):b(1,,B)}𝒳\mathcal{MB}_{\mathcal{X}}=\left\{\left(x_{b},y_{b}\right):b\in\left(1,...,B\right)\right\}\subset\mathcal{X}, 𝒰={(ub):b(1,,B)}𝒰\mathcal{MB}_{\mathcal{U}}=\left\{\left(u_{b}\right):b\in\left(1,\ldots,B\right)\right\}\subset\mathcal{U}
      Output: Classification model f:d{1,L}f:\mathbb{R}^{d}\rightarrow\left\{1,...L\right\}
      Parameters : θ\boldsymbol{\theta} (Parameters of Wide ResNet-2828-22 and ABC)

1:while Training do
2:     for b=1b=1 to BB do
3:         α(xb)=\alpha\left(x_{b}\right)= Augment(xb)\left(x_{b}\right)
4:         α(ub)=\alpha\left(u_{b}\right)= WeakAugment(ub)\left(u_{b}\right)
5:         𝒜k(ub)=\mathcal{A}_{k}\left(u_{b}\right)=StrongAugment(xb)k{}_{k}\left(x_{b}\right), k=1,2k=1,2
6:         Predicted class distribution for α(xb)=ps(y|α(xb))\alpha\left(x_{b}\right)=p_{s}\left(y|\alpha\left(x_{b}\right)\right)
7:         Generate 0/10/1 mask M(xb)M\left(x_{b}\right).
8:         Soft pseudo label qbq_{b} == ps(y|α(ub))p_{s}\left(y|\alpha\left(u_{b}\right)\right)
9:         if max(qb)0.95\max\left(q_{b}\right)\geq 0.95 then
10:              Predicted class distribution for 𝒜k(ub)=ps(y|𝒜k(ub)),k=1,2\mathcal{A}_{k}\left(u_{b}\right)=p_{s}\left(y|\mathcal{A}_{k}\left(u_{b}\right)\right),k=1,2
11:              Generate 0/10/1 mask M(ub)M\left(u_{b}\right).
12:         end if
13:         Loss from the backbone LbackL_{back} += backbone(α(xb),α(ub),𝒜k(ub))\left(\alpha\left(x_{b}\right),\alpha\left(u_{b}\right),\mathcal{A}_{k}\left(u_{b}\right)\right)
14:     end for
15:     Calculate the classification loss LclsL_{cls} and consistency regularization loss LconL_{con}.
16:     Total Loss Ltotal=Lcls+Lcon+LbackL_{total}=L_{cls}+L_{con}+L_{back}
17:     Δ𝜽𝜽Ltotal\Delta\boldsymbol{\theta}\propto\nabla_{\boldsymbol{{\theta}}}L_{total}, 𝜽𝜽+Δ𝜽\quad\boldsymbol{\theta}\leftarrow\boldsymbol{\theta}+\Delta\boldsymbol{\theta}
18:end while

Appendix C Two types of class imbalance for the considered datasets

Refer to caption          Refer to caption
 (a) Long-tailed imbalance           (b) Step imbalance
Figure 6: Long-tailed imbalance and step imbalance

Two types of class imbalance for the considered datasets are illustrated in Figure 6. For both types of imbalance, we set γ=100\gamma=100, N1=1000N_{1}=1000, and β=20%\beta=20\%. In Figure 6 (b), we can see that each minority class has a very small amount of data. Existing SSL algorithms can be significantly biased toward majority classes under step imbalanced settings.

Appendix D Specification of the confidence threshold τ\tau

Table 6: Mean and standard deviation (STD) of validation accuracy during the last 50 epochs
ReMixMatch+ABC CIFAR-10-LT,   γ=100\gamma=100,   β=20%\beta=20\%
τ\tau 1 0.98 0.95 0.9 0.85 0.8 0.75 0.7
Mean and STD 78.9, 0.36 81.8, 0.34 82.3, 0.2 81.3, 0.32 81.5, 0.39 81.2, 0.63 80.0, 2.87 79.0, 5.76

In general, the confidence threshold τ\tau should be set high enough, but not too high. If τ\tau is low, training becomes unstable because many misclassified unlabeled data points would be used for training. However, if τ\tau is too high, most of the unlabeled data points would not be used for consistency regularization. Based on these insights, we set τ\tau as 0.95 in our experiments. We confirmed via experiments that this value of τ\tau enabled high accuracy as well as stability. Specifically, we conducted experiments on CIFAR-1010-LT for the main setting while changing the value of τ\tau. We measured the validation accuracy of ReMixMatch+ABC during the last 50 epochs (1 epoch=500 iterations) of training and calculated the mean and standard deviation (STD) of these values. As can be seen from Table 6, the proposed algorithm achieved the highest mean and lowest STD of the validation accuracy when τ\tau was 0.95. When τ\tau was set higher or lower than 0.95, the mean of the validation accuracy decreased. In particular, as the value of τ\tau decreased from 0.95, the STD increased rapidly, indicating instability of the training.

Appendix E Further details of the experimental setup

We describe further details of the experimental setup. To train the ReMixMatch, we gradually increased the coefficient of the loss associated with the unlabeled data points, following [18]. We found that without this gradual increase, the validation loss of the ReMixMatch did not converge. To train the FixMatch, we used the labeled dataset once more as an unlabeled dataset by removing the labels for the experiments using CIFAR-100100 following the previous study [29], but not for the experiments using CIFAR-1010 and SVHN, because it did not improve the performance. We followed the default settings for the ReMixMatch [3] and FixMatch [29], unless mentioned otherwise.

To train the ABC, we also gradually decreased the parameter of ()\mathcal{B\left(\cdot\right)} for calculating the classification loss in the experiments using CIFAR-1010 and SVHN under the step imbalanced setting. This prevents unstable training by allowing each labeled data point of the majority classes to be more frequently used for training.

Appendix F Geometric mean (G-mean) of class-wise accuracy under the main setting

Table 7: Performance comparison using G-mean for the main setting
CIFAR-1010-LT SVHN-LT CIFAR-100100-LT
Algorithm γ=100\gamma=100, β=20%\beta=20\% γ=100\gamma=100, β=20%\beta=20\% γ=20\gamma=20, β=40%\beta=40\%
FixMatch [29] 62.062.0 87.387.3 38.538.5
w/ CReST+PDA [34] 74.474.4 88.688.6 42.342.3
w/ DARP [18] 71.571.5 87.687.6 40.440.4
w/ DARP+cRT [18] 76.776.7 89.889.8 47.047.0
w/ ABC 80.5 91.8 49.0
ReMixMatch [3] 62.562.5 89.589.5 41.241.2
w/ CReST+PDA [34] 72.272.2 90.790.7 43.143.1
w/ DARP [18] 71.971.9 89.789.7 42.542.5
w/ DARP+cRT [18] 77.977.9 92.092.0 48.348.3
w/ ABC 81.9 93.8 50.8

To evaluate whether the proposed algorithm performs in a balanced way for all classes, we also measured the performance for the main setting using the geometric mean (G-mean) of class-wise accuracy with correction to avoid zeroing. We set the hyperparameter for the correction to avoid zeroing as 1%\%, which indicates that the minimum class-wise accuracy is 1%\%. The results in Table 7 demonstrates that the proposed algorithm performed in a balanced way.

Appendix G Experimental results on SVHN and CIFAR-100100 under various settings

For the experiments using SVHN with γ=150\gamma=150 and β=20%\beta=20\%, the solution of the convex optimization problem of ReMixMatch+DARP+cRT for refining the pseudo labels did not converge, and thus we could not measure the performance. The experimental results for SVHN and CIFAR-100100 under various settings showed the same trend as those for CIFAR-1010, which is described in Section 4.2 of the main paper.

Table 8: Overall accuracy/minority-class-accuracy on SVHN under various settings
SVHN-LT
Algorithm γ=100\gamma=100, β=10%\beta=10\% γ=100\gamma=100, β=30%\beta=30\% γ=50\gamma=50, β=20%\beta=20\% γ=150\gamma=150, β=20%\beta=20\%
FixMatch [29] 88.588.5±0.25\pm 0.25 / 80.380.3±0.42\pm 0.42 88.788.7±0.36\pm 0.36 / 80.780.7±0.65\pm 0.65 91.191.1±0.18\pm 0.18 / 85.385.3±0.28\pm 0.28 85.685.6±0.17\pm 0.17 / 74.674.6±0.43\pm 0.43
w/ CReST + PDA [34] 89.289.2±0.43\pm 0.43 / 81.781.7±0.95\pm 0.95 89.989.9±0.36\pm 0.36 / 83.083.0±0.37\pm 0.37 91.791.7±0.86\pm 0.86 / 87.687.6±0.53\pm 0.53 86.786.7±0.89\pm 0.89 / 76.776.7±1.70\pm 1.70
w/ DARP + cRT [18] 89.389.3±0.33\pm 0.33 / 83.983.9±0.47\pm 0.47 90.790.7±0.28\pm 0.28 / 84.884.8±0.37\pm 0.37 92.192.1±0.30\pm 0.30 / 87.787.7±0.44\pm 0.44 88.088.0±0.74\pm 0.74 / 80.180.1±1.88\pm 1.88
w/ ABC 92.3±0.38\pm 0.38 / 88.7±0.92\pm 0.92 92.3±0.34\pm 0.34 / 88.3±0.49\pm 0.49 93.5±0.17\pm 0.17 / 90.7±0.25\pm 0.25 91.2±0.15\pm 0.15 / 86.2±0.15\pm 0.15
ReMixMatch [3] 89.289.2±0.17\pm 0.17 / 81.781.7±0.41\pm 0.41 90.790.7±0.15\pm 0.15 / 84.584.5±0.46\pm 0.46 92.492.4±0.21\pm 0.21 / 87.887.8±0.48\pm 0.48 88.688.6±0.16\pm 0.16 / 80.480.4±0.42\pm 0.42
w/ CReST + PDA [34] 89.889.8±0.12\pm 0.12 / 83.083.0±0.08\pm 0.08 91.291.2±0.17\pm 0.17 / 85.385.3±0.24\pm 0.24 93.393.3±0.02\pm 0.02 / 90.090.0±0.36\pm 0.36 88.888.8±0.41\pm 0.41 / 80.780.7±0.82\pm 0.82
w/ DARP + cRT [18] 91.791.7±0.26\pm 0.26 / 86.686.6±0.45\pm 0.45 93.293.2±0.08\pm 0.08 / 89.389.3±0.21\pm 0.21 93.693.6±0.41\pm 0.41 / 90.490.4±0.52\pm 0.52 -/ -
w/ ABC 93.2±0.64\pm 0.64 / 92.2±0.44\pm 0.44 94.4±0.37\pm 0.37 / 93.3±0.32\pm 0.32 94.7±0.35\pm 0.35 / 93.5±0.56\pm 0.56 93.2±0.46\pm 0.46 / 91.8±0.79\pm 0.79
Table 9: Overall accuracy/minority-class-accuracy on CIFAR-100100 under various settings
CIFAR-100100-LT
Algorithm γ=20\gamma=20, β=20%\beta=20\% γ=20\gamma=20, β=50%\beta=50\% γ=10\gamma=10, β=40%\beta=40\% γ=30\gamma=30, β=40%\beta=40\%
FixMatch [29] 46.146.1±0.23\pm 0.23 / 26.626.6±0.34\pm 0.34 52.352.3±0.54\pm 0.54 / 34.734.7±0.80\pm 0.80 57.457.4±0.15\pm 0.15 / 44.844.8±0.17\pm 0.17 47.647.6±0.09\pm 0.09 / 27.627.6±0.21\pm 0.21
w/ CReST + PDA [34] 46.746.7±0.49\pm 0.49 / 29.329.3±0.54\pm 0.54 52.752.7±0.06\pm 0.06 / 37.437.4±0.37\pm 0.37 57.357.3±0.23\pm 0.23 / 47.547.5±0.22\pm 0.22 48.548.5±0.06\pm 0.06 / 30.030.0±0.04\pm 0.04
w/ DARP + cRT [18] 48.948.9±0.11\pm 0.11 / 33.533.5±0.17\pm 0.17 55.955.9±0.43\pm 0.43 / 43.543.5±1.28\pm 1.28 59.059.0±0.40\pm 0.40 / 50.450.4±1.09\pm 1.09 51.351.3±0.29\pm 0.29 / 36.436.4±0.50\pm 0.50
w/ ABC 49.7±0.40\pm 0.40 / 34.6±0.76\pm 0.76 58.3±0.74\pm 0.74 / 46.7±1.12\pm 1.12 61.6±0.15\pm 0.15 / 53.0±0.26\pm 0.26 53.6±0.35\pm 0.35 / 38.8±0.69\pm 0.69
ReMixMatch [3] 49.049.0±0.29\pm 0.29 / 29.929.9±0.42\pm 0.42 54.454.4±0.13\pm 0.13 / 37.837.8±0.12\pm 0.12 59.559.5±0.20\pm 0.20 / 47.147.1±0.42\pm 0.42 51.051.0±0.11\pm 0.11 / 32.032.0±0.50\pm 0.50
w/ CReST + PDA [34] 49.449.4±0.32\pm 0.32 / 31.831.8±0.15\pm 0.15 54.454.4±0.21\pm 0.21 / 38.638.6±0.35\pm 0.35 58.858.8±0.08\pm 0.08 / 47.647.6±0.24\pm 0.24 51.951.9±0.34\pm 0.34 / 33.533.5±0.69\pm 0.69
w/ DARP + cRT [18] 50.250.2±0.40\pm 0.40 / 35.235.2±0.55\pm 0.55 54.654.6±1.75\pm 1.75 / 44.844.8±2.09\pm 2.09 59.459.4±1.04\pm 1.04 / 52.152.1±0.71\pm 0.71 52.852.8±0.24\pm 0.24 / 38.438.4±0.30\pm 0.30
w/ ABC 52.5±0.10\pm 0.10 / 38.5±0.42\pm 0.42 59.3±0.66\pm 0.66 / 49.5±1.02\pm 1.02 63.5±0.29\pm 0.29 / 57.1±0.06\pm 0.06 55.4±0.46\pm 0.46 / 42.8±0.67\pm 0.67

Appendix H Experimental results on SVHN and CIFAR-100100 under the step imbalanced setting

Experimental results for SVHN and CIFAR-100100 under the step imbalanced setting showed the same tendency as that for CIFAR-1010, which is described in Section 4.2 of the main paper.

Table 10: Overall accuracy/minority-class-accuracy on SVHN under step imbalanced setting
SVHN-Step,   γ=100\gamma=100,   β=20%\beta=20\%
Algorithm w/ - w/ CReST + PDA [34] w/ DARP + cRT [18] w/ ABC
FixMatch [29] 79.879.8±1.34\pm 1.34 / 61.561.5±2.76\pm 2.76 86.686.6±0.19\pm 0.19 / 76.376.3±0.23\pm 0.23 85.985.9±0.28\pm 0.28 / 74.374.3±0.37\pm 0.37 91.2±0.15\pm 0.15 / 85.6±0.35\pm 0.35
ReMixMatch [3] 82.782.7±0.42\pm 0.42 / 67.467.4±0.81\pm 0.81 85.985.9±0.13\pm 0.13 / 73.973.9±0.16\pm 0.16 90.590.5±1.13\pm 1.13 / 84.384.3±1.86\pm 1.86 91.3±1.61\pm 1.61 / 89.8±0.95\pm 0.95
Table 11: Overall accuracy/minority-class-accuracy on CIFAR-100100 under step imbalanced setting
CIFAR-100100-Step,   γ=20\gamma=20,   β=40%\beta=40\%
Algorithm w/ - w/ CReST + PDA [34] w/ DARP + cRT [18] w/ ABC
FixMatch [29] 46.746.7±0.29\pm 0.29 / 15.015.0±0.26\pm 0.26 49.949.9±0.24\pm 0.24 / 26.726.7±0.41\pm 0.41 50.750.7±0.61\pm 0.61 / 28.828.8±2.04\pm 2.04 54.7±0.06\pm 0.06 / 32.1±0.12\pm 0.12
ReMixMatch [3] 47.347.3±0.12\pm 0.12 / 16.516.5±1.06\pm 1.06 48.548.5±0.18\pm 0.18 / 19.219.2±0.35\pm 0.35 53.653.6±0.28\pm 0.28 / 35.035.0±1.04\pm 1.04 56.0±0.46\pm 0.46 / 38.3±0.55\pm 0.55

Appendix I Floating point operations per second (FLOPS) of each algorithm

As we mentioned in Section 4.3 of the main paper, computation cost required for the algorithms combined with DARP increased as the number of classes or the amount of data increased. In contrast, computation cost required for the proposed algorithm did not significantly increased because the whole training procedure can be carried out using minibatches. FLOPS of FixMatch+CReST and ReMixMatch+CReST are the same as those of FixMatch and ReMixMatch, but the algorithms combined with CReST required iterative re-training with a labeled set expanded by adding unlabeled data points with pseudo labels. We measured FLOPS using Nvidia Tesla-V100. For the experiments on CIFAR-1010 and CIFAR-100100, we used only one GPU, whereas we used four GPUs in parallel for the experiments on LSUN.

Table 12: FLOPS of each algorithm
Algorithm CIFAR-1010 CIFAR-100100 LSUN
FixMatch [29] 14.514.5 iter/sec 14.714.7 iter/sec 2.62.6 iter/sec
FixMatch+DARP [18] 12.012.0 iter/sec 6.36.3 iter/sec 0.40.4 iter/sec
FixMatch+ABC 11.211.2 iter/sec 11.011.0 iter/sec 2.02.0 iter/sec
ReMixMatch [3] 6.96.9 iter/sec 6.96.9 iter/sec 1.31.3 iter/sec
ReMixMatch+DARP [18] 6.36.3 iter/sec 4.54.5 iter/sec 0.30.3 iter/sec
ReMixMatch+ABC 5.85.8 iter/sec 5.65.6 iter/sec 0.90.9 iter/sec

Appendix J Further qualitative analysis and quantitative comparison

Figure 7 (b) presents biased predictions of FixMatch [29], a recent SSL algorithm, trained on CIFAR-1010 with the amount of Class 0 being 100 times more than that of Class 9 as depicted in Figure 7 (a). In contrast, Figure 7 (c) presents that the class distribution of the predicted labels became more balanced using the FixMatch+ABC trained on the same dataset. These results are consistent with those in Figure 1 of the main paper.

Refer to caption Refer to caption Refer to caption
 (a) Class-imbalanced training set      (b) FixMatch       (c) FixMatch+ABC
Figure 7: Predictions on a class-balanced test set of CIFAR-1010 using FixMatch (b) and the FixMatch+ABC (c) trained on a class-imbalanced training set (a).

Because the use of the 0/10/1 mask for the ABC plays a similar role of re-sampling techniques, we compare the representations of proposed algorithm with those of SMOTE (oversampling technique) [7]+CNN, and random undersampling [13]+CNN. Figure 8 (a), (b) and (c) present the t-SNE representations obtained using SMOTE+CNN, undersampling+CNN, and ABC only. Because re-sampling techniques can only be applied to labeled data, they cannot be combined with the SSL algorithms, and thus they were combined with CNN instead. SMOTE+CNN and undersampling+CNN learned less separable representations than the ABC only. These results show that using the 0/10/1 mask instead of re-sampling techniques is more effective because we could utilize unlabeled data. In addition, the 0/1 mask enabled the ABC to be combined with the backbone, so that the ABC could use the high-quality representations learned by the backbone as shown in Figure 8 (d).

Refer to caption Refer to caption Refer to caption Refer to caption
 (a) SMOTE [7]+CNN (b) Undersampling [13]+CNN (c) ABC only (d) ReMixMatch+ABC
Figure 8: t-SNE of the representations of the CIFAR-1010 test set using re-sampling+CNN, ABC only, and ReMixMatch+ABC trained on CIFAR-1010-LT, γ=100\gamma=100, β=20%\beta=20\%

We also compared the performance of the proposed algorithm with those of SMOTE+CNN and undersampling+CNN. The results in Table 13 show the importance of using unlabeled data for training and using the high-quality representations obtained from backbone.

Table 13: Performance of each algorithm in Figure 8 and FixMatch+ABC. The algorithms were trained on CIFAR-1010-LT, γ=100\gamma=100, β=20%\beta=20\% and tested on the test set of CIFAR-1010.
Performance of each algorithm in Figure 8 and FixMatch+ABC Overall Minority
ReMixMatch+ABC 82.4\mathbf{82.4} 75.7\mathbf{75.7}
FixMatch+ABC 81.181.1 72.072.0
Without training backbone (ABC only) 68.768.7 56.256.2
SMOTE+CNN 60.860.8 46.146.1
Undersampling+CNN 48.748.7 55.855.8

Figure 9 presents the confusion matrices of FixMatch, FixMatch+DARP+cRT, and FixMatch+ABC trained on CIFAR-1010-LT, γ=100\gamma=100, β=20%\beta=20\%. Similar to Figure 4 of the main paper, FixMatch and FixMatch+DARP+cRT often misclassified test data points in the minority classes (e.g., classes 88 and 99 into classes 0 and 11). In contrast, FixMatch+ABC classified the test data points in the minority classes with higher accuracy, and produced a significantly more balanced class-distribution than FixMatch and FixMatch+DARP+cRT.

Refer to caption Refer to caption Refer to caption Refer to caption
 (a)FixMatch (b)FixMatch+DARP+cRT (c) FixMatch+ABC
Figure 9: Confusion matrices of the predictions on the test set of CIFAR-1010

Figure 10 presents the confusion matrices of the predictions on the unlabeled data. Similar to Figure 9 and Figure 4 of the main paper, FixMatch+ABC and ReMixMatch+ABC classified the unlabeled data points in the minority classes with higher accuracy, and produced a significantly more balanced pseudo labels than other algorithms. By using these balanced pseudo labels for training, the proposed algorithm could make a more balanced prediction on the test set.

Refer to caption Refer to caption Refer to caption
 (a)FixMatch (b)FixMatch+DARP+cRT (c) FixMatch+ABC
 Refer to caption Refer to caption Refer to caption
 (d)ReMixMatch (e)ReMixMatch+DARP+cRT (f) ReMixMatch+ABC
Figure 10: Confusion matrices of the predictions on the unlabeled data of CIFAR-1010

Appendix K Detailed comparison between the end-to-end training of the proposed algorithm and decoupled learning of representations and a classifier

Although FixMatch+DARP+cRT and ReMixMatch+DARP+cRT also use the representations learned by ReMixMatch [3] and FixMatch [29], they showed worse performance than the proposed algorithm. The performance gap between FixMatch(ReMixMatch)+DARP+cRT and the proposed algorithm results from the different characteristics of the ABC versus DARP+cRT as follows. First, whereas DARP+cRT decouples learning of representations and training of a classifier, the ABC is trained end-to-end interactively with representations that the backbone learns. Second, DARP+cRT does not use unlabeled data for training of its classifier after representations learning is finished, while the ABC is trained with unlabeled data to conduct consistency regularization so that decision boundaries can be placed in a low density region. To verify these reasons, we compare the validation loss graphs of the algorithms based on end-to-end training and decoupled learning of representations and a classifier in Figure 11. We recorded the validation loss of 100 epochs after the representations were fixed, where 1 epoch was set as 500 iterations. For the proposed algorithm, we recorded the validation loss of the last 100 epochs. In Figure 9 (a) and (b), we can see that the validation loss of the algorithms based on decoupled learning of representations and a classifier tended to increase after a few epochs. The validation loss was reduced by conducting consistency regularization (C/R) using unlabeled data, but it still tended to increase. In the case of ReMixMatch+DARP+cRT+C/R* and FixMatch+DARP+cRT+C/R*, which do not fix the representations (algorithms marked with *), high-quality representations learned by the backbone were gradually replaced by the representations learned with a re-balanced classifier, which caused overfitting on minority classes. We can observe a similar tendency in Figure 9 (c) under the supervised learning setting. In contrast, the validation loss of ReMixMatch+ABC, FixMatch+ABC, and the proposed algorithm under supervised learning setting decreased steadily and achieved the lowest validation loss. The performances of the algorithms based on end-to-end training and decoupled learning of representations and a classifier are summarized in Table 14.

Refer to caption Refer to caption Refer to caption
 (a)With ReMixMatch (b)With FixMatch (c) Supervised setting
Figure 11: Validation loss graphs of algorithms based on end-to-end training and decoupled learning of representations and a classifier on the test set of CIFAR-1010. The algorithms in (a) and (b) were trained on the train set of CIFAR-1010-LT with γ=100\gamma=100, β=20%\beta=20\%, N1=1000N_{1}=1000, and the algorithms in (c) were trained on the training set of CIFAR-1010-LT with γ=100\gamma=100, β=100%\beta=100\%, N1=5000N_{1}=5000. C/R and * in the graphs indicate consistency regularization and non-fixed representations, respectively.
Table 14: Performance of the algorithms based on end-to-end training and decoupled learning of representations and a classifier. The algorithms were trained on the training set described in the caption of Figure 11. The algorithms were tested on the test set of CIFAR-1010.
Performance of the algorithms based on end-to-end training versus decoupled learning Overall Minority
Under the semi-supervised learning setting
ReMixMatch+ABC (end-to-end training) 82.4\mathbf{82.4} 75.7\mathbf{75.7}
ReMixMatch+DARP+cRT+C/R* (Decoupled learning) 80.680.6 71.471.4
ReMixMatch+DARP+cRT+C/R (Decoupled learning) 79.579.5 70.470.4
ReMixMatch+DARP+cRT (Decoupled learning) 78.578.5 66.466.4
FixMatch+ABC (end-to-end training) 81.1\mathbf{81.1} 72.0\mathbf{72.0}
FixMatch+DARP+cRT+C/R* (Decoupled learning) 80.380.3 71.671.6
FixMatch+DARP+cRT+C/R (Decoupled learning) 78.778.7 68.368.3
FixMatch+DARP+cRT (Decoupled learning) 78.178.1 66.666.6
Under the supervised learning setting
End-to-End training of CNN with the ABC (end-to-end training) 84.9\mathbf{84.9} 80.6\mathbf{80.6}
cRT* (Decoupled learning of representations and the classifier of CNN) 81.181.1 79.679.6
cRT (Decoupled learning of representations and the classifier of CNN) 80.080.0 79.979.9

Appendix L Ablation study for FixMatch [29] + ABC on CIFAR-10

Results in Table 15 show a similar tendency as that for ReMixMatch+ABC in Section 4.5 of the main paper.

Table 15: Ablation study for FixMatch+ABC on CIFAR-1010-LT, γ=100\gamma=100, β=20%\beta=20\%
Ablation study Overall Minority
FixMatch+ABC (proposed algorithm) 81.1\mathbf{81.1} 72.0\mathbf{72.0}
Without gradually decreasing the parameter of ()\mathcal{B}\left(\cdot\right) for consistency regularization 80.280.2 70.170.1
Without consistency regularization for the ABC 76.276.2 60.960.9
Without using the 0/1 mask for the consistency regularization loss LconL_{con} 74.974.9 58.858.8
Without using the 0/1 mask for the classification loss LclsL_{cls} 77.177.1 62.762.7
Without using the confidence threshold τ\tau for consistency regularization 79.279.2 67.867.8
Using hard pseudo labels for consistency regularization 78.878.8 68.068.0
Without training backbone (ABC only) 68.768.7 56.256.2
Training the ABC with a re-weighting technique 80.380.3 70.570.5
Decoupled training of the backbone and ABC 77.477.4 65.065.0