Texture-guided Saliency Distilling for Unsupervised Salient Object Detection
Abstract
Deep Learning-based Unsupervised Salient Object Detection (USOD) mainly relies on the noisy saliency pseudo labels that have been generated from traditional handcraft methods or pre-trained networks. To cope with the noisy labels problem, a class of methods focus on only easy samples with reliable labels but ignore valuable knowledge in hard samples. In this paper, we propose a novel USOD method to mine rich and accurate saliency knowledge from both easy and hard samples. First, we propose a Confidence-aware Saliency Distilling (CSD) strategy that scores samples conditioned on samples’ confidences, which guides the model to distill saliency knowledge from easy samples to hard samples progressively. Second, we propose a Boundary-aware Texture Matching (BTM) strategy to refine the boundaries of noisy labels by matching the textures around the predicted boundaries. Extensive experiments on RGB, RGB-D, RGB-T, and video SOD benchmarks prove that our method achieves state-of-the-art USOD performance. Code is available at www.github.com/moothes/A2S-v2.
1 Introduction
Unsupervised Salient Object Detection (USOD) methods aim to correctly localize and precisely segment salient objects simultaneously without using manual annotations. Compared to the supervised methods, USOD methods can easily adapt to more practical scenarios (e.g., industrial or medical images) where a large number of labeled images may be very hard to collect. Moreover, USOD methods also can assist some related methods for other tasks, e.g., object recognition [47, 17] and object detection [12, 38]. However, diverse objects, complex backgrounds, and other challenging conditions bring severe challenges to USOD methods.
Most Deep Learning-based (DL-based) methods [68, 73, 35, 31, 22, 58] base on the saliency cues extracted by traditional SOD methods (Fig. 1-c and 1-d). These handcrafted features related cues are employed as pseudo labels to train deep networks under certain constraints, e.g., binary cross-entropy (BCE) loss. However, saliency cues by traditional methods usually shift away from target objects, especially in complex scenes. Moreover, conventional constraints, such as BCE loss, works well on fully-supervised SOD methods, but is suboptimal when fitting the noisy labels for unsupervised methods (Fig. 1-e). Recently, Zhou et al. [80] addressed the first issue by extracting saliency cues (Fig. 1-f) based on a unsupervisedly pre-trained network (Fig. 1-g) instead of using traditional methods. During training, they focus on learning reliable saliency knowledge from easy samples, but ignore latent knowledge in hard samples. The main reason is that hard samples may be wrongly-labeled and corrupt the fragile saliency knowledge learned in the early training phase. Therefore, to leverage hard samples, we argue that all samples should be employed in a meaningful order (i.e., from high reliable to low reliable), which is crucial for mining accurate knowledge from noisy labels. Trained by such a strategy, the network can mine valuable knowledge from hard examples without corrupting the knowledge learned from easy samples.








Deep networks can learn to localize salient regions from noisy labels [80], but still struggle to find the precise boundaries of target objects. Generally, the appearance around saliency boundary has a similar texture as in saliency map. Therefore, matching the textures between different maps can serve as a guidance for producing reasonable saliency boundaries. We will demonstrate that above strategies are applicable to multimodal data besides RGB image, including depth map, thermal image, and optical flow.
Based on the above analysis, we propose a novel framework to tackle the Unsupervised Salient Object Detection (USOD) tasks. Specifically, two strategies are proposed to mine saliency knowledge from noisy saliency labels. First, we propose a Confidence-aware Saliency Distilling (CSD) scheme that scores samples with noisy labels conditioned on samples’ confidences. Then, our CSD guides the network to learn saliency knowledge from easy samples to more complex ones progressively by employing an adaptive loss conditioned on the training progress. Second, we propose a Boundary-aware Texture Matching (BTM) strategy to refine the saliency boundaries of noisy labels by matching the textures around the predicted boundaries. During training, the predicted saliency boundaries are shifting toward surrounding edges in the appearance space of the whole image. Finally, guided by above two mechanisms, our method can produce high-quality pseudo labels to train generalized saliency detectors. Extensive experiments on RGB, RGB-D, RGB-T, and video SOD benchmarks prove that our method achieves state-of-the-art performance compared to existing USOD methods.
The main contributions of our novel USOD method are:
-
1.
We propose a Confidence-aware Saliency Distilling (CSD) to mine rich and accurate saliency knowledge from noisy labels, which breaks through the limitation that existing methods cannot utilize hard samples.
-
2.
We propose a Boundary-aware Texture Matching (BTM) to refine the boundary of the predicted saliency maps by matching textures in different spaces.
-
3.
Extensive experiments on RGB, RGB-D, RGB-T and video SOD benchmarks prove that our method achieves state-of-the-art USOD performance.
2 Related Works
2.1 Supervised Salient Object Detection
Researchers have developed a large family of fully-supervised Salient Object Detection (SOD) algorithms [32, 62, 16, 5, 57, 76, 75, 20, 74, 42, 79, 81, 7] in the past decades. Ronneberger et al. [44] proposed a U-shape structure that progressively upsamples and concatenates the smaller features to the larger ones. To ease the annotation burden, Zhang et al. [72] relabeled the DUTS-TR dataset [55] with scribbles and leveraged an edge detection model for boundary localization. Yu et al. [67] proposed a local coherence loss to find precise boundary based on scribbles annotations.
Multimodal SOD tasks aim at using other modality data to improve the SOD performance, such as depth map, thermal image and optical flow. Recently, abundant methods [61, 15, 51, 48, 82, 8, 18, 43, 4, 49, 50] were proposed. Most of these methods employ a two-stream encoder-decoder structure to aggregate multi-level information in multimodal data. To reduce the annotating cost, Zhao et al. [78] proposed a video SOD dataset with scribbles annotations to indicate the location of salient objects.
Although the above methods have achieved significant performance, they require numerous human annotations for training, which are expensive to collect.

2.2 Unsupervised Salient Object Detection
Traditional SOD methods [28, 23, 84, 64] extracted saliency cues from images by modeling the correlation of hand-crafted features. Inspired by the centralization prior, Jiang et al. [23] considered the distances between boundary superpixels and non-boundary superpixels as saliency scores. Yan et al. [64] employed a tree-structure graphical model to compute the saliency results based on several over-segmented maps. The above methods fail to accurately localize salient objects because the global information in hand-crafted features are not representative enough.
Existing DL-based USOD methods can be divided into two pipelines based on the method used to extract saliency cues from images. First, most USOD methods [68, 73, 35, 71, 31, 58] focused on refining the coarse saliency cues extracted by several traditional SOD methods. For example, Zhang et al. [68] weighted these saliency cues by combining both intra-image and inter-image fusion streams. Zhang et al. [73] designed a noise modeling module to deal with noises in these saliency cues. Nguyen et al. [35] used these saliency cues as labels to train multiple deep networks. Ji et al. [22] refined the saliency maps extracted by traditional SOD methods to produce more accurate saliency predictions. Second, to prevent the localization errors caused by traditional SOD methods, Zhou et al. [80] proposed a novel framework that converts the activation maps of a pre-trained network to high-quality pseudo labels.
In this paper, we propose two novel mechanisms to tackle USOD tasks and explore the potential of improving the quality of pseudo labels by using multimodal data.
3 Our Approach
In our method, we propose two novel strategies to mine accurate saliency knowledge based on the noisy activation maps generated by a deep network , as shown in Fig. 2.
Activation map generation. Following previous work [80], we generate an activation map for input image using a deep network :
(1) |
where is the spatial mean of . Specifically, we employ the ResNet-50 [19] pre-trained by MoCo-v2 [6] as our encoder, which is trained without extra manual annotations. In our decoder, four SE blocks [21] integrate multiple encoder features into . After that, we set to ensure that (1) noisy labels are adaptive to input images when using a fixed threshold and (2) coexisting of positive and negative samples. Next, features are summed over the channel dimension to produce one-channel activation map. Finally, the function produces the final saliency scores, and the function identifies the regions with more corner pixels as background.

Training strategy. The activation maps produced by network perceive some discriminative regions in input images [80], but still are low-quality because of widespread noises. To improve the quality, we employ three strategies to train the network by:
(2) |
where all are hyperparameters. is the loss for the predictions of a different scale. First, we propose a Confidence-aware Saliency Distilling (CSD) scheme to excavate valuable saliency knowledge from simple examples to more complex ones. Second, we propose a Boundary-aware Texture Matching (BTM) strategy to align the boundaries in appearance and the predicted saliency maps. In addition, a multi-scale consistency loss ensures our method to produce consistent predictions for multi-scale inputs.
3.1 Confidence-aware Saliency Distilling
The primary challenge for USOD methods is how to localize salient regions. Existing DL-based USOD methods [68, 35, 80] are based on noisy saliency cues from traditional SOD methods [85, 23, 28, 84] or pre-trained networks [6, 19]. Following [80], we employ the activation maps produced by a pre-trained network as our saliency cues instead of the saliency predictions of traditional SOD methods. The main reasons are two-fold: (1) traditional methods require extra computation loads; (2) network can focus on mining saliency knowledge instead of fitting the inductive bias of traditional methods. We binarize the saliency cues as the initial labels using a fixed threshold of 0.5, while Eqn. 1 ensures that these labels remain adaptive to the input image.
To generate high-quality pseudo labels, USPS [35] uses these initial labels to train deep networks like other fully-supervised methods. However, the simple training strategy in fully-supervised methods is suboptimal for unsupervised methods. Specifically, we define pixels with saliency scores close to 0.5 as hard examples because of their low confidences. In the noisy labels, hard examples are likely to be wrongly labeled. Therefore, it is difficult to learn robust saliency knowledge from these hard samples using traditional loss functions (e.g., BCE loss). To this end, A2S [80] learns reliable saliency knowledge from easy samples. However, for hard examples, the saliency knowledge hidden in noises is not fully explored. In summary, a customized strategy that organizes the samples in a more meaningful order is crucial for USOD methods.
To address the above problem, we propose a Confidence-aware Saliency Distillation (CSD) scheme that scores samples with noisy labels conditional on their confidences and training progress. Concretely, easy samples contain reliable knowledge, and thus can assist our method to learn reliable saliency knowledge. On the contrary, the saliency knowledge in hard samples is hidden within noises, and may corrupt the fragile saliency patterns in the early stages of network training. Inspired by self-paced learning [2], we dynamically adjust the gradients for samples by introducing a factor , which is linearly increasing from 0 to 1 as training proceeding. Our loss can be formulated as:
(3) |
where is the number of pixels, and is the saliency prediction of pixel . Notice that noisy label G is omitted in Eqn. 3 because it is generated based on . Thus, we can calculate the confidence score of pixel in a simpler way as: , where the constant value 0.5 is related to in Eqn. 1. The partial derivative of our over is:
(4) |
where for negative and positive values, respectively.
For an intuitive comparison, we draw the gradient landscapes of different losses in Fig. 3, including , , [80] and our . The gradient of other losses are consistent throughout the training process, while our varies with the training process. Specifically, in the beginning, our assigns low gradients to hard samples to learn reliable saliency knowledge from easy samples. As training proceeding, the gradients of hard samples are increasing to mine more valuable saliency knowledge.

3.2 Boundary-aware Texture Matching
In general, the appearance of saliency boundary has a similar texture as in saliency prediction. Therefore, matching the textures between different maps can guide our method to produce reasonable saliency scores. This strategy is also applicable to other modalities, such as depth map, thermal image, and optical flow, as shown in Fig. 4. By aggregating the rich appearance information in multimodal data, we can provide a more generalized guidance for our USOD method.
Based on the above analysis, we propose a Boundary-aware Texture Matching (BTM) strategy to align the saliency boundaries and image edges by matching their textures. First of all, similar to the LBP feature [36], we extract texture vectors from saliency predictions and appearance, respectively. For saliency predictions, the texture vector for the -th pixel is computed by , where is neighborhoods around pixel . For appearance information, it may be one of RGB image, optical flow, depth map or thermal image. To extract more distinctive features from multimodal data, we formulate the texture vector as , where is a hyperparameter and means the appearance data of modal . It is noteworthy that a small difference between two pixels produces a small element in , but a large element in . Therefore, is defined as the matching penalty for considering pixel as saliency boundary, contrary to similarity. Finally, the formula of our is:
(5) |
where is the binary boundary mask of saliency prediction.
An intuitive example is shown in Fig. 5. For the predicted boundary pixels, we find some nearby pixels that differ significantly in saliency scores. In general, the appearances between boundary and these pixels are also different because they locate within salient objects and backgrounds, respectively. If not the case, we expect the saliency score of this boundary pixel to be close to those nearby pixels, so that the boundary will shift in the opposite direction. After an iterative training process, our method can align the predicted saliency boundaries with image edges.

3.3 Multi-scale Consistency
Salient objects are consistent in multi-scale inputs. Thus, we resize input images to a reference scale and encourage our method to produce consistent predictions by:
(6) |
where and are saliency predictions of different scales.
3.4 Training a Generalized Detector
Following previous USOD methods [35, 31, 80], we generate pseudo labels based on saliency predictions through and use them to train an extra saliency detector with the IOU loss. For RGB SOD, we use the same detector as the previous method [80]. For multimodal SOD tasks, we employ the MIDD [50] network as our detector. To ensure a fair comparison with existing SOD methods, we train the extra detector only using task-specific data and corresponding pseudo labels.
Methods | Year | Sup. | ECSSD | MSRA-B | DUT-O | PASCAL-S | DUTS-TE | HKU-IS | ||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Trained on MSRA-B | ||||||||||||||||||||
SBF [68] | 2017 | U | .812 | .878 | .087 | .867 | .929 | .058 | .611 | .771 | .106 | .711 | .795 | .131 | .627 | .785 | .105 | .805 | .895 | .074 |
MNL [73] | 2018 | U | .874 | .906 | .069 | .881 | .932 | .053 | .683 | .821 | .076 | .792 | .846 | .091 | – | – | – | .874 | .932 | .047 |
USPS [35] | 2019 | U | .875 | .903 | .064 | .896 | .938 | .042 | .715 | .839 | .069 | .770 | .828 | .107 | .730 | .840 | .072 | .880 | .933 | .043 |
DCFD [31] | 2022 | U | .880 | .900 | .064 | .903 | .938 | .041 | .731 | .838 | .064 | .773 | .830 | .105 | .744 | .832 | .068 | .887 | .926 | .044 |
A2S [80] | 2022 | U | .888 | .911 | .064 | .902 | .941 | .041 | .719 | .841 | .069 | .790 | .838 | .106 | .750 | .860 | .065 | .887 | .937 | .042 |
Ours | – | U | .902 | .923 | .056 | .912 | .948 | .036 | .731 | .851 | .065 | .803 | .848 | .099 | .767 | .871 | .061 | .891 | .939 | .041 |
Trained on DUTS-TR | ||||||||||||||||||||
MINet [37] | 2020 | F | .924 | .953 | .033 | .903 | .948 | .038 | .756 | .873 | .055 | .842 | .899 | .064 | .828 | .917 | .037 | .908 | .961 | .028 |
LDF [59] | 2020 | F | .930 | .951 | .034 | .902 | .944 | .037 | .773 | .881 | .052 | .853 | .903 | .062 | .855 | .929 | .034 | .914 | .960 | .028 |
KRN [63] | 2021 | F | .931 | .951 | .032 | .911 | .950 | .036 | .793 | .893 | .050 | .851 | .894 | .068 | .865 | .934 | .033 | .920 | .961 | .027 |
WSSA [72] | 2020 | W | .870 | .917 | .059 | .869 | .929 | .049 | .703 | .845 | .068 | .785 | .855 | .096 | .742 | .869 | .062 | .860 | .932 | .047 |
MFNet [41] | 2021 | W | .844 | .889 | .084 | .872 | .923 | .059 | .621 | .784 | .098 | .756 | .824 | .115 | .693 | .832 | .079 | .839 | .919 | .058 |
SCW [67] | 2021 | W | .900 | .931 | .049 | .898 | .940 | .040 | .758 | .862 | .060 | .827 | .879 | .080 | .823 | .890 | .049 | .896 | .943 | .038 |
EDNS [71] | 2020 | U | .872 | .906 | .068 | .880 | .932 | .051 | .682 | .821 | .076 | .801 | .846 | .097 | .735 | .847 | .065 | .874 | .933 | .046 |
SelfMask [46] | 2022 | U | .856 | .920 | .058 | .844 | .925 | .050 | .668 | .815 | .078 | .774 | .856 | .087 | .714 | .848 | .063 | .819 | .915 | .053 |
DCFD [31] | 2022 | U | .888 | .915 | .059 | .888 | .930 | .045 | .710 | .837 | .070 | .795 | .860 | .090 | .764 | .855 | .064 | .889 | .935 | .042 |
Ourss1 | – | U | .847 | .912 | .057 | .849 | .925 | .054 | .622 | .773 | .111 | .763 | .840 | .093 | .676 | .814 | .082 | .819 | .914 | .053 |
Ours | – | U | .916 | .938 | .044 | .904 | .944 | .039 | .745 | .863 | .061 | .830 | .882 | .074 | .810 | .901 | .047 | .902 | .947 | .037 |
Oursmm | – | U | .913 | .942 | .043 | .899 | .945 | .039 | .737 | .855 | .064 | .831 | .890 | .073 | .803 | .899 | .048 | .898 | .946 | .037 |

4 Experiments
Implementation details.
All experiments are implemented on a single GTX 1080 Ti GPU. The batch size is 8 and input images are resized to . The reference scale is randomly selected from . Only horizontal flipping is employed as our data augmentation strategy. We train our method for 20 epochs using the SGD optimizer with an initial learning rate of 0.1, which is decayed linearly. , , , and are set to 200, 1, 0.05 and 1, respectively. We train the extra detectors for 10 epochs using the SGD optimizer with a learning rate of 0.005.
Methods | Year | Sup. | RGBD135 | NJUD | NLPR | SIP | ||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
DSA2F [48] | 2021 | F | .899 | .958 | .021 | .901 | .937 | .039 | .897 | .953 | .024 | – | – | – |
SPNet [82] | 2021 | F | .927 | .984 | .013 | – | – | – | .903 | .959 | .021 | .893 | .931 | .043 |
CCFE [30] | 2022 | F | .911 | .964 | .020 | .914 | .953 | .032 | .907 | .962 | .021 | .889 | .923 | .047 |
DSU [22] | 2022 | U | .767 | .895 | .061 | .719 | .797 | .135 | .745 | .879 | .065 | .619 | .774 | .156 |
Ours | – | U | .834 | .945 | .037 | .787 | .829 | .109 | .834 | .924 | .043 | .716 | .784 | .124 |
Oursmm | – | U | .877 | .946 | .029 | .862 | .908 | .060 | .852 | .931 | .034 | .873 | .925 | .051 |
Methods | Year | Sup. | DAVSOD | DAVIS | SegV2 | FBMS | ||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
PCSA [18] | 2020 | F | .556 | .749 | .077 | .794 | .922 | .022 | .789 | .931 | .019 | .783 | .868 | .042 |
TENet [43] | 2020 | F | .595 | .773 | .067 | .821 | .941 | .017 | – | – | – | .851 | .915 | .026 |
STVS [4] | 2021 | F | .563 | .764 | .080 | .812 | .940 | .022 | .835 | .950 | .016 | .821 | .903 | .042 |
WSVSOD [78] | 2021 | W | .492 | .710 | .103 | .731 | .900 | .036 | .711 | .909 | .031 | .736 | .840 | .084 |
Ours | – | U | .534 | .747 | .084 | .751 | .913 | .042 | .751 | .914 | .033 | .732 | .794 | .100 |
Oursmm | – | U | .547 | .762 | .085 | .756 | .908 | .037 | .808 | .927 | .021 | .795 | .876 | .060 |
Methods | Year | Sup. | VT5000 | VT1000 | VT821 | ||||||
---|---|---|---|---|---|---|---|---|---|---|---|
MIED [49] | 2020 | F | .761 | .880 | .050 | .853 | .928 | .030 | .760 | .877 | .050 |
MIDD [50] | 2021 | F | .801 | .899 | .043 | .882 | .942 | .027 | .805 | .898 | .045 |
APNet [83] | 2021 | F | .821 | .918 | .035 | .885 | .951 | .021 | .818 | .912 | .034 |
CCFE [30] | 2022 | F | .859 | .937 | .030 | .906 | .963 | .018 | .857 | .934 | .027 |
Ours | – | U | .810 | .904 | .046 | .885 | .939 | .031 | .805 | .900 | .043 |
Oursmm | – | U | .807 | .903 | .047 | .881 | .939 | .032 | .805 | .899 | .044 |
Datasets. In our experiments, we follow the prevalent settings of different SOD tasks. Specifically, for RGB SOD, we use the training subsets of MSRA-B [33] or DUTS [55] to train our method, respectively. ECSSD [45], PASCAL-S [29], HKU-IS [27], DUTS-TE [55], DUT-O [65], and the testing subset of MSRA-B are employed for evaluation. For RGB-D SOD, we choose 2185 samples from NLPR [39] and NJUD [24] as the training set. RGBD135 [9], SIP [14] and the testing subsets of NJUD and NLPR are employed for evaluation. For RGB-T SOD, 2500 images in VT5000 [52] are for training, while VT1000 [53], VT821 [54] and the rest 2500 images in VT5000 are for testing. For video SOD, we choose the training splits of DAVIS [40] and DAVSOD [15] to train our method. Moreover, we randomly select 5 frames of each video in DAVSOD to avoid overfitting. The testing splits of DAVIS, DAVSOD, SegV2 [26] and FBMS [3] are employed for evaluation.
Metrics. We adopt three criteria for evaluation, including ave-, Mean Absolute Error () and E-Measure () [13]. Specifically, where is set to 0.3 [1]. The ave- is the scores by setting the threshold as two times mean values. is the absolute error between predictions and ground truth. measures the global statistics and local pixel matching information.
4.1 Results on RGB SOD
As shown in Tab. 1, we compare the proposed method with fully-supervised methods, MINet [37], LDF [59], and KRN [63], weakly-supervised methods, WSSA [72], MFNet [41], and SCW [67], and unsupervised methods, SBF [68], MNL [73], USPS [35], EDNS [71], A2S [80], DCFD [31], and SelfMask [46]. We list our results with different settings: (1) saliency results without extra detector and post-processing, denoted as “Ourss1”; (2) our full method as “Ours”; (3) training our method using multimodal data, while only DUTS-TR or MSRA-B datasets with pseudo labels are employed to train the extra detector (“Oursmm”).
Either training on MSRA-B or DUTS-TR, the proposed method achieves significant improvements compared to existing USOD methods. Moreover, our unsupervised method is competitive to recent weakly-supervised SCW [67] and fully-supervised methods KRN [63]. Furthermore, the results of “Ourss1” prove that our method extracts precise saliency knowledge from training samples, but somehow weaken when deploying on unseen images. In addition, the similar results of “Oursmm” and “Ours” imply that extra multimodal data may not bring significant improvements to RGB SOD task. As proved in [56], the large-scale DUTS-TR dataset is saturated for training saliency detectors, even in the unsupervised case.
A qualitative comparison is illustrated in Fig. 10. Overall, our saliency predictions are much more precise than other unsupervised methods. For example, in the first example, our method and two fully-supervised methods well segment target object with precise boundaries, while other methods fail to capture the tiny salient object.
4.2 Results on Multimodal SOD
We conduct more experiments on multimodal tasks, including RGB-D, RGB-T and video SOD. Noted that we pre-compute the optical flow for each video frame as an extra modality. To verify the improvements brought by multimodal data, we build two sets for training: (1) Task-specific data (“Ours”). For example, for RGB-T SOD task, we only use 2500 RGB-thermal image pairs for training as prevalent RGB-T SOD methods [49, 50, 83]; (2) Multimodal data from the training sets of four SOD tasks (“Oursmm”).
Overall, our method reports state-of-the-art performance on all multimodal SOD tasks. Specifically, for the RGB-D SOD task in Tab. 2, our method surpasses the latest unsupervised method DSU [22]. For video and RGB-T SOD tasks in Tab. 3 and 4, to the best of our knowledge, our method is the first unsupervised method and is competitive to supervised methods. Furthermore, our method trained on multimodal data (“Oursmm”) produce high-quality pseudo labels for all multimodal SOD datasets simultaneously and has achieved performance improvements on most tasks.
Methods | RGB | RGB-D | VSOD | RGB-T | ||||
---|---|---|---|---|---|---|---|---|
Ours | .917 | .038 | .746 | .110 | .557 | .099 | .932 | .029 |
Oursrgb | .917 | .040 | .804 | .068 | .613 | .089 | .920 | .030 |
Oursmm | .918 | .040 | .826 | .062 | .637 | .079 | .923 | .029 |
4.3 Ablation Study
Label quality comparison.
In Tab. 5, we exhibit the scores of pseudo labels on multimodal SOD datasets. Except for the above two training sets, we employ an additional set that collects all RGB images from four tasks to train our method, denoted as “Oursrgb”. Overall, “Oursmm” reports more generalized results on all datasets. Moreover, training on task-specific data (“Ours”) slightly surpasses the performance of “Oursmm” on the training sets of RGB and RGB-T tasks. Since our method is trained without ground truth, fitting a hybrid dataset may cause a slight performance drop on some subsets. In addition, the comparison between “Oursrgb” and “Oursmm” prove that multimodal data significantly improves the quality of the generated pseudo labels. For an intuitive comparison, we show some examples in Fig. 7. Trained with task-specific data, our method precisely localizes salient regions but fails to segment the whole objects. With more training data, the network perceives fine-grained concepts of target objects, resulting in more complete segmentations.
Impact of initial saliency cues. Our method is based on the activation maps of a pre-trained network [19, 6] instead of traditional methods [23, 84, 28] as other DL-based USOD methods [68, 73, 35, 58, 31]. We compare the performance of these two types of designs in Tab. 6. Overall, the activation map is worse before training, whereas better than traditional methods [23, 28, 84] after training. The activation maps are dynamic during training, such that the network strengthens the learned saliency knowledge incrementally. On the contrary, using traditional methods, the network learns from fixed labels throughout the training process, and thus fits the biased knowledge in those traditional methods. In addition, traditional methods introduce extra computation loads, while the pre-trained network is indispensable for all DL-based USOD methods.

Saliency cues | Before training | After training | ||||
---|---|---|---|---|---|---|
MC [23] | .810 | .877 | .145 | .889 | .914 | .053 |
DSR [28] | .780 | .867 | .118 | .886 | .901 | .059 |
RBD [84] | .803 | .883 | .109 | .891 | .912 | .053 |
Ours | .451 | .659 | .353 | .914 | .940 | .039 |
Effectiveness of loss functions. We conduct ablation studies on various combinations of loss functions in Tab. 7 and more details are introduced as follows.
In ablation study A, our method receives continuous improvements by appending new losses ( and ) or replacing the original to our . This experiment prove that these three losses can assist the network mine more detailed saliency knowledge from different perspectives. More importantly, these losses are supplementary to each other, such that their combination can collaborate to produce high-quality pseudo labels for SOD datasets.
In ablation study B, our loss can provide precise saliency localization information and thus achieves the best performance among all competitors. Specifically, and fail to mine detailed and reliable saliency knowledge from noisy saliency cues. [80] improves the robustness of the learned saliency knowledge by focusing on easy samples, and thus surpasses the loss. [2] also employs a dynamic schedule, but its binary weights filter out latent saliency knowledge in hard samples.
Tag | Loss | |||
A1 | [80] (Baseline) | .882 | .915 | .071 |
A2 | [80] + | .891 | .921 | .066 |
A3 | + | .908 | .937 | .050 |
A4 | + + (Ours) | .917 | .945 | .038 |
B1 | + + | .558 | .540 | .208 |
B2 | + + | .895 | .928 | .044 |
B3 | [80] + + | .903 | .934 | .042 |
B4 | [2] + + | .910 | .938 | .041 |
B5 | + + (Ours) | .917 | .945 | .038 |
C1 | + + | .735 | .819 | .120 |
C2 | + [67] + | .906 | .923 | .052 |
C3 | + + | .907 | .926 | .051 |
C4 | + + (Ours) | .917 | .945 | .038 |
In ablation study C, we compare our with [67] and two variants: (1) using L1 distance for the texture features in appearance space, denoted as ; (2) removing the boundary masks in , denoted as . In general, our outperforms other variants. Specifically, the L1 distance in causes the texture features to be not distinctive enough, resulting in ambiguous saliency boundaries. focuses more on the adjacent pixels so that the boundaries are more susceptible to a limited region rather than a larger patch. Moreover, using , edges within objects or backgrounds may corrupt the learned saliency knowledge.
Pre-training | |||
---|---|---|---|
Supervised | .915 | .942 | .039 |
Unsupervised (Ours) | .917 | .945 | .038 |
Supervised or unsupervised pre-training? Under the unsupervised setting, whether we can use an encoder with supervised pre-training is controversial. For the ImageNet dataset [11], the class labels of images usually indicate the category of the most salient object. It means that the supervised encoder receives extra saliency knowledge from these manual labels. Thus, for completely unsupervised SOD, we employ unsupervised MoCo-v2 to initialize our encoder. As listed in Tab. 8, the performance of our method when using supervised encoder is comparable.
5 Conclusion
In this paper, we propose an Unsupervised Salient Object Detection (USOD) method guided by two novel mechanisms. First, we propose a Confidence-aware Saliency Distilling (CSD) to learn saliency knowledge from easy samples to hard ones with noisy labels progressively. Second, we propose a Boundary-aware Texture Matching (BTM) to make the location of saliency prediction more accurate. As a result, the proposed method produces high-quality pseudo labels to train saliency detectors. Experiments on RGB, RGB-D, RGB-T, and video SOD benchmarks prove that our method outperforms existing USOD methods.
References
- [1] Radhakrishna Achanta, Sheila Hemami, Francisco Estrada, and Sabine Susstrunk. Frequency-tuned salient region detection. In 2009 IEEE conference on computer vision and pattern recognition, pages 1597–1604. IEEE, 2009.
- [2] Yoshua Bengio, Jérôme Louradour, Ronan Collobert, and Jason Weston. Curriculum learning. In Proceedings of the 26th annual international conference on machine learning, pages 41–48, 2009.
- [3] Thomas Brox and Jitendra Malik. Object segmentation by long term analysis of point trajectories. In European conference on computer vision, pages 282–295. Springer, 2010.
- [4] Chenglizhao Chen, Guotao Wang, Chong Peng, Yuming Fang, Dingwen Zhang, and Hong Qin. Exploring rich and efficient spatial temporal interactions for real-time video salient object detection. IEEE Transactions on Image Processing, 30:3995–4007, 2021.
- [5] Peijia Chen, Jianhuang Lai, Guangcong Wang, and Huajun Zhou. Confidence-guided adaptive gate and dual differential enhancement for video salient object detection. In 2021 IEEE International Conference on Multimedia and Expo (ICME), pages 1–6. IEEE, 2021.
- [6] Xinlei Chen, Haoqi Fan, Ross Girshick, and Kaiming He. Improved baselines with momentum contrastive learning. arXiv preprint arXiv:2003.04297, 2020.
- [7] Zixuan Chen, Huajun Zhou, Jianhuang Lai, Lingxiao Yang, and Xiaohua Xie. Contour-aware loss: Boundary-aware learning for salient object segmentation. IEEE Transactions on Image Processing, 30:431–443, 2020.
- [8] Xiaolong Cheng, Xuan Zheng, Jialun Pei, He Tang, Zehua Lyu, and Chuanbo Chen. Depth-induced gap-reducing network for rgb-d salient object detection: An interaction, guidance and refinement approach. IEEE Transactions on Multimedia, 2022.
- [9] Yupeng Cheng, Huazhu Fu, Xingxing Wei, Jiangjian Xiao, and Xiaochun Cao. Depth enhanced saliency detection method. In Proceedings of international conference on internet multimedia computing and service, pages 23–27, 2014.
- [10] Marius Cordts, Mohamed Omran, Sebastian Ramos, Timo Rehfeld, Markus Enzweiler, Rodrigo Benenson, Uwe Franke, Stefan Roth, and Bernt Schiele. The cityscapes dataset for semantic urban scene understanding. In Computer Vision and Pattern Recognition, 2016.
- [11] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pages 248–255. Ieee, 2009.
- [12] Wenhui Diao, Xian Sun, Xinwei Zheng, Fangzheng Dou, Hongqi Wang, and Kun Fu. Efficient saliency-based object detection in remote sensing images using deep belief networks. IEEE Geoscience and Remote Sensing Letters, 13(2):137–141, 2016.
- [13] Deng-Ping Fan, Cheng Gong, Yang Cao, Bo Ren, Ming-Ming Cheng, and Ali Borji. Enhanced-alignment Measure for Binary Foreground Map Evaluation. In International Joint Conference on Artificial Intelligence (IJCAI), pages 698–704, 2018. http://dpfan.net/e-measure/.
- [14] Deng-Ping Fan, Zheng Lin, Zhao Zhang, Menglong Zhu, and Ming-Ming Cheng. Rethinking rgb-d salient object detection: Models, data sets, and large-scale benchmarks. IEEE Transactions on neural networks and learning systems, 32(5):2075–2089, 2020.
- [15] Deng-Ping Fan, Wenguan Wang, Ming-Ming Cheng, and Jianbing Shen. Shifting more attention to video salient object detection. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 8554–8564, 2019.
- [16] Mengyang Feng, Huchuan Lu, and Errui Ding. Attentive feedback network for boundary-aware salient object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 1623–1632, 2019.
- [17] Carola Figueroa Flores, Abel Gonzalez-Garcia, Joost van de Weijer, and Bogdan Raducanu. Saliency for fine-grained object recognition in domains with scarce training data. Pattern Recognition, 94:62–73, 2019.
- [18] Yuchao Gu, Lijuan Wang, Ziqin Wang, Yun Liu, Ming-Ming Cheng, and Shao-Ping Lu. Pyramid constrained self-attention network for fast video salient object detection. In Proceedings of the AAAI conference on artificial intelligence, volume 34, pages 10869–10876, 2020.
- [19] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
- [20] Qibin Hou, Ming-Ming Cheng, Xiaowei Hu, Ali Borji, Zhuowen Tu, and Philip HS Torr. Deeply supervised salient object detection with short connections. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 3203–3212, 2017.
- [21] Jie Hu, Li Shen, and Gang Sun. Squeeze-and-excitation networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 7132–7141, 2018.
- [22] Wei Ji, Jingjing Li, Qi Bi, Chuan Guo, Jie Liu, and Li Cheng. Promoting saliency from depth: Deep unsupervised rgb-d saliency detection. arXiv preprint arXiv:2205.07179, 2022.
- [23] Bowen Jiang, Lihe Zhang, Huchuan Lu, Chuan Yang, and Ming-Hsuan Yang. Saliency detection via absorbing markov chain. In Proceedings of the IEEE international conference on computer vision, pages 1665–1672, 2013.
- [24] Ran Ju, Ling Ge, Wenjing Geng, Tongwei Ren, and Gangshan Wu. Depth saliency based on anisotropic center-surround difference. In 2014 IEEE international conference on image processing (ICIP), pages 1115–1119. IEEE, 2014.
- [25] Seungho Lee, Minhyun Lee, Jongwuk Lee, and Hyunjung Shim. Railroad is not a train: Saliency as pseudo-pixel supervision for weakly supervised semantic segmentation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 5495–5505, 2021.
- [26] Fuxin Li, Taeyoung Kim, Ahmad Humayun, David Tsai, and James M Rehg. Video segmentation by tracking many figure-ground segments. In Proceedings of the IEEE international conference on computer vision, pages 2192–2199, 2013.
- [27] Guanbin Li and Yizhou Yu. Visual saliency based on multiscale deep features. In CVPR, 2015.
- [28] Xiaohui Li, Huchuan Lu, Lihe Zhang, Xiang Ruan, and Ming-Hsuan Yang. Saliency detection via dense and sparse reconstruction. In Proceedings of the IEEE international conference on computer vision, pages 2976–2983, 2013.
- [29] Yin Li, Xiaodi Hou, Christof Koch, James M. Rehg, and Alan L. Yuille. The secrets of salient object segmentation. In CVPR, 2014.
- [30] Guibiao Liao, Wei Gao, Ge Li, Junle Wang, and Sam Kwong. Cross-collaborative fusion-encoder network for robust rgb-thermal salient object detection. IEEE Transactions on Circuits and Systems for Video Technology, 2022.
- [31] Xiangru Lin, Ziyi Wu, Guanqi Chen, Guanbin Li, and Yizhou Yu. A causal debiasing framework for unsupervised salient object detection. In Thirty-sixth AAAI conference on artificial intelligence, 2022.
- [32] Nian Liu, Junwei Han, and Ming-Hsuan Yang. Picanet: Learning pixel-wise contextual attention for saliency detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 3089–3098, 2018.
- [33] Tie Liu, Zejian Yuan, Jian Sun, Jingdong Wang, Nanning Zheng, Xiaoou Tang, and Heung-Yeung Shum. Learning to detect a salient object. IEEE Transactions on Pattern analysis and machine intelligence, 33(2):353–367, 2010.
- [34] Wenfeng Luo, Meng Yang, and Weishi Zheng. Weakly-supervised semantic segmentation with saliency and incremental supervision updating. Pattern Recognition, 115:107858, 2021.
- [35] Duc Tam Nguyen, Maximilian Dax, Chaithanya Kumar Mummadi, Thi Phuong Nhung Ngo, Thi Hoai Phuong Nguyen, Zhongyu Lou, and Thomas Brox. Deepusps: Deep robust unsupervised saliency prediction with self-supervision. arXiv preprint arXiv:1909.13055, 2019.
- [36] Timo Ojala, Matti Pietikainen, and Topi Maenpaa. Multiresolution gray-scale and rotation invariant texture classification with local binary patterns. IEEE Transactions on pattern analysis and machine intelligence, 24(7):971–987, 2002.
- [37] Youwei Pang, Xiaoqi Zhao, Lihe Zhang, and Huchuan Lu. Multi-scale interactive network for salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9413–9422, 2020.
- [38] Prashant W Patil, Subrahmanyam Murala, Abhinav Dhall, and Sachin Chaudhary. Msednet: multi-scale deep saliency learning for moving object detection. In 2018 IEEE International Conference on Systems, Man, and Cybernetics (SMC), pages 1670–1675. IEEE, 2018.
- [39] Houwen Peng, Bing Li, Weihua Xiong, Weiming Hu, and Rongrong Ji. Rgbd salient object detection: A benchmark and algorithms. In European conference on computer vision, pages 92–109. Springer, 2014.
- [40] Federico Perazzi, Jordi Pont-Tuset, Brian McWilliams, Luc Van Gool, Markus Gross, and Alexander Sorkine-Hornung. A benchmark dataset and evaluation methodology for video object segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 724–732, 2016.
- [41] Yongri Piao, Jian Wang, Miao Zhang, and Huchuan Lu. Mfnet: Multi-filter directive network for weakly supervised salient object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 4136–4145, 2021.
- [42] Xuebin Qin, Zichen Zhang, Chenyang Huang, Chao Gao, Masood Dehghan, and Martin Jagersand. Basnet: Boundary-aware salient object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 7479–7489, 2019.
- [43] Sucheng Ren, Chu Han, Xin Yang, Guoqiang Han, and Shengfeng He. Tenet: Triple excitation network for video salient object detection. In European Conference on Computer Vision, pages 212–228. Springer, 2020.
- [44] Olaf Ronneberger, Philipp Fischer, and Thomas Brox. U-net: Convolutional networks for biomedical image segmentation. In International Conference on Medical image computing and computer-assisted intervention, pages 234–241. Springer, 2015.
- [45] Jianping Shi, Qiong Yan, Li Xu, and Jiaya Jia. Hierarchical image saliency detection on extended cssd. IEEE transactions on pattern analysis and machine intelligence, 38(4):717–729, 2015.
- [46] Gyungin Shin, Samuel Albanie, and Weidi Xie. Unsupervised salient object detection with spectral cluster voting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3971–3980, 2022.
- [47] Ali Shokoufandeh, Ivan Marsic, and Sven J Dickinson. View-based object recognition using saliency maps. Image and Vision Computing, 17(5-6):445–460, 1999.
- [48] Peng Sun, Wenhu Zhang, Huanyu Wang, Songyuan Li, and Xi Li. Deep rgb-d saliency detection with depth-sensitive attention and automatic multi-modal fusion. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 1407–1417, 2021.
- [49] Zhengzheng Tu, Zhun Li, Chenglong Li, Yang Lang, and Jin Tang. Multi-interactive encoder-decoder network for rgbt salient object detection. arXiv e-prints, pages arXiv–2005, 2020.
- [50] Zhengzheng Tu, Zhun Li, Chenglong Li, Yang Lang, and Jin Tang. Multi-interactive dual-decoder for rgb-thermal salient object detection. IEEE Transactions on Image Processing, 30:5678–5691, 2021.
- [51] Zhengzheng Tu, Yan Ma, Zhun Li, Chenglong Li, Jieming Xu, and Yongtao Liu. Rgbt salient object detection: A large-scale dataset and benchmark. IEEE Transactions on Multimedia, 2022.
- [52] Zhengzheng Tu, Yan Ma, Zhun Li, Chenglong Li, Jieming Xu, and Yongtao Liu. Rgbt salient object detection: A large-scale dataset and benchmark. IEEE Transactions on Multimedia, 2022.
- [53] Zhengzheng Tu, Tian Xia, Chenglong Li, Xiaoxiao Wang, Yan Ma, and Jin Tang. Rgb-t image saliency detection via collaborative graph learning. IEEE Transactions on Multimedia, 22(1):160–173, 2019.
- [54] Guizhao Wang, Chenglong Li, Yunpeng Ma, Aihua Zheng, Jin Tang, and Bin Luo. Rgb-t saliency detection benchmark: Dataset, baselines, analysis and a novel approach. In Chinese Conference on Image and Graphics Technologies, pages 359–369. Springer, 2018.
- [55] Lijun Wang, Huchuan Lu, Yifan Wang, Mengyang Feng, Dong Wang, Baocai Yin, and Xiang Ruan. Learning to detect salient objects with image-level supervision. In CVPR, 2017.
- [56] Wenguan Wang, Qiuxia Lai, Huazhu Fu, Jianbing Shen, Haibin Ling, and Ruigang Yang. Salient object detection in the deep learning era: An in-depth survey. arXiv preprint arXiv:1904.09146, 2019.
- [57] Wenguan Wang, Shuyang Zhao, Jianbing Shen, Steven CH Hoi, and Ali Borji. Salient object detection with pyramid attention and salient edges. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 1448–1457, 2019.
- [58] Yifan Wang, Wenbo Zhang, Lijun Wang, Ting Liu, and Huchuan Lu. Multi-source uncertainty mining for deep unsupervised saliency detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 11727–11736, 2022.
- [59] Jun Wei, Shuhui Wang, Zhe Wu, Chi Su, Qingming Huang, and Qi Tian. Label decoupling framework for salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 13025–13034, 2020.
- [60] Yichen Wei, Fang Wen, Wangjiang Zhu, and Jian Sun. Geodesic saliency using background priors. In European conference on computer vision, pages 29–42. Springer, 2012.
- [61] Hongfa Wen, Chenggang Yan, Xiaofei Zhou, Runmin Cong, Yaoqi Sun, Bolun Zheng, Jiyong Zhang, Yongjun Bao, and Guiguang Ding. Dynamic selective network for rgb-d salient object detection. IEEE Transactions on Image Processing, 30:9179–9192, 2021.
- [62] Zhe Wu, Li Su, and Qingming Huang. Cascaded partial decoder for fast and accurate salient object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 3907–3916, 2019.
- [63] Binwei Xu, Haoran Liang, Ronghua Liang, and Peng Chen. Locate globally, segment locally: A progressive architecture with knowledge review network for salient object detection. In Proceedings. of the AAAI Conference On Artificial Intelligence, pages 1–9, 2021.
- [64] Qiong Yan, Li Xu, Jianping Shi, and Jiaya Jia. Hierarchical saliency detection. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 1155–1162, 2013.
- [65] Chuan Yang, Lihe Zhang, Huchuan Lu, Xiang Ruan, and Ming-Hsuan Yang. Saliency detection via graph-based manifold ranking. In CVPR, 2013.
- [66] Chuan Yang, Lihe Zhang, Huchuan Lu, Xiang Ruan, and Ming-Hsuan Yang. Saliency detection via graph-based manifold ranking. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3166–3173, 2013.
- [67] Siyue Yu, Bingfeng Zhang, Jimin Xiao, and Eng Gee Lim. Structure-consistent weakly supervised salient object detection with local saliency coherence. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI). AAAI Palo Alto, CA, USA, 2021.
- [68] Dingwen Zhang, Junwei Han, and Yu Zhang. Supervision by fusion: Towards unsupervised learning of deep salient object detector. In Proceedings of the IEEE International Conference on Computer Vision, pages 4048–4056, 2017.
- [69] Jianming Zhang and Stan Sclaroff. Saliency detection: A boolean map approach. In Proceedings of the IEEE international conference on computer vision, pages 153–160, 2013.
- [70] Jianming Zhang, Stan Sclaroff, Zhe Lin, Xiaohui Shen, Brian Price, and Radomir Mech. Minimum barrier salient object detection at 80 fps. In Proceedings of the IEEE international conference on computer vision, pages 1404–1412, 2015.
- [71] Jing Zhang, Jianwen Xie, and Nick Barnes. Learning noise-aware encoder-decoder from noisy labels by alternating back-propagation for saliency detection. In European conference on computer vision, pages 349–366. Springer, 2020.
- [72] Jing Zhang, Xin Yu, Aixuan Li, Peipei Song, Bowen Liu, and Yuchao Dai. Weakly-supervised salient object detection via scribble annotations. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 12546–12555, 2020.
- [73] Jing Zhang, Tong Zhang, Yuchao Dai, Mehrtash Harandi, and Richard Hartley. Deep unsupervised saliency detection: A multiple noisy labeling perspective. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 9029–9038, 2018.
- [74] Pingping Zhang, Dong Wang, Huchuan Lu, Hongyu Wang, and Xiang Ruan. Amulet: Aggregating multi-level convolutional features for salient object detection. In Proceedings of the IEEE International Conference on Computer Vision, pages 202–211, 2017.
- [75] Pingping Zhang, Dong Wang, Huchuan Lu, Hongyu Wang, and Baocai Yin. Learning uncertain convolutional features for accurate saliency detection. In ICCV, 2017.
- [76] Xiaoning Zhang, Tiantian Wang, Jinqing Qi, Huchuan Lu, and Gang Wang. Progressive attention guided recurrent network for salient object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 714–722, 2018.
- [77] Jia-Xing Zhao, Jiang-Jiang Liu, Deng-Ping Fan, Yang Cao, Jufeng Yang, and Ming-Ming Cheng. Egnet: Edge guidance network for salient object detection. In Proceedings of the IEEE International Conference on Computer Vision, pages 8779–8788, 2019.
- [78] Wangbo Zhao, Jing Zhang, Long Li, Nick Barnes, Nian Liu, and Junwei Han. Weakly supervised video salient object detection. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 16826–16835, 2021.
- [79] Xiaoqi Zhao, Youwei Pang, Lihe Zhang, Huchuan Lu, and Lei Zhang. Suppress and balance: A simple gated network for salient object detection. In European Conference on Computer Vision, pages 35–51. Springer, 2020.
- [80] Huajun Zhou, Peijia Chen, Lingxiao Yang, Xiaohua Xie, and Jianhuang Lai. Activation to saliency: Forming high-quality labels for unsupervised salient object detection. IEEE Transactions on Circuits and Systems for Video Technology, 2022.
- [81] Huajun Zhou, Xiaohua Xie, Jian-Huang Lai, Zixuan Chen, and Lingxiao Yang. Interactive two-stream decoder for accurate and fast saliency detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9141–9150, 2020.
- [82] Tao Zhou, Huazhu Fu, Geng Chen, Yi Zhou, Deng-Ping Fan, and Ling Shao. Specificity-preserving rgb-d saliency detection. In Proceedings of the IEEE/CVF international conference on computer vision, pages 4681–4691, 2021.
- [83] Wujie Zhou, Yun Zhu, Jingsheng Lei, Jian Wan, and Lu Yu. Apnet: Adversarial learning assistance and perceived importance fusion network for all-day rgb-t salient object detection. IEEE Transactions on Emerging Topics in Computational Intelligence, 2021.
- [84] Wangjiang Zhu, Shuang Liang, Yichen Wei, and Jian Sun. Saliency optimization from robust background detection. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2814–2821, 2014.
- [85] Wenbin Zou and Nikos Komodakis. Harf: Hierarchy-associated rich features for salient object detection. In Proceedings of the IEEE international conference on computer vision, pages 406–414, 2015.
Method | Training set | Input | Encoder | Pre-train | Saliency cues | Train time |
EDNS [71] | DUTS-TR | VGG-16 | F-IN | [84, 66, 60] | 8h | |
DCFD [31] | DUTS-TR | – | ResNet-50 | F-IN | [28] | – |
Ours | DUTS-TR | ResNet-50 | U-IN | No | 4.5h | |
SBF [68] | MSRA-B | VGG-16 | F-IN | [70, 69, 64] | 3h | |
MNL [73] | MSRA-B | ResNet-101 | F-IN | [84, 28, 23, 64] | 4h | |
USPS [35] | MSRA-B | ResNet-101 | F-CS | [84, 28, 23, 64] | 30h | |
DCFD [31] | MSRA-B | – | ResNet-101 | F-CS | [28] | – |
A2S [80] | MSRA-B | ResNet-50 | U-IN | No | 1h | |
Ours | MSRA-B | ResNet-50 | U-IN | No | 1.3h |
Appendix A Setup of USOD methods.
As listed in Tab. Texture-guided Saliency Distilling for Unsupervised Salient Object Detection, our method achieves better performance under disadvantage settings. Specifically, the input of our method is small than most USOD methods, such as for MNL [73] and for USPS [35]. Moreover, we use ResNet-50 [19] as backbone, which is a weakened version of ResNet-101 used in many USOD methods [73, 35, 31]. As for pre-training, most existing methods employed the encoders pre-trained with manual annotations of some close-related datasets, such as ImageNet [11] for object recognition and Cityscape [10] for semantic segmentation. Such setting indicates that they benefit from the semantic knowledge of manual annotations, which violates the semantic-agnostic definition of the SOD task. On the contrary, the encoder of our method is pre-trained without using any human annotation. It means that no semantic knowledge is involved in the whole training process, which accords with the semantic-agnostic definition. Last, even excluding the additional time of existing methods [71, 68, 73, 35] to extract salience cues using traditional methods, the training time of our method is much less than that of most previous methods.
Image



GT



A1



A2



A3



A4



Appendix B Qualitative comparison of different loss.
In our manuscript, we exhibit the quantitative results of different losses in ablation study A. Here, we provide a qualitative comparison in Fig. 8. Our baseline A1 can accurately localize salient objects in images, but loses many details. Trained using the proposed losses, the network mines more detailed saliency knowledge progressively and thus precisely predicts the saliency boundaries.




Image
Initial
Iter 5
Iter 10




Iter 15
Iter 20
After CRF
GT
Appendix C Visualization of the learned saliency.
In, Fig. 9, we visualize the learned saliency maps in our method during the training process. In the initial stage, our method is able to precisely localize the salient object based on the initial saliency cues, however, some small patches may still be misclassified. After subsequent tuning process, our method can learn more precise saliency knowledge and thus produce a high-quality pseudo label.
Appendix D Effect of hyperparameters.
The performance of our framework is affected by several hyperparameters, including , , , , and . We vary these hyperparameters and exhibit their results in Tab. 10. The results prove that the values , , and work best in practice. Our framework is robust to , and reports the best performance for . Moreover, our framework achieves comparable performance for various and values within . Furthermore, our framework is sensitive to . Although we observe robust performance for , outside this range (e.g., ) seems to induce significant performance drops.
Parameter | Value | |||
---|---|---|---|---|
100 | .911 | .932 | .046 | |
150 | .914 | .937 | .043 | |
200 | .917 | .945 | .038 | |
250 | .915 | .942 | .039 | |
300 | .914 | .943 | .039 | |
0.5 | .911 | .937 | .042 | |
0.7 | .915 | .942 | .039 | |
1 | .917 | .945 | .038 | |
1.2 | .914 | .943 | .038 | |
1.5 | .913 | .941 | .039 | |
0.01 | .869 | .915 | .054 | |
0.03 | .910 | .938 | .042 | |
0.05 | .917 | .945 | .038 | |
0.07 | .914 | .945 | .039 | |
0.09 | .908 | .942 | .040 | |
0.5 | .915 | .943 | .038 | |
0.75 | .915 | .945 | .038 | |
1 | .917 | .945 | .038 | |
1.25 | .915 | .943 | .039 | |
1.5 | .914 | .941 | .039 | |
3 | .913 | .937 | .042 | |
k | 5 | .917 | .945 | .038 |
7 | .914 | .943 | .039 |
Appendix E Loss for training extra saliency detectors.
For fully-supervised SOD methods, there are many choices for the loss functions, such as BCE loss [20, 74], BCE+IOU loss [77, 37], CTLoss [81, 7], BIS(BCE+IOU+SSIM) loss [42]. We employ these losses to train our saliency detector with the generated pseudo labels, as exhibited in Tab. 11. In summary, training our detector with IOU loss achieves the best results compared to other losses. BCE and CTLoss provide pixel-wise supervised signals, which means that training with these losses is easy to overfit the noises and thus degrade the generalization ability of our detector. Similarly, SSIM is based on regional statistics and thus is sensitive to noisy regions in pseudo labels. Unlike the above losses, IOU is robust to pixel-level or region-level noises because it is based on global statistics of saliency predictions.
Loss | DUT-OMRON | ECSSD | ||||
---|---|---|---|---|---|---|
BCE | .708 | .834 | .066 | .891 | .924 | .047 |
BCE+IOU | .726 | .846 | .065 | .894 | .923 | .048 |
BIS | .716 | .838 | .067 | .886 | .919 | .049 |
CTLoss | .743 | .862 | .061 | .907 | .914 | .057 |
IOU | .745 | .863 | .061 | .916 | .938 | .044 |
Training set | ||||
---|---|---|---|---|
LDF [59] | DUTS-TR | .296 | .508 | .315 |
Ours | DUTS-TR | .530 | .664 | .309 |
Ours* | DUTS-TR+X-ray | .924 | .943 | .056 |





Image
GT
LDF
Ours
Ours*
Appendix F Necessity of unsupervised SOD.
Ideally, a supervised class-agnostic SOD model can handle all scenarios, whereas is hard to obtain in practical. First, SOD methods trained on datasets with limited classes may not perform well on unseen classes, even if class labels are not used during training. Second, SOD methods trained on a certain style of images (e.g., natural images) do not perform well on other styles of images (e.g., medical images). To prove this point, we show the results of supervised LDF [59] and our unsupervised method on chest X-ray images 111https://lhncbc.nlm.nih.gov/LHC-downloads/downloads.html in Tab. 12 and Fig. 10. It proves that SOD methods trained on existing SOD datasets perform poorly on X-ray images, while our unsupervised method can achieve significant performance without using extra human annotations for specific scenarios. Third, many weakly-supervised methods leverage USOD methods as pre-processing or auxiliary loss, such as [25, 34]. In addition, our method can be considered as a novel self-supervised learning paradigm.