This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Learning Mask-aware CLIP Representations for
Zero-Shot Segmentation

Siyu Jiao1,2,3, Yunchao Wei1,2,3,  Yaowei Wang3,  Yao Zhao1,2,3,  Humphrey Shi 4

1 Institute of Information Science, Beijing Jiaotong University
2 Beijing Key Laboratory of Advanced Information Science and Network
3 Peng Cheng Laboratory  4 Picsart AI Research (PAIR)
jiaosiyu99@bjtu.edu.cn
Work done during an internship at Picsart AI Research (PAIR).
Abstract

Recently, pre-trained vision-language models have been increasingly used to tackle the challenging zero-shot segmentation task. Typical solutions follow the paradigm of first generating mask proposals and then adopting CLIP to classify them. To maintain the CLIP’s zero-shot transferability, previous practices favour to freeze CLIP during training. However, in the paper, we reveal that CLIP is insensitive to different mask proposals and tends to produce similar predictions for various mask proposals of the same image. This insensitivity results in numerous false positives when classifying mask proposals. This issue mainly relates to the fact that CLIP is trained with image-level supervision. To alleviate this issue, we propose a simple yet effective method, named Mask-aware Fine-tuning (MAFT). Specifically, Image-Proposals CLIP Encoder (IP-CLIP Encoder) is proposed to handle arbitrary numbers of image and mask proposals simultaneously. Then, mask-aware loss and self-distillation loss are designed to fine-tune IP-CLIP Encoder, ensuring CLIP is responsive to different mask proposals while not sacrificing transferability. In this way, mask-aware representations can be easily learned to make the true positives stand out. Notably, our solution can seamlessly plug into most existing methods without introducing any new parameters during the fine-tuning process. We conduct extensive experiments on the popular zero-shot benchmarks. With MAFT, the performance of the state-of-the-art methods is promoted by a large margin: 50.4% (+ 8.2%) on COCO, 81.8% (+ 3.2%) on Pascal-VOC, and 8.7% (+4.3%) on ADE20K in terms of mIoU for unseen classes. Code is available at github.com/jiaosiyu1999/MAFT.git.

1 Introduction

Semantic segmentation, one of the most widely researched topics in computer vision, has achieved remarkable success [3; 39; 15; 16] with the development of deep learning techniques [14]. However, traditional segmentation models are only capable of segmenting a few predefined categories within a closed vocabulary [7; 2; 22; 21], which is much smaller than the number of categories used by humans to describe the real world. Therefore, zero-shot segmentation [31; 1; 10; 13] is introduced to segment objects using arbitrary categories described by texts.

Recently, large-scale visual-language pre-training models (e.g. CLIP [28] and ALIGN [17]) have shown impressive transferability in recognizing novel categories, leading to their increased adoption for tackling the challenging zero-shot segmentation task [6; 33; 20; 27]. A mainstream solution follows the "frozen CLIP" paradigm, which executes the zero-shot segmentation with two steps: 1) first employing a Proposal Generator to produce class-agnostic mask proposals and 2) then leveraging a frozen pre-trained CLIP to classify each mask proposal via similarity matching in the aligned image-text feature space. While acceptable results are obtained, we reveal that these approaches overlook a crucial issue, i.e. the frozen CLIP is insensitive to different mask proposals and tends to produce similar predictions for various proposals of the same image.

To better illustrate the above-mentioned issue, we show several examples in Fig. 1. We use MaskFormer [5] to generate a series of mask proposals and select three typical ones. When using frozen CLIP for classification, we observe that it correctly classifies the high-quality swan proposal p1p_{1}. However, for the other two proposals p2p_{2} and p3p_{3}, which respectively contain only shape information of swan and both regions of swan and river, the frozen CLIP produces similar predictions compared to p1p_{1}. This is reasonable since CLIP is trained by image-text pairs, making it insensitive to pixel-level information (e.g. background noise), and resulting in numerous false positives. Based on the above observations, we consider that an expected CLIP for zero-shot segmentation task should 1) be sensitive to different mask proposals, 2) not compromise its original transferability on novel classes.

To this end, we introduce a Mask-aware CLIP Fine-tuning method (dubbed MAFT). To make CLIP sensitive to different mask proposals, we devise an Image-Proposals CLIP Encoder (IP-CLIP Encoder), which utilizes mask proposals to perform masked Multihead Attention [5; 4]. This design enables the model to handle arbitrary numbers of images and proposals simultaneously. The mask-aware loss is proposed to minimise the distance between the IoU score of mask proposals and the classification score of IP-CLIP Encoder, prompting IP-CLIP Encoder to differentiate various proposals. Besides, to preserve CLIP’s zero-shot transferability, we utilize a frozen CLIP as a teacher network to facilitate fine-tuning. This is achieved by aligning the outputs of the frozen CLIP and IP-CLIP Encoder through self-distillation loss. By performing MAFT, several advantages are provided: 1) Fine-tuning is efficient since only a few mask proposals need to be classified. 2) Compared to pixel-level fine-tuning, mask-aware fine-tuning hardly alters the structure of CLIP itself, preserving its maximum transferability. 3) Mask-aware fine-tuning of CLIP is released from the segmentation module, making it plug-and-play and applicable to any "frozen CLIP" approaches. As shown in Fig. 1, the mask-aware CLIP can well distinguish different proposals and provide proper classification scores for both seen (river) and unseen (swan) classes.

We evaluate our MAFT on three commonly used zero-shot segmentation benchmarks: COCO-Stuff [2], Pascal-VOC [7], and ADE20K [40]. Extensive experiments show that MAFT works well with various zero-shot segmentation methods. In particular, by plugging MAFT, the state-of-the-art approach FreeSeg [27] achieves superior performance on COCO-Stuff (42.2% \rightarrow 50.4%), Pascal-VOC (78.6% \rightarrow 81.8%) and ADE20K (4.4% \rightarrow 8.7%) in terms of mIoU of unseen classes. Furthermore, we conduct experiments in a open-vocabulary setting, where MAFT enhances the performance of A-847 [40], A-150 [40], PC-459 [24], PC-59 [24] and PAS-20 [7] datasets by +3.0%, +11.2%, +6.4%, +19.1% and +4.4%, respectively. Notably, our approach outperforms the freezing CLIP counterpart and establishes new state-of-the-art results on all datasets.

Refer to caption
Figure 1: Comparison between the frozen CLIP and our mask-aware CLIP for proposal classification. Regions of proposals are highlighted with green. The frozen CLIP classifies p1p_{1}, p2p_{2}, and p3p_{3} as swan class and produces similar predictions, although these proposals contain different regions of the image. After the MAFT, the mask-aware CLIP can produce proper scores for different proposals.

2 Related Work

Zero-Shot Segmentation [29] is established to break the restriction of categories and perform segmentation on unseen classes. Earlier works SPNet [31] learn a joint pixel and vocabulary concept embedding space, ZS5 [1] utilizes a generative model to generate pixel-level features based on word embeddings of unseen classes, CaGNet [10] incorporates context information for better feature generation. Recent approaches take the advent of large-scale visual-language models (e.g. CLIP [28] and ALIGN [17]) to leverage rich alignment features from image-text pairs. [34] uses CLIP to generate pseudo-labels for single-image segmentation. STRICT [25] obtains pixel-level pseudo-labels from CLIP for unlabeled pixels and proposes a self-training strategy to capture latent information on unseen classes. LSeg [8] trains a CNN model to compute per-pixel image embeddings and use CLIP text embeddings as a classifier. [32] employs contrastive supervision to learn segmentation masks from text.

Concurrently, recent works [6; 33; 27; 20; 9] follow the "frozen CLIP" paradigm for zero-shot segmentation, they first generate a series of mask proposals and then utilize CLIP [28] or ALIGN [17] to classify them. ZSSeg and OVSeg [33; 20] train CLIP adapters to boost performance. FreeSeg[27] simultaneously uses semantic, instance, and panoptic labels and performs fusion training. OpenSeg[9] takes extra images with image-level supervision (e.g. captions) to scale up training data.

Pre-trained model fine-tuning is widely used for transferring pre-trained knowledge to downstream tasks, e.g. segmentation. However, this strategy may not work well for data-limited tasks like few-shot learning and zero-shot learning due to the daunting overfitting problem. To address this problem and transfer pre-trained knowledge to data-limited tasks, [43; 42; 12; 33; 20; 27] propose to learn text prompts or image prompts by using (a few) annotated images from target dataset. SVF [30] fine-tunes only a few parameters in the pre-trained image encoder to adapt pre-trained knowledge to few-shot segmentation. [38; 37] use contrastive learning to avoid catastrophic forgetting. Alternatively, many outstanding approaches in data-limited tasks [23; 35; 36; 6; 33] choose to freeze the parameters of pre-trained models to maintain the transferability.

Specific to the task of zero-shot/ open-vocabulary segmentation, mainstream approaches use frozen CLIP to avoid overfitting. Recently, MaskCLIP [41] conducts adequate experiments to fine-tune CLIP for open-vocabulary segmentation but has failed. While this attempt is meaningful and appreciated, it is believed that the failure is due to the large domain gap between pixel-level and image-level tasks. This motivates us further research fine-tuning CLIP to be mask-aware (region-level task).

3 Preliminary

Problem Setting. Zero-shot segmentation aims at training a segmentation model capable of segmenting novel objects using text descriptions. Given two category sets CseenC_{seen} and CunseenC_{unseen} respectively, where CseenC_{seen} and CunseenC_{unseen} are disjoint in terms of object categories (CseenCunseen=C_{seen}\cap C_{unseen}=\emptyset). The model is trained on CseenC_{seen} and directly tested on both CseenC_{seen} and CunseenC_{unseen}. Typically, CseenC_{seen} and CunseenC_{unseen} are described with semantic words (e.g. sheep, grass).

Revisiting the "frozen CLIP" paradigm. The "frozen CLIP" approaches [6; 33; 27; 20] execute zero-shot segmentation in two steps: mask proposals generation and mask proposals classification. In the first step, these approaches train a Proposal Generator to generate NN class-agnostic mask proposals (denoting as MM, MN×H×WM\in\mathbb{R}^{N\times H\times W}) and their corresponding classification scores (denoting as ApA^{p}, ApN×|Cseen|A^{p}\in\mathbb{R}^{N\times|C_{seen}|}). MaskFormer [5] and Mask2Former [4] are generally used as the Proposal Generator since the Hungarian matching [19] in the training process makes the mask proposals strongly generalizable. In the second step, NN suitable sub-images (IsubI_{sub}) are obtained by merging NN mask proposals and the input image. IsubI_{sub} is then fed into the CLIP Image Encoder to obtain the image embedding (EIE^{I}). Meanwhile, text embedding (ETE^{T}) is generated by a CLIP Text Encoder. The classification score (Ac,AcN×CA^{c},A^{c}\in\mathbb{R}^{N\times C}) predicted by CLIP is calculated as:

Aic=Softmax(exp(1τsc(EiT,EI))i=0Cexp(1τsc(EiT,EI))),i=[1,2,C]A^{c}_{i}=\mathrm{Softmax}(\frac{\exp(\frac{1}{\tau}s_{c}(E^{T}_{i},E^{I}))}{\sum_{i=0}^{C}\exp(\frac{1}{\tau}s_{c}(E^{T}_{i},E^{I}))}),i=[1,2,...C] (1)

where τ\tau is the temperature hyper-parameter. sc(EiT,EI)=EiTEI|EiT||EI|s_{c}(E^{T}_{i},E^{I})=\frac{E^{T}_{i}\cdot E^{I}}{|E^{T}_{i}||E^{I}|} represents the cosine similarity between EiTE^{T}_{i} and EIE^{I}. CC is the number of classes, with C=|Cseen|C=|C_{seen}| during training and C=|CseenCunseen|C=|C_{seen}\cup C_{unseen}| during inference. Noting that CLIP is frozen when training to avoid overfitting.

To further enhance the reliability of AcA^{c}, the classification score of the Proposal Generator (ApA^{p}) is ensembled with AcA^{c} since ApA^{p} is more reliable on seen classes. This ensemble operation is wildly used in "frozen CLIP" approaches. The pipeline of "frozen CLIP", as well as the merge and ensemble operations, are described in detail in the Appendix.

Although "frozen CLIP" approaches have achieved promising results, it is clear that directly adopting an image-level pre-trained CLIP for proposal classification can be suboptimal. A frozen CLIP usually produces numerous false positives, and the merge operation may destroy the context information of an input image. In view of this, we rethink the paradigm of the frozen CLIP and explore a new solution for proposal classification.

4 Methodology

Refer to caption
Figure 2: Overview of the Mask-Aware Fine-tuning (MAFT). In IP-CLIP Encoder, we modify the CLIP Image Encoder, and apply the mask proposals as attention bias in Multihead Attention from the LthL^{th} layer. The final projection unit is an MLP module used for reshaping the channels of FclsF_{cls}. w.o. MM denotes IP-CLIP Encoder processes image without utilizing mask proposals (MM). Mask-aware Loss is designed to train CLIP to be mask-aware, while Self-distillation Loss is designed to maintain the transferability. Only the IP-CLIP Encoder is trained (orange part), the Proposal Generator and the CLIP Text Encoder are frozen (blue part).

We introduce Mask-Aware Fine-tuning (MAFT), a method for learning mask-aware CLIP representations. Within MAFT, we first propose the Image-Proposal CLIP Encoder (IP-CLIP Encoder) to handle images with any number of mask proposals simultaneously (Sec. 4.1). Then, mask-aware loss and self-distillation loss are introduced to fine-tune the IP-CLIP Encoder and make it distinguishable for different mask proposals while maintaining transferability (Sec. 4.2). The complete diagram of the MAFT is shown in Fig. 2, we use the ViT-B/16 CLIP model for illustration.

4.1 Image-Proposal CLIP Encoder (IP-CLIP Encoder)

IP-CLIP Encoder aims to process arbitrary numbers of images and mask proposals simultaneously. We draw inspiration from MaskFormer [4; 5], which uses attention-masks in Multihead Attention and provides the flexibility for accepting any number of queries and features of different masked regions. Accordingly, we apply mask proposals as attention-masks in Multihead Attention and designate independent classification queries for each mask proposal.

In the IP-CLIP Encoder shown in Fig. 2, we denote the features propagate between Transformer layers as FiF^{i}, where i=[1,212]i=[1,2...12]. We can express FiF^{i} as Fi=[Fclsi;Ffeati],(1+hw)×dF^{i}=[F^{i}_{cls};~{}F^{i}_{feat}],\in\mathbb{R}^{(1+hw)\times d}, here 11 represents a class-embedding vector (FclsiF^{i}_{cls}), hwhw represents the number of the flattened image features (FfeatiF^{i}_{feat}). To obtain the classifications of all mask proposals simultaneously, we repeat FclsiF^{i}_{cls} at layer LL NN times, where NN is the number of mask proposals, denoting the repeated class-embedding vectors as FclsiF^{i*}_{cls}. We can express the modified features (FiF^{i*}) as Fi=[Fclsi;Ffeati],(N+hw)×dF^{i*}=[F^{i*}_{cls};~{}F^{i}_{feat}],\in\mathbb{R}^{(N+hw)\times d}.

Propagation of FiF^{i}, where i=[1,2,L]i=[1,2,...L]. We consider that CLIP’s classification significantly relies on context information. In the first LL Transformer layers, the propagation of FiF^{i} is the same as in standard CLIP. Specifically, FclsiF^{i}_{cls} utilizes cross-attention with all pixels within FfeatiF^{i}_{feat}, effectively retaining the context information.

In the subsequent 12L12-L Transformer layers, the propagation of FiF^{i*} can be partitioned into two parts: the propagation of FclsiF^{i*}_{cls} and the propagation of FfeatiF^{i}_{feat}.

Propagation of FclsiF^{i*}_{cls}. We use FclsiF^{i*}_{cls}[nn] and MM[nn] to represent the position nn in FclsiF^{i*}_{cls} and MM, where n=[1,2N]n=[1,2...N]. It is expected FclsiF^{i*}_{cls}[nn] computes Multihead Attention for the positions where MM[nn]=1=1 and itself. To achieve this, we construct an attention bias BN×(N+hw)B\in\mathbb{R}^{N\times(N+hw)} as follows:

B(i,j)={0,ifM^(i,j)=1,ifM^(i,j)=0,M^=[I(N,N);Flat(M)]B_{(i,j)}=\left\{\begin{aligned} 0&,\mathrm{if}~{}{\hat{M}}_{(i,j)}=1\\ -\infty&,\mathrm{if}~{}{\hat{M}}_{(i,j)}=0\\ \end{aligned}\right.,~{}~{}~{}\hat{M}=[\mathrm{I}(N,N);~{}\mathrm{Flat}(M)] (2)

here I(N,N)\mathrm{I}(N,N) denotes NthN^{th} order identity matrix, Flat\mathrm{Flat}(\cdot) denotes the flatten operation. M^\hat{M} is an intermediate variable for better representation. Therefore, a masked Multihead Attention is used for propagating FclsiF^{i*}_{cls} :

Fcls(i+1)=Softmax(Que(Fclsi)Key(Fi)Td+B)Val(Fi)F^{(i+1)*}_{cls}=\mathrm{Softmax}(\frac{\mathrm{Que}(F^{i*}_{cls})\mathrm{Key}(F^{i*})^{T}}{\sqrt{d}}+B)\mathrm{Val}(F^{i*}) (3)

where Que()\mathrm{Que}(\cdot), Key()\mathrm{Key}(\cdot), and Val()\mathrm{Val}(\cdot) denote linear projections, dd is the hidden dimension of FiF^{i*}. Notably, We omit the MLP Layer and Layer Normalizations in Transformer layers to simplify the representation in Eq. 3 and Eq. 4.

Propagation of FfeatiF^{i}_{feat}. A standard Multihead Attention is used for propagating FfeatiF^{i}_{feat} :

Ffeati+1=Softmax(Que(Ffeati)Key(Ffeati)Td)Val(Ffeati)F^{i+1}_{feat}=\mathrm{Softmax}(\frac{\mathrm{Que}(F^{i}_{feat})\mathrm{Key}(F^{i}_{feat})^{T}}{\sqrt{d}})\mathrm{Val}(F^{i}_{feat}) (4)

Therefore, for any given mask proposal MM[nn], the corresponding class-embedding FclsiF^{i*}_{cls}[nn] only performs Multihead Attention with FfeatiF^{i}_{feat} where MM[nn]=1=1 and FclsiF^{i*}_{cls}[nn]. The propagation of FfeatiF^{i}_{feat} remains undisturbed by attention-masks. Compared with the frozen CLIP, IP-CLIP Encoder leverages context information effectively and reduces computational costs.

4.2 Objective

IP-CLIP Encoder with CLIP pre-trained parameters remains challenging in distinguishing different mask proposals, e.g., when the proposals contain more background regions than foreground objects, IP-CLIP may tend to classify them into the foreground categories. To overcome this limitation, we introduce mask-aware loss and self-distillation loss to fine-tune the IP-CLIP Encoder to be mask-aware without sacrificing transferability.

We conduct the mask-aware loss function (ma\mathcal{L}_{ma}) on AcA^{c}. The goal is to assign high scores to high-quality proposals and low scores to low-quality proposals in AcA^{c}. Concretely, we use the Intersection over Union (IoU) score obtained from ground-truth and align it with the AcA^{c} to prompt CLIP to become mask-aware. Assuming there are kk classes in ground-truth, we can generate kk binary maps of ground-truth and calculate the IOU score (SIoUS_{IoU}) with NN mask proposals. We identify a discrepancy between the maximum values of AcA^{c} and SIoUS_{IoU}. The maximum value of AcA^{c} tends to approach 1, whereas the maximum value of SIoUS_{IoU} ranges from 0.75 to 0.99. This inconsistency can hinder the alignment between these two metrics. Therefore, we introduced a min-max normalization technique for SIoUS_{IoU} as follows:

SIoUnorm=SIoUmin(SIoU)max(SIoU)min(SIoU),SIoUK×NS_{IoU}^{norm}=\frac{S_{IoU}-min(S_{IoU})}{max(S_{IoU})-min(S_{IoU})},S_{IoU}\in\mathbb{R}^{K\times N} (5)

Meanwhile, we select kk pre-existing classes in AcA^{c} (Aselectc,AselectcK×NA^{c}_{select},A^{c}_{select}\in\mathbb{R}^{K\times N}), and employ SmoothL1SmoothL1 Loss to align it with SIoUnormS_{IoU}^{norm}. Therefore, ma\mathcal{L}_{ma} can be formulated as follows:

ma(Aselectc,SIoUnorm)=SmoothL1(Aselectc,SIoUnorm)\mathcal{L}_{ma}(A^{c}_{select},S_{IoU}^{norm})=\mathrm{SmoothL1}(A^{c}_{select},S_{IoU}^{norm}) (6)
SmoothL1(x,y)={0.5(xy)2,if|xy|<1|xy|0.5,otherwise\mathrm{SmoothL1}(x,y)=\left\{\begin{aligned} 0.5\cdot(x-y)^{2}&,~{}~{}~{}\mathrm{if}~{}|x-y|<1\\ |x-y|-0.5&,~{}~{}~{}\mathrm{otherwise}~{}\\ \end{aligned}\right. (7)

In addition to ma\mathcal{L}_{ma}, we also introduce a self-distillation loss dis\mathcal{L}_{dis} to maintain CLIP’s transferability and alleviate overfitting on CseenC_{seen}. Within dis\mathcal{L}_{dis}, we use a frozen CLIP as the teacher net, the IP-CLIP as the student net for self-distillation. The predictions of the frozen CLIP and IP-CLIP are expected to be the same when no mask is included. Denoting the output of the frozen CLIP as ATA_{T}, and the output of the fine-tuned IP-CLIP without masks as ASA_{S}. We use SmoothL1SmoothL1 Loss to minimize the difference as follows:

dis(AS,AT)=SmoothL1(AS,AT)\mathcal{L}_{dis}(A_{S},A_{T})=\mathrm{SmoothL1}(A_{S},A_{T}) (8)

It is important to note that when processing an image through IP-CLIP without mask proposals, the resulting ASA_{S} is a matrix with dimensions C×1\mathbb{R}^{C\times 1}. Therefore, the final loss function can be formulated as: =ma+λdis\mathcal{L}=\mathcal{L}_{ma}+{\lambda}{\mathcal{L}_{dis}}, where we set the constant λ\lambda to 1 in our experiments. The mask-aware fine-tuning process is efficient as we only perform a few iterations (less than 1 epoch).

5 Experiments

Table 1: Comparison with state-of-the-art methods in zero-shot segmentation. mIoUs and mIoUu denote the mIoU(%) of seen classes and unseen classes.
  Method COCO-Stuff Pascal-VOC ADE20K
mIoUs mIoUu hIoU mIoUs mIoUu hIoU mIoUs mIoUu hIoU
SPNet[31] 34.6 26.9 30.3 77.8 25.8 38.8 - - -
ZS5[1] 34.9 10.6 16.2 78.0 21.2 33.3 - - -
CaGNet[10] 35.6 13.4 19.5 78.6 30.3 43.7 - - -
STRICT[25] 35.3 30.3 32.6 82.7 35.6 73.3 - - -
\cdashline1-10[0.8pt/2pt] ZegFormer[6] 36.7 36.2 36.4 90.1 70.6 79.2 17.4 5.1 7.9
ZegFormer +MAFT 36.4 -0.3 40.1 +3.9 38.1 +1.7 91.5 +1.4 80.7 +10.1 85.7 +6.5 16.6 -0.8 7.0 +1.9 9.8 +1.9
\cdashline1-10[0.8pt/2pt] ZSSeg[33] 40.4 36.5 38.3 86.6 59.7 69.4 18.0 4.5 7.2
ZSSeg +MAFT 40.6 +0.2 40.1 +3.6 40.3 +2.0 88.4 +1.8 66.2 +6.5 75.7 +6.3 18.9 +0.9 6.7 +2.2 9.9 +2.7
\cdashline1-10[0.8pt/2pt] FreeSeg[27] 42.4 42.2 42.3 91.9 78.6 84.7 22.3 4.4 7.3
FreeSeg +MAFT 43.3 +0.9 50.4 +8.2 46.5 +4.2 91.4 -0.5 81.8 +3.2 86.3 +1.6 21.4 -0.9 8.7 +4.3 12.4 +5.1
 
Table 2: Results on representative methods [6; 33; 27] with/without MAFT. Here we remove the ensemble operation, and only maintain CLIP classifier results.
  Method COCO-Stuff Pascal-VOC ADE20K
mIoUs mIoUu hIoU mIoUs mIoUu hIoU mIoUs mIoUu hIoU
ZegFormer[6] 18.5 23.0 20.5 81.4 76.8 79.0 5.1 2.6 3.5
ZegFormer +MAFT 35.1 +16.6 31.6 +7.6 33.3 +12.7 87.6 +6.2 79.9 +3.1 83.5 +4.5 15.8 +10.8 7.0 +4.4 9.8 +6.3
\cdashline1-10[0.8pt/2pt] ZSSeg[33] 20.6 27.4 23.6 82.0 71.2 76.2 5.9 2.8 3.9
ZSSeg +MAFT 36.1 +15.5 35.9 +8.3 36.0 +12.4 87.1 +5.1 76.1 +4.9 81.2 +5.0 17.2 +11.3 7.2 +4.4 10.2 +6.3
\cdashline1-10[0.8pt/2pt] FreeSeg[27] 22.3 29.3 25.3 87.4 74.7 80.5 6.5 2.8 3.9
FreeSeg +MAFT 40.1 +17.8 49.7 +20.4 44.4 +19.1 90.4 +3.0 84.7 +10.0 87.5 +7.0 21.3 +14.8 8.7 +5.9 12.2 +8.3
 

5.1 Setting

Dataset. We first follow [1; 11; 26; 6; 33] to conduct experiments on three popular zero-shot segmentation benchmarks, Pascal-VOC, COCO-Stuff and ADE20K, to evaluate our method. Then, we evaluate MAFT on the open-vocabulary setting [20; 33], i.e., training on COCO-Stuff and testing on ADE20K (A-847, A-150), Pascal-Context (PC-459, PC-59), and Pascal-VOC (PAS-20). More details of the dataset settings are provided in the Appendix.

Evaluation Metrics. To quantitatively evaluate the performance, we follow standard practice [1; 31; 10; 25; 6; 33; 27], adopt mean Intersection over Union (mIoU) to respectively evaluate the performance for seen classes (IoUs) and unseen classes (IoUu). We also employ the harmonic mean IoU (hIoU) among the seen and unseen classes to measure comprehensive performance.

Methods. Three representative methods are used to verify the generality of MAFT. We unify the three methods into the same framework, with all methods using ResNet101 as the backbone of Proposal Generator and ViT-B/16 CLIP model for a fair comparison.

  • ZegFormer (CVPR 2022) [6] is an early adopter of the "frozen CLIP" paradigm. It uses MaskFormer as Proposal Generator and employs an ensemble operation to improve the confidence of the results.

  • ZSSeg (ECCV 2022) [33] uses MaskFormer as Proposal Generator and introduces learnable prompts to improve classification accuracy, which significantly affects the subsequent methods. ZSSeg also adopts a self-training strategy, this strategy is excluded from all methods for a fair comparison.

  • FreeSeg (CVPR 2023) [27] represents the state-of-the-art method, unifies semantic, instance, and panoptic segmentation tasks and uses annotations from all three tasks for fusion training. We retrain FreeSeg with only the semantic annotations to ensure fairness.

Implementation details. We employ ResNet101 as backbone of the Proposal Generator and ViT-B/16 CLIP model. The training process consists of two stages. For the first stage, we follow the official code of ZegFormer, ZSSeg and FreeSeg for model training. For the second stage, we fine-tune IP-CLIP Encoder with MAFT. We take the batch size of 16 and set CLIP input image size to 480×\times480. The optimizer is AdamW with a learning rate of 0.00001 and weight decay of 0.00001. The number of training iterations is set to 100 for Pascal-VOC, 1000 for COCO-Stuff and 5000 for ADE20K.

5.2 Comparisons with State-of-the-art Methods

Table 3: Comparison with state-of-the-art methods on the open-vocabulary setting. mIoU is used to evaluate the performance. * denotes additional training data is used.
  A-847 A-150 PC-459 PC-59 PAS-20
SPNet[31] - - - 24.3 18.3
ZSSeg[1] - - - 19.4 38.3
LSeg+[8] 2.5 13.0 5.2 36.0 59.0
OVSeg[20] 7.1 24.8 11.0 53.3 92.6
OpenSeg* [9] 8.8 28.6 12.2 48.2 72.2
\cdashline1-6[0.8pt/2pt] FreeSeg[27] 7.1 17.9 6.4 34.4 85.6
FreeSeg +MAFT 10.1 +3.0 29.1 +11.2 12.8 +6.4 53.5 +19.1 90.0 +4.4
 

In this section, three representative methods are used [6; 33; 27] to evaluate the effectiveness of MAFT. We compare three representative methods with MAFT and frozen CLIP. Additionally, we compare the results with previous state-of-the-art methods [31; 1; 10; 25].

Comparisons in the zero-shot setting. In Tab. 2, MAFT remarkably improves the performance. MAFT promotes the state-of-the-art performance by + 8.2% on COCO, + 3.2% on Pascal, and +4.3% on ADE20K in terms of mIoU for unseen classes. It is important to note that the results for seen classes are mainly based on ApA^{p} rather than AcA^{c} due to the ensemble operation in [6; 33; 27] (Details in Sec. 3). Therefore, the effect of MAFT on the seen classes is relatively insignificant.

Comparisons without ensemble strategy. To better showcase the performance gains from MAFT, we removed the ensemble operation in [6; 33; 27] and presented the results in Tab. 2. It can be seen that the performance of different methods is significantly improved after applying MAFT. In particular, the state-of-the-art method FreeSeg achieves hIoU improvements of 19.1%, 7.0%, and 8.3% on COCO, VOC2012 and ADE20K datasets.

Comparisons in the open-vocabulary setting. We further evaluated the transferability of MAFT in the open-vocabulary setting [20; 33], using FreeSeg as a baseline for comparison. Results are shown in Tab. 3. Compared with OVSeg [20] and OpenSeg [9], FreeSeg achieves suboptimal performance. However, the proposed MAFT enhances the performance of A-847, A-150, PC-459, PC-59 and PAS-20 by 3.0%,11.2%, 6.4%, 19.1% and 4.4%, and outperforms OpenSeg on all five datasets.

5.3 Ablation Study

Table 4: Ablations on COCO dataset. GFLOPs in (a) is used to measure the computation of CLIP Image Encoder. The best results are highlighted with red, and the default settings are highlighted with gray background.
  mIoUs mIoUu hIoU GFLOPs
FreeSeg 22.3 29.3 25.3 1127.0
+ IP-CLIP 29.4 +7.1 36.2 +6.9 32.4 +7.1 53.4
+ ftft (ma\mathcal{L}_{ma}) 39.9 +17.6 47.1 +17.8 43.1 +17.8 53.4
+ ftft (ma\mathcal{L}_{ma} + dis\mathcal{L}_{dis}) 40.1+17.8 49.7+20.4 44.4+19.0 53.4
 
(a) Ablation on components of MAFT. ftft denotes the mask-aware fine-tining
  mIoUs mIoUu hIoU
L1L_{1} 38.6 +16.3 45.8 +16.5 41.8 +16.5
L2L_{2} 40.0 +17.7 45.8 +16.5 42.7 +17.4
SmoothL1SmoothL_{1} 39.9 +17.6 47.1 +17.8 43.1 +17.8
KLKL 40.9 +18.6 41.8 +12.5 41.3 +16.0
 
(b) Ablation on mask-aware loss ma\mathcal{L}_{ma}. dis\mathcal{L}_{dis} is removed.
  mIoUs mIoUu
500 iters 37.7 47.0
1k iters 40.0 49.7
2k iters 41.1 47.6
3k iters 41.4 46.5
4k iters 41.5 46.1
5k iters 42.0 45.7
 
(c) Ablation of the training iterations
  mIoUs mIoUu hIoU
None 40.6 44.7 42.5
+ cls.cls. 40.7 44.7 42.7
+ pos.pos. 40.6 44.9 42.8
+ mlpmlp 40.3 48.7 44.1
+ conv.conv. 40.0 49.7 44.3
+ proj.proj. 40.2 49.1 44.2
 
(d) Ablation of the frozen units in CLIP
  mIoUs mIoUu hIoU
0 39.3 46.4 42.6
2 39.2 46.4 42.5
4 39.5 46.6 42.6
6 40.0 47.8 43.6
8 40.0 49.7 44.3
10 39.9 45.7 42.6
 
(e) Ablation of the start mask attention layer LL

We conduct ablation studies on various choices of designs of our MAFT to show their contribution to the final results in Tab. 4. FreeSeg is used as the baseline model and ensemble operation is removed.

Component-wise ablations. To understand the effect of each component in the MAFT, including the IP-CLIP Encoder and the fine-tuning strategy (ma\mathcal{L}_{ma}, dis\mathcal{L}_{dis}), we start with standard FreeSeg and progressively add each design. (Tab. 4a). FreeSeg uses frozen CLIP and yields inferior performance due to CLIP’s mask-unaware property (1st1^{st} row). Then, IP-CLIP Encoder obtains rich context information and greatly reduces the omputational costs, resulting in an improvement of 7.1% on seen classes and 6.9% on unseen classes. However, mask-aware is not accomplished at this point. Using only ma\mathcal{L}_{ma} for fine-tuning CLIP produces decent performance (the 3rd3^{rd} result). The introduction of dis\mathcal{L}_{dis} (the 4th4^{th} result) maintains transferability while learning mask-aware representations, which further enhances the performance on unseen classes by 2.6%.

Effect of different ma\mathcal{L}_{ma}. Mask-aware Loss ma\mathcal{L}_{ma} is an essential component of MAFT. In Tab. 4b, we investigate how different loss functions (L1L1, L2L2, SmoothL1SmoothL1 and KLKL Loss) impact performance, here we remove dis\mathcal{L}_{dis} for analysis. Results show SmoothL1SmoothL1 Loss boosts performance on CunseenC_{unseen} to 47.1% (+17.8%), KLKL Loss provides +12.5% improvement on CseenC_{seen}, but only +11.8% on CunseenC_{unseen}, manifesting KLKL Loss compromises the model of transferability comparing with SmoothL1SmoothL1 Loss.

Training iterations. Tab. 4c examines the impact of training iterations. Increasing the number of iterations leads to gradual improvement of IoUs, but it also results in significant overfitting on unseen classes. Therefore, we choose to fine-tune 1k iterations to maximize the zero-shot ability.

Frozen units in CLIP. We also explore the impact of fine-tuning units within IP-CLIP Encoder. As illustrated in Fig. 2, IP-CLIP Encoder comprises convolution layers (dubbed as conv.conv.), class embedding (cls.cls.), Transformer layers, final projection (proj.proj.) and positional embedding (pos.pos., not shown in Fig. 2). We start with fine-tuning the entire IP-CLIP Encoder, and then freezing each unit sequentially, as specified in Tab. 4d. We only freeze MLPMLP in the Transformer layers (dubbed as mlpmlp). Compared with fine-tuning the entire IP-CLIP Encoder, the performance of mIoUu is improved by 5.0% when freezing conv.conv., cls.cls., pos.pos. and mlpmlp.

Start mask attention layer. Tab. 4e presents the results of the start mask attention layer (LL). We observe a significant improvement in the performance of unseen classes by +3.4% when the value of LL increases from 0 to 8. This could be attributed to the fact that starting masked Multihead Attention later enables FclsiF^{i*}_{cls} to gain more context information. However, the performance significantly drops when L=10L=10 (from 49.7% to 45.7%), which may be due to the loss of mask-aware property.

5.4 Extending MAFT with SAM

Table 5: Comparison with SAM. We use SAM-H as the proposal generator.
  Pascal-VOC COCO-Stuff
mIoUs mIoUu hIoU mIoU mIoUs mIoUu hIoU mIoU
SAM 85.1 86.7 85.9 85.5 43.1 43.3 43.2 42.1
SAM + MAFT 91.0 +5.9 88.6 +1.9 89.8 +3.9 90.4 +4.9 43.4 +0.3 51.5 +8.2 47.1 +3.9 44.1 +2.0
 
(a) Results in zero-shot segmentation.
  A-847 A-150 PC-459 PC-59 PAS-20
SAM 7.1 17.9 6.4 34.4 85.6
SAM + MAFT 10.1 +3.0 29.1 +11.2 12.8 +6.4 53.5 +19.1 90.0 +4.4
 
(b) Results in open-vocabulary segmentation.

We explore using the Segment Anything Model [18] (SAM) as the proposal generator. We evaluate the performance with SAM-H using an original CLIP (dubbed SAM\mathrm{SAM}) or a mask-aware fine-tuned CLIP (dubbed SAM+MAFT\mathrm{SAM+MAFT}). In fact, SAM can be seamlessly integrated into our framework as the proposal generator. The results are shown in Tab. 5. Experiments are conducted under both zero-shot setting and open-vocabulary setting.

It can be observed that SAM+MAFT\mathrm{SAM+MAFT} obtains significant improvement over SAM\mathrm{SAM} under both settings. Besides, SAM+MAFT\mathrm{SAM+MAFT} also surpasses FreeSeg+MAFT\mathrm{FreeSeg+MAFT} on all benchmarks. Particularly, in the zero-shot setting (Pascal-VOC), SAM+MAFT\mathrm{SAM+MAFT} outperforms FreeSeg+MAFT\mathrm{FreeSeg+MAFT} by 6.8% in terms of mIoUu. This enhancement can be attributed to the stronger generalization capabilities of SAM for unseen classes.

5.5 Extending MAFT with more Vision-Language Models

Table 6: Comparison with more Vision-Language Models.
  backbone A-847 A-150 PC-459 PC-59 PAS-20
OVSeg [20] ViT-L 9.0 29.6 12.4 55.7 94.5
FreeSeg [27] 8.5 21.0 7.6 33.8 86.4
FreeSeg + MAFT 12.1 +3.6 32.0 +11.0 15.7 +8.1 58.5 +24.7 92.1 +5.7
\cdashline1-7[0.8pt/2pt] FreeSeg [27] Res50 5.3 15.5 5.4 28.2 87.1
FreeSeg + MAFT 8.4 +3.1 27.0 +11.5 9.9 +4.5 50.8 +22.6 89.0 +1.9
 

In order to demonstrate the efficacy and robustness of MAFT, we conduct experiments using stronger (CLIP-ViT-L) and ResNet-based (CLIP-Res50) Vision-Language Models. The open-vocabulary results are shown in Tab. 6, we also include the results of OVSeg with CLIP-ViT-L for comparison.

CLIP-ViT-L. According to Tab. 6, FreeSeg with a standard CLIP-ViT-L model (dubbed FreeSeg\mathrm{FreeSeg}) still can not achieve satisfactory results. However, by integrating our MAFT (dubbed FreeSeg+MAFT\mathrm{FreeSeg+MAFT}), the segmentation results are remarkably enhanced, thus establishing new state-of-the-art benchmarks.

CLIP-Res50. Our MAFT can easily adapted into ResNet-based models. Specifically, we modified the AttentionPool2d\mathrm{AttentionPool2d} unit within CLIP-R50 Image Encoder. The mask proposals are introduced as attention bias (BB) in Multihead Attention, with FclsF_{cls} being repeated N times. Notably in CLIP-R50, FclsF_{cls} is obtained via GlobalAveragePooling\mathrm{GlobalAveragePooling} performing on FfeatF_{feat}. The results are presented in Tab. 6. The performance on all 5 datasets is improved by a large margin. FreeSeg+MAFT\mathrm{FreeSeg+MAFT} with CLIP-R50 achieves competitive results with some CLIP-ViT-B-based methods according to Tab. 3.

5.6 Qualitative Study

Visualizations of typical proposals. Fig. 3 shows frozen CLIP and mask-aware CLIP classifications of typical proposals, including high-quality proposals of foreground (p1p_{1}, p4p_{4}), high-quality proposals of background (p3p_{3}, p6p_{6}), a proposal with background noise (p2p_{2}), and a proposal containing part of the foreground (p5p_{5}). The proposal regions are highlighted in green or yellow.
Several observations can be obtained: (1) The frozen CLIP provides good predictions for p1p_{1} and p4p_{4}. (2) The frozen CLIP assigns p2p_{2} as catcat and p5p_{5} as horsehorse, with scores even higher than p1p_{1}, p4p_{4}, indicating the frozen CLIP cannot distinguish proposals containing information on the same objects. (3) The frozen CLIP fails to give correct predictions for p3p_{3} and p6p_{6}, which may be due to the lack of context information. (4) Our mask-aware CLIP gives good predictions for high-quality proposals (p1p_{1}, p3p_{3}, p4p_{4}, p6p_{6}) and provides suitable predictions for p2p_{2} and p5p_{5}.

Qualitative analysis. We show some visual examples in Fig. 4. Some segmentation results of FreeSeg contain background noise (e.g. the 1st1^{st} & 2nd2^{nd} row, 3rd3^{rd} column) or contain only part of the objects (3rd3^{rd} row, 3rd3^{rd} column). In ADE20K-847 dataset, too many classes may lead to the anticipated results (last row, 3rd3^{rd} column) with the frozen CLIP. Using a mask-aware CLIP to learn mask-aware representations can significantly improve these segmentation results, as evident from the last column.

More visual samples are shown in the Appendix.

6 Conclusion

In this paper, we rethink the "frozen CLIP" paradigm in zero-shot segmentation and propose Mask-Aware Fine-Tune (MAFT) for fine-tuning CLIP. Firstly, IP-CLIP Encoder is proposed to handle images with any number of mask proposals. Then, ma\mathcal{L}_{ma} and dis\mathcal{L}_{dis} are designed for fine-tuning CLIP to be mask-aware without sacrificing its transferability. MAFT is plug-and-play and can be applied to any "frozen CLIP" approach. Extensive experiments well demonstrate the performance of various zero-shot segmentation methods is improved by plugging MAFT.

Limitations. Our MAFT introduces a CLIP fine-tining framework to the research of zero-shot segmentation. However, the classification ability for novel classes is still limited by pre-trained vision-language models. How to further narrow this limitation is our future research focus.

Refer to caption
Figure 3: Visualizations of typical proposals & top 5 AcA^{c} by frozen CLIP and mask-aware CLIP.
Refer to caption
Figure 4: Qualitative results. The models are trained with COCO-Stuff and directly tested on VOC2012, COCO, and ADE20K.

Appendix

Here we introduce technical details of the "frozen CLIP" approaches in Sec. A. The dataset settings are shown in Sec. B. Moreover, we provide additional experiments in Sec. C, and additional qualitative results in Sec. D.

Appendix A Technical details of the "frozen CLIP" approaches

Refer to caption
Figure 5: Overview of the "decoupling-paradigm".

Fig. 5 presents an overview of the "frozen CLIP" approach. During training, a standard MaskFormer or Mask2Former is used as Proposal Generator to generate NN mask proposals (MM, MN×h×wM\in\mathbb{R}^{N\times h\times w}) and classification score (ApA^{p}, ApN×|Cseen|A^{p}\in\mathbb{R}^{N\times|C_{seen}|}). During testing, the input image is merged with MM to obtain NN sub-images (IsubI_{sub}, IsubN×h^×w^I_{sub}\in\mathbb{R}^{N\times\hat{h}\times\hat{w}}). These sub-images are fed into a frozen CLIP to get the CLIP classification score (AcA^{c}, AcN×|CseenCunseen|A^{c}\in\mathbb{R}^{N\times|C_{seen}\cup C_{unseen}|}). Here CseenC_{seen} and CunseenC_{unseen} represent a set of seen classes and unseen classes. An ensemble operation is used to ensemble ApA^{p} and AcA^{c} for the final prediction. The merge and the ensemble operations will be introduced in detail in following:

Merge operation. To generate appropriate sub-images based on mask proposals, [6] presents three different merge operations: 1) mask, 2) crop, 3) mask &\& crop. Through experimentation, they demonstrate that the mask &\& crop option yields the best results. Figure 6 provides an example of these operations. It’s worth noting that all sub-images are resized to h^×w^\hat{h}\times\hat{w}, here h^\hat{h} and w^\hat{w} typically take a value of 224, which is the default input size of CLIP Image Encoder. Although acceptable results can be obtained with the merge operation, it involves repeatedly feeding images into CLIP, which leads to significant computational redundancy.

Ensemble operation. Comparatively, ApA^{p} provides higher confidence classification scores for the seen classes, and AcA^{c} provides higher confidence classification scores for the unseen classes. Therefore, an ensemble of ApA^{p} and AcA^{c} achieves better results. The ensemble operation can be formulated as:

A^(c)={Ap(c)λAc(c)(1λ),cCseenAc(c)λ,cCunseen\hat{A}(c)=\left\{\begin{aligned} &A^{p}(c)^{\lambda}\cdot A^{c}(c)^{(1-\lambda)}~{},~{}c\in C^{seen}\\ &A^{c}(c)^{\lambda}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{},~{}c\in C^{unseen}\\ \end{aligned}\right. (9)

here a geometry mean of ApA^{p} and AcA^{c} is calculated (dubbed as A^\hat{A}), and the contribution of both classification scores is balanced by λ\lambda. As per literature [6; 33; 27], λ\lambda usually takes values from 0.6 to 0.8. Therefore, the final output (OO, O|CseenCunseen|×h×wO\in\mathbb{R}^{|C_{seen}\cup C_{unseen}|\times h\times w}) can be obtained by matrix multiplication: O=A^TMO=\hat{A}^{T}\cdot M. With the ensemble operation, the classification results of seen classes primarily depend on ApA^{p}, whereas the classification results of unseen classes mainly rely on AcA^{c}.

Refer to caption
Figure 6: Comparison among three merge operations.

Appendix B Dataset

We follow [1; 11; 26; 6; 33] to conduct experiments on three benchmarks of the popular zero-shot setting, Pascal-VOC, COCO-Stuff and ADE20K, to evaluate the performance of MAFT. Additionally, we evaluate MAFT on the open-vocabulary setting [20; 33], i.e., training on COCO-Stuff and testing on ADE20K, Pascal-Context, and Pascal-VOC.

  • COCO-Stuff: COCO-Stuff is a large-scale semantic segmentation dataset that includes 171 classes. For the zero-shot setting [6; 33; 27], it is divided into 156 seen classes for training and 15 unseen classes for testing. For the open-vocabulary setting, all 171 classes are used for training.

  • Pascal-VOC: There are 10582 images for training and 1,449 images for testing. For the zero-shot setting, Pascal-VOC is split into 15 seen classes and 5 unseen classes. For the open-vocabulary setting, all 20 classes are used for evaluation (dubbed as PAS-20).

  • ADE20K: ADE20K contains 25k images for training and 2k images for validation. For the zero-shot setting, we follow [6] to choose 847 classes present in both training and validation sets, and split them into 572 seen and 275 unseen classes. For the open-vocabulary setting, we use two settings of ADE20K: 150 classes (dubbed as A-150) and 847 classes (dubbed as A-847).

  • Pascal-Context is an extensive dataset of Pascal-VOC 2010. Two versions are used for open-vocabulary setting, one with 59 frequently used classes (dubbed as PC-59) and another with the whole 459 classes (dubbed as PC-459).

Appendix C Additional experiments

C.1 Analysis of the Upper Bound of MAFT

Table 7: Upper Bound analysis.
  COCO-Stuff
mIoUs mIoUu hIoU mIoU
MAFT 43.3 50.4 46.5 43.9
Upper Bound 77.2 82.1 79.6 77.6
 

Considering the mask-aware loss may be limited if the quality of proposals is too bad, we conducted an evaluation of the Upper Bound while using Mask2Former as the proposal generator. The results are presented in Tab. 7. Specifically, we replace AcA^{c} by SIoUS_{IoU} (IoU score between binary ground-truth masks and proposals) during inference, and multiply proposals with SIoUS_{IoU} to obtain the segmentation result. This result can be regarded as the Upper Bound for the given proposals. Notably, the Upper Bound achieves satisfactory results (77.6 % mIoU), indicating Mask2Former is capable of providing high-quality proposals in most cases. Additionally, there is still a large gap between the current performance and the Upper Bound (\approx 30% mIoU), which suggests that our MAFT has enormous potential for improvement, whereas we have achieved state-of-the-art performance.

C.2 Analysis of the Self-Training (ST) strategy

Table 8: Analysis of the self-training (ST) strategy.
  Pascal-VOC COCO-Stuff
mIoUs mIoUu hIoU mIoU mIoUs mIoUu hIoU mIoU
MAFT 91.4 81.8 86.3 89.0 43.3 50.4 46.5 43.9
MAFT + ST 90.0 -1.4 86.3 +4.5 88.1 +1.8 89.1 +0.1 44.1 +0.8 55.2 +4.8 49.0 +2.5 45.0 +1.1
 

Several previous approaches [25; 33; 41] adopt the Self-Training (ST) strategy to enhance performance. we conduct experiments to investigate the application of ST into our method. Specifically, we use the existing FreeSeg+MAFT\mathrm{FreeSeg+MAFT} model to generate pseudo-labels for unseen classes on the training data, and then re-train FreeSeg\mathrm{FreeSeg} with the pseudo-labels. Results are shown in Tab. 8.

The improvement of ST on the unseen category is significant (Pascal: 81.8 % \rightarrow 86.3%, COCO: 50.4% \rightarrow 55.2%) in terms of mIoUu. However, it’s essential to highlight the applicability of ST is limited by a crucial requirement: unseen classes need to be obtained during training. This requirement poses significant limitations on generalizing ST to various scenarios, e.g., open-vocabulary settings, since images of unseen classes may not be obtained during training.

Appendix D Visualization

Refer to caption
Figure 7: Visualizations of typical proposals & top 5 Ac by frozen CLIP and mask-aware CLIP. The correct classes are highlighted in red.
Refer to caption
Figure 8: Qualitative results on A-847, using 847 class names in ADE20K to generate text embeddings.
Refer to caption
Figure 9: Qualitative results on A-150, using 150 class names in ADE20K to generate text embeddings.
Refer to caption
Figure 10: Qualitative results on PC-459, using 459 class names in Pascal-Context to generate text embeddings.
Refer to caption
Figure 11: Qualitative results on PC-59, using 59 class names in Pascal-Context to generate text embeddings.
Refer to caption
Figure 12: Qualitative results on Pascal-VOC, using 20 class names in Pascal-VOC to generate text embeddings.
Refer to caption
Figure 13: Qualitative results on COCO, using 171 class names in COCO-Stuff to generate text embeddings.

We provide more qualitative results, including typical proposals and top-5 AcA^{c} (Fig. 7), as well as examples of models trained on COCO-Stuff and text on A-847 (Fig. 8), A-150 (Fig. 9), PC-459 (Fig. 10), PC-59 (Fig. 11), Pascal-VOC (Fig. 12), and COCO-Stuff(Fig. 13).

Typical Proposals and Top-5 AcA^{c}. Fig. 7 shows frozen CLIP and mask-aware CLIP classifications of typical proposals. In the 2nd column, we provide high-quality proposals of thing classes. Both the frozen CLIP and mask-aware CLIP provide high classification scores for the correct classes. In the 3rd column, we provide proposals that only contain part of the objects (row 1-3), and proposals containing more than 1 class (row 4). The mask-aware CLIP provides more proper results compared to the frozen CLIP. In the 4th column, we provide some high-quality background proposals. The frozen CLIP typically gives incorrect predictions, but the mask-aware CLIP assigns high scores for the correct classes.

Qualitative Analysis. Fig. 8,9,10,11,12,13 show segmentation results on Pascal-VOC, COCO-Stuff, ADE20K. In Pascal-VOC dataset (Fig. 12), which only contains 20 thing classes, the FreeSeg+MAFT\mathrm{FreeSeg+MAFT} model tends to assign background regions to the similar thing classes, e.g., "train" in row 1, "pottedplant" in row3-4. "boat" in row 8. In A-847, A-150, PC-459, PC-59 and COCO-Stuff datasets, both seen classes and unseen classes exist in the input images, the FreeSeg+MAFT\mathrm{FreeSeg+MAFT} model generates better segmentation results compared to FreeSeg\mathrm{FreeSeg}.

References

  • [1] Maxime Bucher, Tuan-Hung Vu, Matthieu Cord, and Patrick Pérez. Zero-shot semantic segmentation. Advances in Neural Information Processing Systems, 32, 2019.
  • [2] Holger Caesar, Jasper Uijlings, and Vittorio Ferrari. Coco-stuff: Thing and stuff classes in context. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 1209–1218, 2018.
  • [3] Liang-Chieh Chen, George Papandreou, Iasonas Kokkinos, Kevin Murphy, and Alan L Yuille. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE transactions on pattern analysis and machine intelligence, 40(4):834–848, 2017.
  • [4] Bowen Cheng, Anwesa Choudhuri, Ishan Misra, Alexander Kirillov, Rohit Girdhar, and Alexander G Schwing. Mask2former for video instance segmentation. arXiv preprint arXiv:2112.10764, 2021.
  • [5] Bowen Cheng, Alex Schwing, and Alexander Kirillov. Per-pixel classification is not all you need for semantic segmentation. Advances in Neural Information Processing Systems, 34, 2021.
  • [6] Jian Ding, Nan Xue, Gui-Song Xia, and Dengxin Dai. Decoupling zero-shot semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 11583–11592, 2022.
  • [7] Mark Everingham, SM Ali Eslami, Luc Van Gool, Christopher KI Williams, John Winn, and Andrew Zisserman. The pascal visual object classes challenge: A retrospective. International journal of computer vision, 111:98–136, 2015.
  • [8] Golnaz Ghiasi, Xiuye Gu, Yin Cui, and Tsung-Yi Lin. Open-vocabulary image segmentation. arXiv preprint arXiv:2112.12143, 2021.
  • [9] Golnaz Ghiasi, Xiuye Gu, Yin Cui, and Tsung-Yi Lin. Scaling open-vocabulary image segmentation with image-level labels. In Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23–27, 2022, Proceedings, Part XXXVI, pages 540–557. Springer, 2022.
  • [10] Zhangxuan Gu, Siyuan Zhou, Li Niu, Zihan Zhao, and Liqing Zhang. Context-aware feature generation for zero-shot semantic segmentation. In Proceedings of the 28th ACM International Conference on Multimedia, pages 1921–1929, 2020.
  • [11] Zhangxuan Gu, Siyuan Zhou, Li Niu, Zihan Zhao, and Liqing Zhang. Context-aware feature generation for zero-shot semantic segmentation. In Proceedings of the 28th ACM International Conference on Multimedia, pages 1921–1929, 2020.
  • [12] Zixian Guo, Bowen Dong, Zhilong Ji, Jinfeng Bai, Yiwen Guo, and Wangmeng Zuo. Texts as images in prompt tuning for multi-label image recognition. arXiv preprint arXiv:2211.12739, 2022.
  • [13] Kunyang Han, Yong Liu, Jun Hao Liew, Henghui Ding, Yunchao Wei, Jiajun Liu, Yitong Wang, Yansong Tang, Yujiu Yang, Jiashi Feng, et al. Global knowledge calibration for fast open-vocabulary segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023.
  • [14] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 770–778, 2016.
  • [15] Zilong Huang, Xinggang Wang, Lichao Huang, Chang Huang, Yunchao Wei, and Wenyu Liu. Ccnet: Criss-cross attention for semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 603–612, 2019.
  • [16] Zilong Huang, Yunchao Wei, Xinggang Wang, Wenyu Liu, Thomas S Huang, and Humphrey Shi. Alignseg: Feature-aligned segmentation networks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(1):550–557, 2021.
  • [17] Chao Jia, Yinfei Yang, Ye Xia, Yi-Ting Chen, Zarana Parekh, Hieu Pham, Quoc Le, Yun-Hsuan Sung, Zhen Li, and Tom Duerig. Scaling up visual and vision-language representation learning with noisy text supervision. In International Conference on Machine Learning, pages 4904–4916. PMLR, 2021.
  • [18] Alexander Kirillov, Eric Mintun, Nikhila Ravi, Hanzi Mao, Chloe Rolland, Laura Gustafson, Tete Xiao, Spencer Whitehead, Alexander C Berg, Wan-Yen Lo, et al. Segment anything. arXiv preprint arXiv:2304.02643, 2023.
  • [19] Harold W Kuhn. The hungarian method for the assignment problem. Naval research logistics quarterly, 2(1-2):83–97, 1955.
  • [20] Feng Liang, Bichen Wu, Xiaoliang Dai, Kunpeng Li, Yinan Zhao, Hang Zhang, Peizhao Zhang, Peter Vajda, and Diana Marculescu. Open-vocabulary semantic segmentation with mask-adapted clip. arXiv preprint arXiv:2210.04150, 2022.
  • [21] Jiaxu Miao, Xiaohan Wang, Yu Wu, Wei Li, Xu Zhang, Yunchao Wei, and Yi Yang. Large-scale video panoptic segmentation in the wild: A benchmark. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 21033–21043, 2022.
  • [22] Jiaxu Miao, Yunchao Wei, Yu Wu, Chen Liang, Guangrui Li, and Yi Yang. Vspw: A large-scale dataset for video scene parsing in the wild. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 4133–4143, 2021.
  • [23] Juhong Min, Dahyun Kang, and Minsu Cho. Hypercorrelation squeeze for few-shot segmentation. In Proceedings of the IEEE/CVF international conference on computer vision, pages 6941–6952, 2021.
  • [24] Roozbeh Mottaghi, Xianjie Chen, Xiaobai Liu, Nam-Gyu Cho, Seong-Whan Lee, Sanja Fidler, Raquel Urtasun, and Alan Yuille. The role of context for object detection and semantic segmentation in the wild. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 891–898, 2014.
  • [25] Giuseppe Pastore, Fabio Cermelli, Yongqin Xian, Massimiliano Mancini, Zeynep Akata, and Barbara Caputo. A closer look at self-training for zero-label semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2693–2702, 2021.
  • [26] Giuseppe Pastore, Fabio Cermelli, Yongqin Xian, Massimiliano Mancini, Zeynep Akata, and Barbara Caputo. A closer look at self-training for zero-label semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2693–2702, 2021.
  • [27] Jie Qin, Jie Wu, Pengxiang Yan, Ming Li, Ren Yuxi, Xuefeng Xiao, Yitong Wang, Rui Wang, Shilei Wen, Xin Pan, et al. Freeseg: Unified, universal and open-vocabulary image segmentation. arXiv preprint arXiv:2303.17225, 2023.
  • [28] Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. In International conference on machine learning, pages 8748–8763. PMLR, 2021.
  • [29] Amirreza Shaban, Shray Bansal, Zhen Liu, Irfan Essa, and Byron Boots. One-shot learning for semantic segmentation. arXiv preprint arXiv:1709.03410, 2017.
  • [30] Yanpeng Sun, Qiang Chen, Xiangyu He, Jian Wang, Haocheng Feng, Junyu Han, Errui Ding, Jian Cheng, Zechao Li, and Jingdong Wang. Singular value fine-tuning: Few-shot segmentation requires few-parameters fine-tuning. arXiv preprint arXiv:2206.06122, 2022.
  • [31] Yongqin Xian, Subhabrata Choudhury, Yang He, Bernt Schiele, and Zeynep Akata. Semantic projection network for zero-and few-label semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 8256–8265, 2019.
  • [32] Jiarui Xu, Shalini De Mello, Sifei Liu, Wonmin Byeon, Thomas Breuel, Jan Kautz, and Xiaolong Wang. Groupvit: Semantic segmentation emerges from text supervision. CVPR, 2022.
  • [33] Mengde Xu, Zheng Zhang, Fangyun Wei, Yutong Lin, Yue Cao, Han Hu, and Xiang Bai. A simple baseline for open-vocabulary semantic segmentation with pre-trained vision-language model. In Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23–27, 2022, Proceedings, Part XXIX, pages 736–753. Springer, 2022.
  • [34] Nir Zabari and Yedid Hoshen. Semantic segmentation in-the-wild without seeing any segmentation examples. arXiv preprint arXiv:2112.03185, 2021.
  • [35] Gengwei Zhang, Guoliang Kang, Yi Yang, and Yunchao Wei. Few-shot segmentation via cycle-consistent transformer. Advances in Neural Information Processing Systems, 34:21984–21996, 2021.
  • [36] Gengwei Zhang, Shant Navasardyan, Ling Chen, Yao Zhao, Yunchao Wei, Honghui Shi, et al. Mask matching transformer for few-shot segmentation. Advances in Neural Information Processing Systems, 35:823–836, 2022.
  • [37] Zekang Zhang, Guangyu Gao, Zhiyuan Fang, Jianbo Jiao, and Yunchao Wei. Mining unseen classes via regional objectness: A simple baseline for incremental segmentation. Advances in Neural Information Processing Systems, 35:24340–24353, 2022.
  • [38] Zekang Zhang, Guangyu Gao, Jianbo Jiao, Chi Harold Liu, and Yunchao Wei. Coinseg: Contrast inter-and intra-class representations for incremental segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023.
  • [39] Hengshuang Zhao, Jianping Shi, Xiaojuan Qi, Xiaogang Wang, and Jiaya Jia. Pyramid scene parsing network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2881–2890, 2017.
  • [40] Bolei Zhou, Hang Zhao, Xavier Puig, Sanja Fidler, Adela Barriuso, and Antonio Torralba. Scene parsing through ade20k dataset. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 633–641, 2017.
  • [41] Chong Zhou, Chen Change Loy, and Bo Dai. Extract free dense labels from clip. In Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23–27, 2022, Proceedings, Part XXVIII, pages 696–712. Springer, 2022.
  • [42] Kaiyang Zhou, Jingkang Yang, Chen Change Loy, and Ziwei Liu. Conditional prompt learning for vision-language models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 16816–16825, 2022.
  • [43] Kaiyang Zhou, Jingkang Yang, Chen Change Loy, and Ziwei Liu. Learning to prompt for vision-language models. International Journal of Computer Vision, 130(9):2337–2348, 2022.