LCTR: On Awakening the Local Continuity of Transformer for
Weakly Supervised Object Localization
Abstract
Weakly supervised object localization (WSOL) aims to learn object localizer solely by using image-level labels. The convolution neural network (CNN) based techniques often result in highlighting the most discriminative part of objects while ignoring the entire object extent. Recently, the transformer architecture has been deployed to WSOL to capture the long-range feature dependencies with self-attention mechanism and multilayer perceptron structure. Nevertheless, transformers lack the locality inductive bias inherent to CNNs and therefore may deteriorate local feature details in WSOL. In this paper, we propose a novel framework built upon the transformer, termed LCTR (Local Continuity TRansformer), which targets at enhancing the local perception capability of global features among long-range feature dependencies. To this end, we propose a relational patch-attention module (RPAM), which considers cross-patch information on a global basis. We further design a cue digging module (CDM), which utilizes local features to guide the learning trend of the model for highlighting the weak local responses. Finally, comprehensive experiments are carried out on two widely used datasets, i.e., CUB-200-2011 and ILSVRC, to verify the effectiveness of our method.

Introduction
Deep learning based methods have achieved unprecedented success in locating objects under a fully supervised setting (Liu et al. 2016; Bochkovskiy, Wang, and Liao 2020; Sun et al. 2021; Wang et al. 2021). However, these methods rely on a large number of bounding box annotations, which are expensive to acquire. Recently, the research on weakly supervised object localization (WSOL) has gained a significant momentum (Zhou et al. 2016; Zhang et al. 2018a; Gao et al. 2021) since it can learn object localizers using only image-level labels.
The pioneering work (Zhou et al. 2016) aggregated features from classification networks to generate class activation maps (CAM) for object localization. Unfortunately, image classifiers tend to focus only on the most discriminative features to achieve high classification performance. Therefore, the spatial distribution of feature responses may only cover the most discerning regions instead of the whole object range, which limits localization accuracy with large margins, as shown in Figure 1(b).
To address this critical problem, many CAM-based approaches have been proposed, such as graph propagation (Zhu et al. 2017), data augmentation (Kumar Singh and Jae Lee 2017; Yun et al. 2019), adversarial erasing (Zhang et al. 2018a; Choe and Shim 2019; Chen et al. 2021) and spatial relation activation (Xue et al. 2019; Zhang, Wei, and Yang 2020; Guo et al. 2021). However, those approaches do alleviate the partial activation issue, but in a compromised manner — the essential philosophy behind it is first obtaining local features and then attempting to recover the non-salient regions to get full object extent. In fact, the fundamental root of this issue is determined by the intrinsic nature of convolution neural networks (CNNs). The CNN features with the local receptive field only capture small-range feature dependencies.
More recently, the transformer architecture (Vaswani et al. 2017) has been developed in the field of computer vision (Dosovitskiy et al. 2020; Wu et al. 2020; Yuan et al. 2021; Touvron et al. 2021; Jiang, Chang, and Wang 2021), which shows that pure transformers can be as effective in feature extraction for image recognition as CNN-based architectures. Notably, transformers with multi-head self-attention capture long-range dependencies, and retain more detailed information without downsampling operators, which naturally handles the partial activation problem in WSOL. TS-CAM (Gao et al. 2021) proposed token semantic coupled attention map from transformer structure, which captured long-range feature dependency among pixels for WSOL. However, transformer-based methods lack the locality inductive bias inherent to CNNs, ignoring the local information, which leads to weak local feature response on the target object, as shown in Figure 1(c). Therefore, how to precisely mine local features in global representations for WSOL still remains an open problem.
In this paper, we propose a novel Local Continuity TRansformer (LCTR) for discovering entire objects of interest via end-to-end weakly supervised training. The key idea of LCTR is to rearrange local-continuous visual patterns with global-connective self-attention maps, thereby bringing locality mechanism to transformer-based WSOL. To this end, we first propose a relational patch-attention module (RPAM) to construct a powerful patch relation map, which takes advantage of the patch attention maps under the guidance of a global class-token attention map. The RPAM maintains the cross-patch information and models a global representation with more local cues. Second, a cue digging module (CDM) is designed succinctly to induce the model to highlight the weak local features (e.g., blurred object boundaries) by a hide-and-seek manner under a local view. In the CDM, to reward the weak response parts, we propose to employ the erased strategy, and induce the learnable convolutional kernels to be weighted by the weak local features. To validate the effectiveness of the proposed LCTR, we conduct a series of experiments on the challenging WSOL benchmarks.
Collectively, our main contributions are summarized as:
-
•
We propose a simple LCTR for WSOL, which greatly enhances the local perception capability of global self-attention maps among long-range feature dependencies.
-
•
We design a relational patch-attention module (RPAM) by considering cross-patch information, which facilitates global representations.
-
•
We introduce a cue digging module (CDM) that encodes weak local features by learnable kernels to highlight the local details of global representations.
-
•
LCTR achieves new state-of-the-art performance on CUB-200-2011 and ILSVRC dataset with 79.2% and 56.1% Top-1 localization accuracy, respectively.

Related Work
CNN-based Methods for WSOL. WSOL aims to learn object localizers with solely image-level supervision. There are many state-of-the-art methods based on the CNN structure. A representative pipeline of CNN-based WSOL is to aggregate deep feature maps with a class-specific fully connected layer to produce class attention maps (CAMs), from which final predicted bounding boxes are extracted (Zhou et al. 2016). Later on, the last fully connected layer is dropped for simplifying (Hwang and Kim 2016). Unfortunately, CAMs tend to be dominated by the most discriminative object part. Therefore, different extensions (Selvaraju et al. 2017; Chattopadhay et al. 2018; Xue et al. 2019; Zhang, Wei, and Yang 2020) have been proposed to improve the generation process of localization maps in order to recover the non-salient regions. HaS (Kumar Singh and Jae Lee 2017) and CutMix (Yun et al. 2019) adopted a random-erasing strategy from input images to force the classification networks to focus on relevant parts of objects. ACoL (Zhang et al. 2018a) introduced two adversarial classification classifiers to locate different object parts and discovered the complementary regions belonging to the same objects or categories. ADL (Choe and Shim 2019) further promoted the localization maps by applying dropout on multiple intermediate feature maps. Besides the erasing strategy, DANet (Xue et al. 2019) used a divergent activation method to learn better localization maps. SPG (Zhang et al. 2018b) and I2C (Zhang, Wei, and Yang 2020) introduced the constraint of pixel-level correlations into the WSOL network. SPA (Pan et al. 2021) leveraged structure information incorporated in convolutional features for WSOL. Some other methods (e.g., GC-Net (Lu et al. 2020), PSOL (Zhang, Cao, and Wu 2020), SPOL (Wei et al. 2021) and SLT-Net (Guo et al. 2021)) divided WSOL into two independent sub-tasks, including classification and the class-agnostic localization.
These studies alleviate the problem by extending from local activations to global ones in an implicit way, which is difficult to balance the image classification and the object localization. In fact, CNNs are prone to capture partial semantic features with local receptive fields, which belongs to the principal problem of CNNs. The problem of how to explore global cues from local receptive fields still exists. In this paper, we introduce a transformer-based structure, where the local-continuity and long-range feature dependencies can be simultaneously activated.
Transformer-based Methods for WSOL. The transformer model (Vaswani et al. 2017) is proposed to handle sequential data in the field of natural language processing. Recent studies also reveal its effectiveness for computer vision tasks (Dosovitskiy et al. 2020; Beal et al. 2020; Carion et al. 2020; Zheng et al. 2021; Hu et al. 2021). Since the local information extracted by the CNNs is deficient, various methods adopt the self-attention mechanism to capture the long-range feature dependencies. ViT (Dosovitskiy et al. 2020) applied the pure transformer directly to sequences of image patches for exploring spatial correlation on the image classification task. DETR (Carion et al. 2020) employed a transformer encoder-decoder architecture for the object detection task. As a pioneered work in WSOL, TS-CAM (Gao et al. 2021) proposed a semantic coupling strategy based on Deit (Touvron et al. 2021) to fuse the patch tokens with the semantic-agnostic attention map to achieve semantic-aware localization results, which has inspired many scholars to study transformer in weakly supervised object localization.
Despite of the progress, TS-CAM is relatively rough as it single-mindedly tries to get the long-range features but ignores the local information. Compared with the existing methods, our LCTR retains long-dependent features while mining for detailed feature cues based on the transformer structure for WSOL.
Methodology
We first present the overview of the proposed LCTR, then give a detailed description of RPAM and CDM, and finally incorporate them with the transformer structure in a joint optimization framework, as shown in Figure 2.
Overview
In accordance with the long-range info-preserving ability of the transformer architecture, LCTR is designed to offer precise localization maps for WSOL. We denote the input images as , where indicates the label of the image , and are the number of images and classes, respectively. We first split into same-sized patches , where denotes the dimension of each patch. We set , and , where is the width/height of a patch, and denote image height and width. For simplicity, we omit the mini-batch dimension. A learnable class token is embedded into the patches. These patches are flattened and linearly projected before being fed to sequential transformer blocks, which can be formulated as:
(1) |
where denotes the input of the first transformer block, is the position embedding and is a linear projection. In particular, the proposed RPAM is employed in each transformer block to obtain a patch relation map , which aggregates cross-patch information on a global basis.
Denote as the output feature of the last transformer block. We reshape and apply the proposed CDM for further highlighting weak local responses. After that we obtain the feature map . Finally, the are fed to a global average pooling (GAP) layer (Lin, Chen, and Yan 2013) followed by a softmax layer to predict the classification probability . The loss function is defined as
(2) |
During testing, we extract the object map from according to the predicted class and obtain the final localization map by element-wise multiplication, given as
(3) |
The is then resized to the same size as the original images by linear interpolation. For a fair comparison, we apply the same strategy detailed in CAM (Zhou et al. 2016) to produce the object bounding boxes.
Relational Patch-Attention Module
The proposed relational patch-attention module (RPAM) (Figure 3) strengthens the global feature representation from two stages: First, we utilize the attention vectors of the class token in the transformer block to generate a global class-token attention map. To fully exploit the feature dependencies of the transformer structure, we then use all the attention vectors of the patches containing the correlation between local features to generate a patch relation map under the guidance of the class-token attention map.
In the -th transformer block, we hypothesize that the output feature map is . The attention matrix of multi-head self-attention module in the block is formulated as:
(4) |
where and denote the queries and keys projected by of self-attention operation in -th transformer block, respectively. represents the number of head and is a transpose operator.
At this point, we first take the average operator to based on heads to obtain . Then, the class-token attention vector is extracted from . The reveals how much each patch contributes to the object regions for image classification. Unfortunately, this map simply captures the global interactions of the class token to all patches, while ignoring the cross-patch correlations, which affects the modeling of local features. To remedy it, we take advantage of the patch attention map in to structure a patch relation vector under the guidance of . The learns the correlation between each patch but couldn’t tell which one is more important. Therefore, we weight each patch attention map by multiplying by to obtain a new map . Note that is reshaped () before the multiplication. After that, we squeeze the first dimension of to a vector () by an average operation. The final patch relation map is calculated by
(5) |
where indicates the reshape operator which coverts the vector () to the map ().
The patch relation map obtains the long-range dependencies that depends on the class-token attention vector . Aggregating cross-patch information from the patch attention maps, facilitates better global representations of the object without extra parameters in a simple way.
Cue Digging Module
RPAM considers the cross-patch information by using the class-token attention map from self-attention mechanism block of the transformer structure, but it is vulnerable if the transformer gets a poor class-token attention map. We thus further propose a cue digging module (CDM) to supply the long-range features based on a hide-and-seek manner.
Inspired by erasing-based methods that remove the most discriminative parts of the target object to induce the model to cover the integral extent of the object, we erase the object regions based on the global class-token attention map, leaving the weak response ones and the background. Then by weighting the learnable convolution kernels in the CDM on the basis of them, we shift part of the attention to object regions with weak responses. With the help of weighted kernels, we can highlight the local details as a supplement to the global representations.
Specifically, we convert the patch parts of class-token attention vectors to the map , and apply it to the feature map by spatial-wise multiplication after being reversed. Note that is calculated by . The feature map then passes through a convolutional layer to generate a new feature map . Next, we score the features into parts corresponding to learnable convolution kernels. In particular, we apply two separate operators, the global average pooling and max pooling, to . Then the feature maps are vectorized and sent to a fully connected layer, respectively. Besides, we add them together and apply a sigmoid function. In this light, we obtain the scores as . Finally, passes through a convolutional layer with weighted convolution kernels by the scores for encouraging the model to learn the object regions with weak responses, which can be formulated as
(6) |
where denotes the kernel weights of the convolutional layer, which are initialized with kaiming uniform initialization (He et al. 2015). , represent the width and height of the kernels, respectively.
The convolution layer with the weighted kernels is applied to the global feature map drawn from the transformer structure. Once the loss in Eq. 2 is optimized, the weighted convolution kernels become more sensitive to the features (i.e., the weak response features of object regions) that favor image classification. In this way, the model pays more attention to local cues and forms better global representations for the target object.

Experiments
Experimental Settings
Datasets. We evaluate the proposed methods on two challenging datasets, including CUB-200-2011 (Wah et al. 2011) and ILSVRC (Russakovsky et al. 2015). We only use image-level labels for training. CUB-200-2011 is a fine-grained bird dataset of categories, which contains images for training and for testing. ILSVRC has about million images in the training set and images in the validation set, with a total of different categories.
Evaluation Metrics. Following previous methods (Zhou et al. 2016; Russakovsky et al. 2015), we adopt the Top-1/Top-5 classification accuracy (Top-1/Top-5 Cls.), Top-1/Top-5 localization accuracy (Top-1/Top-5 Loc.) and localization accuracy with known ground-truth class (Gt-k.) as our evaluation metrics. Specifically, Top-1/Top-5 Cls. is correct if the Top-1/Top-5 predicted category contains the correct label. Gt-k. is correct when the intersection over union (IoU) between the ground-truth and the prediction is larger than , and does not consider whether the predicted category is correct. Top-1/Top-5 Loc. is correct when Top-1/Top-5 Cls. and Gt-k. are both correct.
Methods (Yr) | Backbone | Loc. Acc | ||
Top-1 | Top-5 | Gt-k. | ||
CAM (’16) | GoogLeNet | 41.1 | 50.7 | 55.1 |
SPG (’18) | GoogLeNet | 46.7 | 57.2 | - |
RCAM (’20) | GoogLeNet | 53.0 | - | 70.0 |
DANet (’19) | InceptionV3 | 49.5 | 60.5 | 67.0 |
ADL (’19) | InceptionV3 | 53.0 | - | - |
PSOL (’20) | InceptionV3 | 65.5 | - | - |
SPA (’21) | InceptionV3 | 53.6 | 66.5 | 72.1 |
SLT-Net (’21) | InceptionV3 | 66.1 | - | 86.5 |
CAM (’16) | VGG16 | 44.2 | 52.2 | 56.0 |
ADL (’19) | VGG16 | 52.4 | - | 75.4 |
ACoL (’18) | VGG16 | 45.9 | 56.5 | 59.3 |
SPG (’18) | VGG16 | 48.9 | 57.2 | 58.9 |
DANet (’19) | VGG16 | 52.5 | 62.0 | 67.7 |
MEIL (’20) | VGG16 | 57.5 | - | 73.8 |
PSOL (’20) | VGG16 | 66.3 | - | - |
RCAM (’20) | VGG16 | 59.0 | - | 76.3 |
GC-Net (’20) | VGG16 | 63.2 | - | - |
SPA (’21) | VGG16 | 60.2 | 72.5 | 77.2 |
SLT-Net (’21) | VGG16 | 67.8 | - | 87.6 |
TS-CAM (’21) | Deit-S | 71.3 | 83.8 | 87.7 |
LCTR (Ours) | Deit-S | 79.2 | 89.9 | 92.4 |
Methods (Yr) | Backbone | Loc. Acc | ||
Top-1 | Top-5 | Gt-k. | ||
CAM (’16) | VGG16 | 38.9 | 48.5 | - |
ACoL (’18) | VGG16 | 45.8 | 59.4 | 63.0 |
CutMix (’19) | VGG16 | 42.8 | 54.9 | 59.0 |
ADL (’19) | VGG16 | 44.9 | - | - |
I2C (’20) | VGG16 | 47.4 | 58.5 | 63.9 |
MEIL (’20) | VGG16 | 46.8 | - | - |
RCAM (’20) | VGG16 | 44.6 | - | 60.7 |
PSOL (’20) | VGG16 | 50.9 | 60.9 | 64.0 |
SPA (’21) | VGG16 | 49.6 | 61.3 | 65.1 |
SLT-Net (’21) | VGG16 | 51.2 | 62.4 | 67.2 |
CAM (’16) | InceptionV3 | 46.3 | 58.2 | 62.7 |
SPG (’18) | InceptionV3 | 48.6 | 60.0 | 64.7 |
ADL (’19) | InceptionV3 | 48.7 | - | - |
ACoL (’18) | GoogLeNet | 46.7 | 57.4 | - |
DANet (’19) | GoogLeNet | 47.5 | 58.3 | - |
RCAM (’20) | GoogLeNet | 50.6 | - | 64.4 |
MEIL (’20) | InceptionV3 | 49.5 | - | - |
I2C (’20) | InceptionV3 | 53.1 | 64.1 | 68.5 |
GC-Net (’20) | InceptionV3 | 49.1 | 58.1 | - |
PSOL (’20) | InceptionV3 | 54.8 | 63.3 | 65.2 |
SPA (’21) | InceptionV3 | 52.8 | 64.3 | 68.4 |
SLT-Net (’21) | InceptionV3 | 55.7 | 65.4 | 67.6 |
TS-CAM (’21) | Deit-S | 53.4 | 64.3 | 67.6 |
LCTR (Ours) | Deit-S | 56.1 | 65.8 | 68.7 |
Methods (Yr) | Backbone | Cls. Acc | |
---|---|---|---|
Top-1 | Top-5 | ||
CAM (’16) | GoogLeNet | 73.8 | 91.5 |
RCAM (’20) | GoogLeNet | 73.7 | - |
DANet (’19) | InceptionV3 | 71.2 | 90.6 |
ADL (’19) | InceptionV3 | 74.6 | - |
SLT-Net (’21) | InceptionV3 | 76.4 | - |
CAM (’16) | VGG16 | 76.6 | 92.5 |
ACoL (’18) | VGG16 | 71.9 | - |
ADL (’19) | VGG16 | 65.3 | - |
DANet (’19) | VGG16 | 75.4 | 92.3 |
SPG (’18) | VGG16 | 75.5 | 92.1 |
MEIL (’20) | VGG16 | 74.8 | - |
RCAM (’20) | VGG16 | 75.0 | - |
SLT-Net (’21) | VGG16 | 76.6 | - |
TS-CAM (’21) | Deit-S | 80.3 | 94.8 |
LCTR (Ours) | Deit-S | 85.0 | 97.1 |
Implementation Details. We adopt the Deit (Touvron et al. 2021) as the backbone network, which is pre-trained on ILSVRC (Russakovsky et al. 2015). Particularly, we replace the MLP head with our proposed CDM. Finally, a GAP layer and a softmax layer are added on the top of the convolutional layers. The input images are randomly cropped to pixels after being resized to pixels. We adopt AdamW (Loshchilov and Hutter 2017) with =-, =, = and weight decay of -. On CUB-200-2011, we use a batch size of with a learning rate of - to train the model for 80 epochs. For ILSVRC, the training process lasts epochs with a batch size of and a learning rate of -. After meticulous experiments, we set =4 in the CDM. All the experiments are performed with four Nvidia Tesla V100 GPUs using the PyTorch toolbox.

Comparison with the State-of-the-Arts
Localization. We first compare the proposed LCTR with the SOTAs on the localization accuracy on the CUB-200-2011 test set, as illustrated in Table 1. We observe that LCTR outperforms the baseline (i.e., TS-CAM (Gao et al. 2021)) by 7.9% in terms of Top-1 Loc., and is obviously superior to these CNN-based methods. Besides, Table 2 illustrates the localization accuracy on the ILSVRC validation set. It reports 0.4% performance improvement over the state-of-the-art SLT-Net (Guo et al. 2021).
Classification. Table 3 and Table 3 show the Top-1 and Top-5 classification accuracy on the CUB-200-2011 test set and ILSVRC validation set, respectively. For the fine-grained recognition dataset CUB-200-2011, LCTR achieves remarkable performance of 85.0%/97.1% on Top1/Top-5 Acc.. In addition, LCTR obtains comparable results with SLT-Net (Guo et al. 2021) on Top-1 Acc. and surpasses other methods significantly on the ILSVRC validation set. Note that SLT-Net used a separated localization-classification framework, it cannot retain the global information for the objects in the individual classification network. To sum up, the proposed LCTR can greatly improve the quality of object localization while keeping high classification performance.
Methods (Yr) | Backbone | Cls. Acc | |
---|---|---|---|
Top-1 | Top-5 | ||
CAM (’16) | VGG16 | 68.8 | 88.6 |
ACoL (’18) | VGG16 | 67.5 | 88.0 |
I2C (’20) | VGG16 | 69.4 | 89.3 |
MEIL (’20) | VGG16 | 70.3 | - |
RCAM (’20) | VGG16 | 68.7 | - |
SLT-Net (’21) | VGG16 | 72.4 | - |
CAM (’16) | InceptionV3 | 73.3 | 91.8 |
SPG (’18) | InceptionV3 | 69.7 | 90.1 |
ADL (’19) | InceptionV3 | 72.8 | - |
ACoL (’18) | GoogLeNet | 71.0 | 88.2 |
DANet (’19) | GoogLeNet | 63.5 | 91.4 |
RCAM (’20) | GoogLeNet | 74.3 | - |
MEIL (’20) | InceptionV3 | 73.3 | - |
I2C (’20) | InceptionV3 | 73.3 | 91.6 |
SLT-Net (’21) | InceptionV3 | 78.1 | - |
TS-CAM (’21) | Deit-S | 74.3 | 92.1 |
LCTR (Ours) | Deit-S | 77.1 | 93.4 |
Visualization. For qualitative evaluation, Figure 4 visualizes the final localization results of CAM (Zhou et al. 2016) based on the CNNs, TS-CAM (Gao et al. 2021) based on the transformer and our method on CUB200-2011 and ILSVRC datasets. From the results, compared with the CAM, we consistently observe that our method can cover a more complete range of object regions instead of focusing only on the most discriminative ones. In addition, we capture more localized cues than the TS-CAM method, resulting in more accurate localization. For example, the tail regions of the Bank Swallow and the Colobus are ignored by CAM and TS-CAM methods, while our LCTR is able to aggregate more detailed features of the target object, which enhances the local perception capability of global features among long-range feature dependencies. Please refer to the supplementary materials for more visualized localization results of our method.

Applied Mode | Top-1 Loc. | Gt-k. | Top-1 Cls. |
---|---|---|---|
GMP-fc | 75.6 | 88.7 | 84.7 |
GAP-fc | 75.8 | 88.8 | 84.8 |
(GMP+GAP)-fc | 75.8 | 89.2 | 84.9 |
(GMP-fc) + (GAP-fc) | 76.0 | 90.0 | 85.0 |
Top-1 Loc. | Gt-k. | Top-1 Cls. | |
---|---|---|---|
2 | 75.2 | 88.7 | 84.7 |
4 | 76.0 | 90.0 | 85.0 |
8 | 74.3 | 87.8 | 84.4 |
16 | 74.1 | 88.6 | 83.2 |
32 | 74.0 | 88.4 | 82.9 |
Kernel Size () | Top-1 Loc. | Gt-k. | Top-1 Cls. |
---|---|---|---|
73.5 | 88.8 | 82.5 | |
76.0 | 90.0 | 85.0 |
Ablation Studies
First, we visualize the localization maps with different settings in Figure 5. We observe that the RPAM strengthens the global representations of the baseline (Gao et al. 2021), e.g., the tail-feature response of the African chameleon is enhanced, as it considers more cross-patch information. When only using CDM, we find that the local feature details are mined. For example, the abdominal features of the Black grouse are further activated compared to the baseline. By applying both RPAM and CDM, the final localization map (Figure 5 (e)) highlights the full object extent.
Methods | Dataset | RPAM | CDM | Top-1 | Top-1 |
Loc. | Cls. | ||||
TS-CAM | CUB | 71.3 | 80.3 | ||
TS-CAM* | 73.1 | 81.6 | |||
LCTR | CUB | ✓ | 74.0 | 81.6 | |
✓ | 76.0 | 85.0 | |||
✓ | ✓ | 79.2 | 85.0 | ||
TS-CAM | ILSVRC | 53.4 | 74.3 | ||
TS-CAM* | 53.0 | 74.0 | |||
LCTR | ILSVRC | ✓ | 54.2 | 74.0 | |
✓ | 55.1 | 77.1 | |||
✓ | ✓ | 56.1 | 77.1 |
Next, we explore the concise design of the CDM. From the results on Table 5, we can observe that the mode of using separate fcs with GAP and GMP reports the best performance. These results also verify that GAP and GMP work differently in the CDM. Then, we evaluate the accuracy under different parameters in the CDM, as shown in Table 6. From the experimental results, we observe that the best performance is achieved when . Setting a larger leads to a larger number of parameters and degrades accuracy, which we believe is caused by overfitting. Besides, we examine the impact of the weighted kernel size (i.e., ). Results shown in Table 7 indicate that a kernel size of yields better performance.
Lastly, we investigate the effect with different configurations on the accuracy, as reported in Table 8. On the CUB-200-2011 test set, we can see that RPAM increases the Top-1 Loc. by 0.9% compared with the baseline TS-CAM method. Note that the lightweight RPAM is directly applied in the test phase, so the classification performance remains unchanged. When applying the CDM to the network, we observe an improvement in both classification and localization performance. From this, we believe that the local cues captured by the CDM are important for both two tasks. The best localization/classification accuracy can be achieved when employing both RPAM and CDM. Meanwhile, we conduct the similar experiments on the ILSVRC validation set, which also validate the effectiveness of two modules, as shown in the lower part of Table 8.
Conclusion
In this paper, we propose a novel Local Continuity TRansformer, termed LCTR, for weakly supervised object localization, which induces the model to learn the entire extent of the object with more local cues. We first design a relational patch-attention module (RPAM), considering cross-patch information based on the multi-head self-attention mechanism, which gathers more local patch features for facilitating the global representations. Moreover, we introduce a cue digging module (CDM), which employs a hide-and-seek manner to wake up the weak local features for enhancing global representation learning. Extensive experiments show the LCTR can successfully mine integral object regions and outperform the state-of-the-art localization methods.
Acknowledgments
This work is supported by the National Science Fund for Distinguished Young Scholars (No.62025603), the National Natural Science Foundation of China (No.U1705262, No.62072386, No.62072387, No.62072389, No.62002305, No.61772443, No.61802324 and No.61702136), Guangdong Basic and Applied Basic Research Foundation (No.2019B1515120049), the Natural Science Foundation of Fujian Province of China (No.2021J01002), and the Fundamental Research Funds for the central universities (No.20720200077, No.20720200090 and No.20720200091).
References
- Beal et al. (2020) Beal, J.; Kim, E.; Tzeng, E.; Park, D. H.; Zhai, A.; and Kislyuk, D. 2020. Toward transformer-based object detection. arXiv.
- Bochkovskiy, Wang, and Liao (2020) Bochkovskiy, A.; Wang, C.-Y.; and Liao, H.-Y. M. 2020. Yolov4: Optimal speed and accuracy of object detection. arXiv.
- Carion et al. (2020) Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; and Zagoruyko, S. 2020. End-to-end object detection with transformers. In ECCV.
- Chattopadhay et al. (2018) Chattopadhay, A.; Sarkar, A.; Howlader, P.; and Balasubramanian, V. N. 2018. Grad-cam++: Generalized gradient-based visual explanations for deep convolutional networks. In WACV.
- Chen et al. (2021) Chen, Z.; Cao, L.; Shen, Y.; Lian, F.; Wu, Y.; and Ji, R. 2021. E2Net: Excitative-expansile learning for weakly supervised object Localization. In ACMMM.
- Choe and Shim (2019) Choe, J.; and Shim, H. 2019. Attention-based dropout layer for weakly supervised object localization. In CVPR.
- Dosovitskiy et al. (2020) Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. 2020. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv.
- Gao et al. (2021) Gao, W.; Wan, F.; Pan, X.; Peng, Z.; Tian, Q.; Han, Z.; Zhou, B.; and Ye, Q. 2021. TS-CAM: Token semantic coupled attention map for weakly supervised object localization. ICCV.
- Guo et al. (2021) Guo, G.; Han, J.; Wan, F.; and Zhang, D. 2021. Strengthen learning tolerance for weakly supervised object localization. In CVPR.
- He et al. (2015) He, K.; Zhang, X.; Ren, S.; and Sun, J. 2015. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In ICCV.
- Hu et al. (2021) Hu, J.; Cao, L.; Lu, Y.; Zhang, S.; Wang, Y.; Li, K.; Huang, F.; Shao, L.; and Ji, R. 2021. ISTR: End-to-end instance segmentation with transformers. arXiv.
- Hwang and Kim (2016) Hwang, S.; and Kim, H.-E. 2016. Self-transfer learning for weakly supervised lesion localization. In MICCAI.
- Jiang, Chang, and Wang (2021) Jiang, Y.; Chang, S.; and Wang, Z. 2021. Transgan: Two transformers can make one strong gan. arXiv.
- Kumar Singh and Jae Lee (2017) Kumar Singh, K.; and Jae Lee, Y. 2017. Hide-and-seek: Forcing a network to be meticulous for weakly-supervised object and action localization. In ICCV.
- Lin, Chen, and Yan (2013) Lin, M.; Chen, Q.; and Yan, S. 2013. Network in network. arXiv.
- Liu et al. (2016) Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; and Berg, A. C. 2016. Ssd: Single shot multibox detector. In ECCV.
- Loshchilov and Hutter (2017) Loshchilov, I.; and Hutter, F. 2017. Decoupled weight decay regularization. arXiv.
- Lu et al. (2020) Lu, W.; Jia, X.; Xie, W.; Shen, L.; Zhou, Y.; and Duan, J. 2020. Geometry constrained weakly supervised object localization. In ECCV.
- Pan et al. (2021) Pan, X.; Gao, Y.; Lin, Z.; Tang, F.; Dong, W.; Yuan, H.; Huang, F.; and Xu, C. 2021. Unveiling the potential of structure preserving for weakly supervised object localization. In CVPR.
- Russakovsky et al. (2015) Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. 2015. Imagenet large scale visual recognition challenge. In IJCV.
- Selvaraju et al. (2017) Selvaraju, R. R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; and Batra, D. 2017. Grad-cam: Visual explanations from deep networks via gradient-based localization. In CVPR.
- Sun et al. (2021) Sun, P.; Zhang, R.; Jiang, Y.; Kong, T.; Xu, C.; Zhan, W.; Tomizuka, M.; Li, L.; Yuan, Z.; Wang, C.; et al. 2021. Sparse r-cnn: End-to-end object detection with learnable proposals. In CVPR.
- Touvron et al. (2021) Touvron, H.; Cord, M.; Douze, M.; Massa, F.; Sablayrolles, A.; and Jégou, H. 2021. Training data-efficient image transformers & distillation through attention. In ICML.
- Vaswani et al. (2017) Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A. N.; Kaiser, Ł.; and Polosukhin, I. 2017. Attention is all you need. In NIPS.
- Wah et al. (2011) Wah, C.; Branson, S.; Welinder, P.; Perona, P.; and Belongie, S. 2011. The caltech-ucsd birds-200-2011 dataset.
- Wang et al. (2021) Wang, J.; Song, L.; Li, Z.; Sun, H.; Sun, J.; and Zheng, N. 2021. End-to-end object detection with fully convolutional network. In CVPR.
- Wei et al. (2021) Wei, J.; Wang, Q.; Li, Z.; Wang, S.; Zhou, S. K.; and Cui, S. 2021. Shallow feature matters for weakly supervised object localization. In CVPR.
- Wu et al. (2020) Wu, B.; Xu, C.; Dai, X.; Wan, A.; Zhang, P.; Yan, Z.; Tomizuka, M.; Gonzalez, J.; Keutzer, K.; and Vajda, P. 2020. Visual transformers: Token-based image representation and processing for computer vision. arXiv.
- Xue et al. (2019) Xue, H.; Liu, C.; Wan, F.; Jiao, J.; Ji, X.; and Ye, Q. 2019. Danet: Divergent activation for weakly supervised object localization. In ICCV.
- Yuan et al. (2021) Yuan, L.; Chen, Y.; Wang, T.; Yu, W.; Shi, Y.; Jiang, Z.; Tay, F. E.; Feng, J.; and Yan, S. 2021. Tokens-to-token vit: Training vision transformers from scratch on imagenet. arXiv.
- Yun et al. (2019) Yun, S.; Han, D.; Oh, S. J.; Chun, S.; Choe, J.; and Yoo, Y. 2019. Cutmix: Regularization strategy to train strong classifiers with localizable features. In ICCV.
- Zhang, Cao, and Wu (2020) Zhang, C.-L.; Cao, Y.-H.; and Wu, J. 2020. Rethinking the route towards weakly supervised object localization. In CVPR.
- Zhang et al. (2018a) Zhang, X.; Wei, Y.; Feng, J.; Yang, Y.; and Huang, T. S. 2018a. Adversarial complementary learning for weakly supervised object localization. In CVPR.
- Zhang et al. (2018b) Zhang, X.; Wei, Y.; Kang, G.; Yang, Y.; and Huang, T. 2018b. Self-produced guidance for weakly-supervised object localization. In ECCV.
- Zhang, Wei, and Yang (2020) Zhang, X.; Wei, Y.; and Yang, Y. 2020. Inter-image communication for weakly supervised localization. In ECCV.
- Zheng et al. (2021) Zheng, S.; Lu, J.; Zhao, H.; Zhu, X.; Luo, Z.; Wang, Y.; Fu, Y.; Feng, J.; Xiang, T.; Torr, P. H.; et al. 2021. Rethinking semantic segmentation from a sequence-to-sequence perspective with transformers. In CVPR.
- Zhou et al. (2016) Zhou, B.; Khosla, A.; Lapedriza, A.; Oliva, A.; and Torralba, A. 2016. Learning deep features for discriminative localization. In CVPR.
- Zhu et al. (2017) Zhu, Y.; Zhou, Y.; Ye, Q.; Qiu, Q.; and Jiao, J. 2017. Soft proposal networks for weakly supervised object localization. In ICCV.