RepParser: End-to-End Multiple Human Parsing with Representative Parts
Abstract
Existing methods of multiple human parsing usually adopt a two-stage strategy (typically top-down and bottom-up), which suffers from either strong dependence on prior detection or highly computational redundancy during post-grouping. In this work, we present an end-to-end multiple human parsing framework using representative parts, termed RepParser. Different from mainstream methods, RepParser solves the multiple human parsing in a new single-stage manner without resorting to person detection or post-grouping. To this end, RepParser decouples the parsing pipeline into instance-aware kernel generation and part-aware human parsing, which are responsible for instance separation and instance-specific part segmentation, respectively. In particular, we empower the parsing pipeline by representative parts, since they are characterized by instance-aware keypoints and can be utilized to dynamically parse each person instance. Specifically, representative parts are obtained by jointly localizing centers of instances and estimating keypoints of body part regions. After that, we dynamically predict instance-aware convolution kernels through representative parts, thus encoding person-part context into each kernel responsible for casting an image feature as an instance-specific representation. Furthermore, a multi-branch structure is adopted to divide each instance-specific representation into several part-aware representations for separate part segmentation. In this way, RepParser accordingly focuses on person instances with the guidance of representative parts and directly outputs parsing results for each person instance, thus eliminating the requirement of the prior detection or post-grouping. Extensive experiments on two challenging benchmarks demonstrate that our proposed RepParser is a simple yet effective framework and achieves very competitive performance. We show that it significantly outperforms most two-stage methods and variants of single-stage instance recognition methods.

Introduction
Multiple human parsing (MHP) aims to segment body parts for each person in an image, which is a fundamental yet challenging task in the human-centric intelligence system. Compared to many dense predictions tasks, such as object detection (Ren et al. 2017) and instance segmentation (He et al. 2017; Tian, Shen, and Chen 2020; Yanwei Li and Jia 2021; Yu et al. 2022), it is the arbitrary number of fine-grained body parts that have made MHP much more challenging. In particular, the success of MHP models depends on two key aspects: 1) whether the model can make a correct instance separation, and 2) whether this model can decide what semantic behind each image pixel. Inspired by the success of the human-centric recognition, such as pose estimation (Wang et al. 2021a, 2022; Sun et al. 2019; Wang et al. 2021b), existing methods of multiple human parsing adopt a two-stage strategy, which consists of top-down and bottom-up pipelines. In particular, the top-down pipeline (Fig. 1.(a)) starts with person detection responsible for the first aspect, then an RoI operation is adopted to crop the person from the feature maps or the original image. After that, single-person human parsing is performed for addressing the second aspect. Instead, the bottom-up pipeline (Fig. 1.(b)) firstly segments instance-agnostic body parts responsible for the second aspect, then groups them into instance-aware results for addressing the first aspect. Despite the great progress, previous state-of-the-art methods for multiple human parsing still encounter several challenges, as analyzed below:
The strong coupling of the second stage with the first stage in the two-stage framework significantly hampers high-quality multiple human parsing. Specifically, top-down methods (Ji et al. 2020; Yang et al. 2019; He et al. 2017; Yang et al. 2020) are highly dependent on person detection results while bottom-up methods rely on instance-agnostic part segmentation results. Since person bounding boxes are rectangular, they may contain irrelevant contents such as body parts belonging to other persons. In this way, the performance of human parsing will drop significantly if the person detection performance decreases dramatically. In terms of bottom-up pipeline, existing methods (Gong et al. 2018, 2019; Li et al. 2018; Zhao et al. 2018) predict redundant instance-agnostic body parts. In this way, some body parts may be removed during grouping post-processing due to their low-quality confidential scores. Besides, the processing of assembling body parts (e.g., Hungary algorithm) is often heuristic, making these methods complicated yet inefficient. Overall, the bottleneck of the two-stage frameworks lies in the first stage, as the performance of a model in the first stage decides the upper bound of the entire algorithm.
The above challenges motivate us to rethink two problems: 1) how to design a single-stage pipeline for multiple human parsing, and 2) how to equip this pipeline with the ability to establish a direct mapping from an image to various instance-specific body parts. To handle the above two problems, we present an end-to-end multiple human parsing framework using representative parts, termed RepParser. As illustrated in Fig. 1.(c), the proposed RepParser is designed in an end-to-end manner without resorting to person detection or post-grouping. To this end, RepParser decouples the parsing pipeline into instance-aware kernel generation and part-aware human parsing, which are responsible for instance separation and instance-specific part segmentation, respectively. The core idea is that we empower the parsing pipeline with representative parts, since they are characterized by instance-aware keypoints and can be utilized to dynamically parse each person instance. Specifically, representative parts are obtained by jointly localizing centers of instances and estimating keypoints of body part regions. After that, we dynamically predict instance-aware convolution kernels through representative parts, thus encoding person-part context into each kernel responsible for casting an image feature as an instance-specific representation. Furthermore, a multi-branch structure is adopted to divide each instance-specific representation into several part-aware representations for separate part segmentation. In this way, RepParser accordingly focuses on person instances with the guidance of representative parts and directly outputs parsing results for each person instance, eliminating the need for person detection or body part grouping. In summary, our work has the following contributions:
-
(1)
We propose a novel multiple human parsing pipeline termed RepParser, which eliminates the dependence of prior person detection and avoids heuristic post-grouping operations.
-
(2)
The RepParser is designed in a flexible fashion, as it dynamically encodes person-part contexts into corresponding convolution kernels. To our knowledge, this is the first single-stage method for multiple human parsing and it can inspire related research on fine-grained recognition.
-
(3)
Extensive experiments conducted on two challenging benchmarks demonstrate the effectiveness and generalizability of the proposed method. Moreover, it significantly outperforms most two-stage methods and variants of single-stage instance recognition methods.

Related Work
Multiple Human Parsing
To date, methods of multiple human parsing are based on the two-stage pipeline. Most of them can be divided into two categories: 1) bottom-up paradigm and 2) top-down paradigm. As mentioned above, the bottom-up methods (Gong et al. 2018, 2019; Li et al. 2018; Zhao et al. 2018) regard multiple human parsing as a segment-then-grouping pipeline. The series of the bottom-up methods usually generate redundant human parsing results, leading to high computational costs during post-processing. Compared with bottom-up series, top-down approaches (Ji et al. 2020; Yang et al. 2019; He et al. 2017; Yang et al. 2020; He et al. 2021; Liu et al. 2021) focus on the single-person human parsing problem as they employ person detector to solve the issue of person separation. Furthermore, recent works have developed two versions of top-down framework: the unified top-down model (Yang et al. 2019; He et al. 2017; Yang et al. 2020; Qin et al. 2019) and the separated top-down model (Ji et al. 2020; Ruan et al. 2019; Liu et al. 2019). The difference between the two versions is whether unifying the person detector into a single-person parsing model. For example, Mask-RCNN (He et al. 2017) can be regarded as the first unified top-down approach, where it adopts Faster R-CNN (Ren et al. 2017) to predict a bounding box for each person and extracts region-of-person from detector’s features for performing instance-specific part segmentation. Following this idea, Parsing R-CNN (Yang et al. 2019) and RP R-CNN (Yang et al. 2020) devote to solving the problem of the single-person human parsing and propose new variants of Mask R-CNN via contextual modeling or part re-scoring.
Different from two-stage methods, our work is devoted to designing a novel single-stage pipeline and focuses on instance-aware body part segmentation with representative parts.
Single-stage Instance-level Recognition
Traditional solutions try to build an instance-specific model for instance-level recognition. For example, Tian et al. (Tian, Shen, and Chen 2020) adopt conditional convolutions for one-stage instance segmentation, where each convolution kernel is dynamically generated from a center point of person instance. This design improves the instance segmentation performance while maintaining high efficiency. Moreover, Li et al. (Yanwei Li and Jia 2021) propose a location-aware kernel generation for panoptic scene understanding. Mao et al. (Mao et al. 2021) propose to dynamically generate a keypoint-aware estimator for multi-person pose estimation. Although these different approaches vary in tasks, they all share a characteristic: they focus on the instance-specific convolution kernel generation. However, each convolution kernel generated through existing methods encodes sparse content of an instance (i.e., object center only). Therefore, the generated kernels severely ignore the person-part context which is essential for accurate human parsing, thus leading to suboptimal results as demonstrated in the experimental results.
As a supplement to them, we extend instance-specific modeling to solve the multiple human parsing. Instead of directly deriving from single-stage frameworks that are used in other instance recognition tasks, we propose to parse multiple human instances through representative parts, and encode person-part context into each instance-aware convolution kernels as well as part representation. As a result, the proposed method significantly outperforms variants of single-stage methods applied in other instance recognition tasks.
Methodology
The pipeline of our RepParser is presented in Fig. 2. Given an input image , the goal of multiple human parsing is to localize person instances and segment body parts for each localized person. In particular, it needs to address two issues: 1) how to distinguish each person instance from other instances or background; and 2) how to perform instance-aware parsing without extra operations (i.e., RoI cropping or part grouping). To address these, we propose to parse multiple persons using representative parts. Specifically, RepParser firstly utilizes a backbone network (e.g., ResNet) to obtain an image-level feature with a size of , where indicates the number of channels and denotes the spatial size. Next, a detection branch, which is an FCOS (Tian et al. 2019) head with an object center estimator and a location regressor, is adopted to localize person instances. With the location of person centers, an instance-aware kernel generation branch is used to estimate the representative parts of each person and accordingly generate convolution kernels for each instance. With instance-aware kernels and representative parts, a part-aware parsing module, which is a multi-branch structure, is utilized to generate part-aware features for accurate human parsing. In the following, we describe the details of RepParser.
Representative Parts
As discussed before, bounding boxes and object centers are often used to represent person instances in two-stage methods and single-stage methods. Due to the rectangular shape, a bounding box has a rough global context of a person but cannot account for the semantically important local areas. Instead, the object center only accounts for small local areas, thus ignoring the interrelation among body parts and the global context of an instance. To overcome these limitations, our core idea is that each person instance in an image is represented by representative parts. It is expected that the representative parts can encode the characteristics of each person instance and only focus on the pixels of corresponding body parts. Motivated by this, we propose to dynamically construct representative parts of an instance through keypoints of body parts, as they can reflect the global context of a person (e.g., posture or shape) and semantically salient part areas. Formally, let denotes representative parts of a person, where is the keypoint of the -th part (e.g., face, left-arm, right-arm, and so on), is the number of part categories (e.g., =20 for CIHP dataset). Thus, we parse person instances conditioning on their representative parts, as they not only present characteristics of pose and shape but also reflect person-part relations.
To construct representations of representative parts, we need to localize them by object centers. As shown in Fig. 2, we firstly adopt a feature pyramid network (Lin et al. 2017) to produce multi-scale feature maps from levels 3 to 7. Following FCOS (Tian et al. 2019), we treat each location on the feature maps as a potential instance. Thus, for each location on the feature maps, we estimate the confidential score being a person center and the offsets to representative parts. Based on this, representative parts is calculated by Eq. 1.
(1) |
where are the normalized offsets from the center of a person instance to the center of the representative parts. After that, we construct an initial representation of representative parts by sampling pixel points from the image-level feature . Formally, we denote as the feature of sampled instance point and as the features of representative parts. Next, we employ these sampled representative parts for instance-aware kernel generation and part-aware human parsing.
Instance-aware Kernel Generation
To obtain high-quality instance representations for accurately human parsing, it is expected that the instance-aware convolution kernels are dynamically generated by relying on the characteristics of instances. To achieve this, we propose to generate instance-aware kernels by representative parts, since they encode potential contexts about person-part relations. Instead of directly applying initial representative parts to predict instance-aware kernels, we first re-adjust the representation of the representative parts according to the person-part relations, aiming to dynamically encode the person-part context into the corresponding kernel. Specifically, the re-adjusted representative parts are obtained through Eq. 2:
(2) |
where and are learnable parameters that are respectively responsible for relation estimation and feature re-adjustment. is the standard sigmoid function and means a concatenation operation. The is the estimated relation matrix, where each element in denotes a part being relevant to a person with a confidential score. The denotes the representation of representative parts, which is dynamically generated via person-part interaction. Note that the estimated relational score is continually updated by directly supervised training so that it becomes more accurate from time to time.
Given re-adjusted representative parts and instance representation , we generate two types of convolution kernels through Eq. 3.
(3) |
where and are linear matrices for kernel generation. is used to project an image feature to an instance-aware feature without resolution reduction, while the is responsible for predicting part masks from many part-aware features. Notably, the generated kernels are very compact, as they are designed in a convolution layer with less channels (e.g., 32 for and C for ). We project the image-level feature to the instance-aware feature for human parsing by Eq. 4.
(4) |
where is the convolution operation. Compared with top-down methods, such as Parsing-RCNN consisting of eight convolution layers with 256 channels for instance feature extraction, the generated kernels are much more lightweight.
Part-aware Human Parsing
After handling the issue of person separation by representative parts, we would like to predict an accurate mask for each part. Instead of predicting masks from the instance-aware feature, we first build multiple branches for separate parsing. In each branch, we construct a part-aware representation that can be used to only fire on the pixels of the corresponding part. Specifically, we first divide the instance-aware feature into groups, each of which is responsible for part-specific segmentation. For each group, we construct a geometry map that records relative distances from all pixels to the corresponding representative part, suggesting the salient area of the corresponding part. Formally, we denote all geometry maps as and compute the part-aware representation by Eq. 5.
(5) |
where is the learnable transformation matrix. Next, we utilized dynamically generated kernel to predict mask for each part, which is formalized by Eq. 6.
(6) |
where is the predicted parsing maps and is the standard softmax function. Although it is possible to predict masks by instance-aware feature, we empirically find that using part-aware representation performs better.
Method | Backbone | Epoch | RoI-free | Grouping-free | mIoU | Time (ms) | |||
Bottom-up | |||||||||
PGN (Gong et al. 2018) | - | - | ✓ | 25.3 | 35.5 | 17.6 | 26.9 | 497 | |
MH-Parser (Li et al. 2018) | ResNet-101 | - | ✓ | - | 36.0 | 17.9 | 26.9 | 1486 | |
NAN (Zhao et al. 2018) | - | 80 | ✓ | - | 41.7 | 25.1 | 32.2 | 1037 | |
Top-down | |||||||||
Mask RCNN (He et al. 2017) | ResNet-50 | - | ✓ | - | 33.9 | 14.9 | 25.1 | 243 () | |
Parsing RCNN (Yang et al. 2019) | ResNet-50 | 25 | ✓ | 34.0 | 36.7 | 19.9 | 32.4 | 270 () | |
Parsing RCNN (Yang et al. 2019) | ResNet-50 | 75 | ✓ | 36.1 | 40.5 | 27.4 | 38.3 | 270 | |
SemaTree (Ji et al. 2020) | ResNet-101 | 200 | ✓ | - | 42.5 | 34.4 | 43.5 | 3234 | |
M-CE2P (Ruan et al. 2019) | ResNet-101 | 150 | ✓ | 41.1 | 42.7 | 34.5 | 43.8 | 1107 | |
RP-RCNN (Yang et al. 2020) | ResNet-50 | 75 | ✓ | 37.3 | 45.2 | 40.5 | 39.2 | 394 () | |
Single-stage | |||||||||
DETR∗ (Carion et al. 2020) | ResNet-50 | 25 | ✓ | ✓ | 30.5 | 33.7 | 12.1 | 25.1 | 218 () () |
Deformable DETR∗ (Zhu et al. 2021) | ResNet-50 | 25 | ✓ | ✓ | 33.4 | 34.8 | 14.2 | 29.4 | 241 () () |
condInst∗ (Tian, Shen, and Chen 2020) | ResNet-50 | 25 | ✓ | ✓ | 26.2 | 36.5 | 18.7 | 30.1 | 164 () () |
RepParser (Ours) | ResNet-50 | 25 | ✓ | ✓ | 35.9 | 39.4 | 25.5 | 36.8 | 193 () () |
RepParser (Ours) | ResNet-50 | 75 | ✓ | ✓ | 38.3 | 42.3 | 33.7 | 43.4 | 193 () () |
RepParser (Ours) | ResNet-101 | 75 | ✓ | ✓ | 39.7 | 43.0 | 35.6 | 45.2 | 208 () () |
RepParser (Ours) | Swin-S | 75 | ✓ | ✓ | 41.1 | 45.6 | 42.4 | 55.0 | 220 () () |
Method | Backbone | Epoch | RoI-free | Grouping-free | mIoU | Time (ms) | |||
Bottom-up | |||||||||
PGN (Gong et al. 2018) | ResNet-101 | 80 | ✓ | 55.8 | 39.0 | 34.0 | 61.0 | 497 | |
Graphonomy (Gong et al. 2019) | Xception | 100 | ✓ | 58.6 | - | - | - | - | |
Top-down | |||||||||
Mask RCNN (He et al. 2017) | ResNet-50 | 25 | ✓ | 47.7 | 45.2 | 42.0 | 44.0 | 243 () | |
Mask RCNN (He et al. 2017) | ResNet-50 | 75 | ✓ | 51.1 | 47.4 | 49.4 | 49.5 | 243 | |
Parsing RCNN (Yang et al. 2019) | ResNet-50 | 25 | ✓ | 52.8 | 51.2 | 57.2 | 55.4 | 270 () | |
Parsing RCNN (Yang et al. 2019) | ResNet-50 | 75 | ✓ | 56.3 | 53.9 | 63.7 | 60.1 | 270 | |
Unified (Qin et al. 2019) | ResNet-101 | 37 | ✓ | 55.2 | 48.0 | 51.0 | - | - | |
M-CE2P (Ruan et al. 2019) | ResNet-101 | 200 | ✓ | 59.5 | - | - | - | 1107 | |
BraidNet (Liu et al. 2019) | ResNet-101 | 150 | ✓ | 60.6 | - | - | - | - | |
SemaTree (Ji et al. 2020) | ResNet-101 | 200 | ✓ | 60.9 | - | - | - | 3234 | |
RP-RCNN (Yang et al. 2020) | ResNet-50 | 75 | ✓ | 58.2 | 58.3 | 71.6 | 62.2 | 394 () | |
Single-stage | |||||||||
DETR∗ (Carion et al. 2020) | ResNet-50 | 25 | ✓ | ✓ | 48.3 | 43.8 | 39.3 | 44.2 | 218 () () |
Deformable DETR∗ (Zhu et al. 2021) | ResNet-50 | 25 | ✓ | ✓ | 46.4 | 44.0 | 38.5 | 44.0 | 241 () () |
condInst∗ (Tian, Shen, and Chen 2020) | ResNet-50 | 25 | ✓ | ✓ | 49.7 | 47.1 | 46.9 | 48.1 | 164 () () |
RepParser (Ours) | ResNet-50 | 25 | ✓ | ✓ | 52.9 | 51.9 | 57.5 | 55.7 | 193 () () |
RepParser (Ours) | ResNet-50 | 75 | ✓ | ✓ | 56.3 | 53.1 | 61.5 | 59.3 | 193 () () |
RepParser (Ours) | ResNet-101 | 75 | ✓ | ✓ | 57.9 | 54.4 | 64.9 | 61.5 | 208 () () |
RepParser (Ours) | Swin-S | 75 | ✓ | ✓ | 61.7 | 57.2 | 70.4 | 65.8 | 220 () () |
Experiments
Experimental Setup
Datasets:
Our experiments are conducted on two challenging multiple human parsing datasets: MHP-v2 (Zhao et al. 2018) and CIHP (Gong et al. 2018). MHP-v2 is a commonly used dataset for instance-level human parsing. It is split into 15k/5k/5k images for train/val/test. Each image contains an average of three people with 58 body part categories. In addition, the CIHP dataset is currently the largest multiple human parsing dataset, which covers 19 part categories and involves many crowded scenes. It is split into 28k/5k/5k images for train/val/test.
Metrics:
For evaluation, we use many standard metrics to measure the performance of all parsing models, including the Average Precision based on part () and Percentage of Correctly parsed semantic Parts (). We report and . is the average of at different IoU thresholds ranging from 0.1 to 0.9. In particular, means that is calculated at an IoU threshold of 0.5. In terms of instance-agnostic parsing, we report mean Intersection-Over-Union (mIOU) for model evaluation.
Implementation details:
Our RepParser is implemented based on MMDetection (Chen et al. 2019) on eight NVIDIA Tesla V100 GPUs. Following FCOS (Tian et al. 2019), FPN (Lin et al. 2017) is used as the feature extraction network. The weights of all backbones are pre-trained on ImageNet, while the remaining weights are randomly initialized. We train models using SGD/AdamW for convolution-/transformer-based models, respectively. A mini-batch size of 16 is used. Other details are identical to FCOS (Tian et al. 2019).
Main Results
In this section, we compare proposed RepParser with state-of-the-art multiple human parsing methods and report evaluation results summarized from two datasets. In addition to existing two-stage methods that are either based on top-down paradigm or bottom-up paradigm, we also compare RepParser with other representative single-stage methods that are applied for other instance recognition tasks, including DETR (Carion et al. 2020), Deformable DETR (Zhu et al. 2021) and condInst (Tian, Shen, and Chen 2020). Moreover, we re-implement these single-stage methods and train them under the same settings for a fair comparison, since these methods are not evaluated for multiple human parsing in original papers. In addition to standard metrics, we also measure inference time per image for each method on the same hardware if possible. Furthermore, we also report the relative proportion of time costs reduction for investigation of the efficiency of each single-stage model.



MHP-v2: As shown in Table 1, we evaluate RepParser on MHP-v2 validation set and compare it with state-of-the-art multiple human parsing methods. From the results, we find that RepParser achieves very competitive parsing results, which are comparable to or higher than that of state-of-the-art methods. Compared with the previous bottom-up methods, RepParser with ResNet-50 backbone has better performances (38.3% mIoU vs 25.3% mIoU, 42.3% vs 41.7% ) with much fewer time costs (193ms vs 1037ms).
As for the comparison with top-down methods, RepParser achieves competitive parsing performance with fewer time costs. For example, RepParser with ResNet-50 significantly outperforms the Mask-RCNN under the same settings, e.g., 35.9% mIoU vs 33.4% mIoU, 39.4% vs 33.9% and 36.8% vs 26.8% . On two stronger methods using M-CE2P and RP-RCNN, RepParser shows comparable parsing results under the same setting, but has much lower time costs (e.g., 208ms vs 1107ms, 193ms vs 394ms). It is worth noting that many non-unified top-down methods such as SemaTree (Ji et al. 2020) adopt an isolated object detector to detect persons and then crop the RoIs on the original images, thus leading to high inference costs (i.e., from the input image to the parsing results). As illustrated in Fig. 3, the inference time of top-down methods, such as Mask RCNN, Parsing RCNN, and PR-RCNN, dramatically increases as the number of persons linearly increases. However, the RepParser keeps almost constant inference time, since it eliminates the prior detection and each generated kernel is very compact (i.e., ). This suggests that the RepParser could be applied to many complex real-world scenes, such as crowded scene, while keeping stable yet high efficiency. Moreover, another major merit of the RepParser is that it does not rely on bounding boxes. As illustrated in Fig. 4(left), top-down methods only perform human parsing inside the predicted bounding boxes. As a result, some body parts cannot be parsed if the detector yield inaccurate bounding boxes. However, RepParser can well handle the body parts even outside the box (see Fig. 4(right)).
In terms of the comparison with single-stage parsing methods, the RepParser achieves higher parsing performance than that of other single-stage methods, while maintaining competitive efficiency. This indicates that directly deriving single-stage methods from other instance tasks leads to suboptimal results since they severely ignore the instance-part contexts that are essential for accurate multiple human parsing.
CIHP: Similar to experiments conducted on MHP-v2 dataset, we compare RepParser with state-of-the-art methods on CIHP validation set. Corresponding results are listed in Table 2. In line with findings from Tab. 1, RepParser also achieves competitive performance on CIHP validation set. For example, RepParser with ResNet-50 backbone has a better performance than that of bottom-up models and significantly outperforms other single-stage models under the same setting. Moreover, it also performs comparable to the best top-down competitor RP-RCNN and outperforms other top-down competitors by a margin, but requires less computational costs. Some qualitative results are shown in Fig. 5, which clearly demonstrates the effectiveness of our proposed method.
baseline | KG | PF | mIoU | ||
---|---|---|---|---|---|
✓ | 49.7 | 47.1 | 48.1 | ||
✓ | ✓ | 52.5 | 50.7 | 54.0 | |
✓ | ✓ | ✓ | 52.9 | 51.9 | 55.7 |
Ablation Experiments
width | mIoU | |||
---|---|---|---|---|
8 | 52.5 | 51.3 | 56.4 | 54.4 |
16 | 53.0 | 51.6 | 57.3 | 55.4 |
32 | 52.9 | 51.9 | 57.5 | 55.7 |
64 | 53.4 | 51.5 | 57.1 | 55.3 |
depth | mIoU | |||
2 | 52.9 | 51.9 | 57.5 | 55.7 |
3 | 53.1 | 51.7 | 57.1 | 55.4 |
4 | 52.7 | 51.3 | 56.4 | 54.9 |
The effect of representative parts. As discussed before, the representative parts separately contribute to instance-aware kernel generation and part-aware feature generation. Thus, we choose the condInst (Tian, Shen, and Chen 2020) with ResNet-50 backbone as the baseline and gradually incorporate representative parts into the pipeline. The experimental results are summarized in Table 3. From the results, we have the following observations: First, compared with baseline, applying representative parts to predict convolution kernels leads to a significant improvement, achieving 2.8% of mIoU, 3.6% of and 5.9% of higher than that of baseline. This indicates that encoding person-part context into convolution kernel is particularly important for instance-aware feature generation. Second, constructing part-aware representation via representative parts brings a stable improvement, e.g., improving the score of from 50.7% to 51.9%. This suggests that focusing on salient areas derived from representative parts is particularly beneficial for accurate human parsing.
The effect of the kernel scale. In this section, we investigate the effect of the kernel scale. Here, we consider two factors: the width of each generated convolution kernel and the depth of generated convolution kernels. Our baseline consists of two 1x1 convolutions with 32 channels and performs convolution on the 1/8 down-sampling ratio feature maps. We conduct experiments by adjusting the number of channels or varying the number of convolution layers. As reported in Table 4, the performance improves as the width increases, but it seems to be saturated when the width is set as 32. However, increasing depth has a negligible effect on parsing performance. Thus, one can conclude that simply enlarging model capacity has reached the performance bottleneck.
Qualitative results. As shown in Fig. 5, RepParser can produce good parsing results which are comparable to those of two-stage methods. Furthermore, two-stage methods fail to handle identical parts appearing in an intersection of two bounding boxes (see column 4). In contrast, this has a minor effect on RepParser, as it does not rely on bounding boxes. On the other hand, estimated representative parts tend to be located on semantic parts of persons, thus benefiting the instance-aware human parsing. For more details, we refer the reader to supplementary materials.
Conclusion
In this paper, we propose a new single-stage multiple human parsing method termed RepParser, aiming at breaking the limitation of the two-stage pipeline. To achieve this goal, we utilize representative parts to generate instance-aware kernels as well as part-aware representations, thus facilitating instance-aware human parsing. Extensive experiments on two benchmarks prove the effectiveness of our method.
References
- Carion et al. (2020) Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; and Zagoruyko, S. 2020. End-to-End Object Detection with Transformers. In ECCV, 213–229.
- Chen et al. (2019) Chen, K.; Wang, J.; Pang, J.; Cao, Y.; Xiong, Y.; Li, X.; Sun, S.; Feng, W.; Liu, Z.; Xu, J.; et al. 2019. MMDetection: Open mmlab detection toolbox and benchmark. arXiv preprint arXiv:1906.07155.
- Gong et al. (2019) Gong, K.; Gao, Y.; Liang, X.; Shen, X.; Wang, M.; and Lin, L. 2019. Graphonomy: Universal Human Parsing via Graph Transfer Learning. In CVPR.
- Gong et al. (2018) Gong, K.; Liang, X.; Li, Y.; Chen, Y.; Yang, M.; and Lin, L. 2018. Instance-Level Human Parsing via Part Grouping Network. In ECCV.
- He et al. (2021) He, H.; Zhang, J.; Thuraisingham, B.; and Tao, D. 2021. Progressive One-shot Human Parsing. In Proceedings of the AAAI Conference on Artificial Intelligence.
- He et al. (2017) He, K.; Gkioxari, G.; Dollár, P.; and Girshick, R. B. 2017. Mask R-CNN. In ICCV.
- Ji et al. (2020) Ji, R.; Du, D.; Zhang, L.; Wen, L.; Wu, Y.; Zhao, C.; Huang, F.; and Lyu, S. 2020. Learning Semantic Neural Tree for Human Parsing. In ECCV.
- Li et al. (2018) Li, J.; Zhao, J.; Chen, Y.; Roy, S.; Yan, S.; Feng, J.; and Sim, T. 2018. Multi-Human Parsing Machines. In ACM MM, 45–53.
- Lin et al. (2017) Lin, T.-Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; and Belongie, S. 2017. Feature pyramid networks for object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition, 2117–2125.
- Liu et al. (2019) Liu, X.; Zhang, M.; Liu, W.; Song, J.; and Mei, T. 2019. BraidNet: Braiding Semantics and Details for Accurate Human Parsing. In ACM MM, 338–346.
- Liu et al. (2021) Liu, Y.; Zhang, S.; Yang, J.; and Yuen, P. 2021. Hierarchical Information Passing Based Noise-Tolerant Hybrid Learning for Semi-Supervised Human Parsing. In Proceedings of the AAAI Conference on Artificial Intelligence, 2207–2215.
- Mao et al. (2021) Mao, W.; Tian, Z.; Wang, X.; and Shen, C. 2021. FCPose: Fully Convolutional Multi-Person Pose Estimation With Dynamic Instance-Aware Convolutions. In CVPR.
- Qin et al. (2019) Qin, H.; Hong, W.; Hung, W.; Tsai, Y.; and Yang, M. 2019. A Top-Down Unified Framework for Instance-level Human Parsing. In BMVC.
- Ren et al. (2017) Ren, S.; He, K.; Girshick, R. B.; and Sun, J. 2017. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell.
- Ruan et al. (2019) Ruan, T.; Liu, T.; Huang, Z.; Wei, Y.; Wei, S.; and Zhao, Y. 2019. Devil in the Details: Towards Accurate Single and Multiple Human Parsing. In AAAI, 4814–4821.
- Sun et al. (2019) Sun, K.; Xiao, B.; Liu, D.; and Wang, J. 2019. Deep High-Resolution Representation Learning for Human Pose Estimation. In CVPR, 5693–5703.
- Tian, Shen, and Chen (2020) Tian, Z.; Shen, C.; and Chen, H. 2020. Conditional Convolutions for Instance Segmentation. In ECCV.
- Tian et al. (2019) Tian, Z.; Shen, C.; Chen, H.; and He, T. 2019. FCOS: Fully Convolutional One-Stage Object Detection. In ICCV.
- Wang et al. (2021a) Wang, X.; Gao, L.; Dai, Y.; Zhou, Y.; and Song, J. 2021a. Semantic-aware Transfer with Instance-adaptive Parsing for Crowded Scenes Pose Estimation. In ACM MM, 686–694.
- Wang et al. (2021b) Wang, X.; Gao, L.; Song, J.; Guo, Y.; and Shen, H. T. 2021b. AMANet: Adaptive Multi-Path Aggregation for Learning Human 2D-3D Correspondences. IEEE Transactions on Multimedia.
- Wang et al. (2022) Wang, X.; Gao, L.; Zhou, Y.; Song, J.; and Wang, M. 2022. KTN: Knowledge Transfer Network for Learning Multi-person 2D-3D Correspondences. IEEE Transactions on Circuits and Systems for Video Technology.
- Yang et al. (2020) Yang, L.; Song, Q.; Wang, Z.; Hu, M.; Liu, C.; Xin, X.; Jia, W.; and Xu, S. 2020. Renovating Parsing R-CNN for Accurate Multiple Human Parsing. In ECCV.
- Yang et al. (2019) Yang, L.; Song, Q.; Wang, Z.; and Jiang, M. 2019. Parsing R-CNN for Instance-Level Human Analysis. In CVPR.
- Yanwei Li and Jia (2021) Yanwei Li, X. Q. L. W. Z. L. J. S., Hengshuang Zhao; and Jia, J. 2021. Fully Convolutional Networks for Panoptic Segmentation. In CVPR.
- Yu et al. (2022) Yu, X.; Shi, D.; Wei, X.; Ren, Y.; Ye, T.; and Tan, W. 2022. SOIT: Segmenting Objects with Instance-Aware Transformers. In Proceedings of the AAAI Conference on Artificial Intelligence.
- Zhao et al. (2018) Zhao, J.; Li, J.; Cheng, Y.; Sim, T.; Yan, S.; and Feng, J. 2018. Understanding Humans in Crowded Scenes: Deep Nested Adversarial Learning and A New Benchmark for Multi-Human Parsing. In ACM MM.
- Zhu et al. (2021) Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; and Dai, J. 2021. Deformable DETR: Deformable Transformers for End-to-End Object Detection. In ICLR.
Appendix A: Implementation Details.
In this section, we provide more details about implementations, including the effect of the feature resolution and the details of training schedule.
The effect of the feature resolution
Many works (Sun et al. 2019; Wang et al. 2021b) have demonstrated that higher resolution representation brings better performance for dense prediction tasks. Inspired by this, we investigate which level of the image feature is beneficial for human parsing. Hence, we separately apply generated instance-aware kernels on three different feature maps, which separately are 1/16, 1/8, and 1/4 smaller than the size of the image. Table 5 indicates that the performance drops dramatically if the resolution of the input feature map is downsampled to 1/16 of the input image. We conjecture the possible reason behind this is that the human parsing task requires pixel-level understanding and high-resolution feature maps preserve more visual contents. However, larger resolution leads to a high computation burden. Besides, generating parsing results from the image feature at the 1/4 scale of the image size brings minor gains, when compared it with the counterpart at 1/8 scale of the image size. Thus, we choose 1/8 as the default setting for a better trade-off between accuracy and speed.
Ratio | mIoU | |||
---|---|---|---|---|
1/16 | 51.9 | 49.2 | 51.8 | 52.4 |
1/8 | 52.9 | 51.9 | 57.5 | 55.7 |
1/4 | 53.5 | 51.6 | 57.3 | 55.7 |
Details of training schedule
In general, a good initialization of models leads to better performance. Thus we explore the impact of initialization methods on multiple human parsing. As shown in Tab 6, human parsing methods pre-trained on the COCO keypoint dataset can improve by 1%. It indicates that a good initialization will lead to better parsing results.
Initialization | mIoU | |||
---|---|---|---|---|
ImageNet | 52.9 | 51.9 | 57.5 | 55.7 |
COCO | 54.1 | 52.8 | 59.8 | 57.4 |
Appendix B: More Qualitative Results
In this section, we provide additional qualitative results of our RepParser on CIHP val set, including estimated representative parts and failure cases.


More Qualitative results.
More qualitative results of RepParser are shown in Fig. 6. RepParser can well handle many challenging scenes with occlusions, scale variations, etc. Besides, the estimated centers of representative parts reflect salient parts of each person instance.
More Failure cases.
Some failure cases of our proposed RepParser are shown in Fig. 7. From the visualization results, RepParser can not parse some regions with extreme cases, such as the confusing part regions and dramatic pose variations. We can observe that RepParser fails to distinguish some part regions from other person instances (seen in column 1). Moreover, RepParser can not parse some regions due to the dramatic pose change (seen in column 2). Generalizing from these cases, we can find that each failure case of the part region is significantly interfered by other part region or other person instances. To precisely parse these cases, a method must carefully consider the rich details of the person instance and generate a more discriminative feature representation. We hope that our findings can inspire more research on multiple human parsing.