Abstract
Multiple human parsing aims to segment various human parts and associate each part with the corresponding instance simultaneously. This is a very challenging task due to the diverse human appearance, semantic ambiguity of different body parts, and complex background. Through analysis of multiple human parsing task, we observe that human-centric global perception and accurate instance-level parsing scoring are crucial for obtaining high-quality results. But the most state-of-the-art methods have not paid enough attention to these issues. To reverse this phenomenon, we present Renovating Parsing R-CNN (RP R-CNN), which introduces a global semantic enhanced feature pyramid network and a parsing re-scoring network into the existing high-performance pipeline. The proposed RP R-CNN adopts global semantic representation to enhance multi-scale features for generating human parsing maps, and regresses a confidence score to represent its quality. Extensive experiments show that RP R-CNN performs favorably against state-of-the-art methods on CIHP and MHP-v2 datasets. Code and models are available at https://github.com/soeaver/RP-R-CNN.
You have full access to this open access chapter, Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
Multiple human parsing [8, 21, 40] is a fundamental task in multimedia and computer vision, which aims to segment various human parts and associate each part with the corresponding instance. It plays a crucial role in applications in human-centric analysis and potential down-stream applications, such as person re-identification [22, 27], action recognition [5], human-object interaction [1, 7, 29], and virtual reality [14].
Due to the successful development of convolutional neural networks [13, 33], great progress has been made in multiple human parsing. Current state-of-the-art methods can be categorized into bottom-up, one-stage top-down, and two-stage top-down methods. The bottom-up methods [8, 9, 11] regard multiple human parsing as a fine-grained semantic segmentation task, which predicts the category of each pixel and grouping them into corresponding human instance. This series of methods will have better performance in semantic segmentation metrics, but poor in instance parsing metrics, especially easy to confuse adjacent human instances. Unlike bottom-up methods, the one-stage top-down [30, 40] and two-stage top-down methods [16, 24, 32] locate each instance in the image plane, and then segment each human parts independently. The difference between one-stage and two-stage is whether the detector is trained together with the sub-network used to segment the human part in an end-to-end manner. Compared with the bottom-up, the top-down methods are very flexible, which can easily introduce enhancement modules or train with other human analysis tasks (such as pose estimation [37], dense pose estimation [10, 40] or clothing parsing [35]) jointly. So it has become the mainstream research direction of multiple human parsing. But the human parts segmentation of each instance is independent and cannot make full use of context information, so the segmentation of some small scale human parts and human contours still needs to be improved. In addition, it is worth noting that neither bottom-up or top-down methods have a good way to evaluate the quality of predicted instance parsing maps. Resulting in many low-quality results that cannot be filtered.
In this paper, we are devoted to solving the problem of missing global semantic information in top-down methods, and evaluating the quality of predicted instance parsing maps accurately. Therefore, we propose Renovating Parsing R-CNN (RP R-CNN), which introduces a global semantic enhanced feature pyramid network and a parsing re-scoring network to renovate the pipeline of top-down multiple human parsing. The global semantic enhanced feature pyramid network (GSE-FPN) is built on the widely used FPN [23]. We up-sample the multi-scale features generated by FPN to the same scale and fuse them. Using the global human parts segmentation to supervise and generate the global semantic feature, then fusing the global semantic feature with FPN features on the corresponding scales. GSE-FPN encourages the semantic supervision signal to directly propagate to the feature pyramid, so as to strengthen global information of learned multi-scale features. Global semantic enhanced features are passed to the Parsing branch through the RoIAlign [12] operation, ensuring that each independent human instance can still perceive the global semantic information, thereby improving the parsing performance of small targets, human contours, and easily confused categories. On this basis, the parsing re-scoring network (PRSN) is used to sense the quality of instance parsing maps and gives accurate scores. The score of instance parsing map is related to filtering low-quality results and sorting of instances, which is very important in the measurement of method and practical application. However, almost all the top-down methods use the score of detected bounding-box to represents the quality of instance parsing map [32, 40]. This will inevitably bring great deviation, because the score of bounding-box can only indicate whether the instance is human or not, while the score of the instance parsing map needs to express the segmentation accuracy of each human part, and there is no direct correlation between them. The proposed PRSN is a very lightweight network, taking the feature map and heat map of each human instance as input, using MSE loss to regress the mean intersection over union (mIoU) between the prediction and ground-truth. During inference, we use the arithmetic square root of predicted mIoU score multiply box classification score as human parsing final score.
Extensive experiments are conducted on two challenging benchmarks, CIHP [8] and MHP-v2 [45], demonstrating that our proposal RP R-CNN significantly outperforms the state-of-the-art for both bottom-up and top-down methods. As shown is Fig. 1, RP R-CNN is more accurate in segmenting small parts and human edges, and the predicted parsing scores can better reflect the quality of instance parsing maps. The main contributions of this work are summarized as follows:
-
A novel RP R-CNN is proposed to solve the issue of missing global semantic information and inaccurate scoring of instance parsing maps in top-down multiple human parsing.
-
We introduce an effective method to improve the multiple human parsing results by fusing global and instance-level human parts segmentation.
-
The proposed RP R-CNN achieves state-of-the-art on two challenging benchmarks. On CIHP val set, RP R-CNN yields 2.0 points mIoU and 7.0 points AP\(^\text {p}_\text {50}\) improvements compared with Parsing R-CNN [40]. On MHP-v2 val set, RP R-CNN outperforms Parsing R-CNN by 13.9 points AP\(^\text {p}_\text {50}\) and outperforms CE2P [32] by 6.0 points AP\(^\text {p}_\text {50}\), respectively.
Our code and models of RP R-CNN are publicly available.
2 Related Work
Multi-scale Feature Representations. Multi-scale feature is widely used in computer vision tasks [19, 23, 25, 38]. Long et al. [25] combine coarse, high layer information with fine, low layer information to generate fine features with high resolution, which greatly promotes the development of semantic segmentation. Lin et al. [25] present the feature pyramid network (FPN), and adopt it in object detection, greatly improve the performance of the small object. FPN is a feature pyramid with high-level semantics throughout, through top-down pathway and lateral connections. With the success of FPN, some researches introduce it into other tasks. Panoptic feature pyramid network (PFPN) [19] is proposed by Kirillov et al. and applied to panoptic segmentation. PFPN up-samples the feature pyramids and fuse them to the same spatial resolution, then a semantic segmentation branch is attached to generate high-resolution semantic features. However, the computation cost of PFPN is too large, and it only has a single scale semantic feature. Our GSE-FPN solves the above problems well, which adopts a lightweight up-sampling method, and use the global semantic feature to enhance the multi-scale feature.
Instance Scoring. Scoring the predicted instance is a challenging question. The R-CNN series [6, 31] of object detection approaches use the object classification score as the confidence of detection results. Recent studies [17, 34, 46] believe that it cannot accurately reflect the consistency between the predicted bounding-box and the ground-truth. Jiang et al. [17] present the IoU-Net, which adopts a IoU-prediction branch to predict the IoU between the predicted bounding box and the corresponding ground truth. Tan et al. [34] propose the Learning-to-Rank (LTR) model to produce a ranking score, which is based on IoU to indicate the ranks of candidates during the NMS step. Huang et al. [15] consider the difference between classification score and mask quality is greater in instance segmentation. They proposed Mask Scoring R-CNN, which uses a MaskIoU head to predict the quality of mask result. Different from these studies, this work analyzes and solves the inaccurate scoring of instance parsing maps for the first time. The proposed PRSN is concise yet effective, and reducing the gap between score and instance parsing quality.
Multiple Human Parsing. Before the popularity of convolutional neural network, some methods [26, 39, 41] using hand-crafted visual features and low-level image decompositions have achieved considerable results on single human parsing. However, limited by the representation ability of features, these traditional methods can not be well extended to multiple human parsing. With the successful development of convolutional neural networks [13, 20, 33] and open source of large-scale multiple human parsing datasets [8, 45], some recent researches [8, 32, 40] have achieved remarkable results in instance-level multiple human parsing. Gong et al. [8] present the part grouping network (PGN), which is a typical bottom-up method for instance-level multiple human parsing. PGN reformulates multiple human parsing as semantic part segmentation task and instance-aware edge detection task, the former is used to assign each pixel as human part and the latter is used to group semantic part into different human instances. Ruan et al. [32] rethink and analyze the problems of feature resolution, global context information and edge details in human parsing task, and propose Context Embedding with Edge Perceiving (CE2P) framework for single human parsing. CE2P is a very successful two-stage top-down method, and wins the 1st places on three tracks in the 2018 2nd LIP Challenge. Parsing R-CNN [40] is proposed by Yang et al., which is a one-stage top-down method for multiple human parsing. Based on the in-depth analysis of human appearance characteristics, Parsing R-CNN has made an effective extension on region-based approaches [6, 12, 23, 31] and significantly improved the performance of human parsing. Our work is based on the Parsing R-CNN framework, firstly introducing the global semantic information and reducing the gap between score and instance parsing quality for the top-down methods.
3 Renovating Parsing R-CNN
Our goal is to solve the issue of missing global semantic information and inaccurate scoring of instance parsing map in top-down multiple human parsing pipeline. In this section, we will introduce the motivation, architecture, and components of RP R-CNN in detail.
3.1 Motivation
In the top-down multiple human parsing pipeline, the network outputs three results: bounding-box, instance parsing map and parsing score. The importance of three outputs to the network performance is different. We take Parsing R-CNN [40] as baseline, and make an upper bound analysis of the three outputs. As shown in Table 1, we replace the predicted bounding-box with ground-truth, the multiple human parsing increases by 2.2 points mIoU [25], but AP\(^\text {p}_\text {50}\) and AP\(^\text {p}_\text {vol}\) [45] decrease. But when we replace the corresponding network output with the ground-truth of parsing map and mIoU, all the evaluation metrics have significant improvements. In particular, after the adoption of ground-truth parsing map, each evaluation metric has increased by about 30 points. These experimental results show that the accuracy of bounding-box has no significant impact on the multiple human parsing performance.However, the predicted parsing map and score still have a lot of room for improvement and not been paid enough attention by current studies. This is the motivation for our work.
3.2 Architecture
As illustrated in Fig. 2, the proposed RP R-CNN involves four components: Backbone, GSE-FPN, Detector (RPN and BBox branch), and Parsing branch with PRSN. The settings of Backbone and Detector are the same as Parsing R-CNN [40]. The GSE-FPN is attached to the Backbone to generate multi-scale features with global semantic information. The Parsing branch consists of GCE module, parsing map output and Parsing Re-Scoring Network.
3.3 Global Semantic Enhanced Feature Pyramid Network
RoIAlign [12] aims to obtain the features of a specific region on the feature map, so that each instance can be processed separately. However, this makes the instance unable to directly perceive the global (context) information in branch. Global representation is crucial for human parsing, because we not only need to distinguish human body and background, but also give each pixel corresponding category through understanding the pose and recognizing the clothes the person wears [9]. Therefore, the information about the environment and objects around the human body is helpful for network learning. Some methods [43] perceive more valuable information by changing the area selected by RoIPool/RoIAlign. Different from these, we hope that by explicitly enhancing the global semantic representation of multi-scale features before RoIAlign. As a concrete example, the proposed GSE-FPN are illustrated in Fig. 3, we adopt group normalization [36] and ReLU activation [28] after each convolutional layer.
High-Resolution Feature. For semantic segmentation, high-resolution feature is necessary for generating high-quality results. Dilated convolution is an effective operation, which is adopted by many state-of-the-art semantic segmentation methods [3, 42, 44]. But dilated convolution substantially increases computation cost, and limits the use of multi-scale features. To keep the efficiency of network, and generate high-resolution features, we extend the multi-scale outputs of FPN [23]. Specifically, we up-sample the FPN generated multi-scale features to the scale of ‘P2’ level by bilinear interpolation, which is 1/4 resolution of the original image. Each feature map is followed by a \(1\times 1\) 256-d convolutional layer for aligning to the same semantic space, then these feature maps are fused together to generate high-resolution features.
Global Semantic Feature. As shown in Fig. 3, we stack four \(3\times 3\) 256-d convolutional layers after the high-resolution features to generate global semantic feature. Such a design is simple enough, but also can improve the representation ability of the network. In fact, we have tried some popular enhancement modules for semantic segmentation tasks, such as PPM [44] and ASPP [3, 4], but experiments show that these modules are not helpful to improve human parsing performance. A \(1\times 1\) C-dimension (C is the category number) convolutional layer is attached to the global semantic feature to predict human part segmentation.
Multi-scale Features Fusion. Through the above structure, we can get high-resolution global semantic feature. It is well known that semantic representation can bring performance gains to bounding-box classification and regression [2, 12]. Therefore, we down-sample the global semantic feature to the scales of P3–P6, and use element-wise sum to fuse them with the same scale FPN features. The generated new features are called global semantic enhanced multi-scale features, and denoted as S2–S6. We follow the Proposals Separation Sampling [40] strategy that the S2–S6 level features are adopted for extracting region features for BBox branch and only S2 level is used for Parsing branch.
3.4 Parsing Re-Scoring Network
Parsing Re-Scoring Network (PRSN) aims to predict accurate mIoU score for each instance parsing map, and can be flexibly integrated into the Parsing branch.
Concise and Lightweight Design. PRSN follows the concise and lightweight design, which will not bring too much computation cost to model training and inference. PRSN receives two inputs, one is the \(N\times 512\times 32\times 32\) dimension parsing feature map, the other is the \(N\times C \times 128 \times 128\) dimension segmentation probability map (N is the number of RoIs, C is the category number). A max pooling layer with stride = 4 and kernel = 4 is adopted to make the probability map has the same spatial scale with parsing feature map. The down-sampled probability map and parsing feature map are concatenated together, then followed by two \(3\times 3\) 128-d convolutional layers. A final global average pooling layer, two 256-d fully connected layers, and MSE loss to regress the mIoU between the predicted instance parsing map and ground-truth.
IoU-Aware Ground-Truth. We define the mIoU between the predicted instance parsing map and matched ground-truth as regression target for PRSN. The common Parsing branch can output the segmentation probability map of each human instance, and calculate the loss with the segmentation ground-truth through a cross entropy function. Therefore, the mIoU between them can be calculated directly in the existing framework. It is worth noting that since the instance parsing ground-truth depends on the predicted region of Bbox branch, there is some deviation from the true location of human instance. However, we find that this deviation does not affect the effect of the predicting parsing score, so we do not make corrections to this deviation.
3.5 Training and Inference
As we introduce new supervision into RP R-CNN, there are some changes in the training and inference phases compared with the common methods [30, 40].
Training. There are three losses for global human parts segmentation and Parsing branch: \({\mathcal {L}}_{\text {sem}}\) (segmentation loss), \({\mathcal {L}}_{\text {par}}\) (parsing loss), \({\mathcal {L}}_{\text {res}}\) (re-scoring loss). The segmentation loss and parsing loss are computed as a per-pixel cross entropy loss between the predicted segmentation and the ground-truth labels. We use the MSE loss as re-scoring loss. We have observed that the losses from three tasks have different scales and normalization policies. Simply adding them degrades the overall performance. This can be corrected by a simple loss re-weighting strategy. Considering the losses of the detection sub-network, the whole network loss \({\mathcal {L}}\) can be written as:
The \({\mathcal {L}}_{\text {rpn}}\) and \({\mathcal {L}}_{\text {bbox}}\) are losses of RPN and BBox branch, each of which is composed of classification loss and box regression loss. By tuning \(\lambda _\text {p}\), \(\lambda _\text {s}\) and \(\lambda _\text {r}\), it is possible to make the network converge to optimal performance.
Inference. For network inference, we select top 100 candidate bounding-boxes per image from the human detection results. These candidates are fed into Parsing branch to predict instance parsing map and mIoU score. However, the mIoU score is only trained by positive samples, which leads to the lack of the ability to suppress negative samples. So we fuse mIoU score \({\mathcal {S}}_{\text {iou}}\) and classification score \({\mathcal {S}}_{\text {cls}}\) to generate the final parsing score \({\mathcal {S}}_{\text {parsing}} = \sqrt{{\mathcal {S}}_{\text {cls}} * {\mathcal {S}}_{\text {iou}}}\). In addition, we also find that the global human parts segmentation results are complementary to the Parsing branch result, the former has higher recall for each foreground, and the latter has better details. Thus, when generating semantic segmentation results, we adopt a new combination strategy, as shown in Fig. 4. We filter out the low quality results based on the \({\mathcal {S}}_{\text {parsing}}\), and then generate instance-level human parts segmentation (b) instance-level. The instance-level (b) was and global human parts segmentation (a) \(global \) do element-wise OR operation to get the final result (c) \(combine \). It is worth noting that if the results of (a) and (b) are different at the same pixel, and both of them are predicted as non-background category, we directly adopt the results of (b). This is because the (b) has a more accurate perception of the human parts inside each instance.
4 Experiments
In this section, we describe experiments on multiple human parsing of RP R-CNN. All experiments are conducted on the CIHP [8] and MHP-v2 [45] datasets. We follow the Parsing R-CNN evaluation protocols. Using mean intersection over union (mIoU) [25] to evaluate the human part segmentation. And using average precision based on part (AP\(^\text {p}\)) [45] as instance evaluation metric(s).
4.1 Implementation Details
Training Setup. All experiments are based on Pytorch on a server with 8 NVIDIA Titan RTX GPUs. We use 16 batch-size (2 images per GPU) and adopt ResNet50 [13] as backbone. The short side of input image is resized randomly sampled from [512, 864] pixels, and the longer side is limited to 1,400 pixels; inference is on a single scale of 800 pixels. Each image has 512 sampled RoIs for Bbox branch and 16 sampled RoIs for Parsing branch. For CIHP dataset, there are 135,000 iterations (about 75 epochs) of the training process, with a learning rate of 0.02 which is decreased by 10 at the 105,000 and 125,000 iteration. For MHP-v2 dataset, the max iteration is half as long as the CIHP dataset with the learning rate change points scaled proportionally.
Parsing R-CNN Re-implementation. In order to better illustrate the advantages of RP R-CNN, we have re-implemented Parsing R-CNN according to the original paper [40]. We find that the training of Parsing R-CNN is not very stable. We solve this issue by adding group normalization [36] after each convolutional layer of Parsing branch. As shown in Table 2, our re-implemented Parsing R-CNN achieves comparable performance with original version both on CIHP and MHP-v2 datasets. Therefore, this work takes our re-implemented Parsing R-CNN as the baseline.
4.2 Ablation Studies
In this sub-section, we assess the effects of different settings and components on RP R-CNN by details ablation studies.
Loss Weights. To combine our GSE-FPN with PRSN in Parsing R-CNN, we need to determine how to train a single, unified network. Previous studies demonstrate that multi-task training is often challenging and can lead to degraded results [18]. We also observe that adding the losses of all tasks directly will not give the best results. But grid searching three hyper-parameters (\(\lambda _\text {p}\), \(\lambda _\text {s}\) and \(\lambda _\text {r}\)) is very inefficient, so we first determine \(\lambda _\text {p}\), and then determine \(\lambda _\text {s}\) and \(\lambda _\text {r}\) separately to improve efficiency. As shown in Table 3, we find that the network performance is the best when \(\lambda _\text {p} = 2.0\). We consider the group normalization layer makes the network convergence more stable, so that a proper large loss weight will bring higher accuracy. Table 4 shows that \(\lambda _\text {s} = 2.0\) is the proper loss weight for semantic segmentation. Although increasing the weight of global human parts segmentation loss will slightly reduce the accuracy of human detection, the overall performance is improved. Table 5 shows that the re-scoring task is not sensitive to the loss weight, and its loss scale is smaller than other losses, so it has no significant impact on the optimization of other tasks in multi-task training. To sum up, we choose \(\lambda _\text {p} = 2.0\), \(\lambda _\text {s} = 2.0\) and \(\lambda _\text {r} = 1.0\) as the loss weights of Eq. (1).
Inference Methods. The combination inference method proposed in Fig. 4 utilizes the complementarity of global human parts segmentation and Parsing branch results, and we give the detailed results in Table 6. The performance of using (a) \(semseg \) or (b) \(parsing \) alone is poor. The combination method (c) \(combine \) can significantly improve the metrics of semantic segmentation, and outperforms the baseline by 2 points mIoU.
Ablations on RP R-CNN. In Table 7, we perform the additional ablations of RP R-CNN on CIHP dataset. We observe that GSE-FPN is very helpful to the global human parts segmentation, which yields 2.0 points mIoU improvement. In addition, the global semantic feature also improves the instance metrics, AP\(^\text {p}_\text {50}\), AP\(^\text {p}_\text {vol}\) and PCP\(_\text {50}\) increase 2.8, 1.2 and 1.2 points respectively. With PRSN, the improvements of instance metrics are very significant, AP\(^\text {p}_\text {50}\) improves 5.7 points, and AP\(^\text {p}_\text {vol}\) improves 3.3 points. With GSE-FPN and PRSN, our proposed RP R-CNN achieves 58.2 mIoU and 71.6 AP\(^\text {p}_\text {50}\) on CIHP. The additional ablations on MHP-v2 dataset is shown in Table 8. Through GSE-FPN and PRSN, the performance of human parsing is also significant improved. Particularly, AP\(^\text {p}_\text {50}\) and AP\(^\text {p}_\text {vol}\) are raised considerably by PRSN, 13.0 points and 4.6 points, respectively.
4.3 Comparison with State-of-the-Arts
We evaluate RP R-CNN on the CIHP and MHP-v2 datasets and compare the results to state-of-the-art including bottom-up and one-stage/two-stage top-down methods, shown in Table 9. On CIHP dataset, our proposed RP R-CNN achieves 58.2 mIoU and 71.6 AP\(^\text {p}_\text {50}\), which surpasses Parsing R-CNN [40] in all respects. Compared with PGN [8], the performance advantage of RP R-CNN is huge, and AP\(^\text {p}_\text {50}\) is even 37.6 points ahead. With longer learning schedule (150 epochs), RP R-CNN achieves compared performance with one-stage top-down methods, e.g. M-CE2P and BraidNet. Even though we have adopted a lighter backbone (ResNet50 vs. ResNet101). Finally, using test-time augmentation, RP R-CNN with RestNet50 achieves state-of-the-art performance on CIHP.
On MHP-v2 dataset, RP R-CNN achieves excellent performance. We can observe that our RP R-CNN outperforms Parsing R-CNN consistently for all the evaluation metrics. And compared with M-CE2P, RP R-CNN yields about 10.8 point AP\(^\text {p}_\text {50}\) and 4.1 points AP\(^\text {p}_\text {vol}\) improvements. With RestNet50 backbone, it gives new state-of-the-art of 45.3 AP\(^\text {p}_\text {50}\), 46.8 AP\(^\text {p}_\text {vol}\) and 43.8 PCP\(_\text {50}\).
4.4 Analysis and Discussion
Effect of Parsing Re-Scoring Network. Figure 5 shows that the correlation between mIoU of the predicted parsing map with the matched ground-truth and the bbox/parsing score. As shown, the parsing score has better correlation with the ground-truth, especially for high-quality parsing map. However, it is difficult to score low-quality parsing map, which is the source of some false positive detections. Therefore, the evaluation of low-quality prediction is still a problem to be solved.
Qualitative results. We visualize multiple human parsing results of RP R-CNN in Fig. 6. We can observe that RP R-CNN has a good applicability to dense crowds and occlusions. In addition, the parsing score predicted by RP R-CNN reflects the quality of the parsing map.
5 Conclusions
In this paper, we proposed a novel Renovating Parsing R-CNN (RP R-CNN) model for solving the issue of missing global semantic information and inaccurate scoring of parsing result in top-down multiple human parsing. By explicitly introducing global semantic enhanced multi-scale features and learning the mIoU between instance parsing map and matched ground-truth, our RP R-CNN outperforms previous state-of-the-art methods consistently for all the evaluation metrics. In addition, we also adopt a new combination strategy, which improves the results of multiple human parsing by global semantic segmentation and instance-level semantic segmentation. We hope our effective approach will serve as a cornerstone and help the future research in multiple human parsing.
References
Chao, Y.W., Wang, Z., He, Y., Wang, J., Deng, J.: HICO: a benchmark for recognizing human-object interactions in images. In: ICCV (2015)
Chen, K., et al.: Hybrid task cascade for instance segmentation. In: CVPR (2019)
Chen, L., Papandreou, G., Schroff, F., Adam, H.: Rethinking atrous convolution for semantic image segmentation. arXiv:1706.05587 (2017)
Chen, L.-C., Zhu, Y., Papandreou, G., Schroff, F., Adam, H.: Encoder-decoder with atrous separable convolution for semantic image segmentation. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11211, pp. 833–851. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01234-2_49
Girdhar, R., Ramanan, D.: Attentional pooling for action recognition. In: NIPS (2017)
Girshick, R.: Fast R-CNN. In: ICCV (2015)
Gkioxari, G., Girshick, R., Dollar, P., He, K.: Detecting and recognizing human-object interactions. In: CVPR (2018)
Gong, K., Liang, X., Li, Y., Chen, Y., Yang, M., Lin, L.: Instance-level human parsing via part grouping network. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11208, pp. 805–822. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01225-0_47
Gong, K., Gao, Y., Liang, X., Shen, X., Lin, L.: Graphonomy: universal human parsing via graph transfer learning. In: CVPR (2019)
Guler, R., Neverova, N., Kokkinos, I.: DensePose: dense human pose estimation in the wild. In: CVPR (2018)
He, H., Zhang, J., Zhang, Q., Tao, D.: Grapy-ML: graph pyramid mutual learning for cross-dataset human parsing. In: AAAI (2020)
He, K., Gkioxari, G., Dollár, P., Girshick, R.: Mask R-CNN. In: ICCV (2017)
He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR (2016)
Hsieh, C.W., Chen, C.Y., Chou, C.L., Shuai, H.H., Liu, J., Cheng, W.H.: FashionOn: semantic-guided image-based virtual try-on with detailed human and clothing information. In: ACM MM (2019)
Huang, Z., Huang, L., Gong, Y., Huang, C., Wang, X.: Mask scoring R-CNN. In: CVPR (2019)
Ji, R., et al.: Learning semantic neural tree for human parsing. arXiv:1912.09622 (2019)
Jiang, B., Luo, R., Mao, J., Xiao, T., Jiang, Y.: Acquisition of localization confidence for accurate object detection. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) Computer Vision – ECCV 2018. LNCS, vol. 11218, pp. 816–832. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01264-9_48
Kendall, A., Gal, Y., Cipolla, R.: Multi-task learning using uncertainty to weigh losses for scene geometry and semantics. In: CVPR (2018)
Kirillov, A., Girshick, R., He, K., Dollár, P.: Panoptic feature pyramid networks. In: CVPR (2019)
Krizhevsky, A., Sutskever, I., Hinton, G.: ImageNet classification with deep convolutional neural networks. In: NIPS (2012)
Li, J., et al.: Multi-human parsing in the wild. arXiv:1705.07206 (2017)
Li, W., Zhu, X., Gong, S.: Harmonious attention network for person re-identification. In: CVPR (2018)
Lin, T., Dollár, P., Girshick, R., He, K., Hariharan, B., Belongie, S.: Feature pyramid networks for object detection. In: CVPR (2017)
Liu, X., Zhang, M., Liu, W., Song, J., Mei, T.: BraidNet: braiding semantics and details for accurate human parsing. In: ACM MM (2019)
Long, J., Shelhamer, E., Darrell, T.: Fully convolutional networks for semantic segmentation. In: CVPR (2015)
Luo, P., Wang, X., Tang, X.: Pedestrian parsing via deep decompositional network. In: ICCV (2013)
Miao, J., Wu, Y., Liu, P., Ding, Y., Yang, Y.: Pose-guided feature alignment for occluded person re-identification. In: ICCV (2019)
Nair, V., Hinton, G.: Rectified linear units improve restricted Boltzmann machines. In: ICML (2010)
Qi, S., Wang, W., Jia, B., Shen, J., Zhu, S.-C.: Learning human-object interactions by graph parsing neural networks. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11213, pp. 407–423. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01240-3_25
Qin, H., Hong, W., Hung, W.C., Tsai, Y.H., Yang, M.H.: A top-down unified framework for instance-level human parsing. In: BMVC (2019)
Ren, S., He, K., Girshick, R., Sun, J.: Faster R-CNN: towards real-time object detection with region proposal networks. In: NIPS (2015)
Ruan, T., et al.: Devil in the details: towards accurate single and multiple human parsing. In: AAAI (2019)
Russakovsky, O., et al.: ImageNet large scale visual recognition challenge. IJCV 115, 211–252 (2015). https://doi.org/10.1007/s11263-015-0816-y
Tan, Z., Nie, X., Qian, Q., Li, N., Li, H.: Learning to rank proposals for object detection. In: ICCV (2019)
Tangseng, P., Wu, Z., Yamaguchi, K.: Retrieving similar styles to parse clothing. TPAMI (2014)
Wu, Y., He, K.: Group normalization. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11217, pp. 3–19. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01261-8_1
Xiao, B., Wu, H., Wei, Y.: Simple baselines for human pose estimation and tracking. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11210, pp. 472–487. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01231-1_29
Xiao, T., Liu, Y., Zhou, B., Jiang, Y., Sun, J.: Unified perceptual parsing for scene understanding. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11209, pp. 432–448. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01228-1_26
Yamaguchi, K., Kiapour, M.H., Ortiz, L.E., Berg, T.L.: Parsing clothing in fashion photographs. In: CVPR (2012)
Yang, L., Song, Q., Wang, Z., Jiang, M.: Parsing R-CNN for instance-level human analysis. In: CVPR (2019)
Yang, W., Luo, P., Lin, L.: Clothing co-parsing by joint image segmentation and labeling. In: CVPR (2014)
Yu, F., Koltun, V.: Multi-scale context aggregation by dilated convolutions. In: ICLR (2016)
Zeng, X., et al.: Crafting GBD-net for object detection. TPAMI 40(9), 2109–2123 (2017)
Zhao, H., Shi, J., Qi, X., Wang, X., Jia, J.: Pyramid scene parsing network. In: CVPR (2017)
Zhao, J., Li, J., Cheng, Y., Feng, J.: Understanding humans in crowded scenes: deep nested adversarial learning and a new benchmark for multi-human parsing. In: ACM MM (2018)
Zhu, B., Song, Q., Yang, L., Wang, Z., Liu, C., Hu, M.: CPM R-CNN: calibrating point-guided misalignment in object detection. arXiv:2003.03570 (2020)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Yang, L. et al. (2020). Renovating Parsing R-CNN for Accurate Multiple Human Parsing. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, JM. (eds) Computer Vision – ECCV 2020. ECCV 2020. Lecture Notes in Computer Science(), vol 12357. Springer, Cham. https://doi.org/10.1007/978-3-030-58610-2_25
Download citation
DOI: https://doi.org/10.1007/978-3-030-58610-2_25
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-58609-6
Online ISBN: 978-3-030-58610-2
eBook Packages: Computer ScienceComputer Science (R0)