Beyond Human Parts -- Dual Part-Aligned Representations
谢威 last edited by 谢威
Beyond Human Parts: Dual Part-Aligned Representations for Person Re-Identification
The source code is here.
Challenges - Misalignment Problem
The significant visual appearance changes caused by:
human pose variation
distinct camera viewpoints ......
relies on manually designed splits of the input image or the feature maps into grid cells or horizontal stripes, based on the assumption that the human parts are well-aligned in the RGB color space
The attention mechanism
tries to learn an attention map over the last output feature map and constructs the aligned part features accordingly
Predicting a set of predefined attributes as useful features to guide the matching process.
Injecting human pose estimation or human parsing results to extract the human part aligned features based on the predicted human key points or semantic human part regions, while the success of such approaches heavily counts on the accuracy of human parsing models or pose estimators.
Most of the previous studies mainly focus on learning more accurate human part representations, while neglecting the influence of potentially useful contextual cues that could be addressed as “non-human” parts.
Beyond these predefined part categories, there still exist many objects or parts which could be critical for person re-identification, but tend to be recognized as background by the pre-trained human parsing models.
Dual Part-Aligned Representation
Accurate Human Parts:
- Human parsing model CE2P  extracts the human part masks and computes the human part-aligned representations for the features from the low-levels to high-levels.
the predicted label map:
(rescaled to be of the same size as the feature map ( is the representation of pixel , essentially the row of , )) represents the human part category of pixel of , which is of different values including human part categories and one background category. confidence maps: , where each confidence map is associated with a human part category (or the background category), is the pixel of .
- the representation of the
human part (分量形式)：
: 所有元素全为 1 的向量，表示将 广播到 : 所有元素全为 1 的向量，表示按行求和 (对每一行求和)
- the human part-aligned feature map
(same size as ):
Coarse Non-human Parts:
The latent part branch learns to predict
coarse confidence maps for all
pixels, is the pixel of :
- the latent part-aligned feature map
The self-attention mechanism [2, 3] learns to group all the pixels belonging to the same latent part together. We also extract the latent non-human part information on the feature maps from the low-levels to the high-levels.
Self-attention can learn to group the similar pixels together without extra supervision (also shown useful in segmentation).
The performance gains from the latent part branch, which is in fact the mixture of the coarse human and non-human part information, is mainly attributed to capturing non-human parts.
Although the latent part masks are learned from scratch, DPB (latent) achieves comparable results with the human part branch in general, which carries more strong prior information of the human parts knowledge, showing the importance of the non-human part context.
Through combining the advantages of the complementary information from both parts, our approach learns to augment the representation of each pixel with the representation of the part (human parts or non-human parts) that it belongs to.
Human part branch and latent part branch are complementary to each other.
Human part masks can eliminate the influence of background regions, while the predicted latent part masks serves as reliable surrogate for the non-human part.
The human part branch adopts off-the-shelf human parsing model to inject structural prior information by capturing the predefined semantic human parts for a person, while the latent part branch adopts a self-attention mechanism to help capture the detailed part categories beyond the injected prior information.
 Ting Liu, Tao Ruan, Zilong Huang, Yunchao Wei, Shikui Wei, Yao Zhao, and Thomas Huang. Devil in the details: Towards accurate single and multiple human parsing. arXiv:1809.05996, 2018. [link]
 Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. In NIPS, 2017. [link]
 Xiaolong Wang, Ross Girshick, Abhinav Gupta, and Kaiming He. Non-local neural networks. In CVPR, 2018. [link]