TY - GEN
T1 - Pose-Guided Occlusion-Aware Network for Occluded Person Re-Identification
AU - Zhao, Xiaokun
AU - Zhang, Longfei
AU - Wu, Xingyong
AU - Ding, Gangyi
N1 - Publisher Copyright:
© 2025 SPIE.
PY - 2025/7/22
Y1 - 2025/7/22
N2 - In real-world applications, person re-identification (ReID) often faces varying degrees of occlusion, resulting in missing target features and feature confusion caused by occluding objects. Existing methods using human body cues can primarily mitigate occlusion caused by objects. However, they often fail to distinguish between multiple people, making it difficult to resolve ambiguities caused by non-target pedestrians occluding the target. To address this problem, we propose PONet, which provides a comprehensive solution to occlusion robustness. Specifically, we design the POC-SE module, which explicitly simulates occlusion scenarios involving different objects and people and extracts discriminative features from key regions of the target pedestrian by exploiting joint pose constraints and occlusion prediction mechanisms. We adopt the Swin Transformer with cross-scale attention as the encoder, which enhances the model's focus on key semantic parts while improving high-resolution feature representation. In addition, we propose the prototypical contrastive loss, which mitigates the intra-class variance problem caused by occlusion within samples by integrating the intra-class distribution. Experimental results demonstrate that PONet achieves advanced performance in occlusion scenarios and holistic ReID benchmark tests, fully validating its effectiveness and robustness. Especially on the Occluded-Duke dataset, our method achieves 71.3% mAP and 82.9% Rank-1 accuracy.
AB - In real-world applications, person re-identification (ReID) often faces varying degrees of occlusion, resulting in missing target features and feature confusion caused by occluding objects. Existing methods using human body cues can primarily mitigate occlusion caused by objects. However, they often fail to distinguish between multiple people, making it difficult to resolve ambiguities caused by non-target pedestrians occluding the target. To address this problem, we propose PONet, which provides a comprehensive solution to occlusion robustness. Specifically, we design the POC-SE module, which explicitly simulates occlusion scenarios involving different objects and people and extracts discriminative features from key regions of the target pedestrian by exploiting joint pose constraints and occlusion prediction mechanisms. We adopt the Swin Transformer with cross-scale attention as the encoder, which enhances the model's focus on key semantic parts while improving high-resolution feature representation. In addition, we propose the prototypical contrastive loss, which mitigates the intra-class variance problem caused by occlusion within samples by integrating the intra-class distribution. Experimental results demonstrate that PONet achieves advanced performance in occlusion scenarios and holistic ReID benchmark tests, fully validating its effectiveness and robustness. Especially on the Occluded-Duke dataset, our method achieves 71.3% mAP and 82.9% Rank-1 accuracy.
KW - Occluded person re-identification
KW - Pose-guided
KW - Prototypical contrastive loss
UR - https://www.scopus.com/pages/publications/105022640412
U2 - 10.1117/12.3073553
DO - 10.1117/12.3073553
M3 - Conference contribution
AN - SCOPUS:105022640412
T3 - Proceedings of SPIE - The International Society for Optical Engineering
BT - Seventeenth International Conference on Digital Image Processing, ICDIP 2025
A2 - Poon, Ting-Chung
A2 - Jiang, Xudong
A2 - Wang, Zhaohui
A2 - Tian, Jindong
PB - SPIE
T2 - 17th International Conference on Digital Image Processing, ICDIP 2025
Y2 - 25 April 2025 through 27 April 2025
ER -