Fractal Feature Modeling: Multi-Dimensional Attention and Spatial Adaptive Relationship Learning for Person Re-Identification
Person Re-identification (person Re-ID), as an intelligent video surveillance technology capable of retrieving the same person from different cameras, may bring about some challenges arising from the changes in the person’s poses, different camera views as well as occlusion. Recently, person Re-ID equipped with the attention mechanism has gradually emerged as one of the most active areas of study in the fields of computer vision and fractal feature modeling applications. Despite the upsurge in related research, existing fractal-attention-based methods still face two major challenges when recognizing different pedestrians in unpredictable realistic environments: (1) the adaptability of a single local attention feature to hostile scenes cannot be guaranteed, and (2) the existing methods originating from attention features usually rely on the line mapping or simple variants, which make it difficult to excavate the association relationships among pedestrians with similar appearance attributes. To address these issues, this paper proposes a simple effective fractal feature modeling method, named multi-dimensional attention and spatial adaptive relationship learning framework (MASARF) to explore the correlation between pedestrian bodies for person Re-ID. The proposed framework encompasses a multi-dimensional fractal-attention feature learning model (MDAM) and a dual-branch graph convolutional model (DGCM). In particular, the MDAM comprises the local and global attention modules, which are used to capture multi-dimensional attention features for each person. Subsequently, the DGCM is used to construct the nonlinear mapping association relationships among the various body regions for each person via a dual-branch graph convolutional optimization strategy. Extensive experiments were conducted using public person Re-ID datasets (Market-1501, DukeMTMC-reid, and CUHK-03). The results demonstrate that the performance of the proposed approach is superior to that of state-of-the-art methods between 2% and 10% at Rank-1 (mAP). Essential differences exist between our method and the existing methods in terms of feature extraction and relationship transformation, which provides the validation of its novelty in the person Re-ID domain.