EconPapers    
Economics at your fingertips  
 

EHAFF-NET: Enhanced Hybrid Attention and Feature Fusion for Pedestrian ReID

Jun Yang, Yan Wang (), Haizhen Xie, Jiayue Chen, Shulong Sun and Xiaolan Zhang
Additional contact information
Jun Yang: Big Data and Internet of Things Research Center, China University of Mining and Technology, Beijing 100083, China
Yan Wang: Big Data and Internet of Things Research Center, China University of Mining and Technology, Beijing 100083, China
Haizhen Xie: Big Data and Internet of Things Research Center, China University of Mining and Technology, Beijing 100083, China
Jiayue Chen: Big Data and Internet of Things Research Center, China University of Mining and Technology, Beijing 100083, China
Shulong Sun: Big Data and Internet of Things Research Center, China University of Mining and Technology, Beijing 100083, China
Xiaolan Zhang: Big Data and Internet of Things Research Center, China University of Mining and Technology, Beijing 100083, China

Mathematics, 2025, vol. 13, issue 4, 1-18

Abstract: This study addresses the cross-scenario challenges in pedestrian re-identification for public safety, including perspective differences, lighting variations, occlusions, and vague feature expressions. We propose a pedestrian re-identification method called EHAFF-NET, which integrates an enhanced hybrid attention mechanism and multi-branch feature fusion. We introduce the Enhanced Hybrid Attention Module (EHAM), which combines channel and spatial attention mechanisms. The channel attention mechanism uses self-attention to capture long-range dependencies and extracts multi-scale local features with convolutional kernels and channel shuffling. The spatial attention mechanisms aggregate features using global average and max pooling to enhance spatial representation. To tackle issues like perspective differences, lighting changes, and occlusions, we incorporate the Multi-Branch Feature Integration module. The global branch captures overall information with global average pooling, while the local branch integrates features from different layers via the Diverse-Depth Feature Integration Module (DDFIM) to extract multi-scale semantic information. It also extracts features based on human proportions, balancing high-level semantics and low-level details. Experiments demonstrate that our model achieves a mAP of 92.5% and R1 of 94.7% on the Market-1501 dataset, a mAP of 85.4% and R1 of 88.6% on the DukeMTMC-reID dataset, and a mAP of 49.1% and R1 of 73.8% on the MSMT17 dataset, demonstrating significant accuracy advantages over several advanced models.

Keywords: attention mechanisms; deep learning; feature extraction; multi-branch feature fusion; pedestrian re-identification (search for similar items in EconPapers)
JEL-codes: C (search for similar items in EconPapers)
Date: 2025
References: View complete reference list from CitEc
Citations:

Downloads: (external link)
https://www.mdpi.com/2227-7390/13/4/660/pdf (application/pdf)
https://www.mdpi.com/2227-7390/13/4/660/ (text/html)

Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.

Export reference: BibTeX RIS (EndNote, ProCite, RefMan) HTML/Text

Persistent link: https://EconPapers.repec.org/RePEc:gam:jmathe:v:13:y:2025:i:4:p:660-:d:1593017

Access Statistics for this article

Mathematics is currently edited by Ms. Emma He

More articles in Mathematics from MDPI
Bibliographic data for series maintained by MDPI Indexing Manager ().

 
Page updated 2025-03-22
Handle: RePEc:gam:jmathe:v:13:y:2025:i:4:p:660-:d:1593017