Aiming to address the issues of low recognition accuracy, limited all-weather usability, and high missed detection rate in RGB-T multimodal pedestrian detection algorithms, this paper proposes an improved illumination-aware weight-based RGB-T pedestrian detection method. Firstly, the mid-term fusion strategy is adopted by changing the fusion point from after five layers of convolution to after the third layer of convolution. This ensures that the fused feature map contains both detailed and semantic features. Secondly, through learnable parameters, the illumination weights wd and wn are adjusted to more reasonably map illumination information to fusion weights, resulting in new illumination weights wr and wt. Finally, adjustments are made to the network structure of the semantic segmentation module by replacing two-stream semantic segmentation with fused single-stream semantic segmentation using new illumination weights wr and wt. The proposed method's detection ability is validated on a cleaned KAIST dataset. Experimental results demonstrate that compared to the original network model, our proposed method reduces average missed detection rate from 15.69% to 13.15% while improving mAP@0.5 by 4.1%, showcasing superior performance.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.