面向红外场景的改进YOLOv8行人检测模型

Improved YOLOv8 Pedestrian Detection Model for Infrared Scenarios

  • 摘要: 针对行人检测中因红外图像分辨率低、背景复杂以及多尺度目标难以识别等问题,本文提出了一种面向红外场景的改进YOLOv8行人检测模型( IRD-YOLOv8n)。首先,采用RetinexFormer作为前置图像增强模块,以恢复受损图像细节并提升质量,确保后续模型能够提取更有效的行人特征。其次,在主干网络中将普通卷积替换成感受野注意力卷积( RFAConv),以提高模型的聚焦能力和特征表达的多样性。最后,在Neck中引入自适应分层特征融合模块(HFF模块),实现了跨层特征的精细化融合,增强模型对多尺度目标的检测性能。在FLIR数据集上进行训练和验证,实验结果显示,改进后的IRD-YOLOv8n模型,对比原YOLOv8模型在漏检率上降低了5%、在mAP@0.5、mAP@0.5:0.95上分别提高了2.4%、2.3%,具有广泛的应用前景。

     

    Abstract: To address the issues of low resolution, complex background, and difficulty in recognizing multi-scale targets in pedestrian detection using infrared images, this paper proposes an improved YOLOv8 pedestrian detection model for infrared scenes (IRD-YOLOv8n). Firstly, the RetinexFormer is adopted as the pre-image enhancement module to restore the details of damaged images and improve their quality, ensuring that the subsequent model can extract more effective pedestrian features. Secondly, the ordinary convolution in the backbone network is replaced with the receptive field attention convolution (RFAConv) to enhance the model's focusing ability and the diversity of feature expression. Finally, an adaptive hierarchical feature fusion module (HFF module) is introduced in the Neck to achieve refined fusion of cross-layer features and enhance the model's detection performance for multi-scale targets. Training and validation were conducted on the FLIR dataset. Experimental results show that the improved IRD-YOLOv8n model reduces the missed detection rate by 5% compared to the original YOLOv8 model, and improves mAP@0.5 and mAP@0.5:0.95 by 2.4% and 2.3% respectively, demonstrating broad application prospects.

     

/

返回文章
返回