基于改进YOLOv8-s的融合图像目标检测方法

Object Detection Method of Fused Images Based on Improved YOLOv8-s

  • 摘要: 针对在红外-可见光融合图像上主流目标检测算法精度低、效果差等问题,提出一种基于改进YOLOv8-s的可见光-红外融合图像目标检测方法。首先,在YOLOv8-s中利用PSA与ACmix创造卷积、多尺度融合与自注意力并行结构并融入Conv卷积层,实现较低模型参数下更好地捕捉长距离依赖关系。其次,加入以DySample主导的多通道上采样级联融合模块,在增加采样丰富度的同时增强对通道间信息提取,提高目标的识别精度和定位框的重合度。最后,参考PGI(Programmable Gradient Information)结构设计并仅在训练时使用辅助学习模块PGIv8,在不增加推理成本情况下获得更好的模型参数与检测效果。在LLVIP和FLIR上利用CDDFuse制作融合图像数据集实验,结果显示mAP@50分别达到98.4%和81.1%,mAP@50:95达到71.1%和44.3%,在上述两者数据集上平均精度均值(mAP@50)相较原始YOLOv8-s提高1.3%与4.3%,基本达到实验预期。

     

    Abstract: To solve the problems of low accuracy and the poor effect of mainstream object detection algorithms on infrared-visible fusion images, a target-detection method based on improved YOLOv8-s is proposed. First, PSA and ACmix were used to create convolutional, multi-scale fusion, and self-attention parallel structures in YOLOv8-s, which were integrated into the Conv convolutional layer to better capture long-distance dependencies under lower model parameters. Second, a multi-channel upsampling cascade fusion module dominated by DySample was added to increase the sampling richness and enhance the information extraction between channels to improve the target recognition accuracy and the coincidence of the positioning frame. Finally, the auxiliary learning module PGIv8, which was designed with reference to the PGI structure and used only during training, obtained better model parameters and detection results without increasing the inference cost. The results show that mAP@50 reached 98.4% and 81.1%, and mAP@50:95 reached 71.1% and 44.3%, respectively, and their average accuracies were 1.3% and 4.3% higher than that of the original YOLOv8-s, respectively, which meets the experimental expectations.

     

/

返回文章
返回