孙月平,孟祥汶,郭佩璇,等. 基于改进YOLOv8n的渔光互补池塘作业船导航中线提取方法[J]. 农业工程学报,2024,40(23):1-11. DOI: 10.11975/j.issn.1002-6819.202408065
    引用本文: 孙月平,孟祥汶,郭佩璇,等. 基于改进YOLOv8n的渔光互补池塘作业船导航中线提取方法[J]. 农业工程学报,2024,40(23):1-11. DOI: 10.11975/j.issn.1002-6819.202408065
    SUN Yueping, MENG Xiangwen, GUO Peixuan, et al. Navigation center line extraction method for fishing and light complementary pond operation boat based on improved YOLOv8n[J]. Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 2024, 40(23): 1-11. DOI: 10.11975/j.issn.1002-6819.202408065
    Citation: SUN Yueping, MENG Xiangwen, GUO Peixuan, et al. Navigation center line extraction method for fishing and light complementary pond operation boat based on improved YOLOv8n[J]. Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 2024, 40(23): 1-11. DOI: 10.11975/j.issn.1002-6819.202408065

    基于改进YOLOv8n的渔光互补池塘作业船导航中线提取方法

    Navigation center line extraction method for fishing and light complementary pond operation boat based on improved YOLOv8n

    • 摘要: 由于渔光互补池塘上方光伏板阵列的遮挡会对北斗/GPS定位信号产生较大影响,导致无人作业船自主导航精度大幅降低,并且传统机器视觉算法鲁棒性较低,图像易受光影变化、池中水草分布和水面障碍物遮挡等因素干扰,导致视觉导航线检测效果不理想。针对上述问题,该研究提出一种基于改进YOLOv8n的渔光互补池塘作业船导航中线提取方法。改进模型首先从提高检测实时性角度出发,将HGNetV2网络作为主干网络,采用组归一化方式(batch normalization,BN)与共享卷积结构,设计轻量化检测头网络,减小模型体积;然后使用SPPF_LSKA模块作为特征融合层,提高模型多尺度特征融合能力;最后采用Wise-IoU(weighted interpolation of sequential evidence for intersection over union)损失函数,提升边界框回归性能和对远处小目标的检测精度。利用改进YOLOv8n检测框坐标提取两侧水泥立柱定位参照点,通过最小二乘法拟合两侧水泥立柱行线,进而使用角平分线提取导航中线。消融试验结果表明,相对于原始YOLOv8n模型,改进YOLOv8n模型的计算量、参数量和模型体积分别下降36.0%、36.8%和32.8%,平均精度均值(mean average precision,mAP)为97.9%,查准率为93.1%,单张图像检测时间为6.8 ms,检测速度提升42.9%。不同模型对比试验表明,改进YOLOv8n模型在较低计算成本的基础上,体现出了良好的实时性与精准度检测性能,具有明显优势。在导航中线定位分析试验中,提取水泥立柱定位参照点与人工观测标记点平均直线误差在0~5 m和5~10 m距离范围内分别为3.69 cm和4.57 cm,提取导航中线与实际导航中线平均直线误差为3.26 cm,准确率为92%。在导航中线实时性试验中,导航中线平均提取速度为22.34 帧/s,满足渔光互补池塘无人作业船导航要求,为后续作业船视觉导航系统研究提供参考。

       

      Abstract: Due to the obstruction caused by the photovoltaic panel arrays above the fishing and light complementary ponds, the BeiDou/GPS positioning signals are significantly affected, leading to a substantial reduction in the autonomous navigation accuracy of unmanned workboats. Additionally, traditional machine vision algorithms have low robustness, and the images are easily affected by changes in light and shadow, the distribution of aquatic plants, and surface obstacles, resulting in suboptimal visual navigation line detection. Aiming at the above problems, an improved YOLOv8n model was proposed in this study to extract the navigation center line in fish and light complementary ponds. First, from the point of view of improving the real-time detection, the HGNetV2 network was used as the backbone network, and batch normalization (BN) and shared convolution structure were used to design a lightweight detection head network and reduce the size of the model. Then the SPPF_LSKA module was used as the feature fusion layer to improve the multi-scale feature fusion capability of the model. Finally, Wise-IoU (weighted interpolation of sequential evidence for intersection over union) loss function was used to improve the bounding box regression performance and the detection accuracy of remote small targets. The improved YOLOv8n detection frame coordinates were used to extract the reference points for the positioning of the cement columns on both sides, the lines of the cement columns on both sides were fitted by the least square method, and the middle line of the navigation was extracted by the angle bisection line. Ablation test results showed that compared with the original YOLOv8n model, the calculation amount, parameter number and model volume of the improved YOLOv8n model were decreased by 36.0%, 36.8% and 32.8%, respectively, and the mean average precision (mAP) was 97.9%. The accuracy was 93.1%, the detection time of single image was 6.8 ms, and the detection speed was increased by 42.9%. In the comparison test of different models, compared with YOLOv5s, YOLOv6, YOLOv7 and YOLOv8, the improved YOLOv8n model exhibited the smallest size and the highest degree of lightweight design, while maintaining a high level of detection accuracy and demonstrating excellent performance in detecting concrete columns. In the navigation center line positioning analysis test, the average linear error between the reference point of the extraction cement column and the manual observation mark was 3.69 cm and 4.57 cm within the range of 0~5 m and 5~10 m, respectively, and the average linear error between the extraction navigation center line and the actual navigation center line was 3.26 cm, with an accuracy of 92%. In the real-time test of navigation midline, compared with the original model, the average extraction speed of navigation midline by the improved YOLOv8n was 117.01 f/s on Windows11 test platform, and the average extraction speed of navigation midline was improved by 38.04%. On Jetson test platform, the average extraction speed of navigation midline was 22.34 f/s, and average extraction speed of navigation midline was increased by 36.38%, which meets the navigation requirements of the unmanned fishing and light complementary pond boat and provides a reference for the subsequent research on the visual navigation system of the operation boat.

       

    /

    返回文章
    返回