岳凯,张鹏超,王磊,等. 基于改进YOLOv8n的复杂环境下柑橘识别[J]. 农业工程学报,2024,40(8):152-158. DOI: 10.11975/j.issn.1002-6819.202401118
    引用本文: 岳凯,张鹏超,王磊,等. 基于改进YOLOv8n的复杂环境下柑橘识别[J]. 农业工程学报,2024,40(8):152-158. DOI: 10.11975/j.issn.1002-6819.202401118
    YUE Kai, ZHANG Pengchao, WANG Lei, et al. Recognizing citrus in complex environment using improved YOLOv8n[J]. Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 2024, 40(8): 152-158. DOI: 10.11975/j.issn.1002-6819.202401118
    Citation: YUE Kai, ZHANG Pengchao, WANG Lei, et al. Recognizing citrus in complex environment using improved YOLOv8n[J]. Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 2024, 40(8): 152-158. DOI: 10.11975/j.issn.1002-6819.202401118

    基于改进YOLOv8n的复杂环境下柑橘识别

    Recognizing citrus in complex environment using improved YOLOv8n

    • 摘要: 针对复杂环境下柑橘果实大量重叠、枝叶遮挡且现有模型参数量大、计算复杂度高等问题,提出了一种基于改进YOLOv8n的柑橘识别模型YOLOv8-MEIN。首先,该研究设计了ME卷积模块并使用它改进YOLOv8n的C2f模块。其次,为了弥补CIoU损失函数在检测任务中泛化性弱和收敛速度慢的问题,使用Inner-CIoU损失函数加速边界框回归,提高模型检测性能。最后,在自建数据集上进行模型试验对比,试验结果表明,YOLOv8-MEIN模型交并比阈值为0.5的平均精度均值mAP0.5值为96.9%,召回率为91.7%,交并比阈值为0.5~0.95的平均精度均值mAP0.5~0.95值为85.8%,模型大小为5.8 MB,参数量为2.87 M。与原模型YOLOv8n相比,mAP0.5值、召回率、mAP0.5~0.95值分别提高了0.4、1.0、0.6个百分点,模型大小和参数量相比于原模型分别降低了3.3%和4.3%,为柑橘的自动化采摘提供技术参考。

       

      Abstract: Automated harvesting holds the promising potential in the informatization and automation of smart agriculture. Manual harvesting can be replaced to improve the high efficiency and cost saving. However, the existing models are limited to the large number of overlapping citrus fruits, branch and leaf occlusion in the complex environments, particularly for the much more parameters and high computational complexity. In this study, an improved YOLOv8-MEIN model was proposed for the citrus recognition using YOLOv8n. Firstly, the more efficient (ME) convolution module was designed to improve the CSPDarknet53 into 2-Stage feature pyramid network (C2f) module of YOLOv8n. The high number of common convolutional parameters and computation were reduced to fully meet the requirements of the real-time detection, due to the conditions of mobile hardware performance. The overall computation was also reduced to decrease the number of parameters in the model. Secondly, the Inner- complete intersection over union (CIoU) loss function was used to accelerate the bounding box regression for the better performance of the model. The weak generalization and slow convergence of the CIoU loss function were also compensated in the detection task. The reason was that the CIoU loss function only considered the shape loss in original YOLOv8, thus degenerating to the IoU loss function, when the target and bounding box were at the same aspect ratio. Finally, the YOLOv8-MEIN model was validated for the citrus detection. The YOLOv5s, YOLOv7-tiny, Faster R-CNN, and YOLOv8n models were selected for the training and testing on the citrus dataset. A comparison was carried out on the self-built citrus dataset under the same number of iteration rounds, in order to verify the performance of the YOLOv8-MEIN model. The YOLOv8-MEIN was improved 5.85 percentage points in the recall, 14.3 percentage points in the mean average precision (mAP) mAP0.5 metrics, and 34.8 percentage points in mAP0.5~0.95 metrics, compared with the two-stage Faster R-CNN, where the model size was only 5.1% of Faster R-CNN. Compared with the lightweight networks of YOLO series, such as YOLOv5s, YOLOv7-tiny, and YOLOv8n, the number of parameters was reduced by 59.1%, 52.3%, and 4.4%, respectively, whereas, the mAP0.5~0.95 values were improved by 0.7, 4.0, and 0.6 percentage points, respectively, indicating the outstanding mAP0.5. The experimental results show that the YOLOv8-MEIN model was achieved in the higher detection accuracy and efficiency with the smallest model volume and the lower computational cost. More convenient model was found for the transplantation and application suitable for the real-time target detection of citrus. The size of model and the number of parameters were reduced by 3.3 and 4.3 percentage points, respectively, compared with the mainstream target detections. The improved model was fully met the requirements for the fast and accurate recognition of citrus fruits in the complex environments. The finding can provide the technical references for the automated harvesting of citrus fruits.

       

    /

    返回文章
    返回