陈子文,杨宇帆,张海腾,等. 青花椒田间场景分割与导航线提取方法[J]. 农业工程学报,2024,40(22):1-10. DOI: 10.11975/j.issn.1002-6819.202404148
    引用本文: 陈子文,杨宇帆,张海腾,等. 青花椒田间场景分割与导航线提取方法[J]. 农业工程学报,2024,40(22):1-10. DOI: 10.11975/j.issn.1002-6819.202404148
    CHEN Ziwen, YANG Yufan, ZHANG Haiteng, et al. Methods of green Sichuan pepper field scene segmentation and navigation line extraction[J]. Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 2024, 40(22): 1-10. DOI: 10.11975/j.issn.1002-6819.202404148
    Citation: CHEN Ziwen, YANG Yufan, ZHANG Haiteng, et al. Methods of green Sichuan pepper field scene segmentation and navigation line extraction[J]. Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 2024, 40(22): 1-10. DOI: 10.11975/j.issn.1002-6819.202404148

    青花椒田间场景分割与导航线提取方法

    Methods of green Sichuan pepper field scene segmentation and navigation line extraction

    • 摘要: 精准分割青花椒田场景并提取导航路径,使农机能够自主行进,是青花椒机械化智能作业的重要环节。该研究对重庆市江津区3个青花椒种植示范基地青花椒不同生长阶段进行田间图像采集,并进行数据增强,构建青花椒田间场景数据集。针对青花椒田间场景复杂的问题,提出一种对道路、树干、树冠、天空及背景5类场景进行语义分割的轻量化网络——Mobile-Unet,该网络以U-Net为基础语义分割网络,采用MobileNetV2对U-Net特征提取网络进行替换,并将改进的MobileNetV2 8层5次下采样结构与U-Net网络结构相匹配,使用LeakyReLU激活函数避免训练过程中神经元死亡的问题。基于青花椒田场景分割结果,提出基于道路和树干双重特征的导航线提取方法。试验表明,使用数据增强和Dice Loss损失函数可以明显提升模型预测精度;相比Fast-Unet和BiseNet轻量网络Mobile-Unet在测试集上可获得更高的分割精度,像素准确率、类别平均像素准确率和平均交并比分别为91.15%、83.34%和70.51%,相较于U-Net,识别精度略有下降,但模型复杂度得到明显改善,内存占用量下降92.17个百分点,推理速度提升近9倍。对100张测试集图片进行导航线提取试验,与仅基于道路特征导航线提取方法相比,道路和树干双重特征提取成功率由76%上升至91%,基于道路和树干特征提取导航线的平均航偏角偏差分别为2.6°和6.7°,满足田间导航的精度要求。研究结果可为青花椒田间视觉导航算法的研究提供有效参考。

       

      Abstract: The traditional green Sichuan pepper industry heavily relies on manual labor for planting, field management, and picking. Nevertheless, the increasing aging of the rural population and the consistent decrease in the young workforce have led to a yearly surge in labor costs. Accurate segmentation of the green Sichuan pepper field scene and extraction of the navigation path are crucial steps in enabling agricultural machinery to operate autonomously, thereby advancing the intelligence of agricultural machinery in green Sichuan pepper fields. In this paper, field images were collected from three green Sichuan pepper planting demonstration bases in Jiangjin District, Chongqing Municipality, spanning various stages of green Sichuan pepper planting. A total of 400 images were gathered, with the dataset and test set divided according to a 3:1 ratio. Utilizing the open-source annotation tool Labelme, the images were annotated to construct a navigation dataset between the rows of green Sichuan peppers, followed by data enhancement. Given the complexity of green Sichuan pepper field scenes, a lightweight network, Mobile-Unet, is proposed for the semantic segmentation of five scene types: road, trunk, tree, sky and background. This network takes U-Net as its base semantic segmentation framework and incorporates MobileNetV2 as the feature extraction network. To adapt MobileNetV2 for semantic segmentation, the last three layers of the original MobileNetV2 network were omitted, and its 8-layer 5-times downsampling structure was aligned with the U-Net architecture. Additionally, the LeakyReLU activation function was employed in the convolutional units to circumvent the issue of neuron death during training. Based on the segmentation results of the green Sichuan pepper field scene and its distinctive features, this paper introduces a navigation line extraction approach that incorporates dual characteristics of roads and tree trunks. Experimental results demonstrate that enhancing the dataset and utilizing Dice Loss as the loss function effectively enhanced the model's prediction accuracy. Compared to the two lightweight networks, Fast-Unet and BiseNet, Mobile-Unet could achieve higher segmentation accuracy on the test set, with pixel accuracy (PA) of 91.15%, mean pixel accuracy (MPA) of 83.34%, and mean intersection over union (MIoU) of 70.51%. Compared to U-Net, the recognition accuracy is slightly reduced, but the model's complexity is significantly reduced, with a 92.17 percentage point decrease in memory occupation, and the inference speed is nearly 9 times faster. Additionally, tests conducted on 100 test set images for navigation line extraction, it was observed that the success rate of dual-feature extraction incorporating both road and trunk features increased from 76% to 91%, as compared to the method solely relying on road features. The average yaw angle deviation of the navigation line extracted using road contour and tree trunk features was 2.6° and 6.7°, respectively, satisfying the accuracy requirements for field navigation. The research content of this paper offers a valuable reference for exploring visual navigation algorithms in green Sichuan pepper fields.

       

    /

    返回文章
    返回