ZHANG Rihong, Ou Jianshuang, LI Xiaomin, LING Xuan, ZHU Zheng, HOU Binfa. Lightweight algorithm for pineapple plant center detection based on improved an YoloV4 model[J]. Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 2023, 39(4): 135-143. DOI: 10.11975/j.issn.1002-6819.202210133
    Citation: ZHANG Rihong, Ou Jianshuang, LI Xiaomin, LING Xuan, ZHU Zheng, HOU Binfa. Lightweight algorithm for pineapple plant center detection based on improved an YoloV4 model[J]. Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 2023, 39(4): 135-143. DOI: 10.11975/j.issn.1002-6819.202210133

    Lightweight algorithm for pineapple plant center detection based on improved an YoloV4 model

    • Abstract: Spraying is one of the main physiological and biochemical processes for flower induction in pineapple production. However, manual spraying cannot fully meet large-scale production in recent years, due to the very high labor intensity. Fortunately, automatic spraying can be expected for the flower induction of pineapple during this time. Among them, it is a high demand to accurately detect the center position of pineapple plants during spraying. In this study, a hybrid network model was proposed to accurate recognize the pineapple plant center using an improved YoloV4 deep learning. The conventional YoloV4 model shared an outstanding detect speed more suitable for the agricultural robot. However, the large number of parameters occupied a huge amount of memory in the original model, leading to more computing power in the device. Therefore, it was necessary for a much lighter and more flexible network structure in the reasoning work, particularly with the shorter running time in the actual field operation. The lightweight was then realized using the network framework of the YoloV4 model. Firstly, the number of network layers and feature map channels in the backbone were reduced using the GhostNet network structure as the backbone to replace the CSP-Darknet53, which enhanced the efficiency of the feature extraction. The GhostNet presented a Ghost module for building efficient neural architectures, in order to reduce the computational costs of recent deep neural networks. As such, the Ghost bottlenecks performed better in the lightweight structures of the entire network. Secondly, the depthwise separable convolution was introduced into the Neck network. The convolution set was integrated to reduce the number of parameters with a high level of feature extraction. Besides, the lightweight attention mechanism (called CBAM) was added between the backbone and the neck. Attention weights were then put on the channel and spatial dimensions for a better connection in the channel and space, in order to extract the effective features of the target. Finally, the quantity of the improved model (named Yolov4-GHDW) parameters was reduced by 70%, compared with the original. In the dataset-making stage, some operations were selected to process the collected images for the better generalization ability of the training model, including the vertical flip, adding Gaussian blur, adding salt and pepper noise, and extracting green features. Among them, 2530 images were divided into 1962, 218, and 350 images for the training set, the validation set, and the test set. The collected image dataset was used for the model training at the workstation, where the loss value of the training set and test set were recorded. The model with the smallest loss value was then selected as the optimal solution after training. The results show that the Yolov4-GHDW model presented a much higher recognition speed, and an excellent extraction performance to the target, compared with the original YoloV4, Faster R-CNN, and CenterNet. Specifically, the average accuracies of target recognition under the dense and sparse planting of pineapple plants were 94.7% and 95.5%, respectively, with only 68.4 MB of model memory. The real-time recognition speed reached 27 frames/s, and the average inference time of each image was 72ms, which was 23% shorter than that of YoloV4. Over all, the improved model can realize the identification and detection of pineapple plant center under complex conditions. The performance can fully meet the location requirement of the equipment during operation. Moreover, the improved model was also deployed in an embedded device (named Jetson Nano) to verify the performance. As a result, 13 frames per second was achieved to fluently operate with low power consumption. In conclusion, the Yolov4-GHDW has a high accuracy and recognition speed with a small memory footprint, indicating the balance between speed and accuracy. The finding can also provide technical support for intelligent and accurate flowering equipment in pineapple production.
    • loading

    Catalog

      /

      DownLoad:  Full-Size Img  PowerPoint
      Return
      Return