Abstract
Wheat diseases have posed a severe threat to the grain quality and yield in wheat production. It is highly required for the automatic, rapid, and accurate identification of wheat diseases on equipment with limited resources, particularly for timely preventive measures. In this study, a lightweight detection was proposed for wheat diseases using an improved YOLOv8n, termed PGCW-YOLOv8. Firstly, a lightweight CPU network (PP-LCNet) was introduced to replace the backbone network of the YOLOv8 structure, in order to reduce the large model weight files. Depthwise separable convolution (DepthSepConv) structure was introduced into the Backbone layer to reduce the parameter quantity, thereby reducing the size of weight files for better detection performance. Secondly, a global attention mechanism (GAM) module was added to the Neck section, in order to enhance the feature extraction and fusion of the network. As such, the model was improved to better focus on the small features of the target disease. The higher detection accuracy of minor lesions was achieved to better understand the important information in the image through a global attention mechanism, thereby accurately identifying the diseases under complex background and lighting conditions. Thirdly, a lightweight content-aware reassembly of features (CARAFE) module was introduced to aggregate the context information within a larger receptive field, in order to improve the detection accuracy of the model. The detailed image information was then effectively preserved using upsampling and downsampling operations. Finally, the Wise-IoU boundary loss function was used instead of the original loss function, in order to enhance the bounding box regression performance of the network model. The position and size of disease features were better learnt to improve the detection of small target diseases. Experimental results show that the improved PGCW-YOLOv8 model reduced the computational complexity (GFLOPs), parameters, and model size by 13.6%, 12.5%, and 11.3%, respectively, in the wheat disease datasets collected in field environments, compared with the original YOLOv8n baseline model. Meanwhile, the precision, mean average precision (mAP), and frames per second (FPS) of the improved model increased by 4.5, 1.9 percentage points, and 23.1%, respectively, compared with the original. A comparison was made under the same experimental conditions with the mainstream deep learning models, such as Faster R-CNN, YOLOv5s, YOLOv7, Yolov7-tiny, YOLOXs, and Edge-YOLO. The improved PGCW-YOLOv8 model shared the highest precision and mAP values, while the lowest computational complexity (GFLOPs), parameters, and model size. Three datasets of comparative experiments indicate that the lightweight PGCW-YOLOv8 network model outperformed the original YOLOv8 and YOLOv5 models, especially in the accurate detection of the wheat disease features under complex background and multi-target situations. The finding can provide a strong reference for the intelligent detection of wheat diseases in real time, particularly for rapid detection applications, such as deployment on unmanned aerial vehicles and mobile terminal equipment.