基于YOLOv11-FSi模型的大豆苗期杂草检测方法

    A method for detecting weeds in soybean seedling stage based on the YOLOv11-FSi model

    • 摘要:
      目的 提出一种改进算法以提升杂草目标的精准检测与识别能力,解决田间大豆苗期杂草识别中存在的实时性不足、目标尺度小及重叠遮挡严重等问题。
      方法 选取YOLOv11n为基线模型,提出一种改进的目标检测方法YOLOv11-FSi。该方法通过引入FSConv对传统C3k2模块进行重构,通过频域与空域双分支的协同交互及特征融合,有效增强模型的小目标特征表达能力。在特征金字塔网络输出端嵌入SEAM通道–空间注意力模块,以强化复杂背景及遮挡条件下的特征表征与目标区分能力。同时,引入Inner-MPDIoU损失函数优化边界框回归,提高小目标定位精度与检测鲁棒性。
      结果 改进模型后精确率、召回率和mAP0.5分别达到78.3%、72.2%和81.7%,较基线模型YOLOv11n分别提升4.2、4.5和4.0个百分点。并且模型参数量仅为2.8 M,计算量为7.8 G,保持了良好的轻量化特性。边缘设备部署试验显示,经TensorRT优化加速后,改进模型的推理速度高达126.58 帧·s-1,满足实时检测的需求。
      结论 本文提出的YOLOv11-FSi模型在保持轻量化特性的同时,实现了对幼苗期大豆杂草的高精度识别,可为大豆田间智能化、精准化除草作业提供技术支撑。

       

      Abstract:
      Objective To propose an improved algorithm to enhance the accurate detection and identification ability of weed target, and solve the issues of insufficient real-time performance, small target scale, as well as severe overlapping occlusion in Glycine max (soybean) seedling weed detection in the field.
      Method This study selected YOLOv11n as the baseline model and proposed an improved target detection method, YOLOv11-FSi. This method reconstructed the traditional C3k2 module by introducing FSConv, and effectively enhanced the model’s small target feature representation ability through the collaborative interaction and feature fusion of frequency and spatial domain branches. A SEAM channel-spatial attention module was embedded at the output of the feature pyramid network to strengthen feature representation and target discrimination under complex backgrounds and occlusion conditions. Simultaneously, the Inner-MPDIoU loss function was introduced to optimize bounding box regression, improving small target localization accuracy and detection robustness.
      Result After the improvement, the precision, recall, and mAP0.5 reached 78.3%, 72.2%, and 81.7%, respectively, representing improvements of 4.2, 4.5, and 4.0 percentage points compared to the baseline model YOLOv11n. Furthermore, the model has only 2.8 M parameters and a computational cost of 7.8 G, maintaining excellent lightweight characteristics. Edge device deployment tests showed that after TensorRT optimization and acceleration, the improved model achieved an inference speed of up to 126.58 frames per second, meeting the requirements for real-time detection.
      Conclusion The proposed YOLOv11-FSi model achieves high-precision identification of weeds in soybean field at the seedling stage, while maintaining lightweight characteristics, providing technical support for intelligent and precise weeding operations in soybean fields.

       

    /

    返回文章
    返回