Weed management is a significant challenge for agronomists, especially in highly dense field environments. The study aims to develop a computer vision-based system for distinguishing between potato plants and weeds in the post-emergence stage within high occlusion and complex backgrounds. To achieve this, the RGB dataset having real images was collected from potato farms. After thorough cleaning, the data augmentation was done. Images are annotated at the pixel level and are made freely available to the research community for future investigations. To accurately segment, detect, and classify the potato plant from the weeds, cutting-edge deep learning techniques, i.e., the Mask RCNN and YOLO version 8 (YOLOv8), are trained. Precision (P), recall (R), and mean average precision (mAP@0.5, mAP@0.50-0.95) were used to evaluate the different model scales. YOLOv8 obtained the mAP@0.50 83.4%, whereas Mask RCNN stood at 79%. Mask RCNN obtained the highest P(0.83), R(0.76), and F1(0.91) values for the weed class. The reported performance metrics indicate that although YOLOv8 slightly outperforms Mask RCNN in overall mAP, Mask RCNN achieves higher precision, recall, and F1 scores for the weed class. This implies that Mask RCNN may be more effective at identifying weeds, which is crucial for effective weed management. These accuracy levels may not be exceptional, but they are a testament to the model's capabilities for weed detection in a highly complex and heavily occluded environment.
Read full abstract