Abstract

Deep learning and computer vision have become emerging tools for diseased plant phenotyping. Most previous studies focused on image-level disease classification. In this paper, pixel-level phenotypic feature (the distribution of spot) was analyzed by deep learning. Primarily, a diseased leaf dataset was collected and the corresponding pixel-level annotation was contributed. A dataset of apple leaves samples was used for training and optimization. Another set of grape and strawberry leaf samples was used as an extra testing dataset. Then, supervised convolutional neural networks were adopted for semantic segmentation. Moreover, the possibility of weakly supervised models for disease spot segmentation was also explored. Grad-CAM combined with ResNet-50 (ResNet-CAM), and that combined with a few-shot pretrained U-Net classifier for weakly supervised leaf spot segmentation (WSLSS), was designed. They were trained using image-level annotations (healthy versus diseased) to reduce the cost of annotation work. Results showed that the supervised DeepLab achieved the best performance (IoU = 0.829) on the apple leaf dataset. The weakly supervised WSLSS achieved an IoU of 0.434. When processing the extra testing dataset, WSLSS realized the best IoU of 0.511, which was even higher than fully supervised DeepLab (IoU = 0.458). Although there was a certain gap in IoU between the supervised models and weakly supervised ones, WSLSS showed stronger generalization ability than supervised models when processing the disease types not involved in the training procedure. Furthermore, the contributed dataset in this paper could help researchers get a quick start on designing their new segmentation methods in future studies.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call