In-field disease symptom detection and localisation using explainable deep learning: Use case for downy mildew in grapevine
Metadatos
Mostrar el registro completo del ítemAutor
Hernández, Inés; Gutiérrez Salcedo, Salvador; Barrio, Ignacio; Íñiguez, Rubén; Tardaguila, JavierEditorial
Elsevier
Materia
Disease detection Computer vision Convolutional neural networks
Fecha
2024-09-26Referencia bibliográfica
Hernández, Inés. et. al. Computers and Electronics in Agriculture 226 (2024) 109478. [https://doi.org/10.1016/j.compag.2024.109478]
Patrocinador
European Union Horizon 2020 FET Open program under Grant agreement ID 828940; Grants 1150/2020 and 591/ 2021 by Universidad de La Rioja and Gobierno de La RiojaResumen
Diseases and pests in agriculture significantly impact crop yield and quality. Downy mildew (Plasmopara viticola)
is a particular noteworthy example in grapevines. Traditional detection methods are laborious, subjective and
time-consuming. Consequently, a technological solution based on artificial intelligence, would provide higher
levels of reproducibility and sampling. The aim of this work was to develop an interpretable, automated method
for detection and localisation of plant disease symptoms under field conditions. Images of the grapevine canopy
were taken in 14 commercial vineyard plots under a range of lightning conditions, including both static and onthe-
go settings. The images were processed using a sliding window, classifying sub-images into areas with and
without downy mildew. Transfer learning, fine-tuning and data augmentation were employed to automate the
classification, comparing convolutional neural networks (CNNs) and vision transformers (ViT). Subsequently, the
trained model was integrated into the sliding window to localise regions within the canopy images exhibiting
symptoms of downy mildew. Model predictions were interpreted using explainable artificial intelligence (XAI)
methods. The EfficientNetV2S model achieved an accuracy of 91 % and an F1-score of 0.92 when classifying
image areas and an Intersection over Union (IoU) of 0.83 when locating symptomatic areas. This method showed
promising results, enabling automatic and explainable detection and localisation of plant diseases in complex
conditions. The straightforward labelling process facilitated adaptation to new conditions, making it suitable for
different crops and diseases. Integration into mobile platforms could enhance disease management and reduce
the spread of pathogens, making a significant advance in agricultural technology.





