Hybrid deep learning approach for grape and apple leaf disease detection using CNN, YOLOv11 and EfficientNet-V2s

Loading...
Thumbnail Image

Supplementary material

Other Title

Authors

Salimon, Sajna

Author ORCID Profiles (clickable)

Degree

Master of Applied Technologies (Computing)

Grantor

Unitec, Te Pūkenga – New Zealand Institute of Skills and Technology

Date

2025

Supervisors

Varastehpour, Soheil
Shakiba, Masoud

Type

Masters Thesis

Ngā Upoko Tukutuku (Māori subject headings)

Keyword

grape vines
apple trees
leaf spots
plant disease
modelling
neural networks
image processing
pattern recognition systems in agriculture

Citation

Salimon, S. (2025). Hybrid deep learning approach for grape and apple leaf disease detection using CNN, YOLOv11 and EfficientNet-V2s (Unpublished document submitted in partial fulfilment of the requirements for the degree of Master of Applied Technologies (Computing)). Unitec, Te Pūkenga - New Zealand Institute of Skills and Technology https://hdl.handle.net/10652/6951

Abstract

RESEARCH QUESTIONS 1. How can the integration of CNN, YOLOv11, and EfficientNet-V2s improve the lo calisation and classification of grape and apple leaf diseases in complex agricultural environments? 2. Can the proposed hybrid model achieve real-time inference and high accuracy while remaining computationally efficient enough for deployment in resource-constrained settings? 3. How does the model perform when subjected to variable field conditions, such as inconsistent lighting, overlapping leaves, and background noise? ABSTRACT Grapevine and apple leaf diseases pose a substantial threat to global fruit production, affecting both crop productivity and quality. While early detection is crucial for effective management, traditional approaches such as manual inspection remain slow, subjective, and unsuitable for large scale deployment. This paper presents a hybrid deep learning architecture that combines classification and detection in a structured, modular pipeline. The proposed system integrates a CNN to distinguish between apple and grape leaves, YOLOv11 for rapid and precise localisation of diseased regions, and EfficientNet-V2s for fine-grained disease classification. The framework was trained and tested on ten publicly available datasets comprising a diverse set of grape and apple leaf images under variable field-like conditions. Strong detection performance was achieved using YOLOv11, with a test mAP0.5 of 0.990 and high precision and recall, confirming its robustness across different visual environments. However, the final classification stage using EfficientNet-V2s yielded moderate results, with a test accuracy of 49.2%, highlighting challenges related to class imbalance and subtle inter-class similarities in visual symptoms. Despite classification limitations, the modular structure of the system enabled effective disease region isolation and showed superior performance when compared to baseline models such as YOLOv8 + CNN and YOLOv5 + ResNet. The findings offer a practical and extensible foundation for automated crop disease monitoring. Future adaptations of this work can support improved classification accuracy, domain generalisation, and real world deployment in resource constrained agricultural environments.

Publisher

Link to ePress publication

DOI

Copyright holder

Author

Copyright notice

All rights reserved

Copyright license

Available online at