DFN-PSAN: Multi-level deep information feature fusion extraction network for interpretable plant disease classification

dc.contributor.authorDai, G.
dc.contributor.authorTian, Z.
dc.contributor.authorFan, J.
dc.contributor.authorSunil, C.K.
dc.contributor.authorDewi, C.
dc.date.accessioned2026-02-04T12:25:35Z
dc.date.issued2024
dc.description.abstractAccurate identification of crop diseases is an effective way to promote the development of intelligent and modernized agricultural production, as well as to reduce the use of pesticides and improve crop yield and quality. Deep learning methods have achieved better performance in classifying input plant disease images. However, many plant disease datasets are often constructed from controlled scenarios, and these deep learning models may not perform well when tested in real-world agricultural environments, highlighting the challenges of transitioning to natural farm environments under the new demand paradigm of Agri 4.0. Based on the above reasons, this work proposes using a multi-level deep information feature fusion extraction network (DFN-PSAN) to achieve plant disease classification in natural field environments. DFN-PSAN adopts the YOLOv5 Backbone and Neck network as the base structure DFN and uses pyramidal squeezed attention (PSA) combined with multiple convolutional layers to design a novel classification network PSAN, which fuses and processes the multi-level depth information features output from DFN and highlights the critical regions of plant disease images with the help of pixel-level attention provided by PSA, thus realizing effective classification of multiple fine-grained plant diseases. The proposed DFN-PSAN was trained and tested on three plant disease datasets. The average accuracy and F1-score exceeded 95.27%. The PSA attention mechanism saved 26% of model parameters, achieving a competitive performance among existing related methods. In addition, this work effectively enhances the transparency of the features of the model attention to plant diseases through t-SNE with SHAP interpretable methods. © 2023 Elsevier B.V.
dc.identifier.citationComputers and Electronics in Agriculture, 2024, 216, , pp. -
dc.identifier.issn1681699
dc.identifier.urihttps://doi.org/10.1016/j.compag.2023.108481
dc.identifier.urihttps://idr.nitk.ac.in/handle/123456789/21476
dc.publisherElsevier B.V.
dc.subjectClassification (of information)
dc.subjectCrops
dc.subjectDeep learning
dc.subjectFarms
dc.subjectFeature extraction
dc.subjectImage classification
dc.subjectLearning systems
dc.subjectPixels
dc.subjectCrop disease
dc.subjectDisease classification
dc.subjectFeatures fusions
dc.subjectImages processing
dc.subjectInformation feature
dc.subjectMultilevel feature
dc.subjectMultilevels
dc.subjectPixel attention
dc.subjectPlant disease
dc.subjectExtraction
dc.subjectcrop yield
dc.subjectdesign
dc.subjectdisease
dc.subjectextraction method
dc.subjectimage processing
dc.subjectpixel
dc.subjectsoftware
dc.titleDFN-PSAN: Multi-level deep information feature fusion extraction network for interpretable plant disease classification

Files

Collections