Next Article in Journal
Emergency Response Using Volunteered Passenger Aircraft Remote Sensing Data: A Case Study on Flood Damage Mapping
Next Article in Special Issue
A Novel, Portable and Fast Moisture Content Measuring Method for Grains Based on an Ultra-Wideband (UWB) Radar Module and the Mode Matching Method
Previous Article in Journal
High Precision Timing with Parabolic Equation Fitting in Narrowband Systems
Previous Article in Special Issue
Crop Water Content of Winter Wheat Revealed with Sentinel-1 and Sentinel-2 Imagery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Classification of Plant Leaf Diseases Based on Improved Convolutional Neural Network

1
School of Electrical Engineering and Automation, Anhui University, Hefei 230601, China
2
Institutes of Physical Science and Information Technology, Anhui University, Hefei 230601, China
3
School of Electrical and Information Engineering, Anhui University of Technology, Ma’anshan 243032, China
*
Authors to whom correspondence should be addressed.
Sensors 2019, 19(19), 4161; https://doi.org/10.3390/s19194161
Submission received: 28 July 2019 / Revised: 18 September 2019 / Accepted: 19 September 2019 / Published: 25 September 2019
(This article belongs to the Special Issue Advanced Sensors in Agriculture)

Abstract

:
Plant leaf diseases are closely related to people’s daily life. Due to the wide variety of diseases, it is not only time-consuming and labor-intensive to identify and classify diseases by artificial eyes, but also easy to be misidentified with having a high error rate. Therefore, we proposed a deep learning-based method to identify and classify plant leaf diseases. The proposed method can take the advantages of the neural network to extract the characteristics of diseased parts, and thus to classify target disease areas. To address the issues of long training convergence time and too-large model parameters, the traditional convolutional neural network was improved by combining a structure of inception module, a squeeze-and-excitation (SE) module and a global pooling layer to identify diseases. Through the Inception structure, the feature data of the convolutional layer were fused in multi-scales to improve the accuracy on the leaf disease dataset. Finally, the global average pooling layer was used instead of the fully connected layer to reduce the number of model parameters. Compared with some traditional convolutional neural networks, our model yielded better performance and achieved an accuracy of 91.7% on the test data set. At the same time, the number of model parameters and training time have also been greatly reduced. The experimental classification on plant leaf diseases indicated that our method is feasible and effective.

1. Introduction

With the rapid development of computer technology, traditional machine learning methods have been applied in plant diseases prediction more and more widely. With the popularity of machine learning algorithms in computer vision, in order to improve the accuracy and speed of diagnostic results, researchers have studied automated plant disease diagnosis based on traditional machine learning algorithms, such as random forest, k-nearest neighbor and support vector machine (SVM) [1,2,3]. Tan et al. established a multi-layer BP neural network model to realize the disease identification of soybean leaves, by calculating the chromaticity values of the leaves [4]. By extracting the color and texture characteristics of grape disease leaves, Tian et al. used a support vector machine (SVM) recognition method which achieved better results than the neural network [5]. Wang et al. developed a discriminant analysis method to identify cucumber lesions, by extracting the color, shape and texture features of leaf lesions, as well as combining with environmental information [6]. Zhang et al. also extracted the color, shape and texture features of lesion after lesion segmentation, and then used them to identify five types of corn leaves by K-nearest neighbor (KNN) classifier [7].
Spraying pesticides is a key step in plant growing. Ron et al. developed a robotic sprayer platform with the collaboration of remote operators for detecting targets and completing spraying tasks. They reported that it can prevent 30% to 35% of crop losses [8]. Autonomous selectivity for spraying pesticides can be used to reduce production costs in agriculture, using an auto-driving robot to detect plants in the crop area and to perform selective injection [9]. Moreover, plant disease detection can be based on spectroscopy and imaging processes. Xie et al. proposed an automatic classifier with multi-level learning features for field crop pests [10]. Sindhuja developed a ground-based sensor system to help monitor the status of plant health and diseases in field conditions [11].
Li et al. proposed an improved deep learning pipeline for automatic localization and count of agricultural crop pests, which integrates a convolutional neural network (CNN) of ZF (Zeiler and Fergus model) and a region proposal network (RPN) with Non-Maximum Suppression (NMS) [12]. Yang et al. proposed a new identification method for rice diseases based on deep convolutional neural networks [13]. The model was trained to identify 10 common rice diseases and tested on a dataset containing 500 natural images of diseased and healthy rice leaves and stems captured from rice fields. Under the 10-fold cross-validation strategy, the CNN-based model achieved an accuracy of 95.48%. Xia et al. proposed a convolutional neural network model to solve the problem of multi-classification of crop insects and achieved a heightened accuracy [14]. Sun et al. proposed an improved convolutional neural network which achieved an accuracy of 99.35% on a test dataset containing 54,306 images, which is composed of 26 diseases in 14 different plants [15]. Moreover, the Pascal VOC (Visual Object Classes) Challenge [16] and the recent ImageNet Large Scale Visual Recognition Challenge (ILSVRC) [17] based on the ImageNet dataset [18], have been widely used as benchmarks for many visualization issues in computer vision, including object classification. In 2012, a convolutional neural network, called AlexNet (Alex’s Network) [19], reduced the top5 error rate of the 1000 categories of classified images to 16.4% on ImageNet.
This work focused on five different convolutional neural network architectures, AlexNet, VGGNet (Visual Geometry Group’s Network) [20], GoogLeNet (Google’s Network) [21], ResNet (Residual Network) [22] and SENet (Squeeze-and-Excitation Network) [23], which participated in ILSVRC from 2012 to 2017, and achieved good classification results. There are still some challenges in plant leaf disease classification, which are as follows:
(1)
Limited by experimental conditions, such as current platform and hardware, a large CNN network will cost a long training time and have a slow convergence rate;
(2)
Long training convergence time will cause the final classification accuracy to decrease.
To shorten long training convergence time, decrease enormous parameters of most current network models, and increase recognizing accuracy, this paper proposes an integrated method. It adopts the inception structure to fuse the extracted high-level features, the Squeeze-and-Excitation module to perform feature re-calibration for weighting the features in the channel of CNN, and global average pooling instead of the fully connected layer. The experimental results show that our method is effective in the classification and identification of plant leaf diseases. Compared with other traditional convolutional neural networks, our model achieved the highest classification accuracy rate of 91.7% on our plant leaf disease dataset.

2. Materials and Methods

2.1. Data Preprocessing and Augmentation

We collected 10 kinds of disease leaf images from a library of plant leaf diseases (https://challenger.ai/), where digital color cameras were used for capturing diseased blade images with a resolution width of 256 and an unfixed length, as shown in Figure 1. Because some types of leaf diseases are confusing and unidentifiable, only 10 types of blade data were selected for our research. As for the images of plant corn, we tried to adopt the image class to verify the identification and showed the generalization ability of different CNN network structures for different types of leaf diseases. The diseased parts of apple and cherry leaves are similar, and the degrees of leaf diseases in different disease levels are also similar. It is more practical for our research. Compared with other types of leaf diseases, these diseased leaves can better reflect the distinguishing ability of disease areas for different CNN structures, and can better compare the ability of different CNN structures in leaf classification.
First, all leaf disease images were adjusted so that the length and width of the image were the same, which were resized to 224 × 224. Resizing images to 224 × 224 before inputting images into different networks is done to adapt different pre-training CNN structures. Then, because some leaf disease types contains less images than others and the collection of leaf disease images are random, images of these disease types were horizontally and vertically flipped. The leaf diseases are Cedar Apple Rust—serious, Cherry Powdery Mildew—general, and Cherry Powdery Mildew—serious. Thus, the leaf disease data set was expanded to prevent redundancy of the data set, ensure the validity of image data, and make the classifier balanced. After the data augmentation, the plant leaf disease dataset contained 6108 images, of which 5588 were for the training set and 520 were for the test set. Table 1 lists the number of images for each disease class.

2.2. Convolutional Neural Network (CNN)-Based Method

2.2.1. CNN Overall Architecture

Our deep learning-based network consists of VGG16 convolutional layers as well as the combination of Squeeze-and-Excitation (SE) module and Inception structure. The first five convolutional layers are based on the VGG16 model for self-learning low-to-high features of training images, where deeper convolutional layers reduce more resolution of feature maps, and extract more abstract high-level features. Then, the max pooling layer is used to filter the noise of the feature maps generated by the previous convolutional layer. Inception structure performs feature fusion, broadens the ability of acquiring features on feature maps, and extracts the best distinguishing features based on multi-dimensional analysis. The embedded SE module, re-calibrating the original features in the channel dimension, is used to replace the fully connected layer with the largest average pooling layer, reduce the training parameters as well as quickening the convergence of the model, and thus improving the classification accuracy of the model. The network structure of the improved model and related parameters are shown in Figure 2 and Table 2, respectively.
The five convolutional layers are based on VGG16 pre-training model, which determines which layers of the original network have to be frozen during the pre-training phase, and which layers are allowed to continue learning at a certain learning rate. Usually, the first several layers are frozen because the low-level features can better adapt to various problems. This work used a stochastic gradient descent optimization method to train the model on our own data set. The initial learning rate was set to 0.001, while momentum and weight attenuation were set to 0.0005 and 0.9, respectively. The Dropout layer [24] was used in our experiments to prevent over-fitting in training and make the model more effective.

2.2.2. GoogLeNet’s Inception

Inception module is the main component of GoogLeNet network. The Inception structure embeds multi-scale information and gathers features from different receptive fields to improve identification performance. It maintains the sparse structure, increases the depth and broadens the width of the network, therefore it reduces not only over-fitting but also free parameters. Figure 3 shows that the Inception module uses three different convolution kernels, 1 × 1 convolution, 3 × 3 convolution, 5 × 5 convolution as well as a 3 × 3 max pooling layer. It extracts three different scale features to increase the diversity of features, involving both macroscopic features and microscopic features. The purpose of the pooling layer is to preserve the primitive input information. The module splices the extracted features in the channel dimension and outputs a multi-scale feature map by concatenating these convolutional and pooling layers together.

2.2.3. Global Average Pooling (GAP)

The fully connected network has always been the standard configuration of the CNN network. However, too many parameters in the fully connected layer slows down the training speed of the network and makes it easy to be overfitting. The idea of global average pooling (GAP) [25] is to globally average the entire pixels of each feature map, and get an output for each feature map. The vector that is composed of these output features will be directly sent to softmax for classification. Figure 4 shows the comparison of the fully connected layer and the global averaged pooled layer.

2.2.4. Squeeze-and-Excitation Module

Figure 5 is a schematic diagram of the SE module, which omits the previous series of convolutions in the original SE module. Given an input X, the number of feature channels is C. Unlike the traditional CNN, three operations are taken to recalibrate previously obtained features.
The first one is Squeeze operation. Suppose that the inputs are X = (X1, X2, …, XC), XCRH×W. Formally, a statistic zRC is generated by shrinking X through its spatial dimensions H × W. The c-th element of Z is calculated by:
Z c = F s q ( X c ) = 1 W × H i = 1 W j = 1 H X c ( i , j ) .
Therefore, the Squeeze operation converts the input of H × W × C into an output of 1 × 1 × C , corresponding to the Fsq operation in Figure 5. The result of this step is equivalent to the numerical distribution of the C feature maps of the layer, or global information. The output Zc can be thought of as the description of a set of local descriptors for the entire channel map.
The second operation is the Excitation operation. It can represent the convolution and activation operations, which employ a simple gating mechanism with a sigmoid activation:
S = F e x ( Z , W ) = σ ( g ( Z , W ) ) = σ ( W 2 δ ( W 1 Z ) ) ,
where, δ refers to the ReLu function, and the output Z can be thought of as a set of local descriptors for the entire channel map, W 1 R C r × C and W 2 R C × C r . In order to control the complexity and generalization of the model, the embedding mechanism of the model is parameterized by two nonlinear fully connected layers.
Finally, a reweight operation regards the weight of the output of Excitation as the importance of each feature channel after feature selection, and then weights previous features by channel weighting to complete the pair in the channel dimension. The output of the block is obtained by rescaling X with the activations s:
x ˜ C = F s c a l e ( x C , s C ) = s C x C ,
where, F s c a l e ( x c , s c ) refers to channel-wise multiplication between the scalar sC and the feature map xCRH×W, and x ˜ = | x ˜ 1 , x ˜ 2 , x ˜ C | .
The SE module can be embedded in the Inception and standard network architecture of ResNet, as shown in Figure 6. Figure 6 is a combination structure of the SE module and the Inception module.

3. Experiments and Results

The experiments were performed on an Ubuntu workstation with CPU i7-8700k and RAM 32G, accelerated by two NVIDIA GTX 1080TI GPUs. All of our experiments were implemented by Caffe, a deep learning open source framework [26]. Moreover, accuracy rate was used to evaluate the performance of network models. The accuracy rate refers to the proportion of the number of corrected positive predictions to that of the whole positive predictions. It can be expressed as:
A c c u r a c y = N T P N T P + N F P ,
where, NTP is the number of corrected positive predictions, and NFP is the number of wrongly positive predictions.

3.1. Effects of the Feature Extraction Network

The most important metric we considered is the average accuracy of the test set. Table 3 lists the experimental accuracy, model size and training time for several commonly used deep learning CNN architectures, as well as the results of our method.
The first observation from Table 3 is that different convolution depths make the trained model produce different classification results on the test set. In general, more convolutional layers can learn more complex features from original images. Shallow CNNs such as AlexNet achieved an accuracy of 0.894 on the test set, while the deep networks VGG16, VGG19, ResNet-50, and Inceptionv2 yielded accuracies of 0.905, 0.903, 0.901 and 0.903 on the test set, respectively. Compared to other networks, our network is relatively shallow, but achieves higher accuracy on the test set. One possible reason is in that shallow network has a relatively good generalization compared to deep ones. The other reason is because of the use of the Inception module to broaden the network and combine the multi-scale feature information, as well as the use of the SE module to merge the feature channel into the Inception module and thus weighted and recalibrated features. As a result, our network achieved a maximum accuracy of 91.7% on the test set. Figure 7 shows the trends of accuracy of different CNN models on the test set.

3.2. Comparison of Model Size for Different Network Models

From the comparison of model size of different models in Table 2, we can get an intuitive observation that the larger the size of the CNN model, the more parameters the CNN had, and the longer the training time. The size of the AlexNet, VGG16, and VGG19 training models were 217 MB, 537.2 MB, and 558.4 MB, respectively. The large model size is because the last three layers of these network structures are all fully connected, which causes the number of the trained network model size to be larger than that of other deep learning models. On the contrary, GoogLeNet, Inceptionv2, and Inceptionv3 with Inception structure greatly reduce the size amount to 47.1 MB, 45.1 MB, and 87.3 MB, respectively. The size of our model is 57.3 MB, which is greatly reduced compared with VGG16 and VGG19. The reason is that our model used the Inception structure and the global average pooling instead of the last three-layer fully connected layer. This structure can avoid the requirement of a large number of weight parameters, reduce the size of the CNN model and solve the problem of large memory occupancy and slow convergence in the training CNN model.

3.3. Comparison of Training Time for Different Network Models

The general CNN model linearly converts all extracted feature maps into 4096-dimensional feature vectors after convolutional and pooled layers, and classifies leaf diseases by softmax layer. Table 4 shows the training time of the forward propagation and backpropagation processes for different CNN models and improved models. As can be seen from Table 4, our model performs a forward propagation rate of 0.038 s, which means that the time required to test a picture is 0.038 s. Compared with other CNN models, our model has a faster advantage in the forward propagation time.

3.4. Loss Function and Confusion Matrix of Our Network

From Figure 8a, it can be concluded that our model tends to converge (blue curve), and the final accuracy rate is stable at 91.7% (orange curve), achieving a better classification result. Accuracy is an unreliable performance metric for evaluating the classification model because it can produce misleading results when the sample numbers of different classes in the data set are unevenly distributed. Moreover, the average accuracy of all categories is an accurate indicator for the model on the test set. In other words, the categories that are difficult to classify will be improved by the easily classified categories. The confusion matrix is the degree to which a classification model is accurate for each classification category. From the confusion matrix in Figure 9b, we can conclude that for some difficult-to-classify plant leaf diseases, the classification accuracy of such single-category on the test set is low, because the diseased region in each leaf is too small and the number of different grades of leaf disease is different. Therefore, it is difficult to be classified and identified by model. For instance, leaves of “Cherry Powdery Mildew – general” and those of “Cherry Powdery Mildew – serious” are difficult to be classified, because most of the regions in these leaves are very similar. The confusion matrix of the last experiment showed that the accuracy of disease recognition for corn is 100%, which did not interfere with other types of leaf diseases in classification.

3.5. Visualization of Feature Extraction

Figure 9 visualizes a list of feature extractions after different layers of our network. The visualization of the network model can help us to intuitively understand the classification model. The ideal feature map of CNN should be sparse and contain typical local information. Through the visualization of the model, we can understand what features each layer of CNN learns, which can be used to adjust network parameters to improve the accuracy of the model. As a result, it provides a better understanding of how the CNN network learns the characteristics of the input image by visualizing various convolutional layers. We found that the features learned by CNN are hierarchical. The higher the level is, the more the specific features are presented. Moreover, the higher the dimensional feature maps correctly classify the images, the greater the effect presents. Specifically, the deep layer (Figure 9(7) or (8)) presents some edge corners and abstract features of colors, and the shallow feature map (Figure 9(1) or (2)) responds to the color information of the corners and other edges. The feature map of the middle layer (Figure 9(3), (4), (5), or (6)) has more complex invariance, captures similar textures, and has more layers for feature extraction. The high-level feature map shows the salient pose of the entire image after the extraction of the high-level abstract features.

4. Conclusions

This paper proposed an improved structure of convolutional neural networks for the identification and classification of a large dataset of different plant leaf diseases. Based on the traditional five-layer convolutional model of VGG16, the final, fully connected layer of VGG16 was replaced with Inception and SE modules, which can improve the classification accuracy of the model on the plant leaf disease dataset. Moreover, the global pooling layer can shorten the training time and parameter memory requirements, and also improve the generalization ability of the model. As a result, our method achieved the highest classification accuracy of 91.7% on the test set of plant leaf diseases. Compared with some other CNN methods, it has better adaptability to the change of image spatial position, showing better robustness to identify different diseases of various plant leaves, not limited to different diseases of the same plant.

Author Contributions

Conceptualization, P.C.; Data curation, J.H. and D.Z.; Funding acquisition, P.C.; Investigation, B.W.; Methodology, J.Z.; Project administration, P.C.; Software, J.Z. and D.Z.; Supervision, P.C. and D.Z.; Validation, B.W.; Writing–original draft, J.H.; Writing–review and editing, D.Z., P.C., B.W. and J.Z.

Funding

This work was supported by the National Natural Science Foundation of China (Nos. 61672035, 61472282, and 61872004), Educational Commission of Anhui Province (No. KJ2019ZD05), Anhui Province Funds for Excellent Youth Scholars in Colleges (gxyqZD2016068), the fund of Co-Innovation Center for Information Supply and Assurance Technology in AHU (ADXXBZ201705), and Anhui Scientific Research Foundation for Returness.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Es-Saady, Y.; Massi, I.E.; Yassa, M.E.; Mammass, D.; Benazoun, A. Automatic Recognition of Plant Leaves Diseases Based on Serial Combination of Two SVM Classifiers. In Proceedings of the 2nd International Conference on Electrical and Information Technologies, Xi’an, China, 2–4 December 2016; pp. 561–566. [Google Scholar]
  2. Gavhale, M.K.R.; Gawande, U. An Overview of the Research on Plant Leaves Disease Detection Using Image Processing Techniques. IOSR J. Comput. Eng. 2014, 16, 10–16. [Google Scholar] [CrossRef]
  3. Wang, G.; Sun, Y.; Wang, J.X. Automatic Image Based Plant Disease Severity Estimation Using Deep Learning. Comput. Intell. Neurosci. 2017, 1–8. [Google Scholar] [CrossRef] [PubMed]
  4. Tan, F.; Ma, X.D. The Method of Recognition of Damage by Disease and Insect Based on Laminae. J. Agric. Mech. Res. 2009, 6, 41–43. [Google Scholar]
  5. Tian, Y.W.; Li, T.L.; Li, C.H. Method for Recognition of Grape Disease Based on Support Vector Machine. Trans. Chin. Soc. Agric. Eng. 2007, 23, 175–180. [Google Scholar]
  6. Wang, X.F.; Zhang, S.W.; Wang, Z. Recognition of Cucumber Diseases Based on Leaf Image and Environmental Information. Trans. Chin. Soc. Agric. Eng. 2014, 30, 148–153. [Google Scholar]
  7. Zhang, S.W.; Shang, Y.J.; Wang, L. Plant Disease Recognition Based on Plant Leaf Image. J. Anim. Plant Sci. 2015, 25, 42–45. [Google Scholar]
  8. Ron, B.; Yael, E. Human-robot collaborative site-specific sprayer. J. Field Robot. 2017, 34, 1519–1530. [Google Scholar]
  9. David, R.; Javier, M.M.; Emir, M. 3D Imaging with a Sonar Sensor and an Automated 3-Axes Frame for Selective Spraying in Controlled Conditions. J. Imaging 2017, 3, 9. [Google Scholar] [Green Version]
  10. Xie, C.; Wang, R.; Zhang, J.; Chen, P.; Dong, W.; Li, R.; Chen, T.; Chen, H. Multi-level learning features for automatic classification of field crop insects. Comput. Electron. Agric. 2018, 152, 233–241. [Google Scholar] [CrossRef]
  11. Sindhuja, S.; Ashish, M.; Reza, E. A review of advanced techniques for detecting plant diseases. Comput. Electron. Agric. 2010, 72, 1–13. [Google Scholar]
  12. Li, W.; Chen, P.; Wang, B.; Xie, C. Automatic Localization and Count of Agricultural Crop Pests Based on an Improved Deep Learning Pipeline. Sci. Rep. 2019, 9, 7024. [Google Scholar] [CrossRef] [PubMed]
  13. Yang, L.; Shu, J.Y.; Nian, Y.Z.; Yu, R.L.; Yong, Z. Identification of Rice Diseases Using Deep Convolutional Neural Networks. Neurocomputing 2017, 267, 378–384. [Google Scholar]
  14. Xia, D.; Chen, P.; Wang, B.; Zhang, J.; Xie, C. Insect detection and classification based on improved convolutional neural network. Sensors 2018, 18, 4169. [Google Scholar] [CrossRef] [PubMed]
  15. Sun, J.; Tan, W.J.; Mao, H.P.; Wu, X.H.; Chen, Y.; Wang, L. Identification of Leaf Diseases of Various Plants Based on Improved Convolutional Neural Network. Agric. Eng. Newsp. 2017, 19, 209–215. [Google Scholar]
  16. Mark, E.; Luc, V.G.; Christopher, K.I.; Williams, J.; Winn, A.Z. The Pascal Visual Object Classes (VOC) Challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar]
  17. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A Large Scale Hierarchical Image Database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  18. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. 2012, 115, 211–252. [Google Scholar] [CrossRef]
  19. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet Classification with Deep Convolutional Neural Networks. Available online: http://papers.nips.cc/paper/4824-imagenet-classification-with-deep-convolutional-neural-networks.pdf (accessed on 25 September 2019).
  20. Simonyan, K.; Zisserman, A.; Bengio, Y.; LeCun, Y. (Eds.) Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the 3rd International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  21. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.E.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going Deeper with Convolutions. In Proceedings of the 28th IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  22. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 770–778. [Google Scholar]
  23. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. arXiv 2017, arXiv:1709.01507. [Google Scholar]
  24. Srivastava, N.; Hinton, G.E.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  25. Lin, M.; Chen, Q.; Yan, S. Network in network. arXiv 2013, arXiv:1312.4400. [Google Scholar]
  26. Jia, Y.; Shelhamer, E.; Donahue, J.; Karayev, S.; Long, J.; Girshick, R.; Guadarrama, S.; Darrell, T. Caffe: Convolutional Architecture for Fast Feature Embedding; Cornell University: Ithaca, NY, USA, 2014. [Google Scholar]
Figure 1. Sample images of 10 leaf diseases. (1) Apple healthy (AH); (2) Apple Scab general (ASG); (3) Apple Scab serious (ASS); (4) Apple Frogeye Spot (AFS); (5) Cedar Apple Rust genera (CARG)l; (6) Cedar Apple Rust serious (CARS); (7) Cherry healthy (CH); (8) Cherry Powdery Mildew general (CPMG); (9) Cherry Powdery Mildew serious (CPMS); (10) Corn healthy (CH).
Figure 1. Sample images of 10 leaf diseases. (1) Apple healthy (AH); (2) Apple Scab general (ASG); (3) Apple Scab serious (ASS); (4) Apple Frogeye Spot (AFS); (5) Cedar Apple Rust genera (CARG)l; (6) Cedar Apple Rust serious (CARS); (7) Cherry healthy (CH); (8) Cherry Powdery Mildew general (CPMG); (9) Cherry Powdery Mildew serious (CPMS); (10) Corn healthy (CH).
Sensors 19 04161 g001
Figure 2. The structure of the proposed convolutional neural network (CNN).
Figure 2. The structure of the proposed convolutional neural network (CNN).
Sensors 19 04161 g002
Figure 3. Inception structure model.
Figure 3. Inception structure model.
Sensors 19 04161 g003
Figure 4. Comparison of the fully connected layer and the global averaged pooled layer.
Figure 4. Comparison of the fully connected layer and the global averaged pooled layer.
Sensors 19 04161 g004
Figure 5. Squeeze-and-Excitation (SE) module.
Figure 5. Squeeze-and-Excitation (SE) module.
Sensors 19 04161 g005
Figure 6. The combined structure of the SE module and the Inception structure.
Figure 6. The combined structure of the SE module and the Inception structure.
Sensors 19 04161 g006
Figure 7. Trends in the accuracy of different CNN models on the test set.
Figure 7. Trends in the accuracy of different CNN models on the test set.
Sensors 19 04161 g007
Figure 8. Trend graph of the loss function (a) and confusion matrix (b). (1) Apple healthy (AH); (2) Apple Scab general (ASG); (3) Apple Scab serious (ASS); (4) Apple Frogeye Spot (AFS); (5) Cedar Apple Rust genera (CARG)l; (6) Cedar Apple Rust serious (CARS); (7) Cherry healthy (CH); (8) Cherry Powdery Mildew general (CPMG); (9) Cherry Powdery Mildew serious (CPMS); (10) Corn healthy (CH).
Figure 8. Trend graph of the loss function (a) and confusion matrix (b). (1) Apple healthy (AH); (2) Apple Scab general (ASG); (3) Apple Scab serious (ASS); (4) Apple Frogeye Spot (AFS); (5) Cedar Apple Rust genera (CARG)l; (6) Cedar Apple Rust serious (CARS); (7) Cherry healthy (CH); (8) Cherry Powdery Mildew general (CPMG); (9) Cherry Powdery Mildew serious (CPMS); (10) Corn healthy (CH).
Sensors 19 04161 g008
Figure 9. Visualization of feature map from each layer for a sample leaf. (1) conv1_1 (see Table 2), (2) conv3_1, (3) conv5_1, (4) inception_ 1 × 1, (5) inception_ 3 × 3, (6) inception_ 5 × 5, (7) inception_ pool, (8) pool7.
Figure 9. Visualization of feature map from each layer for a sample leaf. (1) conv1_1 (see Table 2), (2) conv3_1, (3) conv5_1, (4) inception_ 1 × 1, (5) inception_ 3 × 3, (6) inception_ 5 × 5, (7) inception_ pool, (8) pool7.
Sensors 19 04161 g009
Table 1. Statistics of the plant leaf disease dataset.
Table 1. Statistics of the plant leaf disease dataset.
ClassNumber of Training Images (Before Data Augmentation)Number of Training Images (After Data Augmentation)Number of Testing Images
Apple healthy11851185169
Apple Scab general84484430
Apple Scab serious59659622
Apple Frogeye Spot42742761
Cedar Apple Rust general14214220
Cedar Apple Rust serious4016011
Cherry healthy59859885
Cherry Powdery Mildew general16264835
Cherry Powdery Mildew serious15361233
Corn healthy37637654
Total45235588520
Table 2. Related parameters of the convolutional neural network (CNN)-based model.
Table 2. Related parameters of the convolutional neural network (CNN)-based model.
TypeSize/StrideOutput Size
Conv1 (Convolutional layer 1)3 × 3/164 × 224 × 224
Pool1/max3 × 3/164 × 112 × 112
Conv23 × 3/1128 × 112 × 112
Pool2/max3 × 3/1128 × 56 × 56
Conv33 × 3/1256 × 56 × 56
Pool3/max3 × 3/1256 × 28 × 28
Conv43 × 3/1512 × 28 × 28
Pool4/max3 × 3/1512 × 14 × 14
Conv53 × 3/1512 × 14 × 14
Pool5/max3 × 3/1512 × 7 × 7
Pool6/max3 × 3/1512 × 3 × 3
Inception-256 × 3 × 3
Pool7/ave3 × 3/1256 × 1 × 1
Dropout-256 × 1 × 1
Linear-10 × 1 × 1
Softmax-10
Table 3. Classification accuracy (in percent) comparison of CNN-Based models.
Table 3. Classification accuracy (in percent) comparison of CNN-Based models.
CNNAccuracyModel SizeTraining Time
AlexNet0.894217 MB1140.15 s
GoogLeNet0.89847.1 MB332.228 s
VGG160.905537.2 MB1960.2 s
VGG190.903558.4 MB5411.31 s
ResNet-500.90194.3 MB2101.19 s
Inceptionv20.90345.1 MB2187.3 s
Inceptionv30.90187.4 MB6438.72 s
Inceptionv40.89165 MB5787.9 s
SENet0.875220.8 MB1794.78 s
Our method0.91757.3 MB961.1 s
Table 4. Comparison of training time for different CNN models.
Table 4. Comparison of training time for different CNN models.
CNNForward PassBackward PassTotal Time
AlexNet0.052 s0.061 s1140.15 s
GoogLeNet0.013 s0.019 s332.228 s
VGG160.047 s0.014 s1960.2 s
VGG190.167 s0.371 s5411.31 s
ResNet-500.096 s0.114 s2101.19 s
Inceptionv20.102 s0.117 s2187.3 s
Inceptionv30.301 s0.342 s6438.72 s
Inceptionv40.258 s0.321 s5787.9 s
SENet0.116 s0.179 s1794.78 s
Our method0.038 s0.053 s961.1 s

Share and Cite

MDPI and ACS Style

Hang, J.; Zhang, D.; Chen, P.; Zhang, J.; Wang, B. Classification of Plant Leaf Diseases Based on Improved Convolutional Neural Network. Sensors 2019, 19, 4161. https://doi.org/10.3390/s19194161

AMA Style

Hang J, Zhang D, Chen P, Zhang J, Wang B. Classification of Plant Leaf Diseases Based on Improved Convolutional Neural Network. Sensors. 2019; 19(19):4161. https://doi.org/10.3390/s19194161

Chicago/Turabian Style

Hang, Jie, Dexiang Zhang, Peng Chen, Jun Zhang, and Bing Wang. 2019. "Classification of Plant Leaf Diseases Based on Improved Convolutional Neural Network" Sensors 19, no. 19: 4161. https://doi.org/10.3390/s19194161

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop