Next Article in Journal
Land-Streamer vs. Conventional Seismic Data for High-Resolution Near-Surface Surveys
Next Article in Special Issue
Density Estimates as Representations of Agricultural Fields for Remote Sensing-Based Monitoring of Tillage and Vegetation Cover
Previous Article in Journal
Application of Geoinformation Systems for Assessment of Effective Integration of Renewable Energy Technologies in the Energy Sector of Ukraine
Previous Article in Special Issue
Guava Disease Detection Using Deep Convolutional Neural Networks: A Case Study of Guava Plants
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cucumber Leaf Diseases Recognition Using Multi Level Deep Entropy-ELM Feature Selection

1
Department of Computer Science, HITEC University Taxila, Taxila 47080, Pakistan
2
College of Computer Engineering and Sciences, Prince Sattam bin Abdulaziz University, Al-Kharj 16273, Saudi Arabia
3
Department of Computer & Software Engineering, CEME NUST Rawalpindi, Rawalpindi 46000, Pakistan
4
Institute of Information Technology, Quaid-i-Azam University, Islamabad 44000, Pakistan
5
Department of Computer Science & Engineering, Ewha Womans University, Seoul 03760, Korea
6
Department of Computer Science, Hanyang University, Seoul 04763, Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(2), 593; https://doi.org/10.3390/app12020593
Submission received: 19 December 2021 / Revised: 4 January 2022 / Accepted: 5 January 2022 / Published: 7 January 2022
(This article belongs to the Special Issue Sustainable Agriculture and Advances of Remote Sensing)

Abstract

:
Agriculture has becomes an immense area of research and is ascertained as a key element in the area of computer vision. In the agriculture field, image processing acts as a primary part. Cucumber is an important vegetable and its production in Pakistan is higher as compared to the other vegetables because of its use in salads. However, the diseases of cucumber such as Angular leaf spot, Anthracnose, blight, Downy mildew, and powdery mildew widely decrease the quality and quantity. Lately, numerous methods have been proposed for the identification and classification of diseases. Early detection and then treatment of the diseases in plants is important to prevent the crop from a disastrous decrease in yields. Many classification techniques have been proposed but still, they are facing some challenges such as noise, redundant features, and extraction of relevant features. In this work, an automated framework is proposed using deep learning and best feature selection for cucumber leaf diseases classification. In the proposed framework, initially, an augmentation technique is applied to the original images by creating more training data from existing samples and handling the problem of the imbalanced dataset. Then two different phases are utilized. In the first phase, fine-tuned four pre-trained models and select the best of them based on the accuracy. Features are extracted from the selected fine-tuned model and refined through the Entropy-ELM technique. In the second phase, fused the features of all four fine-tuned models and apply the Entropy-ELM technique, and finally fused with phase 1 selected feature. Finally, the fused features are recognized using machine learning classifiers for the final classification. The experimental process is conducted on five different datasets. On these datasets, the best-achieved accuracy is 98.4%. The proposed framework is evaluated on each step and also compared with some recent techniques. The comparison with some recent techniques showed that the proposed method obtained an improved performance.

1. Introduction

Agriculture is one of the most important research topics globally nowadays [1]. Agriculture is a significant source of income and the economy of a country is based on the quality and yields of crops [2]. Cucumber is an important vegetable and during the year 2020, the global cucumber planting area was around 2.25 million hectares and the global yield was 90.35 million tons [3]. The production of crops is highly threatened by diseases and failure to identify and prevent cucumber diseases causes a reduction of cucumber vegetable yield and quality. The failure in early diagnosis causes significant economic losses to growers. Therefore, the rapid diagnosis of crops diseases helps to increase the quality and yield and also increases the national economy [4].
Mostly, identification is accomplished using typical methods like seeing through naked eyes or through a microscope [5]. The results of the manual visual estimation are generally unreliable while the microscopic assessments are generally time-consuming and costly. Most of the agriculturists of underdeveloped countries are illiterate [6]. They are compelled to return those charges along with other expenses like pesticides and fertilizer. The cucumber diseases like anthracnose, powdery mildew, downy mildew, and cucumber mosaic can destroy a large number of crops and the result will be a huge loss and vegetable deficiency [7]. Significant work has been done to accomplish a method that can boost the fastness and accuracy of the process. The methods necessarily contained some sort of computerization [8]. A large number of techniques presented until now are based on digital image processing and machine learning to identify the crops' diseases and achieve the desired output [9].
Image processing has many applications in the domain of computer vision such as medical imaging [10], agriculture [11], and named a few more. Agriculture is a hot application of image processing for the identification and classification of crops and plant diseases [12]. Although detection of cucumber abnormalities and then classifying them using image processing techniques is a critical task due to some sequence of steps [13]. A computerized method consists of some important steps such as preprocessing of original leaf images, detection of the infected region, feature extraction using handcrafted methods, and finally reduction and classification. Recognition of diseased portions in images is the key factor as it can influence the design and performance of the classification algorithms [14]. However, the error in the detection of the infected region extracted the irrelevant features that reduces the recognition accuracy.
Deep learning (DL) [15] is a hot research topic nowadays [16] and is employed everywhere for the detection and recognition tasks for several applications [17] such as biometric [18], image classification [19], surveillance [20], medical [21], and agriculture [22]. The researcher of computer vision introduced many techniques using machine learning and deep learning for plants diseases recognition [23]. Hussain et al. [24] introduced a deep learning technique for the identification of multiple cucumber leaf diseases. They extract deep learning features through two fine-tuned deep models including VGG19 and Inception V3. Both fine-tuned models were trained on the selected dataset using the transfer learning approach. The main advantage of training through TL is to save memory and time. The extracted features were fused by implementing the parallel maximum fusion technique to get the maximum information of each trained image. In the end, a whale optimization algorithm (WOA) was applied to select the robust features and perform classification. The purpose of feature selection is to get the best features because, in the fusion process, a few redundant features were also added. They achieved a maximum of 96.5% accuracy on the selected leaf dataset. In [13], researchers built an automated detection classification model for cucumber leaf diseases. In the first phase, pre-processing was performed to enhance the local contrast of images and to make the infected region more visible. This step makes the infected region clearer that later helped in the accurate segmentation using a novel Sharif saliency-based (SHSB) technique. Then researchers fused the proposed saliency method with active contour segmentation to improve the segmentation accuracy that later extracts the relevant features. In the feature extraction phase, they utilized VGG-19 and VGG-M pre-trained models. The extracted features were refined through three parameters including local entropy, local interquartile range, and local standard deviation. In the final classification, the best accuracy of 98.08% was achieved on multi-class SVM. The strength of this work was less computational time that can be useful for a real-time computerized system. Wang et al. [3] introduced a deep learning-based technique for the recognition of cucumber leaf diseases under complex backgrounds. They fused DeepLabV3+ and U-Net models instead of a single network. In the first step, DeepLabV3+ was used to segment the leaves from the images. Then the diseased area was segmented using U-Net. The fused models give better accuracy than the accuracy reported by the individual models. Researchers in [25], introduced a model for the identification of crop diseases in real-world images. The proposed trilinear convolutional neural network utilized bilinear pooling. In the laboratory environment, the proposed technique achieved 99.99% accuracy and in the real-world environment, the obtained accuracy is 84.11%. Kianat et al. [7] proposed a hybrid system for the recognition of cucumber diseases. In the pre-processing step, the data augmentation was applied using different angles to increase the image count in the dataset. In this step, contrast stretching was also performed to visually improve the images. The features were extracted from binary robust invariant scalable keypoints (BRISK), histogram of gradient (HOG), and features from the accelerated segmented test (FAST). Initially, the irrelevant features were eliminated by utilizing the probability distribution-based entropy (PDbE) technique. Then features were fused using the serial-based method and implemented Manhattan distance-controlled entropy (MDcE) method was to select the robust features. The proposed model achieved maximum accuracy of 93.5%. These techniques faced a major challenge of irrelevant feature extraction that were tried to be resolved through feature selection techniques [26].
Visual inspection of crops was carried out by farmers and agriculture experts. This evaluation process is exhausting, time-consuming, and highly subjective. The development of computer vision systems to identify, recognize, and classify disease-affected crops will keep humans out of the equation, allowing for unbiased, accurate disease-infection decisions [1]. An automatic classification system consists of various steps as mentioned above. Preprocessing is an important step, the aim is to remove noise and improve the quality of original images that later helps in important feature extraction. The extracted features from the refined images are used for the training of deep learning models that are further employed for feature extraction and classification. The key problems which are considered in this work are (i) training a deep learning model on an imbalanced dataset gives the high priority in the prediction to higher numbers of sample class; (ii) disease spots and background objects differ in appearance; (iii) changes in the shape, color, texture, and origin of the disease; (iv) irrelevant and redundant features extraction, and (v) choosing the superlative features for the classification.
In this article, our major focus is to design an automated computerized method for cucumber leaf diseases recognition using deep learning and Entropy-ELM-based best feature selection. The recent methods focused on the infected region identification and then employed for feature extraction; however, the error in the identification step misleads the irrelevant feature extraction that later reduces the classification accuracy. Our major contributions are:
(i)
Four mathematical functions such as horizontal flip, vertical flip, rotate 45, and rotation 60 are implemented for the sake of data augmentation. Later, four deep learning models are fine-tuned and trained on the augmented dataset.
(ii)
Deep learning features are extracted from the average pooling layer instead of the fully connected layer. The extracted deep features are passed to the Softmax classifier and compared the accuracy. Based on the accuracy value, the Densenet201 fine-tuned model is selected for the rest of the process. Moreover, all fine-tuned model features are fused using a new parallel approach.
(iii)
An Entropy-ELM based best feature selection technique is proposed. The proposed technique is applied on both the Densenet201 feature vector and fused vector, that later serially fused for the final classification.
(iv)
To determine which step of the proposed framework is better performed, a comparison is made between all hidden steps.
The rest of the manuscript is organized as follows: a proposed methodology that includes augmentation of the dataset, deep learning-based feature extraction, and Entropy-ELM-based best feature selection, is presented in Section 2. Results are discussed in Section 3 with the help of tables and graphs. Finally, the conclusion of the manuscript is given in Section 4.

2. Proposed Methodology

In this work, an automated framework is proposed for cucumber leaf diseases recognition using deep learning and Entropy-ELM-based best feature selection. The proposed framework is illustrated in Figure 1. In this figure, it is shown that the initial augmentation step is applied to the original images by creating more training data. Then two different phases are utilized. In the first phase, four pre-trained deep models are fine-tuned and selected the best of them based on the accuracy. Features are extracted from the selected fine-tuned model and refined through the Entropy-ELM technique. In the second phase, fused the features of all four fine-tuned models and apply the Entropy-ELM technique, and finally fused with phase 1 selected feature. Finally, the fused features are classified using machine learning classifiers for the final output.

2.1. Dataset Collection and Augmentation

The experiments were performed on the publically available dataset named the Cucumber leaf diseases scan dataset [27]. This dataset consists of six different diseases such as anthracnose, powdery mildew, downy mildew, angular spot, mosaic, and blight. A sample of images are illustrated in Figure 2. Each class has 100 to 150 images originally that are not enough to train a deep learning model. Therefore, we design a simple algorithm (Algorithm 1) for data augmentation that includes four operations such as horizontal flip, vertical flip, rotate 45, and rotate 60. This algorithm is applied to each cucumber disease class and increases the number of images to 2000 in each class. In the later steps, this augmented dataset is utilized for the training of deep models.
Algorithm 1: (Data Augmentation)
Step 1: Input Original Database
Step 2: Consider First Disease Class
Step 3: Count Images of Step 2 (Selected Disease Class)
Step 4: For i = 1 to Total Images of each Class
-
Horizontal Flip and Image Write
-
Vertical Flip and Image Write
-
Rotate 45 and Image Write
-
Rotate 60 and Image Write
Step 5: Repeat Step 2, 3, and 4 for the Rest of the Disease Classes
      End

2.2. Deep Learning Architecture

Four deep learning pre-trained models are employed in this work for feature extraction. The selected models are—VGG16, ResNet50, ResNet101, and DenseNet201. As mentioned in Figure 1, all selected models are initially fine-tuned and then trained through transfer learning using an augmented dataset. A brief description of each deep model is given below.
VGG16 [28] is a pre-trained model that was created by the Visual Geometry Group. This group is a combination of students and teachers focused on Computer Vision at Oxford University. This model is reflected to be one of the best computer vision models in the world. A unique feature of VGG16 is that rather than having numerous hyper-parameters it concentrates on having used identical PL and MPL of 2 × 2 filters of stride 2 and CL of 3 × 3 filters with a stride 1. VGG16 continues the same organization containing Convolutional and Maxpool Layers continuously during the course of the entire structural design. In the end, VGG has 2 Fully Connected Layers afterward a Softmax to output. Due to the fact that the VGG16 has 16 layers with weights, it has the name VGG16. This model was originally trained on an ImageNet dataset having 1000 object classes. The prediction of this model was done by the Softmax layer, defined as:
Θ   =   w 0 x 0 + w 1 x 1 + + w k x k   =   i = 0 k w i x i = w T x
ResNet [29] also known as a Deep Residual Network, have proved to perform with great accuracy and efficiency with a Deep Framework and to create an extra straight pathway for the transmission of data through the network. Within such Deep Systems, the deprivation issue arises because of the rise of Network Layers and the precision begins to dilute which results in its reduction quickly. Backpropagation does not come across the Vanishing Gradient problem when working with RESNET. There are some "Shortcut Connections" that a Residual Network has which are to be equivalent to a regular Convolutional Layer which aids the network to comprehend the Global Features. Then an input x has to be added to the output layer by adding the Shortcut connection, afterward some weight layers below. After the application of these Shortcut Connections, they permitted the network by avoiding the layers which were not beneficial while training. Hence, the output came in an ideal modification of the number of layers to perform rapid training. Mathematical, the output of H   x can be expressed as
H x   = F   x x
A type of Residual Mapping is used to train the weight layers which is expressed as,
  F x   = H   x x
The above-mentioned function F x signifies stacked nonlinear weight layers. Several properties of ResNet50 include the fact that it has 64 kernels including 7 × 7 Convolutional layers. It also includes 16 residual blocks. There are 23 million trainable parameters.
ResNet101 model utilizes Residual links that the angles can stream straightforwardly over to hinder the slopes to get 0 after the utilization of Chain Rule. There are 104 convolutional layers altogether in ResNet101. Alongside, it comprises 33 squares of layers altogether and 29 of these squares utilize past squares yield straightforwardly which is characterized as leftover associations above. Hence the above-mentioned residuals were using such main Operand of Summation (OOS) administrator towards the termination of every square to obtain the contribution of the accompanying squares. Leftover 4 squares get the past square’s yield and apply it to a CL with a channel size of 1 × 1 and a step of 1 after a clump standardization layer, which performs standardization activity and the resultant yield is shipped off the summation administrator at the yield of that block. Mathematically, this model working is defined as follows:
u ( x , t + 1 ) = u ( x , t ) + w ( x , t ) u ( x , t )
T x = 1 2 σ 2 2 x 2 + b x + c T p = 1 2 σ 2 p 2 + i b p + c
T p u ˜ ( p , t ) = d d t u ˜ ( p , t )
u ˜ ( p , t ) = e T p t u ˜ ( p ,   0 )
u ˜ ( p , t ) ( 1 + T p t ) u ˜ ( p ,   0 )
Densenet-201 [30] is a convolutional neural network that is 201 layers deep. In this model, each layer gets feature maps from all preceding layers, the network can be thinner and more compact, resulting in fewer channels. The extra number of channels for each layer is the growth rate k. As a result, it has better computational and memory efficiency. The transition layers between two contiguous dense blocks are 11 Conv followed by 22 average pooling. Within the dense block, feature map sizes are uniform, allowing them to be readily concatenated. A global average pooling is done after the last dense block, and then a softmax classifier is added. The error signal can be transmitted more directly to earlier levels. As previous layers can get direct supervision from the final classification layer, this is a form of implicit deep supervision.

2.3. Transfer Learning Based Feature Extraction

Transfer learning (TL) is a process of reusing a pre-trained model for a new task [31], as illustrated in Figure 3. The ImageNet dataset was used as a source dataset of the pre-trained model. The pre-trained model is fine-tuned and transfer knowledge through the TL concept. In the last, the new fine-tuned model is trained on the augmented cucumber dataset that is utilized for further feature extraction. The features are extracted from the deep layers like FC7 for VGG, Average Pool for ResNet50, ResNet101, and Densenet201. Several hyperparameters are employed during the training process such as 0.0001 learning rate, max epochs are 200, the mini-batch size is 16, and the activation function is sigmoid.

2.4. Entropy-ELM Based Features Selection and Parallel Fusion

Feature selection is an important and hot research topic nowadays [32]. The main purpose of feature selection is to increase the system accuracy and minimize the computational time by focusing on the selection of the most important features [33]. In this work, a new technique is proposed named Entropy-ELM for the best feature selection. This proposed technique worked in the following steps: (i) compute the entropy of input vector; (ii) based on the entropy value, a threshold function is employed that return two vectors—fulfill the threshold value (selected) and not-selected; (iii) ELM [34] employed as a fitness function and selected threshold passed features are utilized as an input. Mathematically, the entropy formulation is defined as follows:
H 1 = k = 1 G P k I d P k
H d i f f , 1 = 0 1 h 1 w 1 d h 1 w d w .
H 1 = H d i f f , 1 + Id ( G ) =   H d i f f , 1 + H 1 , m a x
w h o l e   i m a g e [ Δ w w , y ] k d x d y   1 1 ( Δ w ) k h d Δ w d Δ w = M k    
T =   S e l k                           f o r     F e a t u r e s i   H 1 N o t S e l e c   l         f o r   F e a t u r e s i   < H 1
The detail of this selection process is given in Algorithm 2.
Algorithm 2: (Entropy-ELM)
Step 1: Input Feature Vector N × K  // K is the length of features
Step 2: For i = 1 to N
Step 3: Computer Entropy through Equations (9)–(12)
Step 4: Define Threshold Function as Equation (13)
Step 5: Check Fitness through ELM
Step 6: Evaluate the Accuracy
Step 7: Repeat Step 2–6, until accuracy on the top side
      End
Selected Feature Vector
Finally, the parallel fusion approach is opted to get the fused feature vector. This approach is based on the following three steps. In the first step, get the maximum length feature vector. As we have two feature vectors X and X 1 , where the length of vectors is N × K and N × K 1 , respectively. In the second step, compute the entropy value and perform padding for the lower size feature vector. In the third step, correlation is computed among K and K 1 features for the final fusion. The fused vector is finally utilized for the classification through supervised learning classifiers.
F u s i o n = ψ K , K 1
where K and K 1 X and X 1

3. Experimental Results

The proposed framework is evaluated on the selected cucumber dataset having a ratio of 70:15:15 which means that 70% of the images are utilized to train the model, whereas the 15% for testing and 15% for validation. We combined the testing and validation images and performed testing (30%). All the experimental results are computed with K-Fold cross-validation, whereas the value of K is 10. Several classifiers are implemented as discussed in Table 1. The performance of each classifier is computed through several measures such as recall rate, precision rate, F1-Score, accuracy, and time. The entire framework simulations are conducted on Simulink MATLAB2021a using a Personal Desktop.

3.1. Results

The detailed experimental process of the proposed framework is conducted in this section. The results are computed using the following steps: (i) classification using originally collected dataset on fine-tuned pre-trained models; (ii) classification using augmented dataset on fine-tuned deep models and select the best deep model for the further processing; (iii) best deep model features are refined using a new technique name Entropy-ELM; (iv) fusion of fine-tuned deep model features (augmented dataset), and (v) fused both step features using a parallel approach

3.2. Results on Original Cucumber Dataset

The results of the proposed method on the original cucumber dataset are given in Table 2. In this table, accuracy is computed for each fine-tuned deep model using the original dataset. Fine-tuned VGG16 (F-VGG16) obtained the maximum accuracy of 56.9% on the MG SVM classifier. The fine-tuned ResNet50 and ResNet101 obtained the best accuracy of 58.7 and 55.1% on Cubic SVM and Quadratic SVM, respectively. The fine-tuned Densenet201 deep model obtained an accuracy of 61.9% on Quadratic SVM. Based on these results, it is noticed that the originally collected dataset have several issues like imbalancing and short training data. Using these data, the fine-tuned Densenet201 gives better results for all classifiers.

3.3. Results on Augmented Cucumber Dataset

Experimental results of fine-tuned VGG16 pre-trained model after augmentation are given in Table 3. The best-obtained accuracy is 93.8% on Cubic SVM, whereas the recall rate and precision rates are 93.84 and 93.92%, respectively. The second best-obtained accuracy is 93.6%, which was accomplished on Quadratic SVM, whereas the recall rate and precision rates are 93.66 and 93.72%, correspondingly. The execution time of Linear SVM is better than the rest of the classifiers.
The classification accuracy of fine-tuned ResNet50 on the augmented dataset is given in Table 4. This table presents the highest obtained accuracy on Cubic SVM of 94.6%, whereas the recall and precision rates are 94.36 and 94.46%, correspondingly. The second top accuracy is 94.4% obtained on Quadratic SVM, whereas the recall and precision rates are 94.26 and 94.36%, respectively. Similar to fine-tuned VGG16, the Quadratic SVM executed fast than the rest of the classifiers.
Experimental results of fine-tuned ResNet101 pre-trained model are given in Table 5. The best-obtained accuracy of 97.7% was accomplished on Cubic SVM. The recall and precision rates are 97.7 and 97.7%, correspondingly. The second best-obtained accuracy is 97.2% on Quadratic SVM. The recall and precision rates are 97.24 and 97.32%, correspondingly. In this experiment, the Linear SVM was executed fast than the rest of the selected classifiers.
The classification results of fine-tuned Densenet201 pre-trained model are given in Table 6. In this table, the obtained best accuracy is 98.4% on Cubic SVM. Moreover, the recall and precision rates are 98.44 and 98.5%, correspondingly. Figure 4 illustrated the confusion matrix of Cubic SVM that was utilized for the verification of recall rate. The second best-obtained accuracy is 97.4%, which was accomplished on Quadratic SVM. The computation time of each classifier is also noted and the minimum time is 302 (sec) for LSVM. At the first step comparison among without augmented and augmented datasets, it is noted that the accuracy obtained on the augmented dataset is significantly better. In the second step comparison, it is noted that the fine-tuned DenseNet201 model achieved better results than VGG16, ResNet50, and ResNet101. Based on this analysis, the fine-tuned DenseNet201 is selected for the rest of the experiments.
The fine-tuned deep learning model is selected based on the better accuracy and applied proposed Entropy-ELM feature selection technique. The results are given in Table 7. This presents the best accuracy of 98% on Cubic SVM. The other computed measures are the recall rate which is 98.02, the precision rate at 97.98, and the F1-Score at 98%. The recall rate of Cubic SVM can be also verified through a confusion matrix, illustrated in Figure 5. Compared to the results given in Table 6, it is noted that the accuracy is a bit reduced but on the other side, a huge change occurred in the computation time. The time is also plotted in Figure 6 (FDenseNet201 and Dense Entropy-ELM).
After the selection of the best dense features, in the next step all fine-tuned deep model features are fused using the proposed parallel approach. The results of this experiment are given in Table 8. The best-noted accuracy in this table is 98.2% on Cubic SVM. The recall and precision rates are 97.92 and 98.12%, respectively. Figure 7 illustrated the confusion matrix that can be utilized for the verification of the recall rate. The time of each classifier is also noted and plotted in Figure 6 (Fusion Entropy-ELM). In comparison with the results of Table 6 and Table 7, it is noted that the overall accuracy is improved but the time is more increased than in the Dense Entropy-ELM step.
Finally, the features of Dense Entropy-ELM and Fusion Entropy-ELM are fused using the proposed parallel approach, and the results are given in Table 9. This table presents the best-obtained accuracy of 98.50% on Cubic SVM. The noted precision rate is 98.30, recall rate is 98.36 and F1-Score is 98.48%, respectively. The second best-noted accuracy is 97.5% on Quadratic SVM. The recall rate of Cubic SVM can be verified through a confusion matrix plotted in Figure 8. This figure shows the correct prediction rate of each class in the diagonal. Compared to the results of this experiment with all previous experiments, it is clearly noted that the accuracy is improved and computational time is significantly reduced.

3.4. Discussion

Figure 1 showed the proposed framework that includes a few important steps. This figure illustrated the importance of the data augmentation step. The results without data augmentation having less accuracy than the results obtained after the data augmentation. Moreover, the selection of important features improves the accuracy that is later fused through a parallel approach. This step not only improves the classification accuracy but also reduced the computational time, as plotted in Figure 6. This figure clearly shows that the final fusion step significantly reduced the computational time than the rest of the steps on all classifiers.
In the last, the proposed framework accuracy is compared with recent SOTA techniques, as given in Table 10. The methods mentioned in this table are from the year 2017–2022. Moreover, all the methods mentioned in this table used the same leaf dataset. The recent best accuracy was 98.08% and 96.50% achieved by Khan et al. [13] and Hussain et al. [24]. The other methods such as Lin et al. [35] achieved an accuracy of 96.08% on the same dataset. The proposed framework achieved an accuracy of 98.48% that is improved than the SOTA techniques.

4. Conclusions

Agriculture is a hot topic of research nowadays. In agriculture, deep learning showed significant success from the last decade for the recognition of plant diseases. In this article, a deep learning and Entropy-ELM based framework is proposed for the recognition of cucumber leaf diseases. In the proposed framework, four pre-trained deep models are trained and selected one of them based on the accuracy that is later employed for the selection of best features using the proposed Entropy-Elm technique. In the opposite step, features of all pre-trained models are fused and apply the feature selection technique. In the last, features of both steps are fused and perform classification. The proposed framework is tested on an augmented cucumber leaf dataset and achieved an accuracy of 98.48%. Comparison with the existing techniques showed the proposed framework obtained improved results. From the results, it is concluded that the augmentation process improves the recognition accuracy but also increases the time that was the first limitation of this framework; therefore a feature selection technique is proposed to maintain the accuracy and reduce the computational time. Through feature selection and fusion process, important information is obtained that later improves the classification accuracy. Another limitation of this work was the reduction of a few features that were ignored during the selection process. In the future, EfficientNet deep model will be implemented and features will be refined through the Butterfly metaheuristic algorithm instead of the heuristic search approach [20]. Moreover, reinforcement learning and Graph CNN shall be applied and refined through feature selection algorithms for the better results [38,39,40,41,42].

Author Contributions

Conceptualization, M.A.K., A.A., A.K., S.A., A.B., M.M.I.C., H.-S.Y. and J.C.; methodology, M.A.K., A.A., A.K., S.A., A.B., M.M.I.C., H.-S.Y. and J.C.; software, M.A.K., A.A., A.K., S.A., A.B., M.M.I.C., H.-S.Y. and J.C.; validation, M.A.K., A.A., A.K., S.A., A.B., M.M.I.C., H.-S.Y. and J.C.; formal analysis, M.A.K., A.A., A.K., S.A., A.B., M.M.I.C., H.-S.Y. and J.C.; investigation, M.A.K., A.A., A.K., S.A., A.B., M.M.I.C., H.-S.Y. and J.C.; resources, M.A.K., A.A., A.K., S.A., A.B., M.M.I.C., H.-S.Y. and J.C.; data curation, M.A.K., A.A., A.K., S.A.,A.B., M.M.I.C., H.-S.Y. and J.C.; writing and original draft preparation, M.A.K., A.A., A.K., S.A., A.B., M.M.I.C., H.-S.Y. and J.C.; writing, review and editing, M.A.K., A.A., A.K., S.A., A.B., M.M.I.C., H.-S.Y. and J.C.; visualization, M.A.K., A.A., A.K., S.A., A.B., M.M.I.C., H.-S.Y. and J.C.; supervision, M.A.K., A.A., A.K., S.A., A.B., M.M.I.C., H.-S.Y. and J.C.; project administration, M.A.K., A.A., A.K., S.A., A.B., M.M.I.C., H.-S.Y. and J.C.; funding acquisition, M.A.K., A.A., A.K., S.A., A.B., M.M.I.C., H.-S.Y. and J.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korean government (Ministry of Science and ICT; MSIT) under Grant RF-2018R1A5A7059549.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Iqbal, Z.; Sharif, M.; Shah, J.H.; ur Rehman, M.H.; Javed, K. An automated detection and classification of citrus plant diseases using image processing techniques: A review. Comput. Electron. Agric. 2018, 153, 12–32. [Google Scholar] [CrossRef]
  2. Agarwal, M.; Gupta, S.; Biswas, K. A new conv2d model with modified relu activation function for identification of disease type and severity in cucumber plant. Sustain. Comput. Inform. Syst. 2021, 30, 100473. [Google Scholar] [CrossRef]
  3. Wang, C.; Du, P.; Wu, H.; Li, J.; Zhao, C.; Zhu, H. A cucumber leaf disease severity classification method based on the fusion of DeepLabV3+ and U-Net. Comput. Electron. Agric. 2021, 189, 106373. [Google Scholar] [CrossRef]
  4. Sanju, S.K.; VelammaL, D.B. An Automated Detection and Classification of Plant Diseases from the Leaves Using Image processing and Machine Learning Techniques: A State-of-the-Art Review. Ann. Rom. Soc. Cell Biol. 2021, 25, 15933–15950. [Google Scholar]
  5. Pourdarbani, R.; Sabzi, S.; Rohban, M.H.; Hernández-Hernández, J.L.; Gallardo-Bernal, I.; Herrera-Miranda, I.; García-Mateos, G. One-Dimensional Convolutional Neural Networks for Hyperspectral Analysis of Nitrogen in Plant Leaves. Appl. Sci. 2021, 11, 11853. [Google Scholar] [CrossRef]
  6. Tahir, M.B.; Javed, K.; Kadry, S.; Zhang, Y.-D.; Akram, T.; Nazir, M. Recognition of apple leaf diseases using deep learning and variances-controlled features reduction. Microprocess. Microsyst. 2021, 104027. [Google Scholar] [CrossRef]
  7. Kianat, J.; Sharif, M.; Akram, T.; Rehman, A.; Saba, T. A joint framework of feature reduction and robust feature selection for cucumber leaf diseases recognition. Optik 2021, 240, 166566. [Google Scholar] [CrossRef]
  8. Tariq, U.; Hussain, N.; Nam, Y.; Kadry, S. An Integrated Deep Learning Framework for Fruits Diseases Classification. Comput. Mater. Contin. 2022. [Google Scholar] [CrossRef]
  9. Khirade, S.D.; Patil, A. Plant disease detection using image processing. In Proceedings of the 2015 International Conference on Computing Communication Control and Automation, Pune, India, 26–27 February 2015; pp. 768–771. [Google Scholar]
  10. Rehman, A.; Naz, S.; Khan, A.; Zaib, A.; Razzak, I. Improving coronavirus (COVID-19) diagnosis using deep transfer learning. MedRxiv 2020. [Google Scholar] [CrossRef] [Green Version]
  11. Yasmeen, U.; Tariq, U.; Khan, J.A.; Yar, M.A.E.; Hanif, C.A.; Mey, S.; Nam, Y. Citrus Diseases Recognition Using Deep Improved Genetic Algorithm. Comput. Mater. Contin. 2021, 73, 1–15. [Google Scholar] [CrossRef]
  12. Ramesh, S.; Hebbar, R.; Niveditha, M.; Pooja, R.; Shashank, N.; Vinod, P. Plant disease detection using machine learning. In Proceedings of the 2018 International Conference on Design Innovations for 3Cs Compute Communicate Control (ICDI3C), Bangalore, India, 25–28 April 2018; pp. 41–45. [Google Scholar]
  13. Akram, T.; Sharif, M.; Javed, K.; Raza, M.; Saba, T. An automated system for cucumber leaf diseased spot detection and classification using improved saliency method and deep features selection. Multimed. Tools Appl. 2020, 79, 18627–18656. [Google Scholar]
  14. ur Rehman, Z.; Ahmed, F.; Damaševičius, R.; Naqvi, S.R.; Nisar, W.; Javed, K. Recognizing apple leaf diseases using a novel parallel real-time processing framework based on MASK RCNN and transfer learning: An application for smart agriculture. IET Image Process. 2021, 15, 2157–2168. [Google Scholar] [CrossRef]
  15. Razzak, M.I.; Imran, M.; Xu, G. Efficient brain tumor segmentation with multiscale two-pathway-group conventional neural networks. IEEE J. Biomed. Health Inform. 2018, 23, 1911–1919. [Google Scholar] [CrossRef]
  16. Feizizadeh, B.; Mohammadzade Alajujeh, K.; Lakes, T.; Blaschke, T.; Omarzadeh, D. A comparison of the integrated fuzzy object-based deep learning approach and three machine learning techniques for land use/cover change monitoring and environmental impacts assessment. GISci. Remote Sens. 2021, 58, 1–28. [Google Scholar] [CrossRef]
  17. Feizizadeh, B.; Omarzadeh, D.; Kazemi Garajeh, M.; Lakes, T.; Blaschke, T. Machine learning data-driven approaches for land use/cover mapping and trend analysis using Google Earth Engine. J. Environ. Plan. Manag. 2021, 1–33. [Google Scholar] [CrossRef]
  18. Saleem, F.; Khan, M.A.; Alhaisoni, M.; Tariq, U.; Armghan, A.; Alenezi, F.; Choi, J.-I.; Kadry, S. Human gait recognition: A single stream optimal deep learning features fusion. Sensors 2021, 21, 7584. [Google Scholar] [CrossRef]
  19. Hussain, N.; Kadry, S.; Tariq, U.; Mostafa, R.; Choi, J.-I.; Nam, Y. Intelligent Deep Learning and Improved Whale Optimization Algorithm Based Framework for Object Recognition. Hum. Cent. Comput. Inf. Sci 2021, 11, 34. [Google Scholar]
  20. Khan, S.; Alhaisoni, M.; Tariq, U.; Yong, H.-S.; Armghan, A.; Alenezi, F. Human Action Recognition: A Paradigm of Best Deep Learning Features Selection and Serial Based Extended Fusion. Sensors 2021, 21, 7941. [Google Scholar] [CrossRef]
  21. Muhammad, K.; Sharif, M.; Akram, T.; Kadry, S. Intelligent fusion-assisted skin lesion localization and classification for smart healthcare. Neural Comput. Appl. 2021, 1–16. [Google Scholar] [CrossRef]
  22. Akram, T.; Sharif, M.; Alhaisoni, M.; Saba, T.; Nawaz, N. A probabilistic segmentation and entropy-rank correlation-based feature selection approach for the recognition of fruit diseases. EURASIP J. Image Video Process. 2021, 2021, 1–28. [Google Scholar]
  23. Zhang, S.; Wang, Z. Cucumber disease recognition based on Global-Local Singular value decomposition. Neurocomputing 2016, 205, 341–348. [Google Scholar] [CrossRef]
  24. Hussain, N.; Khan, M.A.; Tariq, U.; Kadry, S.; Yar, M.A.E.; Mostafa, A.M.; Alnuaim, A.A.; Ahmad, S. Multiclass Cucumber Leaf Diseases Recognition Using Best Feature Selection. Comput. Mater. Contin. 2022, 2022, 3281–3294. [Google Scholar] [CrossRef]
  25. Wang, D.; Wang, J.; Li, W.; Guan, P. T-CNN: Trilinear convolutional neural networks model for visual detection of plant diseases. Comput. Electron. Agric. 2021, 190, 106468. [Google Scholar] [CrossRef]
  26. Razzak, I.; Saris, R.A.; Blumenstein, M.; Xu, G. Integrating joint feature selection into subspace learning: A formulation of 2DPCA for outliers robust feature selection. Neural Netw. 2020, 121, 441–451. [Google Scholar] [CrossRef]
  27. Zhang, S.; Wu, X.; You, Z.; Zhang, L. Leaf image based cucumber disease recognition using sparse representation classification. Comput. Electron. Agric. 2017, 134, 135–141. [Google Scholar] [CrossRef]
  28. Qassim, H.; Verma, A.; Feinzimer, D. Compressed residual-VGG16 CNN model for big data places image recognition. In Proceedings of the 2018 IEEE 8th Annual Computing and Communication Workshop and Conference (CCWC), Las Vegas, NV, USA, 8–10 January 2018; pp. 169–175. [Google Scholar]
  29. Theckedath, D.; Sedamkar, R. Detecting affect states using VGG16, ResNet50 and SE-ResNet50 networks. SN Comput. Sci. 2020, 1, 1–7. [Google Scholar] [CrossRef] [Green Version]
  30. Wang, S.-H.; Zhang, Y.-D. DenseNet-201-based deep neural network with composite learning factor and precomputation for multiple sclerosis classification. ACM Trans. Multimed. Comput. Commun. Appl. 2020, 16, 1–19. [Google Scholar] [CrossRef]
  31. Brusa, E.; Delprete, C.; Di Maggio, L.G. Deep Transfer Learning for Machine Diagnosis: From Sound and Music Recognition to Bearing Fault Detection. Appl. Sci. 2021, 11, 11663. [Google Scholar] [CrossRef]
  32. Attique Khan, M.; Sharif, M.; Akram, T.; Kadry, S.; Hsu, C.H. A two-stream deep neural network-based intelligent system for complex skin cancer types classification. Int. J. Intell. Syst. 2021. [Google Scholar] [CrossRef]
  33. Khan, M.A.; Alhaisoni, M.; Tariq, U.; Hussain, N.; Majid, A.; Damaševičius, R.; Maskeliūnas, R. COVID-19 Case Recognition from Chest CT Images by Deep Learning, Entropy-Controlled Firefly Optimization, and Parallel Feature Fusion. Sensors 2021, 21, 7286. [Google Scholar] [CrossRef]
  34. Huang, G.-B.; Zhu, Q.-Y.; Siew, C.-K. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  35. Lin, K.; Gong, L.; Huang, Y.; Liu, C.; Pan, J. Deep learning-based segmentation and quantification of cucumber powdery mildew using convolutional neural network. Front. Plant Sci. 2019, 10, 155. [Google Scholar] [CrossRef] [Green Version]
  36. Ma, J.; Du, K.; Zheng, F.; Zhang, L.; Gong, Z.; Sun, Z. A recognition method for cucumber diseases using leaf symptom images based on deep convolutional neural network. Comput. Electron. Agric. 2018, 154, 18–24. [Google Scholar] [CrossRef]
  37. Zhang, J.; Rao, Y.; Man, C.; Jiang, Z.; Li, S. Identification of cucumber leaf diseases using deep learning and small sample size for agricultural Internet of Things. Int. J. Distrib. Sens. Netw. 2021, 17, 15501477211007407. [Google Scholar] [CrossRef]
  38. Yuan, H.; Tang, G.; Guo, D.; Wu, K.; Shao, X.; Yu, K.; Wei, W. BESS Aided Renewable Energy Supply using Deep Reinforcement Learning for 5G and Beyond. IEEE Trans. Green Commun. Netw. 2021, 11, 1–12. [Google Scholar] [CrossRef]
  39. Yu, K.; Tan, L.; Mumtaz, S.; Al-Rubaye, S.; Al-Dulaimi, A.; Bashir, A.K.; Khan, F.A. Securing Critical Infrastructures: Deep-Learning-Based Threat Detection in IIoT. IEEE Commun. Mag. 2021, 59, 76–82. [Google Scholar] [CrossRef]
  40. Zhang, Q.; Yu, K.; Guo, Z.; Garg, S.; Rodrigues, J.; Hassan, M.M.; Guizani, M. Graph Neural Networks-driven Traffic Forecasting for Connected Internet of Vehicles. IEEE Trans. Netw. Sci. Eng. 2021, 4, 1–9. [Google Scholar] [CrossRef]
  41. Afza, F.; Sharif, M.; Kadry, S.; Manogaran, G.; Saba, T.; Ashraf, I.; Damaševičius, R. A framework of human action recognition using length control features fusion and weighted entropy-variances based feature selection. Image Vis. Comput. 2021, 106, 104090. [Google Scholar] [CrossRef]
  42. Rashid, M.; Sharif, M.; Javed, K.; Akram, T. Classification of gastrointestinal diseases of stomach from WCE using improved saliency-based method and discriminant features selection. Multimed. Tools Appl. 2019, 78, 27743–27770. [Google Scholar]
Figure 1. Proposed framework for cucumber leaf diseases recognition using deep learning and Entropy-ELM.
Figure 1. Proposed framework for cucumber leaf diseases recognition using deep learning and Entropy-ELM.
Applsci 12 00593 g001
Figure 2. Sample images of Cucumber leaf diseases.
Figure 2. Sample images of Cucumber leaf diseases.
Applsci 12 00593 g002
Figure 3. Transfer learning-based training of deep models for cucumber leaf diseases recognition.
Figure 3. Transfer learning-based training of deep models for cucumber leaf diseases recognition.
Applsci 12 00593 g003
Figure 4. Confusion matrix-based representation of Cubic SVM accuracy.
Figure 4. Confusion matrix-based representation of Cubic SVM accuracy.
Applsci 12 00593 g004
Figure 5. Confusion matrix of Cubic SVM after employing feature selection technique.
Figure 5. Confusion matrix of Cubic SVM after employing feature selection technique.
Applsci 12 00593 g005
Figure 6. Comparison of all experiments in terms of testing time.
Figure 6. Comparison of all experiments in terms of testing time.
Applsci 12 00593 g006
Figure 7. Confusion matrix of Cubic SVM after parallel fusion of all selected pre-trained deep features and Entropy-ELM selection.
Figure 7. Confusion matrix of Cubic SVM after parallel fusion of all selected pre-trained deep features and Entropy-ELM selection.
Applsci 12 00593 g007
Figure 8. Confusion matrix of proposed framework of cucumber leaf diseases for Cubic SVM.
Figure 8. Confusion matrix of proposed framework of cucumber leaf diseases for Cubic SVM.
Applsci 12 00593 g008
Table 1. Brief description of selected classifiers.
Table 1. Brief description of selected classifiers.
ClassifiersDetails
LSVMKernel scale: Automatic, Box constraint level: 1
Multiclass method: One-vs-One
QSVMKernel scale: Automatic, Box constraint level: 1
Multiclass method: One-vs-One
CSVMKernel scale: Automatic, Box constraint level: 1
Multiclass method: One-vs-One
MGSVMKernel scale: 45, Box constraint level: 1
Multiclass method: One-vs-One
FKNNNo of neighbor: 10, Distance matric: Euclidean
Distance weight: Equal
Subspace_KNNLearner type: Nearest neighbors, No of learners: 30
Subspace dimensions: 1024
Weighted_KNNNo of neighbor: 10, Distance matric: Euclidean
Distance weight: Squared inverse
Cosine_KNNNo of neighbor: 10, Distance matric: cosine
Distance weight: Equal
Cubic_KNNNo of neighbor: 10, Distance matric: Minkowsi (cubic)
Distance weight: Equal
Medium_KNNNo of neighbor: 10, Distance matric: Euclidean
Distance weight: Equal
Table 2. Classification results on originally selected cucumber dataset without data augmentation step for several fine-tuned deep learning models.
Table 2. Classification results on originally selected cucumber dataset without data augmentation step for several fine-tuned deep learning models.
ClassifierF-VGG16F-ResNet50F-ResNet101F-DenseNet201
Cubic SVM55.658.754.261.8
Quadratic SVM55.154.255.161.9
MG SVM56.94850.760.4
Fine KNN50.748.949.852
Linear SVM51.653.351.158.4
ESD24.947.146.756.4
ES KNN50.755.153.851
WKNN51.647.640.949.5
Cosine KNN47.650.243.649
Medium KNN45.345.336.649.5
Table 3. Classification results of fine-tuned VGG16 deep model after data augmentation.
Table 3. Classification results of fine-tuned VGG16 deep model after data augmentation.
ClassifierRecall Rate (%)Precision Rate (%)Accuracy (%)FNR (%)F1 Score (%)Time (Sec)
Cubic SVM93.8493.9293.86.1693.88300
Quadratic SVM93.6693.7293.66.3493.69242
MG SVM91.7892.0491.88.2291.91463
Fine KNN88.5488.5488.511.4688.54562
Linear SVM90.5690.8890.69.4490.72188
ESD92.9893937.0292.991526
ES KNN88.7288.7488.711.2888.731550
WKNN86.2286.4686.213.7886.341038
Cosine KNN80.881.3680.819.2081.08648
Medium KNN79.4480.979.520.5680.16589
Table 4. Classification results of fine-tuned ResNet50 deep model after data augmentation.
Table 4. Classification results of fine-tuned ResNet50 deep model after data augmentation.
ClassifierRecall Rate (%)Precision Rate (%)Accuracy (%)FNR (%)F1 Score (%)Time (Sec)
Cubic SVM94.3694.4694.65.6494.41964
Quadratic SVM94.2694.3694.45.4494.61387
MG SVM91.3891.791.58.6291.541169
Fine KNN86.686.9686.613.4086.78681
Linear SVM91.1291.48918.8891.30657
ESD90.2290.590.59.7890.36655
ES KNN91.8691.691.78.1491.73600
WKNN79.6681.447820.3480.54748
Cosine KNN82.3882.7282.417.6282.55539
Medium KNN72.6476.672.627.3674.57392
Table 5. Classification results of fine-tuned ResNet101 deep model after data augmentation.
Table 5. Classification results of fine-tuned ResNet101 deep model after data augmentation.
ClassifierRecall Rate (%)Precision Rate (%)Accuracy (%)FNR (%)F1 Score (%)Time (Sec)
Cubic SVM97.797.7697.72.3097.7608
Quadratic SVM97.2497.3297.22.7697.2574
MG SVM94.6494.794.65.3694.61086
Fine KNN94.4294.4494.45.5894.41285
Linear SVM94.3294.6294.35.6894.4513
ESD95.8296.6895.84.1896.22394
ES KNN96.3696.2696.33.6496.34072
WKNN92.1692.4892.37.8492.31501
Cosine KNN86.4686.8486.513.5486.61404
Medium KNN83.4484.4483.516.5683.931342
Table 6. Classification results of fine-tuned Densenet201 deep model after data augmentation.
Table 6. Classification results of fine-tuned Densenet201 deep model after data augmentation.
ClassifierRecall Rate (%)Precision Rate (%)Accuracy (%)FNR (%)F1 Score (%)Time (Sec)
Cubic SVM98.4498.598.41.5698.47355
Quadratic SVM97.497.4697.42.6097.43330
MG SVM95.3295.6295.44.6895.47623
Fine KNN93.4293.4293.46.5893.42734
Linear SVM92.6293.1692.77.3892.89302
ESD96.6296.696.63.3896.611495
ES KNN94.194.0894.15.9094.091923
WKNN92.2692.492.37.7492.33927
Cosine KNN85.4485.9485.314.5685.69807
Medium KNN85.986.985.814.1086.40764
Table 7. Classification results using proposed Entropy-ELM selection approach.
Table 7. Classification results using proposed Entropy-ELM selection approach.
ClassifierRecall Rate (%)Precision Rate (%)Accuracy (%)FNR (%)F1 Score (%)Time (Sec)
Cubic SVM98.0297.9898.01.9898.00116
Quadratic SVM96.5496.6296.53.4696.58135
MG SVM84.281.484.015.8082.78197
Fine KNN94.8294.8494.85.1894.83217
Linear SVM82.8283.1482.817.1882.98133
ESD93.393.7893.06.7093.54201
ES KNN93.1692.293.06.8492.68405
WKNN94.2294.2894.25.7894.25384
Cosine KNN88.1688.9288.411.8488.5493
Medium KNN85.7886.885.814.2286.29204
Table 8. Classification results using proposed parallel features fusion and Entropy-ELM selection of all pre-trained deep models using augmented dataset.
Table 8. Classification results using proposed parallel features fusion and Entropy-ELM selection of all pre-trained deep models using augmented dataset.
ClassifierRecall Rate (%)Precision Rate (%)Accuracy (%)FNR (%)F1 Score (%)Time (Sec)
Cubic SVM97.9298.1298.22.0897.02847
Quadratic SVM97.7497.3897.72.2697.56277
MG SVM94.8495.0694.85.1694.95392
Fine KNN94.8894.994.95.1294.89422
Linear SVM94.769594.85.2494.88475
ESD96.2696.2896.33.7496.27454
ES KNN96.6696.6696.73.3496.66400
WKNN92.7892.8692.87.2292.82534
Cosine KNN86.7881.3286.813.2283.96635
Medium KNN84.8285.684.815.1885.21129
Table 9. Proposed framework classification results using augmented dataset.
Table 9. Proposed framework classification results using augmented dataset.
ClassifierRecall Rate (%)Precision Rate (%)Accuracy (%)FNR (%)F1 Score (%)Time (Sec)
Cubic SVM98.3698.398.51.7498.48111
Quadratic SVM98.197.597.51.9097.80117
MG SVM95.7496.0695.84.2695.90175
Fine KNN94.4294.4294.45.5894.42130
Linear SVM93.0693.6893.26.9493.37103
ESD96.3696.4296.43.6496.39196
ES KNN94.8294.894.85.1894.81277
WKNN92.292.5692.27.8092.38201
Cosine KNN85.4885.7885.414.5285.63142
Medium KNN85.5884.3685.514.4284.97104
Table 10. Comparison with SOTA for cucumber leaf diseases recognition.
Table 10. Comparison with SOTA for cucumber leaf diseases recognition.
MethodsYearAccuracy (%)
Zhang et al. [27]201785.7
Ma et al. [36]201893.4
Lin et al. [35]201996.08
Khan et al. [13]202098.08
Zhang et al. [37]202190.67
Jaweria et al. [7]202193.50
Hussain et al. [24]202296.50
Proposed98.48
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Khan, M.A.; Alqahtani, A.; Khan, A.; Alsubai, S.; Binbusayyis, A.; Ch, M.M.I.; Yong, H.-S.; Cha, J. Cucumber Leaf Diseases Recognition Using Multi Level Deep Entropy-ELM Feature Selection. Appl. Sci. 2022, 12, 593. https://doi.org/10.3390/app12020593

AMA Style

Khan MA, Alqahtani A, Khan A, Alsubai S, Binbusayyis A, Ch MMI, Yong H-S, Cha J. Cucumber Leaf Diseases Recognition Using Multi Level Deep Entropy-ELM Feature Selection. Applied Sciences. 2022; 12(2):593. https://doi.org/10.3390/app12020593

Chicago/Turabian Style

Khan, Muhammad Attique, Abdullah Alqahtani, Aimal Khan, Shtwai Alsubai, Adel Binbusayyis, M Munawwar Iqbal Ch, Hwan-Seung Yong, and Jaehyuk Cha. 2022. "Cucumber Leaf Diseases Recognition Using Multi Level Deep Entropy-ELM Feature Selection" Applied Sciences 12, no. 2: 593. https://doi.org/10.3390/app12020593

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop