<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>November</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Potato Leaf Disease Detection using CNN - A Lightweight Approach</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Abhisek Saha</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Syed Mohammed Musharraf</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Anubhav Dey</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Hiranmoy Roy</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Debotosh Bhattacharjee</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science &amp; Engineering, Jadavpur University</institution>
          ,
          <addr-line>Kolkata-700032</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Information Technology, RCC Institute of Information Technology</institution>
          ,
          <addr-line>Canal South Road, Kolkata 700015</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Netaji Subhash Engineering College</institution>
          ,
          <addr-line>Techno City, Garia, Ranabhutia, West Bengal, Kolkata 700152</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2024</year>
      </pub-date>
      <volume>2</volume>
      <fpage>8</fpage>
      <lpage>29</lpage>
      <abstract>
        <p>Detection of potato leaf diseases at an early stage is of great significance to the agricultural industry. The conventional tactics of disease identification that exist are either unreliable or very complex or costly making them not suitable as viable techniques. However, with a boom in the field of Artificial intelligence, many procedures have come up over the recent years to help solve this problem. Data being the fuel for such procedures, it is very important to source reliable and accurate data for the training purpose of the AI based models. The task of disease detection for potato leaves is quite challenging as the symptoms show a lot of variations depending upon the species, climate and environmental factors. The popular pretrained models used for this purpose are VGG16, Inception V3, ResNet50 which help us to classify diseases of plants. In our research we have tried to build a custom Convolutional Neural Network classification model which is more robust and light weight as compared to the existing approaches. The model is built with a very simple approach and is trained using two standard publicly available datasets namely “PlantVillage” and PLD. The correctness of the suggested model has shown promising and consistent output with accuracy of 99.3% and 99.23%, while implemented on the two datasets respectively. To achieve the said accuracy, we have used image Enhancement algorithm: CLAHE at the preprocessing stage after the data acquisition.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Potato Leaf Disease</kwd>
        <kwd>CNN</kwd>
        <kwd>Image Enhancement</kwd>
        <kwd>Image classification</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Since the dawn of human civilization, agriculture has played a crucial role in transforming people
from roving hunter-gatherers to established citizens [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. It has facilitated the growth of large human
populations by providing a reliable and stable source of nutrition. The history of agriculture is a
long continuum of groundbreaking innovations, evolving rapidly through industrial revolutions and
advancements in modern science, particularly in the 20th and 21st centuries. Unlike a single, definitive
moment, the origin of agriculture in human civilization unfolded over centuries and cannot be precisely
dated. Researchers concur that early Homo Sapiens began transitioning from a nomadic lifestyle to
settling down, domesticating animals, and cultivating cereal seeds during the early Neolithic period,
known as the Neolithic Revolution [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. This shift likely took place as glaciers retreated northward and
the climate warmed, approximately 10,000 years ago—though some estimates place it at 12,000 or even
15,000 years ago. This further led to the development of complex societies as humans learnt the ways of
trade and caused the growth of economy and exchange. From a contemporary standpoint, agriculture
is a very dynamic industry that is essential to meet the basic need of food for enitire population.
In a developing nation like India, agriculture is an important sector of its economy as it contributes
about 15% of total GDP and renders employment to about 60% of its residing population [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. However,
despite being a formidable industry, due to the problem of crop diseases, the sector sufers havoc. Plant
diseases can have a major efect on the leaves, fruits and diferent parts of crops, which degrades
quality of crops and yield [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. This, in turn, contributes to food scarcity and insecurity on a global scale.
It is estimated that crop diseases cause an annual loss of around 16% in global crop yields, making them
a major factor behind famines and rising production costs. According to predictions from the Food
and Agriculture Organization (FAO), there will be 9.5 billion people on the earth by next thirty years
[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], meaning that a 75% increase in food production is necessary to provide a consistent supply of
food. A group of factors that impact plants and their products are diseases and illnesses. As opposed to
disorders, which are mostly caused by factors such as rainfall, temperature, moisture, and nutritional
deficiencies, illnesses are caused by biotic agents such as fungi, bacteria, and algae [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
For the well-being of the crops, early and precise identification of diseases is necessary, so that the
correct cure can be applied in time. Various methods are available for diagnosing plant diseases, with
one of the simplest ones being visual inspection. Conventional diagnostic techniques frequently
depend on the farmer’s knowledge, which can be erratic and untrustworthy. Due to the extensive,
time-consuming process needed and the restricted availability of experts in remote places [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], this
strategy is frequently impracticable. To improve accuracy, researchers have introduced spectrometers
to distinguish between healthy and infected plant leaves [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Another technique involves extracting leaf
DNA using the PCR [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. These methods are complex, costly, and time-consuming, requiring specialized
skills, controlled experimental conditions, and extensive use of crop safety equipments. The role of
artificial intelligence is immensely significant in this aspect. If the training of deep learning based
models, is done on labeled samples, we can perform automated, eficient and accurate leaf disease
detection [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. We have chosen potato as the crop of our concern and our model serves to categorically
classify blight disease both early and late, as well healthy leaf images. Potato (Solanum tuberosum) [10]
is a temperate crop grown under subtropical conditions in India. Soils that are loose, muddy, and sandy
and rich in organic matter are ideal for growing potato crops; alkaline and saline soils are not suited
[11]. In addition to being a great source of fiber and heart disease prevention, potatoes are vital for
overall health. Their high antioxidant content aids in the defense against diseases including excessive
cholesterol and irregular blood sugar levels [12]. The nation has been growing potatoes for more than
three centuries. It is currently one of the most widely grown crops in this nation for vegetable purposes.
Potatoes are a cost-efective food that contributes inexpensive energy to the human diet.
For the purpose of training our mode, we have made use of the publicly available dataset “PlantVillage”
and PLD. Both these datasets contain potato leaves images for both blight diseases, Early and Late as
well as healthy also. Pre-indication of the disease Early Blight can be seen initially towards the base of
the plant, with roughly circular brown colored spots on the leaves and stems. This infection caused by
Fungus can be deadly for tubers, leaves and stems, causing many problems like reduced tuber size, low
produce and crop yield. One more fungal disease that afects potato crops is late blight, which first
manifests as patches on the stems. They spread out very quickly, creating big, dark brown, and black
areas that frequently look oily [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. However, a marked variation in these symptoms can be observed
based on the region it is taken from, the climate, the species etc., making it a dificult task to build a
generalized classifier.
      </p>
      <p>Kamal et. al.[13] developed two models based on MobileNet architecture and applied the architecture
on PlantVillage dataset. They have achieved an accuracy of 97.65% and 98.34% from Reduced MobileNet
and Modified MobileNet respectively for classification of 55 classes in PlantVillage dataset. Liang et.al.
Two diferent types of deep architectures where the first one based is on residual learning and the
second one is based on attention mechanism, have been developed by Karthik and his co-authors
[14]. The model based on the attention mechanism has attained an accuracy of 98% on the public
dataset named plantvillage for the detection of diseased tomato leaf. Khamparia et. al. [15] designed a
hybrid approach of developing an architecture for disease detection of potato leaves using Deep Neural
Network and autoencoders and attained an accuracy of 97.50% on PlantVillage dataset. Islam et. al.
[16] presented an approach which integrates the technique of image processing and machine learning
to classify potato leaf diseases. They used the renowned machine learning algorithm SVM in their
recommended solution and attained an accuracy of 95% over 300 images. For the identification and
classification of cassava illness, Sambasivam and his team[ 17] developed a deep neural network that
was trained on a very small dataset with significant class imbalance. A possible solution to the illness
identification accuracy problem with the least amount of time investment is the Kuan filtered Hough
transformation based reweighted linear program boost classification (KFHT-RLPBC) technique, which
was introduced by Nagarjan and his co-authors[18]. By using the PlantVillage dataset, they achieved
an accuracy of 92%. Geetharamani et.al. [19] recommended a deep Convolutional Neural Network
which can efectively identify and solve the problem of plant leaf diseases. They intended to carry out a
more thorough analysis of the training procedure without utilizing the tagged photos after achieving
an accuracy rate of 96.46% in the PlantVillage dataset. Table 1 discusses the synopsis of related works.
Most of the current methods for crop leaf disease detection have used the popular transfer learning
models, but these models often have a higher degree of parameters which leads to the problem
of computational complexity. Other approaches have implemented custom Convolutional Neural
Networks (CNNs) with reduced parameter counts; however, they generally fall short in achieving
significant amount of accuracy. Thus, there is a need for a more eficient, lightweight model for crop
leaf disease detection.</p>
      <p>In order to detect diseases in potato leaves, we provide a novel lightweight Convolutional Neural
Network (CNN) in this research that can recognize both simple and abstract patterns. The model
architecture, illustrated in Figure 1, comprises convolutional layers, maxpooling layers for edge
feature extraction, batch normalization to normalize the input neuron values, dropout layers to reduce
overfitting, and fully connected layers for classification purpose. We have also used the Contrast
Limited Adaptive Histogram Equalization (CLAHE) technique for image preprocessing.The major
contributions are as follows.</p>
      <p>• The model introduces a lightweight CNN architecture which is both less complex with respect to
number of parameters and highly accurate for leaf disease detection of potato leaves.
• Through the use of convolution and max pooling layers, it is able to capture both intricate patterns
and minute details.
• The addition of CLAHE raises the picture quality, which strengthens the model’s capacity to
accurately detect damaged potato leaves.
• Data augmentation techniques used to make the training and testing datasets larger and more
balanced. The classifier’s capacity for generalization increases with the use of various data
augmentation strategies.</p>
      <p>The study is organized as follows: the proposed CNN architecture is thoroughly explained in Section 2.
The work is concluded in Section 4, while the experiments and comparison findings are presented in
Section 3.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Proposed Methodology</title>
      <p>Three subsections comprise this section: Data Preprocessing, Acquisition of Data, and Classification.</p>
      <sec id="sec-2-1">
        <title>2.1. Acquisition of Data</title>
        <p>
          The images of potato leaf diseases were sourced from two publicly available datasets: the PLD dataset[
          <xref ref-type="bibr" rid="ref5">5</xref>
          ]
and the PlantVillage dataset[21] . Both datasets include two types of blight disease early, late as well
as healthy images . The PlantVillage dataset provided a sum of 2,152 images— 1,000 each for both
kind of blight diseases and 152 for healthy leaves (refer table-4). Owing to the limited quantity and
imbalance of images, additionally, 3,251 photos from Pakistan’s Central Punjab were included from the
PLD dataset.This dataset contains 816 healthy images, 1,303 early blight images, and 1,132 late blight
images after redundancy has been removed (refer to table-5).All photos are saved in uncompressed JPG
style and have RGB color profiles.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Data Preprocessing</title>
        <p>As preproceesing stage, we employed CLAHE [22] algorithm . CLAHE is an image processing technique
designed to enhance image contrast. CLAHE operates on discrete areas of the image, known as tiles,
as opposed to the full image at once, in contrast to conventional histogram equalization. Within each
tile, CLAHE adjusts the contrast adaptively based on the local histogram, allowing it to enhance detail
without overly amplifying noise. After processing each tile, neighboring tiles are merged smoothly to
prevent visible boundaries. CLAHE is designed to prevent excessive contrast amplification by limiting
it. The contrast amplification around each pixel is determined by a slope function transformation. To
control this amplification, CLAHE clips the histogram at a predefined threshold before computing the
cumulative distribution function, efectively constraining the enhancement to avoid noise exaggeration.
By limiting contrast adjustments in uniform areas, CLAHE minimizes noise amplification, producing a
balanced, enhanced image.</p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. Classification</title>
        <p>Image processing, recognition, and classification are the main applications for CNNs, a kind of deep
learning method. The architecture of the human brain served as the model for CNN’s design. The
similarity of both can be understood as human brain has neurons and in the neural networks, it is
the neurons which are the backbone of the entire system. A CNN consists of several layers namely
convolutional, maxpooling, dropout and fully connected .</p>
        <p>The convolution operation for a 2D convolutional layer can be represented as:</p>
        <p>,, = ∑︁ ∑︁ ∑︁ +,+, · ,,, +</p>
        <p>=1 =1 =1
• ,,: Output feature map at position (, ) in channel .
• +,+,: Input feature map at position ( + ,  + ) in channel .
• ,,,: Convolution filter weight of size ( ×  ) for channel  and output channel .
• : Bias term for channel .</p>
        <sec id="sec-2-3-1">
          <title>For max-pooling, the operation can be written as:</title>
          <p>• ,,: Output of max-pooling at position (, ) in channel .</p>
          <p>• +,+,: Input feature map over a pooling window defined by (, ).</p>
          <p>The following is an expression for a Fully connected layer’s output.</p>
          <p>,, = max (+,+,)
,

 = ∑︁ , +</p>
          <p>=1
•  : Output of the th neuron in the layer.
• : Input from the th neuron in the previous layer.
• ,: Weight connecting the th neuron in the previous layer to the th neuron.</p>
          <p>•  : Bias term for the th neuron.</p>
          <p>Softmax function, used to convert final layer’s logits into probabilities, is given by:
 () =</p>
          <p>∑︀=1 
•  (): Softmax output for class .
• : Logit (raw output) for class .</p>
          <p>•  : Total number of classes.</p>
          <p>The funct of these layers is to detect the features like edges and complex patterns. To extract features
and diferent kinds of edges, various types of filters will be used as per the requirement. For Potato
Disease Detection, we used the three well known transfer learning models: VGG16, InceptionV3 and
ResNet-50.</p>
          <p>VGG16 [23] is a deep CNN architecture which has been introduced for image classification in the
year 2014. Its architecture is based on the input size 224x224 pixels of RGB images. It consists of
total 16 layers comprises with 13 convolutional and 3 fully connected layers. Activation function relu
is employed in all the layers. VGG16 is trained on ImageNet dataset and is competent enough for
the classification of images into 1000 classes. categories and detecting objects from 200 classes. The
convolution layer has 3x3 filters with increasing number of filters to detect the complex hierarchal
patterns in the images. Max pooling layers of size 2x2 with a stride of 2 have been used to extract
features that select the maximum valued pixel within each small region. After the feature extraction
layers, there are 2 fully connected layers of 4096 neurons and finally there is a fully connected layer of
1000 neurons for classification purpose.</p>
          <p>The 48-layer deep pretrained CNN InceptionV3 [24] was trained on the ImageNet dataset and can
categorize images into 1000 diferent categories. The network is based on the input size of 299x299
pixels of RGB images. The layers architecture of InceptionV3 consists of Inception modules where
each module is combined of 1x1, 3x3, and 5x5 convolutions. InceptionV3 has fewer no. of parameters
because of factorizing convolutions. A convolution of 5X5 filter can be replaced by two 3x3 filters. In
this context, for a 5x5 filter it requires 25 parameters but for two 3x3 filters it requires 18 parameters. It
will reduce the no. of parameters by 28% without losing the ability to capture patterns by a 5x5 filter.
Because of this light weight architecture, it is computationally eficient to work on.</p>
          <p>ResNet-50 [25] is a popular deep cNN architecture, which is a part of the ResNet (Residual Networks)
family, which was developed to tackle the problem of vanishing gradient in deep networks by introducing
residual blocks. It was first introduced in the year 2016.Convolutional layers, batch normalization, ReLU
activations, and skip (residual) connections make up this 50-layer deep model. 48 convolutional layers,
one max-pooling layer, and one average-pooling layer make up the layers. There are many residual
blocks in each of the four main stages of the model. Each residual block in ResNet-50 has a shortcut
connection that skips single or multiple layers, enabling the gradient to pass back through the network
without vanishing. They consist of three convolutional layers with 1x1, 3x3, and 1x1 convolutions. The
ifnal fully connected layer in ResNet-50 typically has 1,000 output units for 1,000 classes in the ImageNet
dataset, which the model was originally trained on. However, ResNet-50 can be modified to handle
any number of classes by adjusting the number of output units in the final layer. This modification is
common in transfer learning, where the network is adapted to diferent datasets with fewer or more
classes.</p>
          <p>For our experiment of potato disease detection, we propsed a custom CNN which comprises with 20
layers Table-3. The architecture of the proposed CNN model is depicted in following steps. Figure 1
and Figure 2 are representing the model architecture and detailed layerwise flow diagram respectively.
• The images of the datasets PlantVillage and PLD are enhanced using CLAHE for better clarity,
noise reduction and ease of feature extraction.
• Enhanced images undergo augmentation using a variety of techniques that make use of the
dataset and improve the generalization capacity of the model.
• Images are passed to the Input Layer which accepts an input of size 224 X 224 and 3 channels.
• Images are subsequently batch normalized to normalize the output from neurons by
calculating the mean and variance across the mini batch during training, for better convergence and
improvement of training accuracy.
• Normalized images are then fed to our customized feature extractor comprising 5 blocks, each
consisting of Convolution and Maxpooling layers. The Convolution layers are used to capture
minute details for fine feature extraction. The Maxpooling layer on the other hand reduces the
spatial dimensions by selecting a maximum value within a kernel, retaining the most prominent
features. A stride of 1 has been maintained throughout, and activation function ReLU has been
employed with the convolution layers to incorporate the absense of linearity. Only positive
feature values are activated, allowing the model to learn intricate patterns.
• Uniform kernel size of 3x3 has been used for the convolution layers, and all the Maxpooling
layers used are of size 2x2. Block 1 consists of 1 Convolution layer of 32 filters, followed by a
Maxpooling layer. Block 2 consists of 2 convolution layers of 64 filters each and a Maxpooling
layer. Block 3 contains 2 convolution layers of 128 filters each and a Maxpooling layer. Two
convolution layers of 256 filters and a Maxpooling layer is contained in Block 4. In Block 5, the
last and final block, there are 2 convolution layers of 512 filters each and a Maxpooling layer.
• The output from the Feature Extractor is fed into customized Classifier which first flattens the
vectors to a single dimension comprising 8192 neurons, followed by a Dropout layer to avoid
overfitting of the model. After this, dense layers are used which steps down the sizes from 1024
to 256 to 64 neurons consecutively. The dense layers also employ the ReLU activation function.
• The output of the last dense layer, which consists of three neurons representing the three classes,
is then sent through the Softmax activation function to classify the image as either Healthy, Early
Blight, or Late Blight.</p>
          <p>Softmax is generalization of sigmoid function for multiclass classification which generates
probabilistic output between 0 and 1.Relu activation adds non-linearity to the model in each convolution
and max pooling layer by setting just the negative variables to zero while leaving the positive ones
unaltered. Details about hyperparameters are depicted in Table-9.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Experimental Result</title>
      <p>All the experiments have been done using an IntelR CoreTM i5-1135G7 CPU, an NVIDIA Tesla T4 GPU
with 16 GB VRAM, 32 GB of RAM, and a Windows operating system. The deep learning implementation
was carried out using the TensorFlow 2.16.1 framework, Python 3.12.2, to accelerate neural network
operations. The proposed methodology has been implemented using the two public dataset namely PLD
and Plantvillage. Keras, framework of neural networks written in Python, has been used to implement
the model. There is total 2152 images from PlantVillage Dataset and 3251 images of PLD dataset which
have been used. To artificially increase the dataset and flexibility of the model, data augmentation
technique has been employed in the dataset [26]. For both datasets, we have used the same techniques.
We used ImageGenerator class in Keras for augmentation purpose Table 2. We applied the parameters
to rotate the image about an angle of 30 degree and transform the images horizontally, vertically and
zoom into or out in a range of 15% randomly. With the learning rate set to 0.001 and the number of
epochs set to 200, batch size 32 has been utilized Table 9. Categorical cross-entropy is the loss function,
and the Adam optimizer is employed.</p>
      <p>For the Plant Disease detection, we have used the publicly available PlantVillage Dataset [27] and PLD
Dataset. The PlantVillage dataset is a widely available and comprehensive benchmark dataset for crop
leaf disease classification. It includes 54,306 samples across 14 plant species, covering a total of 32 classes.
Among them, 26 classes are from diseased plants, while the remaining 12 classes belongs to healthy
plants.We selected three diferent kinds of potato leaf disease samples—late blight, early blight, and
healthy—from the "PlantVillage dataset" because our study focuses on potato leaf diseases prediction.
In all, 2152 images of potato leaves were used in our experiment; 1000 of these images showed early
blight, 1000 showed late blight, and the remaining 152 showed healthy in Table 4. Potato crops are
susceptible to a fungal ailment called early blight. The PlantVillage dataset has not an adequate number
of images and exhibits an uneven class distribution, so the PLD dataset has been used which has been
created in Pakistan’s Central Punjab region. From that dataset, we have rejected some images of potato
leaves due to redundancy. There is a total of 3251 images of potato leaves used amongst which 1303
pictures are from Early blight section, 816 pictures are from healthy section and 1132 from Late blight
section Table 5.</p>
      <p>The model’s performance is calculated using standard validation metrics. To evaluate how well the
suggested model discriminated, the model’s accuracy, recall, precision, and f1-score were calculated. A
tabular method of displaying the prediction model’s performance is the confusion matrix. The number
of predictions the model made when it properly or erroneously identified the classes is indicated by an
entry in a confusion matrix.</p>
      <p>A classifier’s True Positive (TP) is the number of predictions in which it correctly identifies the positive
class as positive. Conversely, True Negative (TN) refers to the quantity of predictions in which the
classifier correctly identifies the negative class as negative. False Positives (FP) are the quantity of
hypotheses in which the classifier incorrectly predicts the negative class as positive.False Negative (FN)
is the frequency with which the classifier incorrectly predicts the positive class as negative. Accuracy
provides the model’s overall accuracy, or the percentage of all samples that the classifier successfully
classified. Equation (1) can be used to determine accuracy.</p>
      <p>=   (2)</p>
      <p>+</p>
      <p>Recall indicates the percentage of positive samples that the classifier accurately predicted to be
positive. Other names for it include probability of detection, sensitivity, and true positive rate (TPR).
Equation (3) can be used to determine precision.
The F1 score, a mixture of the two measures, is frequently used by practitioners in machine learning
to balance the precision-recall score. It merges recall and precision into one metric. In terms of
mathematics, it is the harmonic mean of recall and precision. It can be computed by using Equation (4).
(3)
(4)
Divyansh[28]
Zhang[29]
Barman[30]
Rozaqi[31]
Proposed Model(Plantvillage dataset)
Proposed Model(PLD dataset)</p>
      <p>Technique
SVM, KNN and Neural Net</p>
      <p>Faster RCNN</p>
      <p>SBCNN</p>
      <p>CNN
CNN
CNN</p>
      <p>Crop
Potato
Tomato
Potato
Potato
Potato
Potato</p>
      <p>A binary classifier’s diagnostic performance can be assessed visually using a ROC curve. Across a
range of threshold values, it compares True Positive Rate (TPR) against the False Positive Rate (FPR).
The FPR shows the percentage of actual negatives that are mistakenly categorized as positives, whereas
the TPR, or sensitivity, shows the percentage of actual positives that are accurately detected.</p>
      <p>The trade-of between sensitivity and specificity is illustrated by the ROC curve. Better performance
is indicated by a model whose curve is closer to the plot’s upper-left corner. This performance is
frequently measured using the Area Under the ROC Curve (AUC), where a value of 0.5 indicates no
discriminatory capacity and a value of 1 indicates flawless classification.</p>
      <p>The confusion matrix for both PlantVillage and PLD dataset is given in Figure 3. If we have a look at
the graphs of the accuracy and loss for both the datasets PLD and PLantVillage 4, then it gives us the
idea of the convergence speed of the model. The training and validation loss graph for both the datasets
PlantVillage and PLD is given in Figure 6. The proposed model has outperformed the 3 transfer learning
models VGG16, ResNet50 and InceptionV3 comprehensively and has achieved a test accuracy of 99.3%
on PlantVillage dataset and 99.23% on the PLD dataset Table 8. The implementation process requires
way less hardware because of the presence of fewer parameters unlike the deep CNN architectures
Table- 6. The model also outperformed the other state-of-art models in terms of accuracy Table 8. The
performance of the proposed model as compared to the other transfer learning models in terms of
Precision, Recall and F1-score is given in Table 7. The two ROC curves for the two datasets PlantVillage
and PLD is given in Figure 7. Thus, the model gives us an eficient way to solve the problem of potato
leaf disease detection with a better accuracy than the pretrained models and less computational hazards.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Conclusion and Future Scope</title>
      <p>Agriculture holds the most important economical aspect of our country as majority of the common
people have been heavily relied on Agriculture. Identifying illnesses that damage economically valuable
crops early on is crucial to preventing farmers from sufering financial losses related to these crops.
Potato is one of the most staple crops and our experiment is based on [26] the detection of the potato
leaf diseases. Our unique Convolutional Neural Network categorizes potato leaves into three groups for
that purpose, blight disease namely early and late along with healthy. Due to presence of fewer layers
and parameters than the other transfer learning and CNN models, it is highly eficient and resourceful
in the computationally constrained environments. It achieved more than 99% precision in PlantVillage,
as well as in the PLD data set.</p>
      <p>The vision of our research is to improve the model’s adaptability and resilience so that it can identify
diseases in a variety of crops besides potatoes. A mobile and web application for the benefit of the
farmer community and contribution to this sector overall will be developed, along with an attempt to
further minimize the parameters to make the model more eficient.</p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgments</title>
      <p>Thanks to the developers of ACM consolidated LaTeX styles https://github.com/borisveytsman/acmart
and to the developers of Elsevier updated LATEX templates https://www.ctan.org/tex-archive/macros/
latex/contrib/els-cas-templates.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <sec id="sec-6-1">
        <title>The author(s) have not employed any Generative AI tools.</title>
        <p>[10] P. Moallem, N. Razmjooy, A multi layer perceptron neural network trained by invasive weed
optimization for potato color image segmentation, Trends in Applied Sciences Research 6 (2012)
445–455. doi:10.3923/tasr.2012.445.455.
[11] N. E. Khalifa, M. Taha, L. Abou El-Magd, A. E. Hassanien, Artificial Intelligence in Potato
Leaf Disease Classification: A Deep Learning Approach, 2021, pp. 63–79. doi: 10.1007/
978-3-030-59338-4_4.
[12] S. Radha, J. Chatterjee, N. Jhanjhi, S. Brohi, Performance of deep learning vs machine learning in
plant leaf disease detection, Microprocessors and Microsystems 80 (2021) 103615. doi:10.1016/j.
micpro.2020.103615.
[13] K. KC, Z. Yin, M. Wu, Z. Wu, Depthwise separable convolution architectures for plant disease
classification, Computers and Electronics in Agriculture 165 (2019). doi: 10.1016/j.compag.
2019.104948.
[14] R. Karthik, M. Hariharan, S. Anand, P. Mathikshara, A. Johnson, R. Menaka, Attention embedded
residual cnn for disease detection in tomato leaves, Applied Soft Computing 86 (2020) 105933.
[15] A. Khamparia, G. Saini, D. Gupta, A. Khanna, S. Tiwari, V. Albuquerque, Seasonal crops disease
prediction and classification using deep convolutional encoder network, Circuits, Systems, and
Signal Processing 39 (2020). doi:10.1007/s00034-019-01041-0.
[16] M. Islam, A. Dinh, K. Wahid, P. Bhowmik, Detection of potato diseases using image segmentation
and multiclass support vector machine, 2017, pp. 1–4. doi:10.1109/CCECE.2017.7946594.
[17] A predictive machine learning application in agriculture: Cassava disease detection and
classification with imbalanced dataset using convolutional neural networks, Egyptian Informatics Journal
22 (2021) 27–34. doi:https://doi.org/10.1016/j.eij.2020.02.007.
[18] N. Deepa, N. Nagarajan, Kuan noise filter with hough transformation based reweighted linear
program boost classification for plant leaf disease detection, Journal of Ambient Intelligence and
Humanized Computing 12 (2021). doi:10.1007/s12652-020-02149-x.
[19] G. G., A. P. J., Identification of plant leaf diseases using a nine-layer deep convolutional
neural network, Computers Electrical Engineering 76 (2019) 323–338. URL: https://www.
sciencedirect.com/science/article/pii/S0045790619300023. doi:https://doi.org/10.1016/j.
compeleceng.2019.04.011.
[20] Q. Liang, S. Xiang, Y. Hu, G. Coppola, D. Zhang, W. Sun, Pd2se-net: Computer-assisted plant
disease diagnosis and severity estimation network, Computers and Electronics in Agriculture 157
(2019) 518–529. doi:10.1016/j.compag.2019.01.034.
[21] D. Hughes, M. Salathe, An open access repository of images on plant health to enable the
development of mobile disease diagnostics through machine learning and crowdsourcing (2015).
[22] P. Musa, F. Rafi, M. Lamsani, A review: Contrast-limited adaptive histogram equalization (clahe)
methods to help the application of face recognition, 2018, pp. 1–6. doi:10.1109/IAC.2018.
8780492.
[23] K. Simonyan, A. Zisserman, Very deep convolutional networks for large-scale image recognition,
arXiv 1409.1556 (2014).
[24] C. Szegedy, V. Vanhoucke, S. Iofe, J. Shlens, Z. Wojna, Rethinking the inception architecture for
computer vision, in: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR),
2016, pp. 2818–2826. doi:10.1109/CVPR.2016.308.
[25] K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition, 2016, pp. 770–778.</p>
        <p>doi:10.1109/CVPR.2016.90.
[26] V. Kukreja, A. Baliyan, V. Salonki, R. Kaushal, Potato blight: Deep learning model for binary and
multi-classification (2021) 967–672. doi: 10.1109/SPIN52536.2021.9566079.
[27] D. Hughes, M. Salathe, An open access repository of images on plant health to enable the
development of mobile disease diagnostics through machine learning and crowdsourcing (2015).
[28] D. Tiwari, M. Ashish, N. Gangwar, A. Sharma, S. Patel, S. Bhardwaj, Potato leaf diseases detection
using deep learning, 2020, pp. 461–466. doi:10.1109/ICICCS48265.2020.9121067.
[29] Y. Zhang, C. Song, D. Zhang, Deep learning-based object detection improvement for tomato
disease, IEEE Access PP (2020) 1–1. doi:10.1109/ACCESS.2020.2982456.
[30] U. Barman, D. Sahu, G. G. Barman, J. Das, Comparative assessment of deep learning to detect
the leaf diseases of potato based on data augmentation (2020). doi:10.1109/ComPE49325.2020.
9200015.
[31] A. J. Rozaqi, A. Sunyoto, Identification of disease in potato leaves using convolutional neural
network (cnn) algorithm, in: 2020 3rd International Conference on Information and Communications
Technology (ICOIACT), 2020, pp. 72–76. doi:10.1109/ICOIACT50329.2020.9332037.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>F.</given-names>
            <surname>Arshad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mateen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hayat</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wardah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Al-Huda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Gu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Al-antari</surname>
          </string-name>
          <string-name>
            <surname>Aisslab</surname>
          </string-name>
          ,
          <article-title>Pldpnet: Endto-end hybrid deep learning framework for potato leaf disease prediction</article-title>
          ,
          <source>Alexandria Engineering Journal</source>
          <volume>78</volume>
          (
          <year>2023</year>
          ). doi:
          <volume>10</volume>
          .1016/j.aej.
          <year>2023</year>
          .
          <volume>07</volume>
          .076.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>N.</given-names>
            <surname>Przulj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Velimirovic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Petrović</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Ilić</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mirosavljević</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Trkulja</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Jovovic</surname>
          </string-name>
          , From the Stone Hoe to Circular Agriculture,
          <year>2024</year>
          , pp.
          <fpage>119</fpage>
          -
          <lpage>168</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>P.</given-names>
            <surname>Dalwadi</surname>
          </string-name>
          ,
          <article-title>An analysis of india's agricultural sector: Challenges and opportunities</article-title>
          ,
          <source>EPRA International Journal of Multidisciplinary Research (IJMR)</source>
          (
          <year>2023</year>
          )
          <fpage>293</fpage>
          -
          <lpage>296</lpage>
          . doi:
          <volume>10</volume>
          .36713/epra13069.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Iqbal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Khan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sharif</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Shah</surname>
          </string-name>
          ,
          <article-title>An automated detection and classification of citrus plant diseases using image processing techniques: A review, Computers and Electronics in Agriculture 153 (</article-title>
          <year>2018</year>
          )
          <fpage>12</fpage>
          -
          <lpage>32</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.compag.
          <year>2018</year>
          .
          <volume>07</volume>
          .032.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>D</given-names>
            <surname>.-J. Rashid</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Khan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ghulam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Almotiri</surname>
          </string-name>
          , M. Al Ghamdi,
          <string-name>
            <surname>K.</surname>
          </string-name>
          <article-title>Masood, Multi-level deep learning model for potato leaf disease recognition</article-title>
          ,
          <source>Electronics</source>
          <volume>10</volume>
          (
          <year>2021</year>
          ). doi:
          <volume>10</volume>
          .3390/ electronics10172064.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S.</given-names>
            <surname>Manzoor</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Manzoor</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Islam</surname>
          </string-name>
          , J. Boudjadar, AgriScanNet-18:
          <article-title>A Robust Multilayer CNN for Identification of Potato Plant Diseases</article-title>
          ,
          <year>2024</year>
          , pp.
          <fpage>291</fpage>
          -
          <lpage>308</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>031</fpage>
          -47724-9_
          <fpage>20</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Y.</given-names>
            <surname>SASAKI</surname>
          </string-name>
          , T. OKAMOTO,
          <string-name>
            <given-names>K.</given-names>
            <surname>Imou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Torii</surname>
          </string-name>
          ,
          <article-title>Automatic diagnosis of plant disease</article-title>
          ,
          <source>JOURNAL of the JAPANESE SOCIETY of AGRICULTURAL MACHINERY 61</source>
          (
          <year>2010</year>
          )
          <fpage>119</fpage>
          -
          <lpage>126</lpage>
          . doi:
          <volume>10</volume>
          .11357/ jsam1937.61.2_
          <fpage>119</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>T.</given-names>
            <surname>Hussain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. P.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Anwar</surname>
          </string-name>
          ,
          <article-title>A quantitative real time pcr based method for the detection of phytophthora infestans causing late blight of potato, in infested soil</article-title>
          ,
          <source>Saudi Journal of Biological Sciences</source>
          <volume>21</volume>
          (
          <year>2014</year>
          )
          <fpage>380</fpage>
          -
          <lpage>386</lpage>
          . URL: https://www.sciencedirect.com/science/article/pii/ S1319562X13000892. doi:https://doi.org/10.1016/j.sjbs.
          <year>2013</year>
          .
          <volume>09</volume>
          .012.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>J.</given-names>
            <surname>Johnson</surname>
          </string-name>
          , G. Sharma,
          <string-name>
            <given-names>S.</given-names>
            <surname>Srinivasan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Masakapalli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Dua</surname>
          </string-name>
          ,
          <article-title>Enhanced fieldbased detection of potato blight in complex backgrounds using deep learning</article-title>
          ,
          <source>Plant Phenomics</source>
          <year>2021</year>
          (
          <year>2021</year>
          )
          <fpage>1</fpage>
          -
          <lpage>13</lpage>
          . doi:
          <volume>10</volume>
          .34133/
          <year>2021</year>
          /9835724.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>