<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>S.A. Moqurrab);</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Binary Rainfall Classification using SMOTE: An Effective Machine Learning Strategy</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Syed Atif Moqurrab</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Abdul Razaque</string-name>
          <email>a.razaque@iitu.edu.kz</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yersain Chinibayev</string-name>
          <email>y.chinibayev@iitu.edu.kz</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Tolganay Chinibayeva</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Forecasting, SMOTE</institution>
          ,
          <addr-line>Data Cleaning, Normalization, Data Balancing, Framework</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>International Information Technology University</institution>
          ,
          <addr-line>Manas St. 34/1, Almaty, 050040</addr-line>
          ,
          <country country="KZ">Kazakhstan</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0003</lpage>
      <abstract>
        <p>Water is critical to human survival. The plants we grow, the animals we boost, and the requirement human body to stay hydrated all depend on water. It is crucial to precisely predict the rainfall for effective utilization of water resources, food productivity, and proper storage of water. Because of recent climate changes, accurate rainfall forecasting has become more complicated than earlier. This paper improves the efficiency and accuracy of rainfall forecasting with the help of data balancing through SMOTE and machine learning. The dataset of twelve years duration was collected from a weather forecasting portal which includes several atmospheric attributes. Preprocessing methodologies are applied first, which include cleaning and normalization of data as well as data balancing using SMOTE. Performance comparison has been made for various machine learning techniques which include Naïve Bayes, MLP SVM, KNN, and Decision. It has been found that Decision Tree outperforms other techniques in terms of forecasting accuracy, precision, recall, and f1 measures. The best accuracy we achieved using the Decision Tree in this research was 99.8% for both rain and no rain classes. Similarly 100% precision,99% recall, and 99% f1measure for no rain class and for rain.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Rain forecasting has applications in water storage and management, flood prevention, agricultural
planning, mobility planning, and many other fields. Accurate rain forecasting is an important and
complex research area. Supervised machine learning techniques have been mostly used in the
literature for this problem. There are various environmental factors such as humidity, wind speed,
pressure, concentrations, and pollutants that have a role in rainfall. In the past, many researchers
have investigated and proposed various methodologies and algorithms for predicting rainfall and
are still engaged in this research area for improved results in terms of efficiency and accuracy
using data mining and machine learning techniques. Machine learning algorithms make use of time
series data by analyzing it for rain prediction.</p>
      <p>
        Time series analysis is an approach for the creation of accurate models with the values of the
variables positioned at periodic intervals [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Reading of time series data supports understanding
of unseen forms of the data and assists in improved examination by using a suitable model for
effective prediction. Time series data is normally gathered over a certain time duration on regular
intervals [
        <xref ref-type="bibr" rid="ref2 ref3 ref4 ref5">2-5</xref>
        ] and can be used for forecasting in multi-domain areas like economic conditions,
stock exchange, and weather, etc. However, weather forecasting with the help of time series data
is a complex job [
        <xref ref-type="bibr" rid="ref6 ref7 ref8">6-8</xref>
        ].
      </p>
      <p>Another method for rainfall forecasting is via statistical methodology, however, it requires lots
of data attributes like local time, seasons, air pressure, cloud conditions, temperature, humidity,
etc. As the nature of rainfall data is non-linear which makes the data noisy and unbalanced, various
techniques need to be applied like data cleaning, normalization, and balancing on it to achieve
higher accuracy in results.</p>
      <p>
        Weather forecasting can help to take necessary measures to avoid human, animal, and
infrastructure losses and to support the development of agriculture, economy, and health of any
country and its people. In this research, we have performed data processing using SMOTE
(Synthetic Over-sampling Technique Minority) and applied various machine learning algorithms
to achieve higher accuracy and efficiency in results as compared to previously performed
experiments by various researchers using data mining techniques [
        <xref ref-type="bibr" rid="ref10 ref11 ref12 ref9">9-12</xref>
        ]. For our experiments, we
have used the rainfall data in Lahore – City of Pakistan for over 12 years (December 2005 to
November 2017) [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
      </p>
      <p>
        For predicting rainfall, a classification framework is applied where the datasets are initially
processed through cleaning, normalization, and balancing. It has been observed that datasets
normally contain inaccurate or omitted values. Using data cleaning, such anomalies can be
removed. Unclean data can lead to a range of issues, including linking errors, model
misspecification, errors in parameter assessment, and wrong examination that in return results in
false conclusions. Whereas, normalization is a process that is frequently used to prepare data for
machine learning. The objective of normalization is to transform the values of numeric columns
into a common scale in the dataset to refer to it, without distorting range differences or losing
information. These pre-processing steps are essential for a smooth classification method with a
high rate of accuracy [
        <xref ref-type="bibr" rid="ref14 ref15">14, 15</xref>
        ]. It has been found that Decision Tree outperforms other techniques
in terms of forecasting accuracy, precision, recall, and f1 measures. The best accuracy we achieved
using the Decision Tree in this research was 99.8% for both rain and no rain classes. Similarly
100% precision,99% recall, and 99% f1measure for no rain class and rain.
      </p>
      <p>The organization of the paper is as follows: Section II discusses related work in the field of
rainfall prediction using machine learning, section III describes the proposed methodology and
techniques adapted, section IV defines the dataset used and its pre-processing, followed by the
experiments performed and results tabulated in Section V. Section VI concludes the paper.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related work</title>
      <p>
        There are numerous methodologies and algorithms for data mining and machine learning to
forecast rainfall [
        <xref ref-type="bibr" rid="ref10 ref11 ref12 ref13 ref14 ref15 ref9">9-15</xref>
        ]. However, we have investigated only those that are closely related to our
approach. Some researchers [
        <xref ref-type="bibr" rid="ref16 ref17">16,17</xref>
        ] have used a neural network model by capturing non-linear
dependencies of past weather modes and future climate states. Some other researchers have used
support vector machines [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] to classify directly for environmental forecasting, however using
SVM, the results get limited in range as compared to neural network methodologies. Few other
techniques have also used Bayesian networks to model and forecast weather [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]. The technique
adapted uses a machine-learning algorithm to find the most prime Bayesian networks and factors
to reduce the computation cost based on different dependencies and the experiments have shown
promising results. In general, in the domain of forecasting and visualization of huge collections of
datasets, SOM (Self-Organizing Map) and Support Vector Machine are the prime machine learning
methodologies.
      </p>
      <p>
        Generally, the experiments for weather forecasting use a two-step approach. First, the dataset
is split into a tiny set of vectors, then these vectors are divided into teams using victimization
clump algorithms. The main objective of hierarchical algorithms is to scale back process prices for
every cluster. The second step provides a rough image for every cluster thus lowering the
prediction, and cost and increasing dependability [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ] as compared to other techniques. The
researchers have used a similar approach in their experiments using a number of different
machine learning algorithms. Comparative analysis has been performed for various machine
learning techniques such as M5 Model Trees, Support Vector Machine, Logistic Regression, Markov
Chain, Radial Basis Neural Network, Genetic Programming, and k-Nearest Neighbor [21] for
rainfall forecasting time series data of 42 towns using numerous climate attributes. The research
verified that machine learning algorithms can perform well compared to the Markov Chain
methodology. There are several other models recommended by the researchers [22-28], however
accurate forecasting examination has not been achieved because of the difficult data structures of
the weather, categorical and dynamic patterns of the weather, noise in data, and dimensionality of
the data. Therefore, there is a requirement for an effective model to predict the weather.
      </p>
      <p>
        The latest study [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] for rain forecast prediction did a comprehensive analysis of binary
classification (Rain and No rain). However, the author uses multiple well-known classifiers with
different data splitting ratios for both the ‘No rain’ and ‘Rain’ classes. Based on their result the ‘No
rain’ class predicted with high precision, recall, and F1 measure respectively as compared to the
‘Rain’ class. To improve the rain class prediction was their future research work. In this paper, we
limit our research to improving the rain class prediction. The details of our proposed methodology
are available in section 3.
      </p>
      <p>Next, we have proposed our methodology that provides higher accuracy in rainfall prediction.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <p>The objective of this research is to compare, validate, verify, and receive higher accuracy in the
result for forecasting rainfall in Lahore - a City in Pakistan using effective techniques, such as
SMOTE and Machine learning.</p>
      <p>The methodology adopts a three-step approach. The first step provides pre-processing on
selected datasets by applying data cleaning, data normalization, and data balancing using SMOTE.
The second step is applying machine learning algorithms to train and classify data. The algorithms
used in our experiments are Naïve Bayes, Logistic Regression, SVM, KNN, MLP, Decision Tree, and
Random Forest. Step three tabulates and evaluates results using accuracy, precession, recall, F1
measure, TP rate, and FP rate. The complete methodology is shown in Figure 1.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Dataset and pre-processing</title>
      <p>Time series models are the premise for any study of the performance of procedures over certain
time period. Time series prediction is a significant region of machine learning [29]. In our
No
methodology, we have used datasets developed using time series models. The dataset includes
many environmental attributes. Table I describes various attribute used, their types and their
units of measurement.</p>
      <p>Typically dataset contains misleading values. Data cleaning process helps in recovering the
missing values. Missing value can create inaccuracy in results. In the cleaning process, we have
replaced the missing values with the mean which is one the most widely used methods in the
literature. Table II shows Valid and Missing records in each attribute with the selected dataset.</p>
      <p>
        Missing values were replaced by mean, and data normalization was applied to maintain the
values in certain boundaries [
        <xref ref-type="bibr" rid="ref1 ref11">1, 11</xref>
        ]. This normalization is performed using Z-Score: a commonly
used mythology for this purpose. The normalization approach deals with the noise via prescribing
the values intervals. However, after this missing value replacement and normalization, the dataset
still contains discrepancies i.e. the data is highly imbalanced. This imbalance means that one class
is represented using a large number of instances whilst the other is represented by a handful
instance [30]. Thus, the data is required to be balanced.
      </p>
      <p>There are many techniques available to balance the distribution of the classification type
variable. In our experiments, we have used SMOTE (Synthetic Minority Oversampling method)
because of its extensive use for data balancing in the literature [30]. SMOTE is a technique that
reduces the effect of getting a few times inside the minority elegance. The strategy includes taking
a subset of records from the minority elegance, intelligently growing new synthetic comparable
times, adding them to the authentic dataset, and using the brand new dataset as a sample in the
schooling procedure for the classifier version [31].</p>
      <sec id="sec-4-1">
        <title>4.1 Classifiers</title>
        <p>4.1.1 Naïve Bayes
Various classifiers were used in this research which is discussed in details in section 4.1.
Naive Bayes (NB) classifier expect that the nearness of a
specific component in a class is
inconsequential to the nearness of some other element. Equation 1 and 2 shows the working of
Bayesian classifier.</p>
        <p>( ⁄ ) =
 ( ⁄ ) ( )</p>
        <p>( )
 ( ⁄ ) =  (  | ) ×  (  | ) × ⋯ ×  ( | ) ×  ( )
 ( ⁄ ) Represents the probability of class (f) given the predicator (Z);
 ( ) Shows the probability of class;
 ( ) Shows the probability of Predicator;
 ( ⁄ ) Represents Likelihood ratio of predicator class.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.1.2 KNN Classifier (KNN)</title>
        <p>To predict new data points, the K-Nearest Neighbor Classifier (KNN) uses a similarity
measures approach. The reason this study uses the KNN algorithm is that it depends entirely on
the resemblance of the characteristics. Selecting the correct value of K is very essential to obtain
ideal outcomes. K's value is the amount of closest neighbors regarded in a vector's classification.




= √∑ = ( −  )</p>
        <p>= ∑ = | −  |
= (∑ = (| −  |) ) ⁄</p>
        <p>Above mentions equations represents the similarity level between two data points’. Yi and Zi
represent “n” data points.
(2)
(3)
(4)
(5)
(6)
(7)</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.1.3 Support Vector Machine (SVM)</title>
        <p>Support Vector Machine "(SVM) is a supervised algorithm for machine learning that can be
used for classification or regression challenges. It is mostly used in classification issues, though.
In this algorithm, each data item is plotted as a point in n-dimensional space (where n is the
number of characteristics you have) with the value of each function being the value of a specific
coordinate. Then, by discovering the hyper-plane that differentiates the two classes very well.</p>
        <p>+  ∑  
  (  ∅ (  ) +  ) ≥  −   
  ≥  ,  =  , … … . 
Where  is a steady capacity, T is a coefficient vector, c is a constant and 
 represents
parameters for the handling of non-input information. The index I marks the instances of N
practice. Note that the class labels are represented by  ∈ ±1 and the independent variables are
represented by   . The kernel is used to convert information into the function space from the</p>
      </sec>
      <sec id="sec-4-4">
        <title>4.1.4 Decision Tree</title>
        <p>input (independent). It should be observed that the greater the  , the greater the penalization of
the mistake.  should, therefore, be carefully selected to prevent overfitting.</p>
        <p>The Decision Tree (DT), another algorithm used in latest anomaly-based IDS studies, is the
same as any tree structure composed of corners, nodes, leaves, etc. Typically, a function and
threshold are applied to a node and the information is divided down the tree where, for instance,
if the information is below a threshold, it goes left and right above a threshold until it ends up in
a final cluster or class [33]. One DT technique is an ID3 algorithm that uses entropy to quantify
data. The entropy is given below.</p>
        <p>Where ( 1,  2, …   ) represents the probabilities of the class labels.</p>
        <p>Gini index is a metric of sample inequality. It has a value of 0 to 1. Gini value index 0 implies
that the sample is completely homogeneous and all components are comparable, whereas Gini
value index 1 implies maximum element inequality. It is the sum of each class's square
probabilities. It is shown as,</p>
        <p>Gini index =  − ∑ =</p>
        <p />
      </sec>
      <sec id="sec-4-5">
        <title>4.1.5 Multilayer Perceptron (MLP)</title>
        <p>A neural network is a sequence of algorithms that attempt to acknowledge fundamental
interactions in a collection of information through a method that mimics the functioning of the
human brain. Neural networks can adapt to altering inputs; therefore, the network produces the
best possible result without redesigning the output requirements [32].
(8)
(9)
n = number of incoming layer inputs
i = counter from 0 to n</p>
      </sec>
      <sec id="sec-4-6">
        <title>4.1.6 Evaluation Metrics</title>
        <p>∫( + ∑ =     )
(10)</p>
        <p>Different metrics are used to evaluate the performance of our proposed model. These are
mentioned below.</p>
        <p>Accuracy =</p>
        <p>+
+
+ +

(11)
(12)
(13)
(14)</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Experiments and results</title>
      <p>Once the data is clean, normalized and balanced, the data is loaded into Weka for analysis and
comparison of various machine learning algorithms. Dataset is split as 30% for testing and 70%
for training. Based on comprehensive experiments results are tabulated. Let’s look into them one
by one.</p>
      <sec id="sec-5-1">
        <title>5.1 Experiment Results of Proposed Method</title>
        <p>The experiment uses 50% percent of data containing “No Rain” class and 50% with “Rain” class
data. Results are tabulated and shown in Table 3. Most of the algorithms have shown better results
with 99% accuracy for both classed “No Rain” and “Rain” respectively, precision recall and F1
Measure. 50:50 Data Balancing Ratio For No Rain and Rain Class Results.</p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2 Comparison of proposed technique with base paper (no rain class)</title>
        <p>The Table 4 shows the comparison of our proposed method with the existing study based on
“No Rain” class. The results based on precisions, Recall and F-1 measure shows that overall our
proposed method improves on average 6%, 0.4% and 4% respectively.</p>
      </sec>
      <sec id="sec-5-3">
        <title>5.3 Comparison of proposed technique with base paper (rain class)</title>
        <p>Similarly in table 5 shows the comparison of our proposed method with the existing study
based on “Rain” class. The results based on precisions, Recall and F-1 measure shows that overall
our proposed method improves on average 58%, 76%, and 72% respectively.</p>
      </sec>
      <sec id="sec-5-4">
        <title>5.4 Critical Analysis</title>
        <p>
          Shabib Aftab, Munir Ahmad [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] used data mining techniques for Lahore rain predication,
experiments showed good results for a no-rain class in terms of precision, recall, and f-measures.
However for rain class these techniques did not perform well and results were not accurate. So
more robust method was needed to solve this problem. In order to improve results for rain class
we proposed SMOTE which balance the data and then this balance data is pass to machine
learning models to train and test the performance. The results based on precisions, Recall and
F1 measure shows that overall our proposed method improves on average 58%, 76%, and 72%
respectively for rain class and precisions, Recall and F-1 measure was improved on average 6%,
0.4% and 4% respectively.
        </p>
        <p>Similarly, figure 5 shows the comparison of accuracy achieved in this research using different
machine learning algorithms. We achieved 95.76% accuracy using Naive Bayes (NB) algorithm.
Similarly, the accuracy we achieved from KNN was 99.15%. 99.15%, 99.8%, and 99.21% accuracy
were achieved using Support Vector Machine (SVM), Decision Tree (DT) and
Multi-layerperceptron (MLP).</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion and future work</title>
      <p>
        In this paper, we have performed rain predication for using seven Machine Learning Algorithms:
Naive Bayes, Support Vector Machine (SVM), Multilayer Perceptron (MLP), K Nearest Neighbor
and Decision Tree. For this purpose, we have used 12 years of time series data from December
2005 to November 2017.The dataset was also used by [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] for experiments and there study showed
that the data they collected produced good results for one class which was no-rain class but rain
class results were not good. The reason was the dataset was imbalance. So in this study data
balancing method name SMOTE was used to balance the dataset and then performed
preprocessing to clean the dataset and replace the missing value by mean. After replacing the
missing values dataset is normalized using Z-score normalization to bring the data into one scale.
After that dataset is divided into two sets one is training having 70% data and another set is testing
having 30% data respectively.
      </p>
      <p>Machine learning algorithms show that data balancing has improved the results. Our
experiment results show that Decision Tree performed well for both the classes in terms of
precision, recall and f1-scores. The proposed model can be recommended for the major cities in
Pakistan for rain prediction in practice. In future we will use deep learning approach like LSTM to
identify the behavior of weather time wise.</p>
    </sec>
    <sec id="sec-7">
      <title>7. References</title>
      <p>[21] S. Cramer, M. Kampouridis, A. A. Freitas, and A. K. Alexandridis (2017). An extensive
evaluation of seven machine learning methods for rainfall prediction in weather
derivatives,” Expert Syst. Appl., vol. 85, pp. 169– 181.
[22] C. Zhang, W.-B. Chen, X. Chen, R. Tiwari, L. Yang and G. Warner, “A Multimodal Data Mining
Framework for Revealing Common Sources of Spam Images”, Journal of multimedia, vol. 4,
no. 5. October, pp. 313-320.
[23] C. Li, M. Zhang, C. Xing and J. Hu. (2011). Survey and Review on Key Technologies of Column</p>
      <p>Oriented Database Systems. Computer Science, vol. 37, no. 12, pp. 1-8.
[24] M. Zhang. (2011). Application of Data Mining Technology in Digital Library”, Journal of</p>
      <p>Computers, vol. 6, no. 4, pp. 761-768.
[25] C.-W. Shen, H.-C. Lee, C.-C. Chou and C.-C. Cheng. (2011). Data Mining the Data Processing</p>
      <p>Technologies for Inventory Management”, Journal of Computers, vol. 6, no. 4, pp. 784-791.
[26] Z. Danping and D. Jin. (2011). The Data Mining of the Human Resources Data Warehouse in</p>
      <p>University Based on Association Rule”, Journal of Computers, vol. 6, no. 1, pp. 139-146.
[27] J. Jiang, B. Guo, W. Mo and K. Fan. (2012). Block-Based Parallel Intra Prediction Scheme for</p>
      <p>HEVC”, Journal of Multimedia, vol. 7, no. 4, pp. 289-294.
[28] S.-Y. Yang, C.-M. Chao, P.-Z. Chen and C.-Hao. (2011). SunIncremental Mining of Closed
Sequential Patterns in Multiple Data Streams”, Journal of Networks, vol. 6, no. 5, pp.
728735.
[29] G.Vamsi Krishna (2015). An Integrated Approach for Weather Forecasting based on Data
Mining and Forecasting Analysis” International Journal of Computer Applications (0975 –
8887) Volume 120 – No.11, pp.26-29.
[30] Nele Verbiest, Enislay Ramento, Chris Cornelis, Francisco Herrera. (2014). Preprocessing
noisy imbalanced datasets using SMOTE enhanced with fuzzy rough prototype selection”.</p>
      <p>Elsevier - Applied Soft Computing Journal. 511–517.
[31] Lina Guzman, DIRECTV, Data sampling improvement by developing SMOTE technique in</p>
      <p>SAS, Paper 3483-2015.
[32] Almiani, M., AbuGhazleh, A., Al‐Rahayfeh, A., &amp; Razaque, A. (2020). Cascaded hybrid
intrusion detection model based on SOM and RBF neural networks. Concurrency and
Computation: Practice and Experience, 32(21), e5233.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Razaque</surname>
            , Abdul, Marzhan Abenova, Munif Alotaibi, Bandar Alotaibi, Hamoud Alshammari, Salim Hariri, and
            <given-names>Aziz Alotaibi.</given-names>
          </string-name>
          (
          <year>2022</year>
          ).
          <article-title>Anomaly detection paradigm for multivariate time series data mining for healthcare</article-title>
          .
          <source>Applied Sciences</source>
          <volume>12</volume>
          , no.
          <volume>17</volume>
          :
          <fpage>8902</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>N.</given-names>
            <surname>Mishra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. K.</given-names>
            <surname>Soni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A. K.</given-names>
            <surname>Upadhyay</surname>
          </string-name>
          (
          <year>2017</year>
          ).
          <article-title>A Comprehensive Survey of Data Mining Techniques on Time Series Data for Rainfall Prediction</article-title>
          ,” vol.
          <volume>11</volume>
          , no.
          <issue>2</issue>
          , pp.
          <fpage>168</fpage>
          -
          <lpage>184</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ahmad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Aftab</surname>
          </string-name>
          , and
          <string-name>
            <surname>S. S. Muhammad.</surname>
          </string-name>
          (
          <year>2017</year>
          ).
          <article-title>Machine Learning Techniques for Sentiment Analysis: A Review,”</article-title>
          <source>Int. J. Multidiscip. Sci. Eng</source>
          ., vol.
          <volume>8</volume>
          , no.
          <issue>3</issue>
          , p.
          <fpage>27</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Razaque</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Amsaad</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hariri</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Almasri</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rizvi</surname>
            ,
            <given-names>S. S.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Frej</surname>
            ,
            <given-names>M. B. H.</given-names>
          </string-name>
          (
          <year>2020</year>
          ).
          <article-title>Enhanced grey risk assessment model for support of cloud service provider</article-title>
          .
          <source>IEEE Access</source>
          ,
          <volume>8</volume>
          ,
          <fpage>80812</fpage>
          -
          <lpage>80826</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>K.</given-names>
            <surname>Lu</surname>
          </string-name>
          and
          <string-name>
            <surname>L. Wang.</surname>
          </string-name>
          (
          <year>2011</year>
          ).
          <source>A Novel Nonlinear Combination Model Based on Support Vector Machine for Rainfall Prediction. 2011 Fourth Int. Jt. Conf. Comput. Sci. Optim</source>
          ., pp.
          <fpage>1343</fpage>
          -
          <lpage>1346</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>W. C.L.</given-names>
            and
            <surname>K</surname>
          </string-name>
          .-W. Chau. (
          <year>2012</year>
          ).
          <source>Prediction of Rainfall Time Series Using Modular Soft Computing Methods. Eng. Appl. Artif. Intell.</source>
          , vol.
          <volume>26</volume>
          , no.
          <issue>852</issue>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>37</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>V. B.</given-names>
            <surname>Nikam</surname>
          </string-name>
          and
          <string-name>
            <surname>B. B. Meshram.</surname>
          </string-name>
          (
          <year>2013</year>
          ).
          <article-title>Modeling Rainfall Prediction Using Data Mining Method: A Bayesian Approach</article-title>
          .
          <source>2013 Fifth Int. Conf. Comput. Intell. Model. Simul.</source>
          , pp.
          <fpage>132</fpage>
          -
          <lpage>136</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Al-Qubaydhi</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alenezi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alanazi</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Senyor</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alanezi</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alotaibi</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          , ... &amp;
          <string-name>
            <surname>Alotaibi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          (
          <year>2022</year>
          ).
          <article-title>Detection of unauthorized unmanned aerial vehicles using YOLOv5 and transfer learning</article-title>
          .
          <source>Electronics</source>
          ,
          <volume>11</volume>
          (
          <issue>17</issue>
          ),
          <fpage>2669</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Shabib</given-names>
            <surname>Aftab</surname>
          </string-name>
          , Munir Ahmad, Noureen Hameed, Muhammad Salman Bashir, Iftikhar Ali, Zahid Nawaz. (
          <year>2018</year>
          ).
          <article-title>Rainfall Prediction in Lahore City using Data Mining Techniques</article-title>
          .
          <source>Int. J. of Advanced Computer Science and Applications</source>
          , Vol.
          <volume>9</volume>
          , No.
          <issue>4</issue>
          , pp.
          <fpage>254</fpage>
          -
          <lpage>260</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ahmad</surname>
          </string-name>
          and
          <string-name>
            <surname>S. Aftab.</surname>
          </string-name>
          (
          <year>2017</year>
          ).
          <article-title>Analyzing the Performance of SVM for Polarity Detection with Different Datasets, Int</article-title>
          .
          <source>J. Mod. Educ. Comput. Sci.</source>
          , vol.
          <volume>9</volume>
          , no.
          <issue>10</issue>
          , pp.
          <fpage>29</fpage>
          -
          <lpage>36</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ahmad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Aftab</surname>
          </string-name>
          , and
          <string-name>
            <surname>S. S. Muhammad.</surname>
          </string-name>
          (
          <year>2017</year>
          ).
          <article-title>Machine Learning Techniques for Sentiment Analysis: A Review,”</article-title>
          <source>Int. J. Multidiscip. Sci. Eng</source>
          ., vol.
          <volume>8</volume>
          , no.
          <issue>3</issue>
          , p.
          <fpage>27</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ahmad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Aftab</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Ali</surname>
          </string-name>
          , and
          <string-name>
            <surname>N. Hameed.</surname>
          </string-name>
          (
          <year>2017</year>
          ).
          <article-title>Hybrid Tools and Techniques for Sentiment Analysis: A Review,”</article-title>
          <source>Int. J. Multidiscip. Sci. Eng</source>
          ., vol.
          <volume>8</volume>
          , no. 3.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>[13] “http://ru8.rp5.ru/Weather_archive_in_Lahore.” [Online]. Available: http://ru8.rp5.ru/Weather_archive_in_Lahore.</mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>C.</given-names>
            <surname>Sivapragasam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Liong</surname>
          </string-name>
          , and
          <string-name>
            <surname>M. Pasha.</surname>
          </string-name>
          (
          <year>2001</year>
          ).
          <article-title>Rainfall and runoff forecasting with SSASVM approach</article-title>
          ,” J. Hydroinformatics, no.
          <source>April</source>
          <year>2016</year>
          , pp.
          <fpage>141</fpage>
          -
          <lpage>152</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>D.</given-names>
            <surname>Isa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. H.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. P.</given-names>
            <surname>Kallimani</surname>
          </string-name>
          , and
          <string-name>
            <surname>R. RajKumar.</surname>
          </string-name>
          (
          <year>2008</year>
          ).
          <article-title>Text Document Preprocessing with the Bayes Formula for Classification Using the Support Vector Machine,”</article-title>
          <source>IEEE Trans. Knowl. Data Eng.</source>
          , vol.
          <volume>20</volume>
          , no.
          <issue>9</issue>
          , pp.
          <fpage>1264</fpage>
          -
          <lpage>1272</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Krasnopolsky</surname>
          </string-name>
          ,
          <string-name>
            <surname>Vladimir</surname>
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>and Michael S. FoxRabinovitz.</surname>
          </string-name>
          (
          <year>2006</year>
          ).
          <article-title>Complex hybrid models combining deterministic and machine learning components for numerical climate modeling and weather prediction</article-title>
          .
          <source>Neural Networks. 19.2</source>
          . pp.
          <fpage>122</fpage>
          -
          <lpage>134</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Lai</surname>
            ,
            <given-names>Loi</given-names>
          </string-name>
          <string-name>
            <surname>Lei</surname>
          </string-name>
          , et al. (
          <year>2004</year>
          ).
          <article-title>Intelligent weather forecast</article-title>
          .
          <source>”Machine Learning and Cybernetics</source>
          ,
          <year>2004</year>
          . Proceedings of 2004 International Conference on. Vol.
          <volume>7</volume>
          . IEEE.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Radhika</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>M. Shashi.</surname>
          </string-name>
          (
          <year>2009</year>
          ).
          <article-title>Atmospheric temperature prediction using support vector machines</article-title>
          . “
          <source>International Journal of Computer Theory and Engineering</source>
          <volume>1</volume>
          .1:
          <fpage>55</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Antonio</surname>
            <given-names>S.</given-names>
          </string-name>
          , et al. (
          <year>2002</year>
          ).
          <article-title>Bayesian networks for probabilistic weather prediction</article-title>
          .
          <source>”15th Eureopean Conference on Artificial Intelligence (ECAI).</source>
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>R. Usha</given-names>
            <surname>Rani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.K.</given-names>
            <surname>Rama Krishna</surname>
          </string-name>
          <string-name>
            <surname>Rao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. Kiran</given-names>
            <surname>Kumar Reddy</surname>
          </string-name>
          (
          <year>2015</year>
          ).
          <article-title>An Efficient Machine Learning Regression Model for Rainfall Prediction</article-title>
          ”
          <source>International Journal of Computer Applications</source>
          (
          <volume>0975</volume>
          -
          <fpage>8887</fpage>
          ) Volume
          <volume>116</volume>
          - No.
          <issue>23</issue>
          , pp.
          <fpage>25</fpage>
          -
          <lpage>30</lpage>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>