<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Utilization of Machine Learning in Recognition of Rocks and Mock-mines by Sonar Chirp Signals</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Yurii Kryvenchuk</string-name>
          <email>yurii.p.kryvenchuk@lpnu.ua</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mykhailo Dmytryshyn</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>CEUR Workshop Proceedings</string-name>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Lviv, Ukraine</string-name>
        </contrib>
      </contrib-group>
      <fpage>0000</fpage>
      <lpage>0002</lpage>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>.1</p>
    </sec>
    <sec id="sec-2">
      <title>3. Methods</title>
      <sec id="sec-2-1">
        <title>3.1. Dataset</title>
        <p>The research utilized the "Connectionist Bench (Sonar, Mines vs. Rocks)" dataset from the UCI
Machine Learning repository. The dataset is a csv file in which sonar patterns are stored. These
patterns result from bouncing sonar signals off a metal cylinder and rocks, each explored across
various angles and conditions. The sonar signals transmitted are frequency-modulated chirps,
ascending in frequency, and were captured from diverse aspect angles—spanning 90 degrees for
the cylinder and 180 degrees for the rock. Each pattern consists of 60 numerical values within
the range of 0.0 to 1.0. These numbers denote the energy within specific frequency bands,
integrated over defined time periods. Notably, the integration aperture for higher frequencies
occurs later in time due to their transmission later in the chirp.</p>
        <p>The labels assigned to each record are "R" for rocks and "M" for mines (metal cylinders). While
the labels exhibit an ascending order corresponding to the aspect angle, they do not directly
encode the angle information.</p>
      </sec>
      <sec id="sec-2-2">
        <title>3.2. Data processing and organization methods</title>
        <p>In the context of the work on "Utilization of Machine Learning in recognition of rocks and
mockmines by sonar chirp signals," Label Encoding is employed to convert the class labels (categories)
into numerical values. For the task of recognizing rocks ("R") and mock-mines ("M") based on
sonar chirp signals, the classes can be encoded into numerical values.</p>
        <p>
          For example, if there is a column with class labels like:
["R", "M", "R", "R", "M", "M", "R", "M", "R", "R"]
Label Encoding can be used to transform these classes into numerical values, for instance:
[
          <xref ref-type="bibr" rid="ref1 ref1 ref1 ref1 ref1 ref1">1, 0, 1, 1, 0, 0, 1, 0, 1, 1</xref>
          ]
        </p>
        <p>Here, "R" has been assigned the value 0, and "M" has been assigned the value 1. This
conversion allows machine learning algorithms to work with the data, as many algorithms
require numerical values for both input and output.</p>
        <p>Label Encoding can be performed using libraries like scikit-learn in Python, utilizing the
LabelEncoder class. This encoding is particularly useful when dealing with categorical data in
machine learning models. Ensemble methods, specifically AdaBoost-Samme, were employed to
leverage the strengths of multiple weak learners. Decision trees, logistic regression, and random
forests were individually used as base classifiers within the ensemble framework to assess the ir
impact on classification accuracy. Various neural network architectures were explored.
Techniques such as dropout and L2 regularization were applied to mitigate overfitting and
enhance generalization performance. The performance of each model was assessed using
accuracy as the result of cross validation score.</p>
      </sec>
      <sec id="sec-2-3">
        <title>3.3. ML Methods</title>
        <p>In this study, a diverse set of machine learning algorithms has been employed to discern patterns
and classify sonar signals. The algorithms chosen demonstrate versatility in handling the
complexity of the data and offer a comprehensive exploration of the recognition task. The
following algorithms have been applied:</p>
      </sec>
      <sec id="sec-2-4">
        <title>3.4. Overfitting</title>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>4. Experiment</title>
      <sec id="sec-3-1">
        <title>4.1. Dataset Preprocessing:</title>
      </sec>
      <sec id="sec-3-2">
        <title>4.2. Evaluation</title>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>5. Results</title>
      <sec id="sec-4-1">
        <title>Algorithm</title>
      </sec>
      <sec id="sec-4-2">
        <title>AdaBoost-Samme - decision tree</title>
      </sec>
      <sec id="sec-4-3">
        <title>AdaBoost-Samme - logistic regression</title>
      </sec>
      <sec id="sec-4-4">
        <title>AdaBoost-Samme - random forest</title>
      </sec>
      <sec id="sec-4-5">
        <title>Decision Tree</title>
      </sec>
      <sec id="sec-4-6">
        <title>Accuracy</title>
        <p>71.12%
79.76%
87.50%
73.52%</p>
      </sec>
      <sec id="sec-4-7">
        <title>Decision Tree - min cost complexity pruning</title>
      </sec>
      <sec id="sec-4-8">
        <title>Gaussian process - Laplace approximation</title>
      </sec>
      <sec id="sec-4-9">
        <title>K-nearest neighbors vote</title>
      </sec>
      <sec id="sec-4-10">
        <title>Logistic Regression</title>
      </sec>
      <sec id="sec-4-11">
        <title>Logistic Regression - L1</title>
      </sec>
      <sec id="sec-4-12">
        <title>Logistic Regression - L2</title>
      </sec>
      <sec id="sec-4-13">
        <title>Logistic Regression - L1 and L2</title>
      </sec>
      <sec id="sec-4-14">
        <title>Multi-layer Perceptron</title>
      </sec>
      <sec id="sec-4-15">
        <title>Multi-layer Perceptron - L2</title>
      </sec>
      <sec id="sec-4-16">
        <title>Neural Network</title>
      </sec>
      <sec id="sec-4-17">
        <title>Neural Network - dropout</title>
      </sec>
      <sec id="sec-4-18">
        <title>Neural Network - L2</title>
      </sec>
      <sec id="sec-4-19">
        <title>Neural Network - dropout and L2</title>
        <p>Random Forest</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>6. Discussions</title>
      <sec id="sec-5-1">
        <title>6.1. Effectiveness of methods</title>
      </sec>
      <sec id="sec-5-2">
        <title>6.2. Comparison with Previous Research</title>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Conclusions</title>
      <p>systems.
[14] W. Gong, J. Tian, J, Liu, Underwater Object Classification Method Based on Depthwise
Separable Convolution Feature Fusion in Sonar Image, Applied Sciences 12 (2022).
https://doi.org/10.3390/app12073268.
[15] Harsh Yadav, Dropout in Neural Networks, 2022.</p>
      <p>URL: https://towardsdatascience.com/dropout-in-neural-networks-47a162d621d9
[16] IBM, What is overfitting?, 2024. URL: https://www.ibm.com/topics/overfitting.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>X.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <surname>R</surname>
          </string-name>
          , Dong,
          <string-name>
            <surname>Z</surname>
          </string-name>
          ,Lv,
          <source>Deep Learning-Based Classification of Raw Hydroacoustic Signal, Journal of Marine Science and Engineering</source>
          <volume>11</volume>
          (
          <year>2023</year>
          ). https://doi.org/10.3390/jmse11010003.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Jason</given-names>
            <surname>Brownlee</surname>
          </string-name>
          ,
          <source>Dropout Regularization in Deep Learning Models with Keras</source>
          ,
          <year>2022</year>
          . URL: https://machinelearningmastery.com
          <article-title>/dropout-regularization-deep-learning-modelskeras/.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Steiniger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kraus</surname>
          </string-name>
          , T. Meisen,
          <article-title>Survey on deep learning based computer vision for sonar imagery</article-title>
          ,
          <source>Engineering Applications of Artificial Intelligence</source>
          <volume>114</volume>
          (
          <year>2022</year>
          ). https://doi.org/10.1016/j.engappai.
          <year>2022</year>
          .105157
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>D.</given-names>
            <surname>Karimanzira</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Renkewitz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Shea</surname>
          </string-name>
          ,
          <source>Object Detection in Sonar Images, Electronics</source>
          <volume>9</volume>
          (
          <year>2020</year>
          ). https://doi.org/10.3390/electronics9071180
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.</given-names>
            <surname>Fernandes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Junior</surname>
          </string-name>
          ,
          <article-title>Deep Learning Models for Passive Sonar Signal Classification of Military Data</article-title>
          ,
          <source>Remote Sensing 14</source>
          , (
          <year>2022</year>
          ). https://doi.org/10.3390/rs14112648
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6] OpenGenus, Advantages and Disadvantages of Logistic Regression,
          <year>2024</year>
          . URL: https://iq.opengenus.org/advantages-and
          <article-title>-disadvantages-of-logistic-regression/.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>IBM</surname>
          </string-name>
          ,
          <article-title>What is a Decision Tree?</article-title>
          ,
          <year>2024</year>
          . URL: https://www.ibm.com/topics/decisiontrees#:~:text=
          <source>A%20decision%20tree%20is%20a,internal%20nodes%20and%20leaf%20 nodes.</source>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A.</given-names>
            <surname>Nagpal</surname>
          </string-name>
          ,
          <source>L1 and L2 Regularization Methods</source>
          ,
          <year>2017</year>
          . URL: https://towardsdatascience.com/l1-and
          <string-name>
            <surname>-</surname>
          </string-name>
          l2
          <string-name>
            <surname>-</surname>
          </string-name>
          regularization
          <article-title>-methods-ce25e7fc831c</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Tian</surname>
          </string-name>
          ,
          <string-name>
            <surname>Y. Zhang,</surname>
          </string-name>
          <article-title>A comprehensive survey on regularization strategies in machine learning</article-title>
          ,
          <source>Information Fusion</source>
          , (
          <year>2021</year>
          ). https://doi.org/10.1016/j.inffus.
          <year>2021</year>
          .
          <volume>11</volume>
          .005
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>J.</given-names>
            <surname>Shubham</surname>
          </string-name>
          ,
          <source>An Overview of Regularization Techniques in Deep Learning</source>
          ,
          <year>2023</year>
          . URL: https://www.analyticsvidhya.com/blog/tag/regularization-in
          <article-title>-deep-learning/.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>L.</given-names>
            <surname>Ruhela</surname>
          </string-name>
          , Droput Rtgularization,
          <year>2023</year>
          . URL: https://ruhelalakshya.medium.com/dropout - regularization-b27885b4c55b.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>N.</given-names>
            <surname>Srivastava</surname>
          </string-name>
          , G. Hinton, (Eds.),
          <article-title>Dropout: A Simple Way to Prevent Neural Networks from Overfitting</article-title>
          ,
          <source>Journal of Machine Learning Research</source>
          <volume>15</volume>
          , (
          <year>2014</year>
          ):
          <fpage>1929</fpage>
          -
          <lpage>1958</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>G.</given-names>
            <surname>Geeks</surname>
          </string-name>
          ,
          <source>Logistic Regression in Machine Learning</source>
          ,
          <year>2023</year>
          . URL: https://www.geeksforgeeks.org/understanding-logistic-regression/.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>