<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>based approach for Photographs and Painting Classification using CNN Model</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Hitesh Kumar Sharma</string-name>
          <email>hkshitesh@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff6">6</xref>
          <xref ref-type="aff" rid="aff7">7</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Tanupriya Choudhury</string-name>
          <email>tanupriya1986@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff6">6</xref>
          <xref ref-type="aff" rid="aff7">7</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sachi Nandan Mohanty</string-name>
          <email>sachinandan09@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff6">6</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Shrabanee</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff6">6</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Swagatika</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
          <xref ref-type="aff" rid="aff6">6</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Satabdi Swain</string-name>
          <email>satabdiswain1@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff5">5</xref>
          <xref ref-type="aff" rid="aff6">6</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>&amp; Engineering, VIT-AP University</institution>
          ,
          <addr-line>Amaravati, Andhra Pradesh</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Adjunct Professor, Dept. of Computer Science, Singidunum University, Serbia and School of Computer Science</institution>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>CNN</institution>
          ,
          <addr-line>Deep Learning, Image Processing, Classification, Machine Learning</addr-line>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Dehradun- 248007</institution>
          ,
          <addr-line>Uttarakhand</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>Dept. of Computer Science and Engineering, Siksha 'O' Anusandhan Deemed to be University</institution>
          ,
          <addr-line>Bhubaneswar</addr-line>
        </aff>
        <aff id="aff5">
          <label>5</label>
          <institution>Dept. of Information Technology, (Mtech IT), Utkal University</institution>
          ,
          <addr-line>Bhubaneswar, Odisha -751004</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
        <aff id="aff6">
          <label>6</label>
          <institution>Odisha-751030</institution>
          ,
          <country country="IN">India</country>
        </aff>
        <aff id="aff7">
          <label>7</label>
          <institution>School of Computer Science, University of Petroleum &amp; Energy Studies (UPES)</institution>
          ,
          <addr-line>Energy Acres, Bidholi</addr-line>
        </aff>
      </contrib-group>
      <fpage>220</fpage>
      <lpage>226</lpage>
      <abstract>
        <p>One of the most powerful technologies for dealing with a wide range of real-world difficulties in the fast-paced world of the twenty-first century is Machine Learning (ML). Both regular and differently-abled persons benefit from machine learning. The Convolutional Neural Network (CNN) has been proposed for a variety of applications such as multimedia processing and so on. Here in this research paper, we have described the way and created a Binary Classification model using CNN for identifying the Paintings and Photographs. Each painting and Photograph have been warped using various procedures such as a convolutional layer, dense layer, and Flatten layers. The model is used for Binary Classifications. The wrap has been done at random on a large dataset for CNN training. We explore the architecture of CNN affects the accuracy of the identification. The proposed Model aims to increase the efficiency and accuracy of the model.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        With the advancement of machine learning algorithms [
        <xref ref-type="bibr" rid="ref12 ref21 ref8">2</xref>
        ], there has been development in the field
of computer vision challenges, staying up to date on deep learning that is assisting in the quick evolution
of AI that is becoming a necessity nowadays. Like how a youngster figures out how to perceive objects,
we need to show a calculation and a great many pictures before it can sum up the information and make
predictions for pictures.
      </p>
      <p>
        CNN is a type of neural network that uses convolutional algorithms to CNN works by taking an
image and assigning a weighting to it depending on the image's many objects. CNN [
        <xref ref-type="bibr" rid="ref13 ref9">5</xref>
        ][9] has mainly
used image classifications [12], such as identifying/classifying photos vs. painting. It also has other
functions, such as image segmentation and signal processing.
XXXX-XXXX-XXXX (A. 5)
      </p>
      <p>2020 Copyright for this paper by its authors.</p>
      <p>A CNN can alternatively be constructed as a U-Net design, which consists of two nearly mirrored
CNNs, in this image size output is similar to the image size input and is used in U-net architecture for
image improvement and image segmentation. In astrophysics, CNNs are used to evaluate radio
telescope data and forecast the most plausible visual representation of the data. One of the hottest IT
topics is computer vision, which is to study the machines' ability to identify the images and videos [4].
Computer vision is utilized in the field of self-driving automobiles, robotics, and facial identification;
for this the profoundly specific method known as CNN has gained monstrous headway in PC vision.
CNN has been fabricated utilizing deep learning AI innovation [13].</p>
    </sec>
    <sec id="sec-2">
      <title>2. Literature Review</title>
      <p>Neural Network Architecture (NNA) was researched as a technique for picture classification [1].
The system is comprised of two arrangements of natural eye imitates just as variety grouping
autoencoding. It included a lot of convoluted photographs, yet as the review advanced, the calculation
continuously further developed the MNIST models. The open-source information base MNIST is used
for the preparation set. It also investigated with the dataset of Street View House Numbers, which gave
a more significant result on the grounds that even natural eyes can't differentiate. The ImageNet
challenge is utilized to survey the adequacy of CNN models for picture classification. Since presenting
AlexNet in 2012, they've made gradual enhancements to the plan, which have expanded execution.
GoogLeNet was presented in 2015 by (Szegedy et al. [14], 2015), which was an enhancement for
AlexNet, generally because of a decrease in the measure of boundaries included. Also, in 2014,
(Simonyan and Zisserman, 2014 [6]) presented the VGGNet, which performed well because of the
organization's profundity. The image classification technique dependent on the construction of CNN
was talked about in the diary (CNN). The grounding was completed by taking out the extra face pictures
from the face picture information so that a specific number of face and non-facial pictures were used
for preparing the research information. The picture order framework uses a bi-scale CNN with 120
prepared information and auto-stage preparing on the Face Detection Data Set and Benchmark (FDDB)
to achieve an 81.6 percent location rate with just six bogus up-sides. In contrast, the current status of
the craftsmanship accomplishes around an 80% discovery rate with 50 bogus up-sides.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Proposed CNN model for Paintings and Photograph Identification</title>
      <p>
        We have investigated and motivated bottom-up strategies for increasing the classification accuracy
of CNN models for the picture classification problem. We have used the Keras and TensorFlow [11]
deep learning libraries [
        <xref ref-type="bibr" rid="ref4">3</xref>
        ] to execute the model in Python. This model is open to the public to use and
improve.
3.1.
      </p>
    </sec>
    <sec id="sec-4">
      <title>Convolutional Neural Network (CNN)</title>
      <p>CNN’s are a sort of profound neural organization that is habitually used to assess visual information.
Clinical picture examination, picture and video acknowledgment, picture arrangement, recommender
frameworks [8], normal language preparation, and monetary time series are mostly regions where CNN
can be utilized. A Facial Recognition framework based on CNN is a Deep Learning calculation [7] that
can take an information picture, allot learnable loads and inclinations to different parts of the picture,
and recognize distinctive looks. Deep learning is the innovative way for this research innovation. The
deep learning concept is delegated to artificial intelligence as it can imitate the human thought in an
intelligent way. Ordinarily, the framework will be pre-stacked with hundreds, if not thousands, of
information to make the 'instructional meeting' more productive and faster. It starts by furnishing some
sort of 'preparing' with every one of the information input.
3.2.</p>
    </sec>
    <sec id="sec-5">
      <title>Photographs and Painting Classification Dataset Specifications</title>
      <p>We have used Siddesh Sambasivam's Photographs and Painting Classification Dataset on Kaggle
for our project [10]. This dataset can be found on Kaggle. There are 7041 images in all, including
paintings and photography. It contains 3010 images of paintings and photographs for validation.
(https://www.kaggle.com/iiplutocrat45ii/painting-vs-photograph-classification-dataset). Because of
the limited computing capacity, we have checked the correctness after 10 Epochs. It has been taken
random data samples and trained our model on the 7041 train dataset before testing it on 3010. The
training accuracy was observed to be 94.1 percent after 10 epochs, while the validation accuracy was
found to be 90.73 percent. Table 2 represents the labels distributions in 0’s and 1’s, and Basic CNN
Model Configuration is represented in Table 2.</p>
    </sec>
    <sec id="sec-6">
      <title>The Architecture of the Proposed CNN Model</title>
      <p>The layered distribution architecture is shown in Figure 3. It consists of three layers such as dense,
Flatten, and Convo2d layer. Model used is Inception_resNet_V2.</p>
      <p>In the Conv2d layer, the activation function that is used is ReLu (Rectified Linear initiation work),
which is, in a way, a piecewise direct capacity that mainly yields positive info and returns 0 for any bad
information. It is a direct numerical capacity that requires some investment to prepare and accomplishes
higher exactness. This initiation work likewise supports the goal of disappearing inclination issues that
are normal with other enactment capacities, for example, Sigmoid or TanH. SoftMax, otherwise called
delicate argmax or standardized outstanding capacity, is one more enactment work utilized for yield.
It's mostly used to change over a network's output to a likelihood distribution over projected yield
classes. Figure 3 shows the layer configuration of the model.</p>
    </sec>
    <sec id="sec-7">
      <title>4. Proposed CNN Model Enactment</title>
      <p>The enactment of the proposed model in the research paper is defined. Kaggle provided the dataset
for Photographs and Paintings. Here the image target size has been set to 512px and rescale it using the
Image Data Generator. Due to a lack of resources, we have used 7041 example photos to train our
model. Figure 3 shows some images that were chosen at random.</p>
    </sec>
    <sec id="sec-8">
      <title>5. Experimental Outcomes</title>
      <p>The proposed model gave the outcomes that are being displayed in image form with great accuracy.
The exhibition of our model is assessed utilizing the accompanying measurable boundaries.</p>
      <p>Accuracy: Precision educates us about the rate regarding positive IDs that were genuinely correct.
Review: Recall educates us about the rate regarding real up-sides that were precisely distinguished.</p>
      <p>Exactness is a measurement used to survey arrangement models. Exactness advises us about the rate
regarding the right expectations made by our model.</p>
      <p>Above Output (Figure 3) depicts the accuracy of the proposed model as a Confusion matrix
representation. Actual findings are represented by vertical columns, whereas forecasted images are
represented by horizontal rows.</p>
      <p>The Accuracy (Figure 3 and Figure 3) is increasing, and loss is decreasing for both Training and
validation datasets. After the 15 epochs, the accuracy will become constant, i.e., 97.1%.</p>
    </sec>
    <sec id="sec-9">
      <title>6. Conclusion</title>
      <p>
        In a nutshell, this research work focuses on the classification of the picture by utilizing deep learning
and the TensorFlow framework. It has two main objectives and one subobjective. The objectives are
indistinguishably knotted to the conclusions, which is proved by us in this research study. Firstly, it can
be quantified that all of the findings gained so far have been extremely impressive. Secondly, this
research also focuses on the CNN, which is mostly useful in the image categorization technologies. To
justify the sub-objective, the CNN technique was further examined in depth, beginning with the
assembly, then the training model, and finally, the image was classified into classes. The epochs in
CNN were allowed to control the accuracy in a way avoiding issues like overfitting. Furthermore,
Python was used as the programming language all through this research work as it is viable with the
TensorFlow structure, which in a way takes into account the plan of the framework which is to be done
all together in Python.
7. References
[
        <xref ref-type="bibr" rid="ref13 ref9">5</xref>
        ] A. V. R, H. F. Mahdi, T. Choudhury, T. Sarkar and B. P. Bhuyan, "Freshness Classification of
Hog Plum Fruit Using Deep Learning," 2022 International Congress on Human-Computer
Interaction, Optimization and Robotic Applications (HORA), 2022, pp. 1-6, doi:
10.1109/HORA55278.2022.9799897.
[6] K Simonyan, A Zisserman, Very deep convolutional networks for large-scale image
recognitionarXiv preprint arXiv:1409.1556, 2014.( https://arxiv.org/pdf/1409.1556.pdf )
[7] Vivank Sharma, B. Valaramathi, K. Santhi, Shobhit Srivastava, Sumit Jahagirdar. "Performance
Analysis of the Classifiers for Optical Character Recognition”, 2019 International Conference on
Intelligent Computing and Control Systems (ICCS), 2019.
[8] R. Biswas et al. “A Framework for Automated Database Tuning Using Dynamic SGA Parameters
and Basic Operating System Utilities”, Database Systems Journal vol. III, no. 4/2012.
[9] R Sille, T Choudhury, P Chauhan, D Sharma - , A Systematic Approach for Deep Learning Based
      </p>
      <p>Brain Tumor Segmentation., Ingénierie des Systèmes d'Information, 2021.
[10] Link : https://www.kaggle.com/iiplutocrat45ii/painting-vs-photograph-classification-dataset (Last
accessed on : 1st May 2022 19:58 IST)
[11] M. A. Abadi, P. Barham, E. Brevdo, Z. Chen, C. Citro, for example, is one of the most well-known
companies in the world. Using Tensorflow, we can do large-scale machine learning on
heterogeneous distributed systems.. preprint arXiv:160304467 arXiv:160304467
arXiv:160304467 arXiv:1603044 (2016).
[12] X. Long, H. Lu, Y. Peng, X. Wang, S. Feng, Image classification based on improved VLAD.</p>
      <p>Multimedia Tools Appl. 75(10), 5533–5555.
[13] Mishra, M., Sarkar, T., Choudhury, T. et al. Allergen30: Detecting Food Items with Possible
Allergens Using Deep Learning-Based Computer Vision. Food Anal. Methods (2022).
https://doi.org/10.1007/s12161-022-02353-9.
[14] Christian Szegedy, Wei Liu, Yangqing Jia, Pierre Sermanet, Scott Reed, Dragomir Anguelov,
Dumitru Erhan, Vincent Vanhoucke, Andrew Rabinovich; Proceedings of the IEEE Conference
on Computer Vision and Pattern Recognition (CVPR), 2015, pp. 1-9</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Adam_Optimizer Layer(Layer1)</surname>
          </string-name>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <article-title>model_ crossentropy_categorical(Layer2)</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>128 Model_filters,</mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>3x3 Filter_size,</mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>ReLU Activation(Layer3)</surname>
          </string-name>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>Max-Pooling2d</surname>
          </string-name>
          _
          <article-title>Layer 2x2 size_kernel(Layer4)</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <source>Model_Dropout_Layer</source>
          <volume>20</volume>
          %
          <article-title>(Layer5)</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <article-title>Model_Conv2D Layer 64 filters,</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>5x5 Filter_Size,</mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <surname>ReLU Activation(Layer6) Model</surname>
          </string-name>
          _
          <article-title>Max-Pooling2d_Layer 2x2 Size_kernel(Layer7)</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <source>Model_Dropout_Layer</source>
          <volume>20</volume>
          %
          <article-title>(Layer8)</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <article-title>Model_Con2D Layer 356 model_filters,</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>5x5 Filter_size,</mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <article-title>Relu_Activation(Layer9) Model_Max-Pooling2d Layer 2x2 kernel_size(Layer10)</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <source>Model_Dropout</source>
          <volume>20</volume>
          %
          <article-title>(Layer11)</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <article-title>Model_Flatten layer 2404 Neurons(Layer12)</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <article-title>Model_Dense Layer 128 Neurons(Layer13) Model_Batch Normalization Relu Activation(Layer14)</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <source>Model_Dropout</source>
          <volume>25</volume>
          %
          <article-title>(Layer15)</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <article-title>Model_Dense Layer 512 Neurons(Layer16) Model_Batch Normalization Relu_Activation(Layer17)</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <source>Model_Dropout</source>
          <volume>20</volume>
          %
          <article-title>(Layer18)</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <article-title>2 classes(Layer19) [1</article-title>
          ]
          <string-name>
            <surname>Krishna</surname>
            ,
            <given-names>M</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Neelima</surname>
            ,
            <given-names>M</given-names>
          </string-name>
          &amp; Mane, Harshali &amp; Matcha,
          <string-name>
            <surname>Venu.</surname>
          </string-name>
          (
          <year>2018</year>
          ).
          <article-title>Image identification using</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          <source>neural networks. 7. 614. 10.14419/ijet.v7i2.7</source>
          .
          <fpage>10892</fpage>
          . [2]
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>G.-B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhu</surname>
            ,
            <given-names>Q.-Y.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Siew</surname>
          </string-name>
          , C.
          <article-title>-</article-title>
          K.
          <article-title>Extreme learning machine: Theory and</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          applications.
          <source>Neurocomputing</source>
          <volume>70</volume>
          ,
          <fpage>489</fpage>
          -
          <lpage>501</lpage>
          (
          <year>2006</year>
          ). [3]
          <string-name>
            <surname>Nguyen</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <article-title>et al</article-title>
          .
          <article-title>ML and DL frameworks and libraries for substantial and ample data mining: A</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          survey.
          <source>Artif. Intell. Rev</source>
          .
          <volume>52</volume>
          ,
          <fpage>77</fpage>
          -
          <lpage>124</lpage>
          (
          <year>2019</year>
          ). [4]
          <string-name>
            <surname>Sharma</surname>
            ,
            <given-names>Hitesh</given-names>
          </string-name>
          &amp; Ahmed, Md &amp; Mor, Anurag &amp; Paul, Gaurav &amp; Gupta,
          <string-name>
            <surname>Prashasti.</surname>
          </string-name>
          (
          <year>2021</year>
          ). Deep
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          <source>Learning Approach for Traffic Signs Detection. 10.13140/RG.2.2.19147.11048.</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>