<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Hyperspectral Image Reconstruction in Remote Sensing: LaplaceGAN Synthesis Coupled with VGG-UNet Classification</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Shikha Sain</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Dr. Monika Saxena</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science</institution>
          ,
          <addr-line>Banasthali Vidyapith, Jaipur ,Rajasthan 304022</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>Enhancing and reconstructing environmental images involve refining visual data to improve quality and reconstructing scenes. In remote sensing, this aids in accurate analysis, contributing to advanced understanding and decisionmaking. This study focuses on advancing hyperspectral image analysis in remote sensing through the design of a deep learning-based model aimed at enhancing and reconstructing environmental images. An integral aspect involves introducing a novel approach using LaplaceGAN to generate synthetic images with high fidelity, building upon real images as a foundational basis. Furthermore, the study proposes the implementation of a specialized VGGUNet architecture tailored for the classification of hyperspectral images, specifically addressing the nuances of remote sensing data. To assess the model's efficacy, a comparative analysis is conducted, pitting the performance of VGG-UNet against alternative methods such as Res-UNet and Faster R-CNN in the context of remote sensing image classification. This research aims to contribute to the field by designing a deep learning model that not only analyzes hyperspectral images comprehensively but also enhances and reconstructs environmental images, thereby advancing the most recent methods for better comprehension and judgement in a range of remote sensing applications.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Hyperspectral images</kwd>
        <kwd>Remote sensing</kwd>
        <kwd>LaplaceGAN</kwd>
        <kwd>VGG-UNet 1</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>A new era in image processing has been brought about by the development of deep
learning. In the realm of environmental science, where the acquisition of high-quality
imagery is crucial for accurate analysis and decision-making, Deep learning model
applications are now the main focus of research and development. Environmental images,
ranging from satellite captures of Earth's surface to ground-level snapshots of ecosystems,
are often beset with challenges that compromise their utility[1]. Factors such as
atmospheric interference, low resolution, and noise frequently degrade the visual quality of
these images, impeding their efficacy in various applications such as environmental
monitoring, land cover classification, and climate change analysis.</p>
      <p>Neural networks with numerous layers are used in deep learning, a type of machine
learning, to automatically learn and extract characteristics from incoming data.
Convolutional neural networks (CNNs) are a particular kind of neural networks intended
for tasks involving images. Have emerged as powerful tools for deciphering complex
patterns within environmental images. Enhanced images contribute to more accurate land
cover classification, facilitate object detection, and improve anomaly detection in
environmental monitoring [2]. In applications such as disaster management, climate
monitoring, and ecological preservation, the availability of high-quality images is
instrumental in informed decision-making.</p>
      <p>However, deploying deep learning models for environmental image enhancement is not
without its challenges. Limited availability of diverse and comprehensive datasets,
interpretability of deep learning models, and computational demands pose significant
hurdles [3]. This research paper seeks to delve into the transformative deep learning's
promise for improving and rebuilding environmental imagery. By examining the most
recent developments, difficulties, and possible uses, this study seeks to add to the growing
conversation on how deep learning may influence environmental image processing in the
future. Researchers and scientists are now investigating the use of environmental remote
sensing, also known as imaging spectroscopy, to find and identify minerals, terrestrial
plants, artefacts, and backdrops, we collect hyper spectral data.</p>
      <p>The following are the research's primary goals:
1. To provide deep learning frameworks for the thorough examination and
categorization of hyperspectral photos acquired by remote sensing.
2. To introduce an innovative approach based on LaplaceGAN for generating
synthetic images with high fidelity, utilizing real images as a basis.
3. To propose the implementation of a VGG-UNet architecture specifically created
specifically for the purpose of classifying hyperspectral photos obtained using
remote sensing technology.
4. Conduct a comparative analysis, pitting the performance of VGG-UNet against
alternative methods such as Res-UNet and Faster R-CNN, to assess their efficacy
in the context of remote sensing image classification.
2. Literature Review</p>
      <p>[4] investigated that deep learning has shown to be one of the most effective machine
learning strategies for a range of inference problems in recent years. [5] investigated the
impact of image motion artifacts on cardiac magnetic resonance (MR) segmentation and
evaluated many methods for concurrently correcting aberration and segmenting heart
chambers. [6] reviewed the state of the art in terms of research on using deep learning to
several picture fusion scenarios, including multi-modal, sharpening, and image fusion from
digital photography.[7]. examined in order to provide a comprehensive summary of recent
advancements in deep learning-based image super-resolution.</p>
      <p>[8] investigated the current approaches to data augmentation, encouraging
advancements, and meta-level choices for data augmentation implementation will all be
covered in this study. [9] outlined a deep learning architecture of U-net-based mapping
paradigm for urban villages. The study area is located in Guangzhou City, China. For
investigation, a building boundary vector file and an eight pan-sharpened band, 0.5-meter
spatial resolution Worldview satellite picture were employed. [10] proposed the use of
deep learning based on augmented convolutional neural networks (CNNs) for the real-time
detection of apple leaf disease. The apple leaf disease dataset (ALDD) is first created in this
study by using techniques for picture interpretation and data augmentation. The dataset
consists of complex photos captured in actual field situations as well as images captured in
laboratories.</p>
      <sec id="sec-1-1">
        <title>Technology</title>
      </sec>
      <sec id="sec-1-2">
        <title>Major Inclusion</title>
      </sec>
      <sec id="sec-1-3">
        <title>Parameters</title>
        <p>[11]
[12]
[13]
[14]
CNN</p>
        <sec id="sec-1-3-1">
          <title>Decision</title>
          <p>Classifier</p>
        </sec>
        <sec id="sec-1-3-2">
          <title>Tree</title>
        </sec>
        <sec id="sec-1-3-3">
          <title>Target-Adapted CNN</title>
        </sec>
        <sec id="sec-1-3-4">
          <title>GIS Tool</title>
        </sec>
        <sec id="sec-1-3-5">
          <title>Utilizing Partial Spectral Convolutions for Image In Dimension painting for Irregular Holes</title>
        </sec>
        <sec id="sec-1-3-6">
          <title>Using data from Landsat-7 Classifier Six</title>
          <p>ETM+, a decision tree Land Use Classes
classifier may be used to
categorize land uses in an
agricultural region</p>
          <p>Utilizing time series DBEST
segmentation to detect shifts
in vegetation patterns</p>
          <p>GIS, a possible tool for Thematic Info
enabling the creation and use
of thematic data to assess the
potentiality of groundwater</p>
        </sec>
        <sec id="sec-1-3-7">
          <title>Image fusion can be Frequency attempted in the frequency Domain domain</title>
        </sec>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>3. Methodology</title>
      <p>3.1. Dataset Description</p>
      <p>In this study, we collected hyper spectral images of remote sensing from below website.
https://www.kaggle.com/datasets/sciencelabwork/hyperspectral-image-sensing
datasetground-truth. The process of hyperspectral remote sensing involves the collection and
examination of a broad spectrum of electromagnetic wavelengths.
3.2. Data Pre-processing</p>
      <p>Data preprocessing involves the application of techniques to handle various aspects of
the data:
•
•</p>
      <p>Data Cleaning: The dataset is inspected for artifacts, noise, and inconsistencies.
The goal of this procedure is to find and fix any problems that might compromise
the accuracy and dependability of the data. Common data cleaning techniques
include removing outliers, addressing missing values, and handling corrupted
data points.</p>
      <p>Data Normalization: Normalization is performed to ensure that all features in
the dataset have a consistent scale. This step is crucial for training machine
learning models because it stops certain characteristics from controlling the
learning process because of variances in their strength.
3.3. LaplaceGAN</p>
      <p>The LaplaceGAN (LapGAN) serve a crucial purpose in environmental image
enhancement and reconstruction by leveraging their generative capabilities to enhance the
photos' clarity, authenticity, and interpretive power. As shown in Fig. 1, the generator and
discriminator are the two main parts of the Laplace Generative Adversarial Network
(LapGAN).</p>
      <p>Within the LapGAN framework, the generator is designed to intake both noise and labels
as input. In this setup, the discriminator not only provides the probability of authenticity
(real or fake) but also specifies the category of the input sample. The intricacies of the model
are elucidated as follows:</p>
      <sec id="sec-2-1">
        <title>Generator</title>
        <p>In the Laplace GAN, the generator processes noise vectors and load pattern labels
encoded in one-hot format to produce synthetic load profiles. The architecture involves a
fully-connected layer for initial mapping, followed by two transposed convolution layers
with a 2 × 2 stride and a 4 × 4 kernel for up-sampling. To stabilize training and ensure
smooth gradient flow, batch normalization layers are applied after each computational step.
The final transposed convolution layer, crucial for shaping generated loads within the same
interval as normalized real loads, uses a Sigmoid activation function.</p>
      </sec>
      <sec id="sec-2-2">
        <title>Discriminator</title>
        <p>The discriminator is used to assess the authenticity and quality of the model-generated
pictures of synthetic environments. Its main objective is to differentiate between produced
and actual pictures. Ensuring the produced images exhibit biologically plausible structures.
Engaging in adversarial training with the generator, the discriminator continually adapts to
maintain its ability to differentiate between real and synthetic data. By providing feedback
to the generator, the discriminator guides the improvement of synthetic images, fostering a
dynamic interplay that refines both components. Ultimately, the discriminator acts as a
critical quality control mechanism, contributing to the production of high-quality,
biologically relevant environmental images through the LapGAN training process.</p>
        <p>The detailed results of the conducted
3.4. Build Model</p>
      </sec>
      <sec id="sec-2-3">
        <title>VGG-UNET</title>
        <p>A deep learning architecture called the Proposed VGG-UNet combines the advantages of
the VGG16 and U-Net models [16] While the U-Net model is an architecture ideally suited
for image segmentation, the VGG16 model is a potent feature extractor as shown in figure.</p>
        <p>Figure 2 Architecture of VGG-UNET</p>
        <p>The encoder and the decoder are the two primary components of the VGG-UNet
architecture. While the decoder creates the segmentation mask, the task of extracting
features from the input image falls to the encoder. The VGG16 model serves as the basis for
the encoder in the proposed VGG-UNet. The max pooling layer comes after each of the 16
convolutional layers that make up the VGG16 model. Five blocks with varying numbers of
filters each comprise the convolutional layers. Every block has filters that are 3 by 3 in size.
The U-Net model serves as the foundation for the proposed VGG-UNet decoder.
3.5. Test-Train Split</p>
        <p>The training set and the test set are the two primary subsets of the dataset. The test set
is used to evaluate how effectively the trained model generalizes to fresh, untested data,
while the training set is used to train the model. A validation set is often constructed in order
to avoid overfitting and fine-tune the model during training.</p>
      </sec>
      <sec id="sec-2-4">
        <title>Train the network:</title>
        <p>Using the train data, the proposed VGG-UNet model and Res-UNet, Faster R-CNN are
trained. The performance of the proposed model is assessed and contrasted with these two
additional Res-UNet, Faster R-CNN architectures.
3.6. Performance Metrics</p>
        <p>Subsequently, the training set is employed for the purpose of training various models,
including VGG-UNet model and Res-UNet, Faster R-CNN. The efficacy of the proposed
algorithm will be evaluated using the performance measures listed below. The accuracy,
sensitivity, precision, and F1-score of the confusion matrix are used to evaluate a
technique's efficacy.</p>
        <p>Accuracy =</p>
        <p>TP + TN
TP + TN + FP + FN</p>
        <p>(1)
Sensitivity = TP</p>
        <p>TP + FN
Pr ecision = TP</p>
        <p>TP + FP
F1− score = 2* Pr ecisison * Re call</p>
        <p>Pr ecision + Re call
(2)
(3)
(4)</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>4. RESULTS</title>
      <p>The confusion matrix displays how well the suggested VGG-UNet classifier in contrast to
established models like Res-UNet and Faster R-CNN for hyperspectral image classification
in remote sensing. The proposed model demonstrated proficiency in identifying 128
hyperspectral and 112 non-hyperspectral images. However, it faced challenges in
accurately classifying non-hyperspectral images, resulting in 18 false positives and a 53%
recall rate for that class. Conversely, it exhibited an 85% recall for true hyperspectral images
but misclassified 22 as non-hyperspectral. Notably, the focus should be on refining the
model to minimize misclassifications of non-hyperspectral images. Despite these
challenges, the proposed architecture outperformed existing models by achieving higher
correct classifications and fewer misclassifications, showcasing its potential for improved
hyperspectral image classification in remote sensing applications.</p>
      <p>Recall</p>
      <p>Precision</p>
      <p>F1-score</p>
      <p>The graphical comparisons highlight the superior performance of the proposed
VGGUNet classifier against established models (Res-UNet and Faster R-CNN) in hyperspectral
image classification for remote sensing. The VGG-UNet consistently outperforms with the
highest accuracy (95.62%), precision (98.14%), recall (93.55%), and F1-score (91.69%).
These findings highlight the VGG-UNet model's remarkable ability to categorizes
hyperspectral pictures reliably, highlighting its usefulness in remote sensing applications.
The thorough assessment based on several metrics confirms that the VGG-UNet is a solid
and trustworthy option for developing hyperspectral image analysis in the remote sensing
domain.</p>
    </sec>
    <sec id="sec-4">
      <title>5. Conclusion.</title>
      <p>Finally, the assessment of the suggested VGG-UNet classifier against established models,
Res-UNet and Faster R-CNN, for hyperspectral image classification in remote sensing has
yielded valuable insights. The study proposes the implementation of a specialized
VGGUNet architecture tailored for the classification of hyperspectral images, specifically
addressing the nuances of remote sensing data. To assess the model's efficacy, a
comparative analysis is conducted, pitting the performance of VGG-UNet against alternative
methods such as Res-UNet and Faster R-CNN in the context of remote sensing image
classification. Complementing these findings, the F1-score, recall, accuracy, and precision
comparisons depicted in the graphs consistently showcased the superior performance of
the VGG-UNet model. With a notably high accuracy of 95.62%, precision at 98.14%, recall at
93.55%, with a remarkable 91.69% F1-score, the VGG-UNet outperformed Res-UNet and
Faster R-CNN. These metrics collectively underscore the VGG-UNet's efficacy in
hyperspectral image classification, reinforcing its potential for enhanced remote sensing
applications. Despite facing challenges, the proposed architecture demonstrated its
superiority through higher correct classifications and fewer misclassifications, emphasizing
its promise for improved hyperspectral image analysis. With a foundation for model
refinement and breakthroughs in accurate hyperspectral image classification for
environmental monitoring and decision-making processes. the studies thorough evaluation
and comparative analyses advance deep learning techniques in remote sensing.
doi: 10.1109/TPAMI.2020.2982166.
[8] C. Shorten and T. M. Khoshgoftaar, “A survey on Image Data Augmentation for Deep</p>
      <p>Learning,” J. Big Data, vol. 6, no. 1, 2019, doi: 10.1186/s40537-019-0197-0.
[9] Z. Pan, J. Xu, Y. Guo, Y. Hu, and G. Wang, “Deep learning segmentation and
classification for urban village using a worldview satellite image based on U-net,”
Remote Sens., vol. 12, no. 10, pp. 1–17, 2020, doi: 10.3390/rs12101574.
[10] P. Jiang, Y. Chen, B. Liu, D. He, and C. Liang, “Real-Time Detection of Apple Leaf
Diseases Using Deep Learning Approach Based on Improved Convolutional Neural
Networks,” IEEE Access, vol. 7, pp. 59069–59080, 2019, doi:
10.1109/ACCESS.2019.2914929.
[11] S. Nie, L. Gu, Y. Zheng, A. Lam, N. Ono, and I. Sato, “Deeply Learned Filter Response
Functions for Hyperspectral Reconstruction,” Proc. IEEE Comput. Soc. Conf. Comput.</p>
      <p>Vis. Pattern Recognit., pp. 4767–4776, 2018, doi: 10.1109/CVPR.2018.00501.
[12] M. Q. Huang, J. Ninić, and Q. B. Zhang, “BIM, machine learning and computer vision
techniques in underground construction: Current status and future perspectives,”
Tunn. Undergr. Sp. Technol., vol. 108, pp. 1–78, 2021, doi:
10.1016/j.tust.2020.103677.
[13] J. Zhao et al., “Deep learning in hyperspectral image reconstruction from single rgb
images—a case study on tomato quality parameters,” Remote Sens., vol. 12, no. 19,
pp. 1–14, 2020, doi: 10.3390/rs12193258.
[14] C. Engineering and S. Attila, “hyperspectral image classification using deep learning,”
2021.
[15] A. J. X. Guo and F. Zhu, “Improving deep hyperspectral image classification
performance with spectral unmixing,” Signal Processing, vol. 183, pp. 1–24, 2021, doi:
10.1016/j.sigpro.2020.107949.
[16] S. Krishnamoorthy, Y. Zhang, S. Kadry, and W. Yu, “Framework to Segment and
Evaluate Multiple Sclerosis Lesion in MRI Slices Using VGG-UNet,” Comput. Intell.
Neurosci., vol. 2022, 2022, doi: 10.1155/2022/4928096.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <given-names>Z.</given-names>
            <surname>Guan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Miao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Mu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Ye</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D.</given-names>
            <surname>Gao</surname>
          </string-name>
          , “
          <article-title>Forest Fire Segmentation from Aerial Imagery Data Using an Improved Instance Segmentation Model,” Remote Sens</article-title>
          ., vol.
          <volume>14</volume>
          , no.
          <issue>13</issue>
          ,
          <year>2022</year>
          , doi: 10.3390/rs14133159.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <given-names>X.</given-names>
            <surname>Han</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L</given-names>
            .
            <surname>Xue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Shao</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Xu</surname>
          </string-name>
          , “
          <article-title>A power spectrum maps estimation algorithm based on generative adversarial networks for underlay cognitive radio networks,” Sensors (Switzerland)</article-title>
          , vol.
          <volume>20</volume>
          , no.
          <issue>1</issue>
          ,
          <year>2020</year>
          , doi: 10.3390/s20010311.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <given-names>N.</given-names>
            <surname>Sharma</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Hefeeda</surname>
          </string-name>
          , “
          <article-title>Hyperspectral reconstruction from RGB images for vein visualization</article-title>
          ,”
          <source>MMSys 2020 - Proc. 2020 Multimed. Syst. Conf.</source>
          , pp.
          <fpage>77</fpage>
          -
          <lpage>87</lpage>
          ,
          <year>2020</year>
          , doi: 10.1145/3339825.3391861.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <surname>K. De Haan</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Rivenson</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Wu</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Ozcan</surname>
          </string-name>
          , “
          <article-title>Deep-Learning-Based Image Reconstruction and</article-title>
          Enhancement in Optical Microscopy,
          <source>” Proc. IEEE</source>
          , vol.
          <volume>108</volume>
          , no.
          <issue>1</issue>
          , pp.
          <fpage>30</fpage>
          -
          <lpage>50</lpage>
          ,
          <year>2020</year>
          , doi: 10.1109/JPROC.
          <year>2019</year>
          .
          <volume>2949575</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Imaging</surname>
          </string-name>
          , vol.
          <volume>39</volume>
          , no.
          <issue>12</issue>
          , pp.
          <fpage>4001</fpage>
          -
          <lpage>4010</lpage>
          ,
          <year>2020</year>
          , doi: 10.1109/TMI.
          <year>2020</year>
          .
          <volume>3008930</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Tian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Jiang</surname>
          </string-name>
          , and J. Ma, “
          <article-title>Image fusion meets deep learning: A survey and perspective</article-title>
          ,” Inf. Fusion, vol.
          <volume>76</volume>
          , no.
          <source>August</source>
          , pp.
          <fpage>323</fpage>
          -
          <lpage>336</lpage>
          ,
          <year>2021</year>
          , doi: 10.1016/j.inffus.
          <year>2021</year>
          .
          <volume>06</volume>
          .008.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chen</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S. C. H.</given-names>
            <surname>Hoi</surname>
          </string-name>
          , “
          <article-title>Deep Learning for Image Super-Resolution: A Survey,”</article-title>
          <source>IEEE Trans. Pattern Anal. Mach</source>
          . Intell., vol.
          <volume>43</volume>
          , no.
          <issue>10</issue>
          , pp.
          <fpage>3365</fpage>
          -
          <lpage>3387</lpage>
          ,
          <year>2021</year>
          ,
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>