<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>DETECTION OF AIRCRAFTS USING RCNN IN REMOTE SENSING IMAGES</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Bhavani Sankar Panda</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kakita Murali Gopal</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Rabinarayan Satpathy</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Giet University</institution>
          ,
          <addr-line>Gunupur ,Odisha</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Sri Sri University</institution>
          ,
          <addr-line>Cuttack, Odisha</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
      </contrib-group>
      <fpage>248</fpage>
      <lpage>257</lpage>
      <abstract>
        <p>Modern remote sensing images have developed, and the resultant of the sophisticated technology makes the detection of objects a tedious task. Detection of aircraft from remote sensing imagery has some challenges that are often not addressed. The image captured by remote sensing technology would have more information about the objects in the scene. Smaller objects around the aircraft and similar unknown objects could easily affect the detection model. And there will be numerous objects in the wide-open space of the airfield and airports. Although several research studies showed aircraft detection from remote sensing images, the models didn't solve the overlap problem. And therefore, the present study proposed a method to resolve the overlap problem using the Region-based convolutional neural network architecture from the YOLOv3 model to detect the aircraft from the remote sensing images. The model would use the Interaction of unit and the ground truth data from the images to select the appropriate bounding boxes for the aircraft. The model was trained, and the results produced the model could perform at an accuracy of 98.5 %. The RCNN model could detect the aircraft appropriately, and the metrics analysis shows the model had less error while detecting the aircraft. The performance comparison with other aircraft detection models showed that the proposed model performed better.</p>
      </abstract>
      <kwd-group>
        <kwd>Aircraft detection</kwd>
        <kwd>Region of interest</kwd>
        <kwd>remote sensing images</kwd>
        <kwd>RCNN 1</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Recently, object detection based on remote sensing images has been developed, and the
necessity of collecting information from remote sensing images has also grown. Especially for the
military and defence applications, the detection of objects in the remote sensing images is important,
and aircraft detection from the remote sensing images is among them. Detecting aircraft from the
airports and airspace could provide crucial information and is often considered an important aspect of
military Intel. Aircraft detection has important aspects, and several research studies have tried to
implement deep learning models from remote sensing images. The remote sensing images with
sophisticated technology can now produce a higher resolution of images. These high-resolution
images have more information about the object on the surface, and the deep learning models couldn't
differentiate the objects. As often, the objects become harder to distinguish. The images could provide
information about the people in the field, so detecting the only aircraft from the airspace and airfield
has become a complex task. To discuss further on the detection of objects from the remote sensing
images, the features are harder to be extracted. The selection of features becomes difficult during the
bounding box creations. Although the studies have discussed this, an appropriate solution was not
found yet. Extracting information from the remote sensing images need to be improved, especially for
the detection of a single object.</p>
      <p>Some earlier studies have shown that the RCNN model could detect aircraft. The Region-based
convolutional neural networks were known for the quick and appropriate selection of features from
the images. Many image processing applications and detection models use the RCNN model to
perform feature extraction and detection. To employ the RCNN models, the dataset has to be properly
annotated.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Literature Review</title>
      <p>The research study on detecting aircraft through remote sensing [1] aimed to resolve the aeroplane
detection, including inappropriate detection of small objects while developing the ROI and proposed
an algorithm to improve aeroplane detection using remote sensing images. The author used multilayer
features and a new, improved non-maximal suppression to detect the aeroplane from the remote
sensing images. The multilayer feature fusion method made the model include features from the
different layers and fuse them to select the appropriate features from the images. The non-maximal
suppression improved the detection box overlap problem and appropriately selected the specified
feature. The proposed method improved the selection and detection of aeroplanes from remote
sensing images. Another study on detecting aeroplanes, [2] using remote sensing images made a deep
learning-based. The region-based convolutional neural network is used for the detection, built from
scratch. The RCNN performs the detection and identification of the aircraft. A huge dataset
comprising remote sensing images of airports and aircraft are used for training, and the validation of
the model was based on the images collected from the Turkey airport. The proposed model displayed
accuracy of 98.34 %. And the model was able to detect the aircraft using bounding boxes, and the
prediction results demonstrated aircraft detection.</p>
      <p>Here the study demonstrated [3] a tomato detection using YOLO-tomato to resolve the overlap
challenges in the detection of tomato. This improved YOLOv3 model with an altered bounding box in
a circular shape, which improved the calculation of non-maximum suppression. And the model
showed effective results and performance compared to other studies. Also, the study adopted [4] a
new method to create bounding boxes based on the height and width of the cluster centres. The
bounding boxes are created based on the Markov chains values, and Intersection over union method is
followed to measure the distance between clusters and individual points. The results showed the
method could converge faster and the performance metrics was better compared to the conventional
YOLOv3 model.</p>
      <p>The study on Platform Screen Doors [5] used a deep learning technique to identify the object and
classify it. And as stated by the Authors it was the first attempt to detect objects on platform screen
doors in metros. The study used 984 images at 600 × 480 pixels and manually labelled for six
different objects. The object detection performance was evaluated by using three famous detection
models, YOLOv3, Single-Shot MultiBox-Detector (SSD), and Center-Net through transfer learning.
The YOLOv3 model was able to detect foreign objects at an exceptional speed which is about 200
FPS (frames per second), and CentreNet produced accurate results consistently. The study on
detecting aircraft [6] used reinforcement learning and a CNN model. The reinforcement learning
adopted here to resolve the challenge in detecting the number of objects from the remote sensing
images. The detection agent by the reinforcement learning accurately detects the aircraft using a
bounding box and performs better. The CNN model is used to detect the features and analyze the
probability of finding the aircraft in the image.</p>
      <p>Moreover, another study on Aircraft detection [7] using a novel method from remote sensing
images depends on the deep-residual network (ResNet) and Super Vector (SV). The ResNet is used to
merge the convolutional features and improves the resolution of features which in turn produces a
better proposal for detection. Meanwhile, SV coding is used to extract the histogram of oriented
gradient (HOG) from the ROI to complete the detection and classification process. The model could
produce better results even in the presence of irregular backgrounds.</p>
      <sec id="sec-2-1">
        <title>2.1 Research Gap</title>
        <p>Although there were enough studies to detect aircraft from the remote sensing images, aircraft
detection is still complex, and the overlapping of other objects from the images is not yet solved
completely. The remote sensing images also has objects that might cause improper detection of
aircraft [2]. The deep learning models have overlap issues and smaller objects around the aircraft.</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2 The objective of the study</title>
        <p>Based on the literature review and the research gap, the following objectives are followed for the
proposed study,
1. To detect aircraft from the remote sensing images using the Region-based Convolutional Neural</p>
        <p>Network model.
2. To use Intersection over a unit and ground truth data parameters to select the appropriate features to
overcome the irrelevant smaller object detection.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Proposed Methodology</title>
      <p>The study proposed an aircraft detection model based on the remote sensing images, using the
YOLOv3 model. This section discusses the overall proposed methodology for the study. First, the
dataset details were discussed compared to earlier studies on detecting aircraft with the images and
then followed by the neural network model adopted for the study, which is YOLOv3 and transfer
learning was adopted here for customized objects.
3.1.</p>
      <sec id="sec-3-1">
        <title>Dataset description</title>
        <p>From the flow diagram, it can be observed that the process starts with a dataset, which was collected
from the kaggle open-source[8]—the dataset comprised of images with and without aircraft. The
images were first initially pre-processed and annotated for the presence of aircraft. The dataset is then
split into two sets with aircraft and without crafts. While splitting the images, it is observed that 8000
images have aircraft, and 24,000 images don't have aircraft. The images were colour (RGB) based,
and the details are listed in table 1,
The selected images for the training were 8000 images with a resolution of 20 x 20 pixels. The
annotations were predefined with the dataset yet verified once manually. Then the images were scaled
to 416 x416 to fit the YOLOv3 architecture.
3.2.</p>
      </sec>
      <sec id="sec-3-2">
        <title>Neural network Architecture</title>
        <p>The Neural network architecture adopted here is the YOLOv3 model, one of the famous object
detection tools among the deep learning model. The reason to choose the YOLOv3 model is that the
earlier studies showed the complexity in detecting the bounding boxes for the aircraft and the
inconsistent accurate results showed that from the remote sensing images, detection of aircraft could
be a challenge without overlapping. And as for the YOLOv3 architecture, the resultant model could
resolve the overlapping issue as the Region based convolution selection were followed. The transfer
learning method is adopted here, and for the final object detection, the final layer functions are based
on detecting only aircraft from the input images. Every other object detection from the conventional
YOLOv3 model is removed.
As for the layers, The RCNN model uses 53 convolutional layers, and there will be no pooling layers.
The model is a complete, fully convolution network model as only convolutional layers are used.
Initially, images were loaded into the convolutional layer with 416 x 416, filtered and got a reduced
to search for the features that are to be detected from the images. Again another res-block was used
two times and then upsampled, producing downscaled images of 52 x 52. Following the up sampling,
the images are further passed onto res-block eight times which in turn produces a reduced image of
about 26 x26. And the repeated convolutions by the res-block number of times and the up sampling
finally detect the set of the regions for the input image.
3.3.</p>
      </sec>
      <sec id="sec-3-3">
        <title>ROI based feature selection</title>
        <p>
          The YOLOv3 model adopts a single-stage object detector for the detection problem. The image is
split into grids equally, as shown in the image. Then each cell would be analyzed to create the
bounding box. The boxes were created based on the confidence parameter, which utilizes the ground
truth and the Interaction of the units from the image and the equation (
          <xref ref-type="bibr" rid="ref1">1</xref>
          ).
(
          <xref ref-type="bibr" rid="ref1">1</xref>
          )
The binary cross-entropy for predicting the object score and truth object score is used as a loss
function as shown in the equation (
          <xref ref-type="bibr" rid="ref2">2</xref>
          ).
 1 = ∑ =0 ∑
 2
        </p>
        <p>
          [ ̂ 
(   ) − (1 −  
 ) 
(1 −    )]
(
          <xref ref-type="bibr" rid="ref2">2</xref>
          )
The coordinates of the bounding box is given by the parameter bx and by, as by the equation (
          <xref ref-type="bibr" rid="ref3">3</xref>
          ) &amp; (
          <xref ref-type="bibr" rid="ref4">4</xref>
          )
where, the Cx and Cy are the offset of the grid cell. The position of the bounding box is given by the tx
and ty.
  = 
  = 
(  ) +  
(  ) +  
(
          <xref ref-type="bibr" rid="ref3">3</xref>
          )
(
          <xref ref-type="bibr" rid="ref4">4</xref>
          )
Where, the width and height of the bounding box is given by the parameter bw and bh, as by the
equation (
          <xref ref-type="bibr" rid="ref5">5</xref>
          ) &amp; (
          <xref ref-type="bibr" rid="ref6">6</xref>
          ) where, the pw and ph are the offset of the grid cell. The position of the bounding
box is given by the tw and th.
  =     
 ℎ =  ℎ  ℎ
(tx^ , ty^ , tw^ , th ^ ) are the truth values and (gx, gy, gw, gh) are the ground truth box parameters in the
equation (
          <xref ref-type="bibr" rid="ref7">7</xref>
          ), (
          <xref ref-type="bibr" rid="ref8">8</xref>
          ), (
          <xref ref-type="bibr" rid="ref9">9</xref>
          ) &amp; (
          <xref ref-type="bibr" rid="ref10">10</xref>
          ).
 2 = ∑ =0 ∑
 2
        </p>
        <p>2
∑ =0 ∑

[(
[(
(  )
 − 
(  )
 − 
(  )
 2
 ) + (
 2
(  ) ) + (
( ℎ)</p>
        <p>(  )
 − 
−</p>
        <p>
          (  ) ))] +
( ℎ) ))] (
          <xref ref-type="bibr" rid="ref11">11</xref>
          )
The model uses two equations for the coordinate prediction to generate the bounding box, equation
(
          <xref ref-type="bibr" rid="ref1">1</xref>
          ) to generate the probability in terms of confidence and equation (
          <xref ref-type="bibr" rid="ref11">11</xref>
          ) uses the square error for the
prediction of coordinates. The above mentioned equations are the representation of the calculations to
create the appropriate bounding box in the given image.
        </p>
        <p>The object detection in the YOLOv3 model was based on three different scales. The three different
scale detection could reduce the detection of smaller objects problem. The 8000 images are used for
training the model, and for the model training 120 batch size and 300 epochs. The model was trained
by using CPU - Intel Core i7 – 7700 @ 2.80 GHz, GTX 1050, 16GB RAM and in Windows 8/Python
3.7 environment.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results and Discussion</title>
      <p>This section discusses the results and the performance analysis for the proposed model to detect the
aircraft. As discussed earlier, the study has used the YOLOv3 model to detect aircraft from remote
sensing images. Initially, the dataset contained both aircraft and no aircraft set of images. The images
with aircraft are filtered as presented in fig. 4 &amp; 5.</p>
      <p>The model was trained with 8000 images using the YOLOv3 model, and the proposed method was
compared with other studies to evaluate the model and its detection ability.</p>
      <p>Above table 2 shows the performance comparison for detecting aircraft from remote sensing
images. From the summarized data, it could be observed that the proposed method has produced 98.5
% accuracy. But compared to the YOLOv2 based model and the study [1], the time taken to detect the
aircraft is less for the proposed model with minor differences. However, the accuracy of detection is
better in the proposed method.
The comparison of metrics FPR, MR and ER are shown in fig 6, and the graphical representation of
accuracy comparison for the proposed model is shown in fig.7.
Proposed 98.5 97.72 97.45 0.0168 0.0129</p>
      <p>The comparison of the conventional CNN model for the proposed RCNN model shown in Table 3
compares confusion matrix metrics. Although the CNN model showed better accuracy, specificity,
and sensitivity, the FP rate and FN rate was better in the proposed RCNN model, indicating the model
has less error in detection. And the graphical representation of the comparison of the false positive
and false negative rates are shown in fig.8</p>
      <p>Precision and the Recall metrics show the model's ability to reproduce the results with consistent
accuracy. Table 4, for the comparison of metrics with previous studies, shows how the model's
performance increased over the years. The proposed model showed a 97.65 % precision and 97.13 %
recall which is good enough for the proposed model. The comparison of the proposed model with the
earlier studies for the different metrics showed us the proposed model for detecting the aircraft from
the remote sensing images.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>The detection of aircraft from the remote sensing images had complex challenges such as detection
of smaller objects and overlapping of objects. Detecting objects from the remote sensing images has
several limitations, which result in the improper detection of the aircraft. To overcome the underlying
issues in aircraft detection, the proposed method adopts the YOLOv3 model to identify the aircraft
from the remote sensing images. Based on the RCNN model and confidence based selection of ROI
they made, the model to choose the appropriate features for the detection of aircraft. This made the
model identify the aircraft effectively from the remote sensing images. The performance comparison
with earlier studies for the proposed model showed that the model performed better in detecting the
aircraft.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <given-names>M.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Ma</surname>
          </string-name>
          , and Y. Han,
          <article-title>"Effective airplane detection in remote sensing images based on multilayer feature fusion and improved nonmaximal suppression algorithm,"</article-title>
          <source>Remote Sensing</source>
          , vol.
          <volume>11</volume>
          , p.
          <fpage>1062</fpage>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>F.</given-names>
            <surname>Ucar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Dandil</surname>
          </string-name>
          , and
          <string-name>
            <given-names>F.</given-names>
            <surname>Ata</surname>
          </string-name>
          ,
          <article-title>"Aircraft detection system based on regions with convolutional neural networks,"</article-title>
          <source>International Journal of Intelligent Systems and Applications in Engineering</source>
          , vol.
          <volume>8</volume>
          , pp.
          <fpage>147</fpage>
          -
          <lpage>153</lpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3. G. Liu,
          <string-name>
            <given-names>J. C.</given-names>
            <surname>Nouaze</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. L. Touko</given-names>
            <surname>Mbouembe</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J. H.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <article-title>"YOLO-tomato: A robust algorithm for tomato detection based on YOLOv3,"</article-title>
          <source>Sensors</source>
          , vol.
          <volume>20</volume>
          , p.
          <fpage>2145</fpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <given-names>L.</given-names>
            <surname>Zhao</surname>
          </string-name>
          and
          <string-name>
            <given-names>S.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>"Object detection algorithm based on improved YOLOv3,"</article-title>
          <source>Electronics</source>
          , vol.
          <volume>9</volume>
          , p.
          <fpage>537</fpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>Y.</given-names>
            <surname>Dai</surname>
          </string-name>
          , W. Liu,
          <string-name>
            <given-names>H.</given-names>
            <surname>Li</surname>
          </string-name>
          , and
          <string-name>
            <given-names>L.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <article-title>"Efficient foreign object detection between PSDs and metro doors via deep neural networks,"</article-title>
          <source>IEEE Access</source>
          , vol.
          <volume>8</volume>
          , pp.
          <fpage>46723</fpage>
          -
          <lpage>46734</lpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Sun</surname>
          </string-name>
          , and
          <string-name>
            <given-names>X.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <article-title>"An aircraft detection framework based on reinforcement learning and convolutional neural networks in remote sensing images,"</article-title>
          <source>Remote Sensing</source>
          , vol.
          <volume>10</volume>
          , p.
          <fpage>243</fpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <given-names>J.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Xiao</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <article-title>"Aircraft detection in remote sensing images based on a deep residual network and super-vector coding,"</article-title>
          <source>Remote Sensing Letters</source>
          , vol.
          <volume>9</volume>
          , pp.
          <fpage>228</fpage>
          -
          <lpage>236</lpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>K.</surname>
          </string-name>
          <article-title>a</article-title>
          . dataset. https://www.kaggle.com/rhammell/planesnet.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <given-names>H.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , J. Zhang, and
          <string-name>
            <given-names>F.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <article-title>"Fast aircraft detection in satellite images based on convolutional neural networks,"</article-title>
          <source>in 2015 IEEE International Conference on Image Processing (ICIP)</source>
          ,
          <year>2015</year>
          , pp.
          <fpage>4210</fpage>
          -
          <lpage>4214</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <given-names>W.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Wang</surname>
          </string-name>
          , and
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>"Object detection in high-resolution remote sensing images using rotation invariant parts based model,"</article-title>
          <source>IEEE Geoscience and Remote Sensing Letters</source>
          , vol.
          <volume>11</volume>
          , pp.
          <fpage>74</fpage>
          -
          <lpage>78</lpage>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <given-names>W.</given-names>
            <surname>Diao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Dou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          , and
          <string-name>
            <given-names>K.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <article-title>"Efficient saliency-based object detection in remote sensing images using deep belief networks,"</article-title>
          <source>IEEE Geoscience and Remote Sensing Letters</source>
          , vol.
          <volume>13</volume>
          , pp.
          <fpage>137</fpage>
          -
          <lpage>141</lpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <given-names>S.</given-names>
            <surname>Ren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Girshick</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <article-title>"Faster r-cnn: Towards real-time object detection with region proposal networks,"</article-title>
          <source>Advances in neural information processing systems</source>
          , vol.
          <volume>28</volume>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <given-names>K.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , S. Ren, and
          <string-name>
            <given-names>J.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <article-title>"Deep residual learning for image recognition,"</article-title>
          <source>in Proceedings of the IEEE conference on computer vision and pattern recognition</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>770</fpage>
          -
          <lpage>778</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>