=Paper=
{{Paper
|id=Vol-2819/session2paper4
|storemode=property
|title=Convolutional Neural Networks for Feature Extraction and Automated Target Recognition in Synthetic Aperture Radar Images
|pdfUrl=https://ceur-ws.org/Vol-2819/session2paper4.pdf
|volume=Vol-2819
|authors=John Geldmacher,Christopher Yerkes,Ying Zhao
}}
==Convolutional Neural Networks for Feature Extraction and Automated Target Recognition in Synthetic Aperture Radar Images==
Convolutional Neural Networks for Feature
Extraction and Automated Target Recognition in
Synthetic Aperture Radar Images
John Geldmacher, Christopher Yerkes, and Ying Zhao, Memeber, IEEE
Abstract—Advances in the development of deep neural networks a synthetic aperture radar (SAR) sensor. Both IR and SAR
and other machine learning algorithms combined with ever more images require a trained imagery analyst to reliably identify
powerful hardware and the huge amount of data available on the targets. A repetitive and time consuming task that currently
internet has led to a revolution in ML research and applications.
These advances present massive potential and opportunity for military requires human expertise, importantly and creatively, is an
applications such as the analysis of Synthetic Aperture Radar (SAR) ideal problem for deep learning. Automated target recognition
imagery. SAR imagery is a useful tool capable of capturing high (ATR) seeks to reduce the total work load of analysts so that
resolution images regardless of cloud coverage and at night. However, their effort can be spent on the more human-centric tasks like
there is a limited amount of publicly available SAR data to train presenting and explaining intelligence to a decision maker.
a machine learning model. This paper shows how to successfully
dissect, modify, and re-architect cross-domain object recognition ATR is also intended to reduce the time from collection to
models such as the VGG-16 model, transfer learning models from exploitation by screening images at machine speeds rather than
the ImageNet, and the k-nearest neighbor (kNN) classifier. The paper manually. SAR ATR is complicated by the available data to
demonstrates that the combinations of these factors can significantly train and assess machine learning models. Unlike other image
and effectively improve the automated object recognition (ATR) classification tasks, there is not a large and freely available
of SAR clean and noisy images. The paper shows a potentially
inexpensive, accurate, transfer and unsurpervised learning SAR ATR amount of training data for researchers. Further, the data that
system when data labels are scarce and data are noisy, simplifying is publicly available only covers a small fraction of the types
the whole recognition for the tactical operation requirements in the of targets an effective SAR ATR system would be required to
area of SAR ATR. identify.
Keywords—k-nearest neighbor, kNN, deep learning, Synthetic
Aperture Radar images, SAR images, transfer learning, VGG-16
II. A DVANTAGES AND C HALLENGES OF S YNTHETIC
A PERTURE R ADAR (SAR) I MAGES , DATA D ESCRIPTION ,
I. I NTRODUCTION AND R ELATED W ORK
The analysis and classification of targets within imagery Synthetic Aperture Radar (SAR) is a radar mounted to a
captured by aerial and space-based systems provides the US moving platform that uses the platform’s motion to approx-
intelligence community and military geospatial intelligence imate the effect of a large antenna. The high resolution that
(GEOINT) personnel with important insights into adversary can be achieved by creating a radar with an effective aperture
force dispositions and intentions. It has also entered the much greater in size than is physically possible allows for
mainstream thanks to openly available tools like Google Earth. radar returns to be processed into images similar to what can
The high resolution of space-based sensors and common be achieved with a camera [19]. SAR imagery provides an im-
use of overhead imagery in everyday life means with the portant tool for the United States Intelligence Community and
exception of decoys and camouflage, an average person is military geospatial intelligence (GEOINT) personnel because
now reasonably capable of identifying objects in electro- of its all-weather, day/night collection capability. Additionally,
optical (EO) imagery. EO images are, however, limited by some wavelengths that SAR imaging systems operate in have
cloud coverage and daylight. About half of the time when a degree of foliage and ground penetrating capability allowing
a satellite in low earth orbit could image a target it will for the detection of buried objects or objects under tree cover
be night, necessitating the use of either an infrared (IR) or that would not be observable by other sensors such as EO
sensors.
This will certify that all author(s) of the above article/paper are employees These important advantages of SAR imaging for GEOINT
of the U.S. Government and performed this work as part of their employment,
and that the article/paper is therefore not subject to U.S. copyright protection. analysts do come with some significant drawbacks inherent to
No copyright. Use permitted under Creative Commons License Attribution SAR images. Because SAR images are not true optical images,
4.0 International (CC BY 4.0). In: Proceedings of AAAI Symposium on they are susceptible to noise generated by constructive and
the 2nd Workshop on Deep Models and Artificial Intelligence for Defense
Applications: Potentials, Theories, Practices, Tools, and Risks, November 11- destructive interference between radar reflections that appear
12, 2020, Virtual, published at http://ceur-ws.org/ as bright or dark spots called “speckle” in the image [19]. Also
J. Geldmacher, C. Yerkes, and Y. Zhao are with the Department of various materials and geometries will reflect the radar pulses
Information Sciences, Naval Postgraduate School, Monterey, CA, USA
C. Yerkes is also with Oettinger School of Science and Technology differently, creating blobs or blurs that can obscure the objects
Intelligence, National Intelligence University, Bethesda, MD, USA physical dimensions. These issues, as well as problems caused
images with the breakthrough feature extraction layers as
demonstrated in convolutional neural networks, produced
generally good results. An SVM method proposed by [27]
achieved 91% accuracy in a five-class test [27], while a
Bayesian classifier reported a 95.05% accuracy in a 10-classes
test [13].
In recent years, the work on classification of SAR imagery
has focused on the use of CNNs. In 2015, Morgan showed
that a relatively small CNN could achieve 92.1% accuracy
across the 10-class of the MSTAR dataset, roughly in line with
the shallow methods previously explored. Morgan’s method
also showed that a network trained on nine of the MSTAR
target classes could be retrained to include a tenth class 10-20
times faster than training a 10-class classifier from scratch.
The ability to more easily adapt the model to changes in
target sets represents an advantage over shallow classification
Fig. 1. Example Photographs and MSTAR Im- techniques [11]. This is especially valuable in a military
ages by Class. Photograph of BMP-2 from https : ATR context given the fluid nature of military operations,
//www.militaryf actory.com/armor/detail.asp?armor id = 50, where changes to the order of battle may necessitate updating
photograph of BTR-70 from https : //military.wikia.org/wiki/BT R −
70. All other photographs and SAR images adapted from the MSTAR a deployed ATR system. Malmgren-Hansen et al., explored
dataset. transfer learning from a CNN pre-trained on simulated SAR
images generated by using ray tracing software and detailed
computer aided design models of target systems. They showed
by Doppler shift in moving objects and radar shadows, make that model performance was improved, especially in cases
the identification and classification of objects in SAR images where the amount of training data was reduced [10]. The
a difficult and tedious task that also requires a well-trained technique of generating simulated SAR images for training
and experienced analyst. could also be valuable in a military SAR ATR context where
The the Moving and Stationary Target Acquisition and an insufficient amount of training data for some systems may
Recognition (MSTAR) data set is a publicly available data set exist.
consisting of synthetic aperture radar images of the following The use of a linear SVM as a replacement for the softmax
10 classes of military vehicles: activation that is typically used for multiclass classifiers in
1) 2S1: former Soviet Union (FSU) self-propelled artillery neural networks has been shown to be potentially more effec-
2) BMP-2: FSU infantry fighting vehicle tive for some classification tasks [22]. Transfer learning from
3) BRDM-2: FSU armored reconnaissance vehicle ImageNet to MSTAR with an SVM classifier was explored by
4) BTR-60: FSU armored personnel carrier [1] in 2018. Their methodology compared the performance of
5) BTR-70: FSU armored personnel carrier an SVM classifier trained on mid-level feature data extracted
6) D7: Caterpillar tractor frequently used in combat engi- from multiple layers from AlexNet, GoogLeNet, and VGG-
neering roles 16 neural networks without retraining the feature extracting
7) T-62: FSU main battle tank network. Although they reported over 99% accuracy when
8) T-72: FSU main battle tank classifying features extracted from mid-level convolutional
9) ZIL-131: FSU general purpose 6x6 truck layers from AlexNet, performance of the SVM on features
10) ZSU-23-4: FSU self-propelled anti-aircraft gun from fully-connected layers did not achieve 80% accuracy. The
11) SLICY: the Sandia Laboratories implementation of best performance reported on from the VGG-16 architecture
cylinders (SLICY). The SLICY consisting of simple was 92.3% from a mid-level convolutional layer, but only
geometric shapes such as cylinders, edge reflectors, and 49.2% and 46.3% from features extracted in the last two fully-
corner reflectors which could be used for calibration connected layers [1].
of sensors or for modeling the propagation of radar
reflections.
III. T RANSFER L EARNING AND F EATURE E XTRACTION
Fig. 1 shows the example photographs and MSTAR images
by class. It demonstrates the difficulties an imagery analyst CNNs require a very large amount of data to train an
would face when identifying targets in SAR imagery. The accurate model and it is not uncommon for data sets with
vehicles that are easily recognizable in photos become blurs in tens or even hundreds of thousands of images to be used
SAR images. Due to its public availability and ease of access when training a model. Transfer learning presents one possible
for researchers, the data set has become the standard for SAR solution when training a CNN on a limited data set by
image Automated Target Recognition (ATR) classification leveraging knowledge from a previously learned source task to
research. aid in learning a new target task [14]. In an image classification
ATR in SAR imagery using “shallow” classification meth- problem, transfer learning works by training a CNN on a data
ods, which are traditional classifiers applied directly to SAR set that has a very large number of images and freezing the
parameters for a certain number of layers and extracting mid-
level feature representations before training further layers and
the final classification layer [7].
ImageNet is an open source labeled image database orga-
nized in a branching hierarchical method of “synonym sets”
or “synsets”. For example, the “tank” synset is found in a
tree going from vehicle to wheeled vehicle to self-propelled
vehicle to armored vehicle to tank. The ImageNet database Fig. 2. The Original VGG-16 Architecture
consists of over 14 million labeled images organized into
over 21,000 synsets. Pre-trained ImageNets are often used in
transfer learning.
Transfer learning is typically used when source and target
tasks are not too dissimilar in order to avoid negative transfer.
Negative transfer occurs when the features learned in the
transfer learning method actually handicap the model perfor-
mance [14]. However, transfer learning becomes more useful
when a curious phenomenon that many deep neural networks
trained on natural images learn similar features across images
from different domains. Fig. 3. VGG-16 with Transfer learning by freezing the first six layers where
the weights taken directly from the ImageNet
Evidence shows that low and mid-level features could
represent basic ATR features in images such as texture,
corners, edges, and color blobs [9], and the low and mid-level two convolutional/pooling blocks frozen for training to take
neural network feature extraction function resembles the actual advantage of the broad feature detection of the pre-trained
biological and human neurons’ function. Low and mid-level network as shown in Fig. 3. Our method, as in Fig. 4, shows
of features extracted from CNNs are likely common across that the dense layer of 1024 features extracted are saved and
even dissimilar data sets. A transfer learning approach between used as the input to a shallow classifier.
different domains is feasible and ATR tasks are evidently In our experiment, the standard VGG-16 model is imple-
successful in cross-domain applications [2], [8]. For example, mented in the Keras application program interface (API) with
the application of transfer learning to remote sensing target TensorFlow as the backend. The ImageNet weights available
detection and classification was studied [16], which showed in the Keras are ported from the Visual Geometry Group at
that a CNN classifier trained on a photographic data set could Oxford University that developed the VGG architecture for
be retrained to perform remote sensing classification of ships ILSVRC-2014 localization and classification tasks [18]. We
at sea with a good performance. also use, Orange, which is an open source data science and
machine learning toolkit that allows users to easily manipulate
IV. SAR ATR: F EATURE E XTRACTION C OMBINED WITH data through a graphical user interface. Orange has several
S HALLOW C LASSIFIERS built-in machine learning algorithms and simplifies the data
A. Multistep Classifier management and pre-processing requirements to allow users
to experiment with approaches to machine learning and data
In practice and in cross-domain applications, very few
science [3].
people train an entire CNN from scratch because it is relatively
A CNN is trained on the 2200 training images with a 20%
rare to have a data set of sufficient size. For this reason, trans-
validation split. The training and test data were both then run
fer learning with feature extraction combined with shallow
through the retrained neural network. The last fully connected
classifiers are suitable choices for SAR images.
layer before the neural network’s output was saved as a 1024-
The network architecture employed in this paper was a
dimensional vector for each image as shown in Fig. 3.
modified VGG-16 architecture [18]. The original VGG-16 ar-
The extracted features run through the Orange workflow are
chitecture is shown in Fig. 2. It consists of two or three linked
pictured in Fig. 5. The precision and recall are used to compare
convolutional/pooling blocks, three fully-connected layers, and
the base CNN performance, kNN, SVM, and random forest
a softmax activation in the end to determine the class label.
The network employs a 3x3 kernel and a stride of one so
that each pixel is the center of a convolutional step. The
architecture has been modified to freeze model weights for
the first two convolutional/pooling blocks (e.g., the first six
layers). The model top has also been replaced with a fully-
connected layer, a dropout layer to mitigate overfitting, and
two final fully-connected layers with a softmax activation for
classification [16]. This is also referred as a modified VGG-
16 architecture or a VGG-16 architecture in this paper. It Fig. 4. The multistep classifier: Extract features from the VGG-16 and then
was initialized with the ImageNet weights and had the first apply a shallow classifier
Fig. 7. Examples of Noisy SAR images
Fig. 5. The Orange Workflow
Fig. 8. Comparison of VGG-16 with the multistep classifiers for added noise
Fig. 6. Comparison of VGG-16 with the multistep classifiers
classifiers. For the kNN classifier, k was set to 11 and weighted perturbation or adversarials’ deliberated manipulations [6],
Euclidian distance was used to determine which class label to [21]. To study the effect, random Gaussian noise with a noise
assign to test images. A sigmoid kernel was used in the SVM factor of 0.2 was added to the images from the data set.
classifier, and the random forest consisted of 10 decision trees. Fig. 7 shows an example of an original SAR image with one
These settings were unchanged in experiments two and three. added noise. The feature extraction from CNN and follow
on shallow classification process was then repeated without
retraining the base model in order to test the robustness of the
B. Results
model. The baseline model (i.e., modified VGG-16 model) was
The baseline model, which is the modified VGG-16, is then retrained on the noisy images for 30 epochs and accuracy
shown in Fig. 3. The modified VGG-16 without transfer was compared.
learning and trained exclusively on MSTAR, resulted in an Neither the neural network nor any of the multistep classi-
average precision and recall of 0.96. The same modified VGG- fiers proved robust enough to handle the addition of random
16 with full transfer learning of the convolutional layers with noise to the images. However, after retraining the kNN and
weights from ImageNet resulted in an average precision and SVM multistep classifiers perform better than the modified
recall of 0.88. Although, the transfer learning approach has VGG-16 with partial transfer learning.
the advantage of converging much more quickly than the CNN
initialized with random weights, the full transfer learning of
all convolutional weights and only retraining the CNN top V. D ISCUSSION
did not match the performance of the non-transfer learning Performance on the SLICY class is of interest because it
approach, suggesting some negative transfer occurs in the later demonstrates the model’s ability to discriminate an invalid
convolutional layers. As shown in Fig. 6, the best performed target from a valid target. All other classes, with the exception
is the modified VGG-16 with partial transfer learning in Fig. 3 of the D7, are former Soviet Union military equipment. The
and resulted an average precision and recall of 0.98. The D7 is a Caterpillar bulldozer. Up-armored versions of the D7
multistep classifier using a kNN classifier in Fig. 4 was able to and related equipment are often used in combat engineering
match the best baseline performance with an average precision roles. In a military context this means they are likely to be a
and call of 0.98, while the SVM and random forest classifiers valid target. As demonstrated by the high precision and recall
fell short of the baseline model’s performance. in this class across all models, valid targets are very rarely
classified as a SLICY (high precision) and the random objects
C. Adding Noise are not being accepted as valid targets (high recall).
As described before, ATR of SAR images are typically The performance of the kNN classifier is also notable since
sensitive to the noise in the images. A CNN is known to the use of an SVM for classification after feature extraction is
be vulnerable to the noisy models both from environmental previously studied; however, little research has been done on
ple) and no supervised learning or no class labels required
approach for SAR ATR. Recently, various learning-to-hash
algorithms [24] are used to approximate the exact nearest
neighbors, which translates a supervised learning problem and
kNN into an index/search problem [15], and simplifies the
whole recognition for the tactical operation requirements in
the area of SAR ATR. If there are no class labels of SAR
available, our multistep classifiers with transfer learning and
kNN can provide an unsupervised classification with a high
accuracy and confidence to match an object which looks like
another object seen before.
Fig. 10 also shows the comparison of kNN with other
supervised learning methods. The kNN method is the best
among all the methods for the average precision and recall,
where classification of the SLICY has a precision of 0.98 and
a recall of 1. The future work is to test on more and different
data sets (e.g., EO and IR data) to validate if the multistep
methods can apply to cross-domain ATR problems.
Fig. 9. VGG-16 TensorFlow architecture layout
VI. C ONCLUSION
Cross-domain transfer learning from photographs to SAR
imagery is effective for training a neural network both for
feature extraction and classification. A retrained neural net-
work can function as an efficient feature extractor for training
a shallow classifier. kNN and SVM classifiers are potentially
useful replacement for softmax activation in a neural network.
Multistep classification methods using a shallow classifier
trained on features extracted from a neural network, outper-
formed the base neural network when tested on noisy data
and as the amount of training data decreases. This is valuable
to improve CNN in a broader machine vision community by
applying feature extraction followed by shallow classifiers for
clean and noisy images. Transfer learning and kNN multistep
classification methods could be significant in terms of setting
up a robust image indexing system with minimum supervised
training and learning required.
Fig. 10. Multistep classifiers results: VGG-6 transfer learning features +
k-means (k=2048) + kNN & other methods in Orange
Currently, the analysis community does not have an estab-
lished standard for the percent of correctly identified targets
by an imagery analyst. Instead the analysis relies on the
user’s experience and confidence in their own work, providing
the performance of kNN rather than a softmax activation for
responses such as “possible main battle tank” or “likely BMP-
neural network output.
2”, and thus a direct comparison to expert-level performance
To further explore the relations of feature extraction, transfer is difficult to establish. Both the baseline model employing
learning, and kNN, we ran an additional experiment where transfer learning and the shallow classifiers using a neural
we first extracted the transfer weights of the first six layers network as a feature extractor performed with a high degree
of the VGG-16 architecture from ImageNet. Since the flatten of accuracy and would be valuable in an operational context
dimension is 32x32x128=131,072, as shown in Fig. 9, we as an aid to GEOINT analysts.
applied the unsupervised learning k-means algorithm to group
the 131,072 dimension into 2048 clusters. The reasoning
here is that the first six layers probably embed the best
features (texture, corners, edges, and color blobs) that can
be used for classification. Finally, We performed kNN and
other supervised learning methods in Orange based on the
2048 dimensional train and test data. Fig. 10 shows the test
data results from Orange for the VGG6-transfer-kmeans-kNN
method with an average precision and recall of 0.93. The six
layers of transfer learning together with k-means and kNN
provide an inexpensive (without GPU or AWS, for exam-
R EFERENCES [20] Stewart, M. (2019, February 26). Simple Introduction to
Convolutional Neural Networks. Towards Data Science.
https://towardsdatascience.com/simple-introduction-to-convolutional-
[1] Al Mufti, M., Al Hadhrami, E., Taha, B., & Werghi, N. (2018). Au-
neural-networks-cdf8d3077bac
tomatic target recognition in SAR images: Comparison between pre-
[21] Sitawarin,C. & Wagner, D. (2019). On the Robustness of Deep K-
trained CNNs in a tranfer learning based approach. 2018 Interna-
Nearest Neighbors. arXiv:1903.08333v1
tional Conference on Artificial Intelligence and Big Data (ICAIBD).
[22] Tang, Y. (2013). Deep Learning using Linear Support Vec-
https://doi.org/10.1109/ICAIBD.2018.8396186
tor Machines. 2013 ICML Challenges in Representation Learning.
[2] Chen, D., Liu, S., Kingsbury, P., Sohn, S., Storlie, C. B., Habermann, E. https://arxiv.org/abs/1306.0239
B., Naessens,J. M., Larson, D. W. & Liu, H. (2019). Deep learning and [23] Timothy D. Ross, Jeff J. Bradley, Lannie J. Hudson, & Michael P.
alternative learning strategies for retrospective real-world clinical data. O’Connor. (1999). SAR ATR: so what’s the problem? An MSTAR
Digital Medicine (2019)2:43 ; https://doi.org/10.1038/s41746-019-0122- perspective. 3721. https://doi.org/10.1117/12.357681
0. [24] Wang, J., Zhang, T., Song, J., Sebe, N., & Shen, H. T. (2017). A
[3] Demšar, J., Zupan, B., Leban, G., & Curk, T. (2004). Orange: From Survey on Learning to Hash. IEEE TRANSACTIONS ON PATTERN
Experimental Machine Learning to Interactive Data Mining. In Knowl- ANALYSIS AND MACHINE INTELLIGENCE.(2017)13:9.
edge Discovery in Databases: PKDD 2004 (Vol. 3202). Springer. [25] Glorot,X., Bordes, A., & Bengio, Y. (2011). Deep Sparse Recti-
https://doi.org/10.1007/978-3-540-30116-5 58 fier Neural Networks. In Geoffrey Gordon, David Dunson, Miroslav
[4] Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., & Fei-Fei, Dudı́k (Eds.), Proceedings of the Fourteenth International Confer-
L. (2009). ImageNet: A large-scale hierarchical image database. ence on Artificial Intelligence and Statistics (pp. 315–323). PMLR.
2009 IEEE Conference on Computer Vision and Pattern Recogition. http://proceedings.mlr.press/v15/glorot11a.html
https://doi.org/10.1109/CVPR.2009.5206848 [26] Yiu, T. (2019, June 12). Understanding Random Forest: How
[5] Gandhi, R. (2018, June 7). Support Vector Machine—Introduction the Algorithm Works and Why it Is So Effective. Towards Data
to Machine Learning Algorithms. Towards Data Science. Science. https://towardsdatascience.com/understanding-random-forest-
https://towardsdatascience.com/support-vector-machine-introduction- 58381e0602d2f3c8
to-machine-learning-algorithms-934a444fca47 [27] Zhao, Q., & Principe, J. (2001). Support vector machines for SAR auto-
[6] Goodfellow, I., Bengio, Y., & Courville, A. (2016). Deep learning. The matic target recognition. IEEE Transactions on Aerospace and Electronic
MIT Press. ISBN: 0262035618 Systems, 37(2), 643–654.
[7] Kang, C., & He, C. (2016). SAR image classification based on the
multi-layer network and transfer learning of mid-level representations.
2016 IEEE International Geoscience and Remote Sensing Symposium
(IGARSS. https://doi.org/10.1109/IGARSS.2016.7729290
[8] Khosravi, P., Kazemi, E., Zhan, Q., Malmsten, J. E., Toschi, M., Zisi-
mopoulos,P., Sigaras, A., Lavery, S., Cooper, L., Hickman, C., Meseguer,
M., Rosenwaks, Z., Elemento, O., Zaninovic, N. & Hajirasouliha, I.
(2019). Deep learning enables robust assessment and selection of hu-
man blastocysts after in vitro fertilization,Digital Medicine (2019)2:21;
https://doi.org/10.1038/s41746-019-0096-y
[9] Liu,L., Chen,J., Fieguth,P., Zhao, G., Chellappa, R. & Pietikäinen, M.
(2019). From BoW to CNN: Two Decades of Texture Representation
for Texture Classification. International Journal of Computer Vision
(2019)127:74–109
[10] Malmgren-Hansen, D., Kusk, A., Dall, J., Nielsen, A., Enghold, R.,
& Skriver, H. (2017). Improving SAR Automatic Target Recognition
Models With Transfer Learning From Simulated Data. IEEE Geoscience
and Remote Sensing Letters, 14(9), 1484–1488.
[11] Morgan, D. (2015). Deep convolutional neural networks for ATR from
SAR imagery. Algorithms for Synthetic Aperture Radar Imagery XXII,
9475. https://doi.org/10.1117/12.2176558
[12] Nielsen, M. (2015). Neural Networks and Deep Learning. Determination
Press. http://neuralnetworksanddeeplearning.com/
[13] O’Sullivan, J. A., DeVore, M. D., Kedia, V., & Millier, M. I. (2001).
SAR ATR performance using a conditionally Gaussian model. IEEE
Transactions on Aerospace and Electronic Systems, 37(1), 91–108.
https://doi.org/10.1109/7.913670
[14] Pan, S., & Yang, Q. (2010). A Survey on transfer learning.
IEEE Transactions on Knowledge and Data Engineering, 22(10).
https://doi.org/10.1109/TKDE.2009.191
[15] Peng, T., Boxberg, M., Weichert, W., Navab, N. Marr, C.
(2019). Multi-task learning of a deep k-nearest neighbour network
for histopathological image classification and retrieval. bioRxiv doi:
https://doi.org/10.1101/661454
[16] Rice, K. (2018). Convolutional Neural Networks For Detec-
tion And Classification Of Maritime Vessels In Electro-Optical
Satellite Imagery [Master’s thesis, Naval Postgraduate School].
http://hdl.handle.net/10945/61255
[17] Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Ma,
S., Huang, Z., Karpathy, A., Kholsa, A., Bernstein, M., Berg, A.,
& Fei-Fei, L. (2015). ImageNet Large Scale Visual Recognition
Challenge. International Journal of Computer Vision, 115, 211–252.
https://doi.org/10.1007/s11263-015-0816-y
[18] Simonyan, K., & Zisserman, A. (2015). Very Deep Convolutional
Networks For Large-Scale Image Recognition. International Conference
on Learning Representations 2015.
[19] Skolnik, M. (1981). Introduction to Radar Systems (Second). McGraw-
Hill, Inc.