<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>An Intelligent Application Development Platform for Service Robots</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Y. Sugawara</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>T. Morita</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>S. Saito</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>T. Yamaguchi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Y. Sugawara and S. Saito are with the Graduate School of Science and Technology, Keio University</institution>
          ,
          <addr-line>Kanagawa, Japan (phone:</addr-line>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2015</year>
      </pub-date>
      <fpage>16</fpage>
      <lpage>20</lpage>
      <abstract>
        <p>-Robots should adjust their actions to suit the surrounding situation. This requires robots to be able to interchange signals to symbols and symbols to signals. We propose an intelligent application development platform based on the Service Oriented Architecture (SOA) and Semantic Web Service (SWS). Our platform has five layers: Service, Process, Module, Knowledge and Data. The first three layers are implemented by using OWL-based Web Service Ontology (OWL-S). In the Module layer, there are two kinds of module: modules for action and modules for recognition. By combining these modules, this architecture can handle symbols and signals. In this study, we applied our method to RobotCafe, which is a cafe where robots work as waiters, to enable the robots to change their greetings to suit the situation when customers come and to assist the services of the cafe.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>I. INTRODUCTION</p>
      <p>
        Service robots are playing an increasing role in daily life
and various fields. However, due to their restricted functions
and movements, they can only undertake specific services;
one robot cannot easily handle many services. Therefore,
robots should cooperate with each other to provide services.
This cooperation is called Multi-Robot Coordination (MRC),
and we approached this issue by applying SWS for MRC[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
      <p>However, if robots cannot sense the surrounding situation,
they cannot provide good service like a human, even if they
cooperate with other robots. So, to make better service robots,
robots should be able to integrate symbols and signals in order
to adjust their activities dynamically by interchanging
symbols to signals and signals to symbols.</p>
      <p>In this paper, we propose an intelligent application
development platform based on SWS and OWL-S in order to
complete a task given by a user by coordinating with other
robots. By using our platform, users can create robot services
easily by combining multiple software modules. Our platform
has five layers: Service, Process, Module, Knowledge and
Data. The first three layers are implemented by using OWL-S.
Processes of OWL-S have inputs and outputs, so a process can
easily be combined with other processes by binding a process’s
input to another’s output. In the Module layer, there are two
kinds of module: modules for action and modules for
recognition. By combining these modules, this architecture can
handle symbols and signals. This case study of RobotCafe
shows how our proposed platform can be used to easily
construct a service that can be changed dynamically to suit the
situation.</p>
    </sec>
    <sec id="sec-2">
      <title>II. RELATED WORKS</title>
      <sec id="sec-2-1">
        <title>A. The Ubiquitous Network Robot Platform</title>
        <p>
          The ubiquitous network robot (UNR)[
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] is a fundamental
technology for providing services by cooperating with robots,
smartphone applications and environment sensors via a
network. This aims to support human activities in various
situations such as shopping malls, hospitals, homes and so on.
The UNR platform is public open source software which helps
users to make UNR systems. This study focuses on how robots
collaborate on tasks with smartphone applications and
environment sensors.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>B. RoboEarth</title>
        <p>
          RoboEarth[
          <xref ref-type="bibr" rid="ref3">3</xref>
          ] provides an infrastructure for sharing
knowledge between robots. In this context, knowledge
includes software components, environmental maps,
knowledge about tasks and object recognition models. In
RoboEarth, robot hardware, software and performance can be
written in a form that software can read by using the Semantic
Robot Description Language, which is based on the Web
Ontology Language (OWL). RoboEarth focuses on not MRC
but sharing knowledge with robots; just a single robot deals
with tasks in RoboEarth.
        </p>
        <p>C. OWL-S</p>
        <p>
          OWL-S[
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] is an ontology for services which is supported
by W3C. Using OWL-S makes it possible for software agents
to discover, invoke, compose and monitor web services.
OWL-S has three main parts: the service profile, the service
model and the service grounding. The service profile describes
the outline of a service, the preconditions for a service, and the
effects of a service. The service profile is used to advertise and
discover services. The service model gives a detailed
description of a service’s operation. The service grounding
provides details on how to interoperate with a service via
messages, and relates OWL-S of a service to the web service’s
WSDL.
        </p>
        <p>These previous works show how easily robot services can
be made; our research focuses on this issue.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>III. SYSTEM ARCHITECTURE In order to make robot services easily, we propose an architecture which has five layers, and which allows users to make robot services easily by just connecting components.</title>
      <sec id="sec-3-1">
        <title>A. System Architecture of Intelligent Application</title>
      </sec>
      <sec id="sec-3-2">
        <title>Development Platform</title>
        <p>This section outlines our proposed system architecture.
Figure 1 shows the hierarchy of the intelligent application
development platform, which consists of five layers.




</p>
        <p>Service layer: This layer is a task in the real world.
Users can make a robot service by using a service or a
combination of some services. A task is realized by a
sequence of a number of processes according to a
workflow.</p>
        <p>Process layer: This layer means a subtask composing
a task in the real world. Processes are realized by some
robots and some sensors. Each process has a flow
which consists of some modules. Processes from more
than one service can be used as a component that
implements the service.</p>
        <p>Module layer: This layer is a minimum software
module, which means a robot’s activity or a sensor’s
recognition. Modules are divided into two kinds:
modules for action and modules for recognition.
Modules for action are software modules that robots
perform. Modules for recognition are software
modules for sensors that recognize signal information
such as age, facial expression, poses and objects.</p>
        <p>
          Knowledge layer: This layer consists of some
ontologies and rules which are used by modules in
order to acquire knowledge about tasks, circumstances
and so on. For example, in this layer, there are the
Japanese Wikipedia Ontology (JWO)[
          <xref ref-type="bibr" rid="ref5">5</xref>
          ], OWL-S,
domain ontologies and domain rules. JWO is a general
ontology made from Wikipedia. JWO is used in dialog
with users. The details of JWO are described in [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ].
OWL-S is used in order to define services, processes,
and modules. Domain ontologies define the input and
output class of services, processes and modules.
Domain rules can change robot activities based on the
situation.
        </p>
        <p>Data layer: The data layer contains data used for
module processing. For example, the data layer
contains environmental maps, instance data, a gesture
database and dialog history.</p>
        <p>
          To realize this architecture, OWL-S is used. Modules are
defined as atomic processes in OWL-S, while services and
processes are defined as composite processes. Each service,
process and module has inputs and outputs. Inputs and outputs
correspond to classes of domain ontologies. Users can
combine modules, processes and services with each other by
binding before one’s output to the next one’s input. Some
modules for action use domain rules written in the Semantic
Web Rule Language (SWRL)[
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] in order to modify a robot’s
activities according to the situation. SWRL is a rule language
based on OWL, and enables domain ontologies to make new
triples when conditions are satisfied by writing If-Then rules.
This realizes the integration of symbols and signals.
        </p>
        <p>This section explains in detail the proposed architecture
based on a case study of the robot service: RobotCafe. In
RobotCafe, we use a robot named “NaoTorso”, which has an
upper body of the humanoid robot “Nao” and a mobile robot
“Turtlebot2”. Figure 2 shows an overview of RobotCafe.
RobotCafe has two services: customer relations and tray
service. The output of the former service binds to the input of
the latter.</p>
        <p>Each service has several processes and the output of the
previous process binds to the input of the next process. Each
process has a sequence of some modules. In this paper, we
focus on the process of greetings in customer relations. The
greeting process has nine modules as shown in Fig. 3, where
rectangles, rounded rectangles and ovals represent the modules
comprising the greeting process, arrows mean flows, and
labels of arrows mean the input and output of the modules. In
the blue rounded rectangles, the signals sensed by cameras are
changed to symbols such as age and the kind of object. In a red
oval, symbols such as face direction are changed to signals
which go forward to the next module.</p>
        <p>
          In the greeting process, a speech robot search is done first
and returns the IP address of the searched robot. According to
this IP address, the robot performs age recognition and object
recognition. In age recognition, the robot estimates the
customer’s age. In the age recognition module, we use the
NAOqi API[
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] of Aldebaran Robotics to enable the robot to
estimate the customer’s age. In object recognition, the robot
recognizes the objects which a customer has. This method can
recognize 200 objects such as umbrellas and suitcases. Then,
in the module which constructs the customer model, a person
model is made according to rules for judging the type of person
based on age by using the estimated age and the recognized
objects. Simultaneously, robot models are constructed in the
“Construction of robot model” module based on the result of
the speech robot search method. By using person models and
robot models, the cafe model is constructed. In the
“Acquisition of greeting contents” module, the contents of
greeting speech are determined by referring to the cafe model
and the rules for judging the contents of greeting speech. Then,
in the “Face direction recognition” module, the processing
waits until the customer’s face is directed at the robot. Finally,
in the speech module, the robot searched in the first module
speaks the content of the greeting speech by using speech
synthesis to convert from text to voice.
        </p>
        <p>In this flow, the class hierarchy and properties of the
RobotCafe ontology shown in Fig. 4 are used as inputs and
outputs of modules. Related to the greeting process, the Person
class, Object class, Robot class and Cafe class are determined
in this ontology. The Person class has subclasses such as Adult
class and Child class, which are used to judge the person type
of the customer. The Object class has subclasses such as
Umbrella class and Baggage class, which are used to
determine the greeting comment based on the customer’s
belongings.</p>
        <p>The Person class has two properties: age property and
hasObject property. The age property’s domain is the Person
class and its range is int type. The hasObject property’s domain
is the Person class and its range is the Object class. The Robot
class has the greeting property, which represents the contents
of greeting speech. The greeting property’s domain is the
Robot class and its range is String type. The Cafe class has two
properties: hasClient property and hasRobot property. The
hasClient property represents the relationship between the cafe
and the client, while the hasRobot property represents the
relationship between the cafe and the robot which is in the cafe.
The hasClient property’s domain is the Cafe class and its range
is the Person class. The hasRobot property’s domain is the
Cafe class and its range is the Robot class.</p>
        <p>Figure 5 shows a part of the SWRL rules for RobotCafe. In
Fig. 5, there are rules for judging the person type based on age
and rules for judging the greeting speech contents. For
example, in the former rules, there are rules that a person has
the YoungChild class when the person is 12 years old or
younger, and has the TeenAge class when the person is 13 to
19 years old. In the latter rules, there are rules such as that a
robot says “Hello. Thank you for coming.” when the customer
belongs to the TeenAge class or Adult class and that a robot
says “Hi. Where are your parents?” when the customer belongs
to the Child class.</p>
        <p>In addition, there are rules related to the customer’s
belongings in rules for judging the greeting speech contents.
For instance, there are the rules that a robot says “Would you
mind putting your umbrella there?” in addition to the usual
comments when the customer has an umbrella, and rules that a
robot says “Could I take care of your baggage?” in addition to
the usual comments when the customer has baggage.
Therefore, by using these rules, the greeting comments can be
tailored to the situation such as the customer’s age and
belongings.</p>
        <p>Table 1 shows the types, names and details of inputs and
outputs of modules composing the greeting process. In this
table, modules whose input or output class is null such as the
IP (Internet Protocol) address of the robot and recognized age
have no RDF graph input or output but basic data type input or
output.
Figure 6 shows the detailed processing of the acquisition
by the greeting contents module. This module inputs RDF
graphs merging the Person model, Robot model and Cafe
model. In this module, rules for judging the greeting speech
contents are used as inputs. Then, this process makes a triple
which defines the greeting contents by using the greeting
property and outputs the triple. Figure 6 shows an example in
which the greeting content is “Hello. Thank you for coming.
Would you mind putting your umbrella there?” for a customer
who belongs to the MiddleAge class and has an umbrella.</p>
        <p>As mentioned above, RobotCafe can be implemented by
using the proposed architecture based on modules, processes,
services, ontologies and rules and can be dynamically changed
according to the circumstances.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>IV. DISCUSSION</title>
      <p>As an example of the robot service using symbols and
signals, we describe the greeting process in RobotCafe. In this
process, we use a robot called “NaoTorso”, which is a
combination of a humanoid and the mobile robot “NaoTorso
with Kobuki”. This robot has an upper body of the humanoid
robot “Nao” and a mobile robot “Turtlebot2”. This robot is
good at dialogue and stable movement.</p>
      <p>Figure 7 shows the greeting process. When a customer
comes, NaoTorso approaches the customer and speaks the
greeting dialogue.</p>
      <p>Figure 8 shows the implementation of person detection.</p>
      <p>
        The left side of Fig. 8 shows an example of person detection,
which is implemented by using Choregraphe[
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] of Aldebaran
Robotics. In Choregraphe, boxes mean each robot activity
such as saying some text, person recognition and so on. Users
can create a robot service by connecting some boxes. We
made RobotCafe by implementing boxes which use web
services corresponding to modules in our intelligent
application development platform.
The upper right side of Fig. 8 shows the result of person
recognition. When a man enters the room, the robot recognizes
him like this. We use Regions with Convolutional Neural
Network features(R-CNN)[
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] to extract object labels and their
locations simultaneously from an input image. R-CNN
preliminary learns feature extractors by training a deep
convolutional neural network (CNN), AlexNet[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], on the
ImageNet ILSVRC dataset[
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. In the inference stage, R-CNN
firstly obtains object proposal regions by a selective search[
        <xref ref-type="bibr" rid="ref13">13</xref>
        ],
then extracts features for each region by using the pre-trained
deep CNN. Finally, those features are classified to one of the
classes of interest or background by pre-trained Support
Vector Machines (SVMs)[
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. We use R-CNN as an object
detector used by NAO.
      </p>
      <p>The lower right side of Fig. 8 shows the appearance of the
person and the robot.</p>
      <p>The greeting process is performed as follows, which
shows the greeting comments spoken by NaoTorso according
to the customer’s situation.
 A child (whether with belongings or not)</p>
    </sec>
    <sec id="sec-5">
      <title>NaoTorso: Hi. Where are your parents?</title>
      <p> An adult without belongings</p>
    </sec>
    <sec id="sec-6">
      <title>NaoTorso: Hello! Thank you for coming.</title>
      <p> An adult with an umbrella</p>
      <p>NaoTorso: Hello! Thank you for coming. Would you mind
putting your umbrella there?
 An adult with some baggage</p>
      <p>NaoTorso: Hello! Thank you for coming. Could I take care
of your baggage?</p>
    </sec>
    <sec id="sec-7">
      <title>V. CONCLUSIONS</title>
      <p>In this paper, we proposed an intelligent application
development platform based on SWS and OWL-S. This
architecture enables users to integrate symbols and signals and
to make robot services easily which can be tailored to the
circumstances. The architecture has five layers: Service,
Process, Module, Knowledge and Data. The upper layer
consists of the combination of components in the lower layer.
In the case study, we showed a greeting service in a cafe which
can be changed according to the customer’s situation.</p>
      <p>In the future, we will create many modules for actions and
modules for recognition in order to achieve different and
difficult tasks. Then, we will create modules for
affect/emotion processing. In addition, we will build the
architecture in order to manage robots and sensors and a
workflow editor which enables users to use this intelligent
application development platform easily. Finally, we will do
more testing in complex scenarios involving many processes.</p>
    </sec>
    <sec id="sec-8">
      <title>ACKNOWLEDGEMENT This work has been supported by JST/CREST research titled A Framework PRINTEPS to Develop Practical Artificial Intelligence.</title>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Mori</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Ogawa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hikawa</surname>
          </string-name>
          , and
          <string-name>
            <given-names>T.</given-names>
            <surname>Yamaguchi</surname>
          </string-name>
          ,
          <article-title>"Multi-robot Coordination Based on Ontologies and Semantic Web Service," in Knowledge Management and Acquisition for Smart Systems</article-title>
          and Services, Gold Coast,
          <year>2014</year>
          , pp.
          <fpage>150</fpage>
          -
          <lpage>164</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M.</given-names>
            <surname>Sato</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kamei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Nishio</surname>
          </string-name>
          , and
          <string-name>
            <given-names>N.</given-names>
            <surname>Hagita</surname>
          </string-name>
          ,
          <article-title>"The Ubiquitous Network Robot Platform: Common platform for continuous daily robotic services,"</article-title>
          <source>in 2011 IEEE/SICE International Symposium on System Integration (SII)</source>
          ,
          <year>Kyoto</year>
          ,
          <year>2011</year>
          , pp.
          <fpage>318</fpage>
          -
          <lpage>323</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>M.</given-names>
            <surname>Waibel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Beetz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Civera</surname>
          </string-name>
          ,
          <string-name>
            <surname>R. D'Andrea</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Elfring</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Galvez-Lopez</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Haussemann</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Janssen</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. M. M. Montiel</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Perzylo</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Schiessle</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Tenorth</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          <string-name>
            <surname>Zweigle</surname>
          </string-name>
          , and R. van de Molengraft,
          <article-title>"RoboEarth,"</article-title>
          <source>IEEE Robotics &amp; Automation Magazine</source>
          , vol.
          <volume>18</volume>
          , no.
          <issue>2</issue>
          , pp.
          <fpage>69</fpage>
          -
          <lpage>82</lpage>
          ,
          <year>June 2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <article-title>[4] 4. OWL-S: Semantic Markup for Web Services</article-title>
          , http://www.w3.org/Submission/OWL-S/#tex2html2
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Japanese</given-names>
            <surname>Wikipedia Ontology</surname>
          </string-name>
          , http://www.wikipediaontology.org.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S.</given-names>
            <surname>Tamagawa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sakurai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Tejima</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Morita</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Izumi</surname>
          </string-name>
          , and T. Yamaguchi, “
          <article-title>Learning a Large Scale of Ontology from Japanese Wikipedia</article-title>
          ,” IEEE/WIC/ACM International Conference on Web Intelligence, pp.
          <fpage>279</fpage>
          -
          <lpage>286</lpage>
          ,
          <year>September 2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>SWRL</given-names>
            <surname>: Semantic Web Rule Language</surname>
          </string-name>
          , http://www.w3.org/Submission/SWRL/
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>NAOqi</surname>
            <given-names>API</given-names>
          </string-name>
          , http://doc.aldebaran.com/2-1/naoqi/index.html
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Choregraphe</surname>
          </string-name>
          , http://doc.aldebaran.com/2-1/software/choregraphe/index.html
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>R.</given-names>
            <surname>Girshick</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Donahue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Darrell</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Malik</surname>
          </string-name>
          , “
          <article-title>Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation,” Computer Vision</article-title>
          and Pattern Recognition, pp.
          <fpage>580</fpage>
          -
          <lpage>587</lpage>
          ,
          <year>June 2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>K.</given-names>
            <surname>Alex</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Sutskever</surname>
          </string-name>
          , and
          <string-name>
            <given-names>E. H.</given-names>
            <surname>Geoffrey</surname>
          </string-name>
          , “
          <article-title>ImageNet Classification with Deep Convolutional Neural Networks</article-title>
          ,
          <source>” Advances in Neural Information Processing Systems</source>
          <volume>25</volume>
          , pp.
          <fpage>1097</fpage>
          -
          <lpage>1105</lpage>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>O.</given-names>
            <surname>Russakovsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Su</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Krause</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Satheesh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Karpathy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Khosla</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bernstein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. C.</given-names>
            <surname>Berg</surname>
          </string-name>
          , and L.
          <string-name>
            <surname>Fei-Fei</surname>
          </string-name>
          ,
          <article-title>“ImageNet Large Scale Visual Recognition Challenge</article-title>
          ,”
          <source>International Journal of Computer Vision</source>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>42</lpage>
          ,
          <year>April 2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>J. R. R.</given-names>
            <surname>Uijlings</surname>
          </string-name>
          , K. E. A. van de Sande, T. Gevers,
          <article-title>and</article-title>
          <string-name>
            <given-names>A. W. M.</given-names>
            <surname>Smeulders</surname>
          </string-name>
          , “
          <article-title>Selective Search for Object Recognition</article-title>
          ,”
          <source>International Journal of Computer Vision</source>
          , no.
          <issue>2</issue>
          , vol.
          <volume>104</volume>
          , pp.
          <fpage>154</fpage>
          -
          <lpage>171</lpage>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>C.</given-names>
            <surname>Cortes</surname>
          </string-name>
          and
          <string-name>
            <given-names>V.</given-names>
            <surname>Vapnik</surname>
          </string-name>
          , “
          <string-name>
            <surname>Support-Vector</surname>
            <given-names>Networks</given-names>
          </string-name>
          ,”
          <article-title>Machine learning</article-title>
          , vol.
          <volume>20</volume>
          , no.
          <issue>3</issue>
          , pp.
          <fpage>273</fpage>
          -
          <lpage>297</lpage>
          ,
          <year>1995</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>