<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A first approach to AGI-based Robot Task Planning</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Michele Thiella</string-name>
          <email>michele.thiella@studenti.unipd.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Elisa Tosello</string-name>
          <email>elisa.tosello@unipd.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Enrico Pagello</string-name>
          <email>enrico.pagello@unipd.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Robot Task Planning, Artificial General Intelligence, OpenCog, Robot manipulation</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Dept. of Information Engineering, University of Padova</institution>
          ,
          <addr-line>Via Gradenigo 6/B, 35131 Padova</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Robotics srl</institution>
          ,
          <addr-line>Contrà Valmerlara 21, 36100 Vicenza</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Current research in robot Task Planning aims to develop solvers which quickly adapt to new assignments and scenarios. To this aim, we extend an existing proto-Artificial General Intelligence system, namely OpenCog, and give it the ability to efectively solve manipulation tasks whose domains contain four actions: pick, place, stack, and unstack. To let OpenCog solve this class of problems, we exploit its modules as the foundation of a Knowledge Base that describes and stores domains, problems, and the interactions between them. Then, we equip the system with a Breadth-First Search algorithm that finds the sequence of actions that solve the assignments. To prove the goodness of our proposal, we include and analyze a manipulation task where a manipulator robot has to interact with a human user to assemble some industrial components. Obtained results show that our system is complete and generic in terms of the domain and problem under evaluation. Future work will improve the achieved computational time and performance.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>(E. Pagello)
solving process and an easy adaptability to new assignments; (iii) storing data in a Knowledge
Base (KB) to facilitate the sharing of lessons learnt.</p>
      <p>
        In this context, we propose to exploit an existing proto-Artificial General Intelligence (AGI)
system: Open Cognition (OpenCog) [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. It includes a flexible Knowledge Representation
(KR) embodied in a scalable knowledge store, a cognitive process scheduler, and a plug-in
architecture for the interaction between cognitive, perceptual, and control algorithms. Such
features let OpenCog adapt to new conditions, understand them, and create behaviors based on
the information learned. To make the system able to solve TP problems, we exploited its modules
as the foundation of a KB that describes and stores domains, problems, and the interactions
between them. Then, we equipped the system of a Breadth-First-Search (BFS) algorithm that
ifnds the sequence of actions that solve an assignment. Our contributions follow:
• the implementation of a BFS algorithm for TP that solves manipulation problems through
the combination of four actions: pick, place, stack, and unstack;
• the implementation of a supra-system that equippes OpenCog with such an algorithm
and all the features necessary for TP;
• a preliminary evaluation of our proposal while solving an assembly task in conjunction
with a human user. Obtained results show that our system is complete and generic in
terms of problem and domain definitions. Future developments will improve the achieved
computational time and performance.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. State of the art</title>
      <p>
        AGI is the ability of an intelligent agent to understand or learn any intellectual task that a
human being can [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. It is a primary goal of several artificial intelligence research.
      </p>
      <p>
        Among others, the OpenAI project [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] applies Deep NNs (DNNs) to translate natural language
into code [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], connects text to images [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], and produce human-like text [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Although OpenAI
achieved valuable results, the NNs approach is expensive and requires an extensive amount of
data to achieve generality. Moreover, it sufers from the adversarial example problem: adding
an imperceptibly small but carefully designed perturbation leads the model to make a wrong
prediction [
        <xref ref-type="bibr" rid="ref10 ref8 ref9">8, 9, 10</xref>
        ]. Finally, NNs are black boxes: while they can approximate any function,
studying their structure won’t give any insight on the function being approximated [11].
      </p>
      <p>For these reasons, we investigated another AGI-oriented approach: OpenCog. This
opensource framework includes a comprehensive model of human-like general intelligence. It
exploits an integrative approach in which multiple AI algorithms cooperate on a common
representational substrate. Such algorithms include DNNs, Probabilistic Logic Theorem
Proving, Evolutionary Learning, and Concept Blending. The system is largely scalable, given the
amount and diversity of data it can contain. Furthermore, current research is improving its
distributivity by developing a decentralized structure composed of many KR databases and their
query/reasoning engines. Agents do not need to locally solve assigned tasks. They can exploit
the available knowledge and eventually augment it with new experiences.</p>
      <p>These features make OpenCog easily extensible to the resolution of TP problems. Indeed, our
own experience on TP has shown that most current planners are still extremely dependent on
the problem, the domain, and the language used to define them [ 12, 13, 14]. Examples include
the Planning Domain Definition Language (PDDL)-based approaches [ 15]. We aim to overcome
these limitations and give generality to the TP solvers.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Our proposal</title>
      <p>
        To provide OpenCog [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] with TP capabilities, we added a Search Algorithm to its modules, and
we implemented a sense-plan-act framework whose sense unit includes both visual perception
and Natural Language Processing (NLP). The exploited OpenCog modules follow:
• Atom Space. It is a KR database composed of Atoms. Atoms are hypergraphs that enclose
any type of information (i.e., data, procedures, etc.) [16]. Atoms refer to both Nodes
(vertices) and Links (edges) of an hypergraph. For example, a Concept Node represents
any physical or abstract concept. Inheritance Nodes specify both intentional (is-a) and
extensional (is-an-instance-of) relationships. A Predicate Node names the predicate of
a relation, where predicates are functions that have arguments, and produce a truth
value as output. An Evaluation Link lets specify the truth value of a predicate on a set of
arguments. Finally, a Query Link specifies a search pattern that can be grounded, solved,
or satisfied. Thus, within an industrial environment, if we want to specify that a snap
ring is a movable object positioned on the top of another snap ring, and that an agent is
holding a bearing sleeve, we use the following Atoms code:
• Relex2Logic (R2L). R2L deduces the predicate-argument structure of natural language
sentences. It first produces a logical representation corresponding to hypergraphs of
Atoms that are associated with the words of these sentences. Then, it structures the
obtained descriptions to express the sentence logic in the form of a hypergraph. The
following example shows some English sentences stored in the form of hypergraphs:
• Pattern Matcher (PM). PM works with hypergraphs in the sense of fast extraction of
specific data and query engine. In detail, it searches the AtomSpace for specific patterns
of Atoms, i.e., hypergraphs with nodes and links of several types. If patterns have “holes”
(i.e., variable locations), PM will “fill in the blanks” [ 17].
      </p>
      <p>In this context, we store four actions in the AtomSpace: pick, place, stack, and unstack. Then,
we add a Search Algorithm able to find the sequence of actions that bring a certain environment
from its initial state to a desired one. The algorithm is based on BFS: each node is an AtomSpace
that depicts the environment at a given time. Each edge out of a node represents one action
executable at that state. Thus, each branch of the tree ends up describing a diferent sequence
of actions, and the tree expansion takes place through the execution of actions using PM. We
encode actions as Query Links, where a Query Link type Atom is defined as follows:
For example, the pick action is a Query Link that contains a pattern describing the prerequisites
needed for an object to be picked up. Executing this rule will search for all atoms that satisfy
these prerequisites within the AtomSpace at the current moment. The result is a list of Atoms
corresponding to the objects that can actually be picked up. Finally, Atoms are pasted into the
second pattern of the rule, resulting in a hypergraph each, describing the pick action performed.</p>
      <p>We decided to encode actions as Query Links for the following reasons:
1. They allow us to specify a search pattern using Variable Nodes, which correspond to the
“holes” introduced above. When we execute a Query Link, PM fills the “holes” assigning
the Variable Nodes with the solution Atoms found by the search.
2. We can create new graphs as an efect of the Query Link execution. These graphs are new
patterns containing the Variable Nodes replaced with the solution from the first point and
inserted directly into the Atom Space.
3. Query Links are imperative: they actually perform the action, as opposed to declarative
Atoms that simply describe it (which can be used in the future to reason about the action
and eventually improve it with experience).</p>
      <p>Based on these reasons, we can perform an action with or without parameters. In the former
case, we can execute the Query Link in two steps following motivations 1 and 2. This mode
allows to understand which objects are involved in the action. In the latter, we add a constraint
to the Query Link search, limiting it to the object passed as a parameter.</p>
      <p>Starting from the root node, the following steps compose our BFS algorithm:
1. If the node is root:
• If the robot is holding an object in the root Atomspace, then two temporary nodes
are created as copies of the root related to the place and stack action.</p>
      <p>• Otherwise, the same nodes are created but related to the pick and unstack actions.
2. If the node is not root:
• Consider that two of the next four actions can be excluded looking at the previous
one. Starting from either place or stack, only pick and unstack can be performed,
and vice versa. Consequently, by looking at the action assigned to the edge incident
on this node, two temporary nodes are created as copies of this node, associating
the actions following the rule just explained.
3. In each temporary node, the assigned action is executed in Without Parameters mode.
4. From the resulting Atomspace of each one, the objects on which the action is applicable
are extracted.
5. For each object found, a new copy node of the root is created. Then, the assigned action
is applied to this node in With Parameters mode.
6. These final nodes become the children of the BFS tree and the edges are labelled with the
respective performed action.
7. The algorithm starts again for each new node, following the breadth-first order of the
queue where nodes are stored.</p>
      <p>This makes the algorithm and the entire system generic with respect to domains, problems, and
actions. To avoid loops of actions (i.e., actions that bring the environment into a state already
encountered in the same branch), we check if each new node has already been encountered
within its branch. In this case, the related sequence of actions created a loop and the expansion
of that branch ends.</p>
      <p>To add extra generality to our proposal and always allow to define the assignment as a
Markov Decision Process (MDP), we include a learning phase, strictly connected to R2L. In
detail, human users can help the system to reconstruct the initial state of the system, from
which the search algorithm will look for a sequence of actions to reach the assigned goal. To
this aim, they can input additional information via English sentences and R2L accurately store
this data as hypergraph. In the same way, human users can request the system to bring the
environment to a certain goal. Users do not need to describe the arrangement of the whole
environment: they can focus on only the states of the objects involved in the task.</p>
      <p>Figure 1 shows the architecture of the obtained framework. A perception module detects
the objects populating the robot’s surroundings. A human user inputs the desired goal and
the information useful to achieve it (e.g., the position of objects not visible for the robot). The
Search Algorithm exploits all available data to find a solution. If a solution exists, the robot will
perform each action of the computed sequence until achieving the assignment.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments and Results</title>
      <p>To prove the goodness of our proposal, we implemented a Robot Operating System (ROS)
[18]based setup where a human operator has to perform an assembly task. Given the task, a
Franka Emika Panda manipulator robot1 helps the human user by identifying the appropriate
industrial components to be assembled, picking them in the correct order, and placing them on
a pre-defined unloading position close to the human user. The help is reciprocal since the scene
is not known beforehand: the human operator helps the robot by giving information about the
position of non-visible objects. In this way, the robot can plan the sequence of manipulation
actions optimal to achieve the task.</p>
      <p>As shown in Figure 3, we mounted the robot on the same workbench where the human
operates. Then, a free table acts as a collector and receives both the incoming pieces and
the final assembly. Finally, a set of bins simulate a warehouse and group the incoming parts
according to their type. We have a Microsoft Kinect One on the top of the robot end-efector,
letting it detect the pieces on the table and inside the bins. To simplify both the perception and
manipulation routines, industrial components are simulated as cubes with an AprilTag2 fiducial
marker attached on their tops. Each marker associates a semantic representation to each cube,
i.e., the name of the industrial component it represents. To simplify collision avoidance, we
assume that the human operator does not move within the robot workplace.</p>
      <p>At the beginning, we assume that all useful components are in the scene, each inside its
corresponding bin. The human helper informs the robot about the task to be performed. The
robot scans its surroundings, detects the useful visible objects, and exploits the associated
1See https://www.franka.de/
2See https://april.eecs.umich.edu/software/apriltag</p>
      <p>Parameters</p>
      <p>Values
Objects
Goal
Additional
Solution
Time
N. Iters.
the solution, the computational time, and the number of iterations needed to find the solution.
arm of Figure 2. Each of its components is simulated by a cube. The snap rings are red cubes, the
bearing sleeves are blue cubes, and the recirculating ball sleeve is a green cube. At the beginning, each
cube is inside its corresponding colored bin. The human user gives the robot some additional info:
”BearingSleeve1 is on BearingSleeve2”, ”SnapRing1 is on SnapRing2”. This information, together with
the retrieved visual data, lets the robot deduce the initial scene and compute an action plan that neatly
brings the components from their initial location to the unloading station.
semantic representations to create a description of the environment as hypergraphs within
the AtomSpace. Then, it interrogates the human operator to retrieve the missing information.
Once known the entire environment configuration, the Search Algorithm looks for a solution
by combining the set of preconditions and efects of the considered actions. The result is a set
of actions that let the robot unload all components on the pre-defined unloading area in the
correct order. If a solution exists, the robot executes the computed actions and gives the human
user all pieces needed to compose the product.</p>
      <p>We performed multiple tests with diferent assembled products and various quantities of
pieces to be assembled. For the sake of brevity, we only report the results obtained when
assembling the 5 items of Table I: two snap rings, a recirculating ball sleeve, and two bearing
sleeves. They should form the component of Figure 2: a piece of an orthogonal torque reaction
arm. Thus, the following goal configuration should be true: “SnapRing1 is on the workbench.
BearingSleeve1 is on SnapRing1. RecirculatingBallSleeve is on BearingSleeve1. BearingSleeve2 is on
RecirculatingBallSleeve. SnapRing2 is on BearingSleeve2.” As shown in the table, as additional
information, the human informs the robot that some parts are not visible because they are
behind other items: “SnapRing1 is on SnapRing2”, “BearingSleeve1 is on BearingSleeve2”. Once
the system has learned the configuration of its surroundings and the final goal, it process the
information as hypergraphs within the AtomSpace until producing a sequence of pick, place,
stack, and unstack actions that solves the human request. The solution follows (see Table I):
1. (unstack (ConceptNode ”SnapRing1”) (ConceptNode ”SnapRing2”))
2. (stack (ConceptNode ”SnapRing1”) (ConceptNode ”workbench”))
3. (unstack (ConceptNode ”BearingSleeve1”) (ConceptNode ”BearingSleeve2”))
4. (stack (ConceptNode ”BearingSleeve1”) (ConceptNode ”SnapRing1”))
5. (pickup (ConceptNode ”RecirculatingBallSleeve”))
6. (stack (ConceptNode ”RecirculatingBallSleeve”) (ConceptNode ”BearingSleeve1”))
7. (pickup (ConceptNode ”BearingSleeve2”))
8. (stack (ConceptNode ”BearingSleeve2”) (ConceptNode ”RecirculatingBallSleeve”))
9. (pickup (ConceptNode ”SnapRing2”))
10. (stack (ConceptNode ”SnapRing2”) (ConceptNode ”BearingSleeve2”))</p>
      <p>Once the Search Algorithm has found the sequence of actions to achieve the assignment, the
robot executes the computed actions, as shown in Figure 3. To find a solution, the system takes
120 seconds on a laptop with Intel(R) Core(TM) i5-9300H CPU, NVIDIA GeForce GTX 1650
Max-Q graphics card and 8GB of RAM. The operating system is installed on an external SSD
connected with the USB 3.1 standard. The computational time decreases exponentially as the
number of objects to be assembled decreases.</p>
      <p>We compared our proposal with Fast-Forward (FF) [19] and SMTPlan [20]. The former takes
0.006 sec to find a solution, the latter takes 18.141 sec. This result may seem discouraging.
Otherwise, we should account that both the algorithms take as input a problem and a domain
formulated using PDDL. Such formulations should faithfully represent the state of the world in
its entirety. In our case, instead, the Search Algorithm is complete, and the overall framework is
independent of the formulation of initial and final configurations. Indeed, the human operator
can inform the system about the observability of the problem and can help deduce the start
configuration of the environment while exploiting the natural language.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusions and Future Work</title>
      <p>In this paper, we extended OpenCog to let the system solve TP problems. Without loss of
generality, we focused on robot manipulation problems solvable through a combination of four
actions: pick, place, stack, and unstack. To let OpenCog solve these problems, we exploited
its modules as the foundation of a KB that describes and stores domains, problems, and the
interactions between them. Then, we equipped the system of a BFS algorithm that finds the
sequence of actions that solve the assignments. We included a preliminary evaluation of our
proposal that asked a manipulator robot to interaction with a human operator to solve an
assembly task. Obtained results prove that our system is complete and generic in terms of the
considered problem and domain.</p>
      <p>Future work includes the integration of additional actions to improve our system’s generality.
Moreover, we will focus on increasing its performance via temporal reasoning. Finally, we will
test our proposal in dynamic environments populated by movable objects.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>Part of this work was supported by MIUR (Italian Minister for Education), under the initiative
Departments of Excellence (Law 232/2016), and by Fondazione Cariverona, under the project
Collaborazione Uomo-Robot per Assemblaggi Manuali Intelligenti (CURAMI).
[11] V. Buhrmester, D. Münch, M. Arens, Analysis of explainers of black box deep neural
networks for computer vision: A survey, CoRR abs/1911.12116 (2019). URL: http://arxiv.
org/abs/1911.12116. arXiv:1911.12116.
[12] F. Ceola, E. Tosello, L. Tagliapietra, G. Nicola, S. Ghidoni, Robot task planning via deep
reinforcement learning: a tabletop object sorting application, in: 2019 IEEE International
Conference on Systems, Man and Cybernetics (SMC), 2019, pp. 486–492. doi:10.1109/SMC.
2019.8914278.
[13] G. Nicola, L. Tagliapietra, E. Tosello, N. Navarin, S. Ghidoni, E. Menegatti, Robotic object
sorting via deep reinforcement learning: a generalized approach, in: 2020 29th IEEE
International Conference on Robot and Human Interactive Communication (RO-MAN),
2020, pp. 1266–1273. doi:10.1109/RO- MAN47096.2020.9223484.
[14] L. Tagliapietra, E. Tosello, E. Menegatti, A planning domain definition language
interpreter and knowledge base for eficient automated planning, in: The 16th International
Conference on Intelligent Autonomous Systems (IAS-16)), 2021.
[15] M. Ghallab, A. Howe, C. Knoblock, D. Mcdermott, A. Ram, M. Veloso, D. Weld, D. Wilkins,
PDDL—The Planning Domain Definition Language, 1998. URL: http://citeseerx.ist.psu.edu/
viewdoc/summary?doi=10.1.1.37.212.
[16] B. Goertzel, Folding and unfolding on metagraphs, CoRR abs/2012.01759 (2020). URL:
https://arxiv.org/abs/2012.01759. arXiv:2012.01759.
[17] F. Baader, T. Nipkow, Term Rewriting and All That, Cambridge University Press, 1998.</p>
      <p>doi:10.1017/CBO9781139172752.
[18] M. Quigley, K. Conley, B. P. Gerkey, J. Faust, T. Foote, J. Leibs, R. Wheeler, A. Y. Ng, Ros:
an open-source robot operating system, in: ICRA Workshop on Open Source Software,
2009.
[19] J. Hofmann, Ff: The fast-forward planning system, AI Magazine 22 (2001) 57–62.
[20] M. Cashmore, M. Fox, D. Long, D. Magazzeni, A compilation of the full pddl+ language into
smt, Proceedings of the International Conference on Automated Planning and Scheduling
26 (2016) 79–87. URL: https://ojs.aaai.org/index.php/ICAPS/article/view/13755.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , P. Khandelwal,
          <string-name>
            <given-names>P.</given-names>
            <surname>Stone</surname>
          </string-name>
          ,
          <article-title>An empirical comparison of pddl-based and asp-based task planners</article-title>
          , CoRR abs/
          <year>1804</year>
          .08229 (
          <year>2018</year>
          ). URL: http://arxiv.org/abs/
          <year>1804</year>
          .08229. arXiv:
          <year>1804</year>
          .08229.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>D.</given-names>
            <surname>Hart</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Goertzel</surname>
          </string-name>
          ,
          <article-title>Opencog: A software framework for integrative artificial general intelligence</article-title>
          ., in: AGI, volume
          <volume>171</volume>
          <source>of Frontiers in Artificial Intelligence and Applications</source>
          , IOS Press,
          <year>2008</year>
          , pp.
          <fpage>468</fpage>
          -
          <lpage>472</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>B.</given-names>
            <surname>Goertzel</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          Pennachin (Eds.),
          <source>Artificial General Intelligence</source>
          , Cognitive Technologies, Springer,
          <year>2007</year>
          . URL: http://dblp.uni-trier.de/db/series/cogtech/354023733.html.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>P.</given-names>
            <surname>Dhariwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Hesse</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Klimov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Nichol</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Plappert</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Radford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Schulman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sidor</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Zhokhov</surname>
          </string-name>
          , Openai baselines, https://github.com/openai/baselines,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>M. C.</surname>
          </string-name>
          et al.,
          <article-title>Evaluating large language models trained on code</article-title>
          ,
          <source>CoRR abs/2107</source>
          .03374 (
          <year>2021</year>
          ). URL: https://arxiv.org/abs/2107.03374. arXiv:
          <volume>2107</volume>
          .
          <fpage>03374</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Radford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. W.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Hallacy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ramesh</surname>
          </string-name>
          , G. Goh,
          <string-name>
            <given-names>S.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Sastry</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Askell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Mishkin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Clark</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Krueger</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Sutskever</surname>
          </string-name>
          ,
          <article-title>Learning transferable visual models from natural language supervision</article-title>
          ,
          <source>CoRR abs/2103</source>
          .00020 (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>T. B. Brown</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Mann</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Ryder</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Subbiah</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Kaplan</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Dhariwal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Neelakantan</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Shyam</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Sastry</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Askell</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Agarwal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Herbert-Voss</surname>
            , G. Krueger,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Henighan</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Child</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Ramesh</surname>
            ,
            <given-names>D. M.</given-names>
          </string-name>
          <string-name>
            <surname>Ziegler</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Winter</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Hesse</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Chen</surname>
            , E. Sigler,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Litwin</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Gray</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Chess</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Clark</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Berner</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>McCandlish</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Radford</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          <string-name>
            <surname>Sutskever</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Amodei</surname>
          </string-name>
          ,
          <article-title>Language models are few-shot learners</article-title>
          , CoRR abs/
          <year>2005</year>
          .14165 (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A.</given-names>
            <surname>Gleave</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Dennis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Kant</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Wild</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Levine</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Russell</surname>
          </string-name>
          ,
          <article-title>Adversarial policies: Attacking deep reinforcement learning</article-title>
          , CoRR abs/
          <year>1905</year>
          .10615 (
          <year>2019</year>
          ). URL: http://arxiv.org/abs/
          <year>1905</year>
          . 10615. arXiv:
          <year>1905</year>
          .10615.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>N.</given-names>
            <surname>Papernot</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>McDaniel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Jha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Fredrikson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z. B.</given-names>
            <surname>Celik</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Swami</surname>
          </string-name>
          ,
          <article-title>The limitations of deep learning in adversarial settings</article-title>
          ,
          <source>in: 2016 IEEE European Symposium on Security and Privacy (EuroS P)</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>372</fpage>
          -
          <lpage>387</lpage>
          . doi:
          <volume>10</volume>
          .1109/EuroSP.
          <year>2016</year>
          .
          <volume>36</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>I. Goodfellow,</surname>
          </string-name>
          <article-title>Attacking machine learning with adversarial examples</article-title>
          ,
          <year>2020</year>
          . URL: https: //openai.com/blog/adversarial-example-research/.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>