<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Spatial Relations in Text-to-Scene Conversion</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Bob Coyne</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Richard Sproat</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Julia Hirschberg</string-name>
          <email>juliag@cs.columbia.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Columbia University</institution>
          ,
          <addr-line>New York NY</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Oregon Health &amp; Science University</institution>
          ,
          <addr-line>Beaverton, Oregon</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Spatial relations play an important role in our understanding of language. In particular, they are a crucial component in descriptions of scenes in the world. WordsEye (www.wordseye.com) is a system for automatically converting natural language text into 3D scenes representing the meaning of that text. Natural language o ers an interface to scene generation that is intuitive and immediately approachable by anyone, without any special skill or training. WordsEye has been used by several thousand users on the web to create approximately 15,000 fully rendered scenes. We describe how the system incorporates geometric and semantic knowledge about objects and their parts and the spatial relations that hold among these in order to depict spatial relations in 3D scenes.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>Spatial relations are expressed either directly or implicitly in a wide range of
natural language descriptions. To represent these descriptions in a 3D scene, one
needs both linguistic and real-world knowledge, in particular knowledge about:
the spatial and functional properties of objects; prepositions and the spatial
relations they convey, which is often ambiguous; verbs and how they resolve to
poses and other spatial relations. For example, to interpret apple in the bowl we
use our knowledge of bowls { that they have interiors that can contain objects.
With di erent objects (e.g., boat in water), a di erent spatial relation is conveyed.</p>
      <p>
        WordsEye [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] is a system for automatically converting natural language text
into 3D scenes representing the meaning of that text. A version of WordsEye has
been tested online (www.wordseye.com) with several thousand real-world users.
We have also performed preliminary testing of the system in schools, as a way to
help students exercise their language skills. Students found the software fun to
use, an important element in motivating learning. As one teacher reported, \One
kid who never likes anything we do had a great time yesterday...was laughing
out loud."
      </p>
      <p>WordsEye currently focuses on directly expressed spatial relations and other
graphically realizable properties. As a result, users must describe scenes in
somewhat stilted language. See Figure 1. Our current research focuses on improving
the system's ability to infer these relations automatically. However, in this
paper, we describe the basic techniques used by WordsEye to interpret and depict
directly expressed spatial relations.</p>
      <p>In Section 2 we describe previous systems that convert natural language
text to 3D scenes and prior linguistic work on spatial relations. In Section 3 we
provide an overview of WordsEye. In Section 4 we discuss the spatial, semantic
and functional knowledge about objects used to depict spatial relations in our
system. We conclude and describe other ongoing and future work in Section 5.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Prior Work</title>
      <p>
        Natural language input has been investigated in some early 3D graphics systems
[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ][
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] including the Put system [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], which was limited to spatial arrangements
of existing objects in a pre-constructed environment. In this system, input was
restricted to an arti cial subset of English consisting of expressions of the form
P ut(X; P; Y ), where X and Y are objects and P is a rigidly de ned spatial
preposition. Work at the University of Pennsylvania's Center of Human
Modeling and Simulation [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], used language to control animated characters in a closed
virtual environment. CarSim [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] is a domain-speci c system that creates
animations from natural language descriptions of accident reports. CONFUCIUS
[
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] is a multi-modal text-to-animation system that generates animations of
virtual humans from single sentences containing an action verb. In these systems
the referenced objects, attributes, and actions are typically relatively small in
number or targeted to speci c pre-existing domains.
      </p>
      <p>
        Spatial relations have been studied in linguistics for many years. One
reasonably thorough study for English is Herskovits [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], who catalogs ne-grained
distinctions in the interpretations of various prepositions.3 For example, she
distinguishes among the various uses of on to mean \on the top of a horizontal
surface" (the cup is on the table), or \a xed to a vertical surface" (the picture
is on the wall). Herskovits notes that the interpretation of spatial expressions
may involve considerable inference. For example, the sentence the gas station is
at the freeway clearly implies more than just that the gas station is located next
to the freeway; the gas station must be located on a road that passes over or
under the freeway, the implication being that, if one proceeds from a given point
along that road, one will reach the freeway, and also nd the gas station.
3 It is important to realize that how spatial relations are expressed, and what kinds
of relations may be expressed varies substantially across languages. Levinson and
colleagues [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] have catalogued profound di erences in the ways di erent languages
encode relations between objects in the world. In particular, the Australian language
Guugu Yimithirr and the Mayan language Tzeltal use absolute frames of reference
to refer to the relative positions of objects. In Guugu Yimithirr, one can locate a
chair relative to a table only in terms of cardinal points saying, for example, that the
chair is north of the table. In English such expressions are reserved for geographical
contexts | Seattle is north of Portland | and are never used for relations at what
Levinson terms the \domestic scale". In Guugu Yimithirr one has no choice, and
there are no direct translations for English expressions such as the chair is in front
of the table.
      </p>
      <p>Eye of the Beholder by Bob Coyne
No Dying Allowed by Richard Sproat
Input text: The silver penny is on the moss
ground. The penny is 7 feet tall. A clown is 2
feet in front of the penny. The clown is facing the
penny.</p>
      <p>Input text: Eight big white washing machines
are in front of the big cream wall. The wall is 100
feet long. The \No Dying Allowed" whiteboard is
on the wall. The whiteboard is one foot high and
ve feet long. The ground is tile. Death is in front
of the washing machines. It is facing southeast.</p>
      <p>
        Death is eight feet tall.
Our current system is an updated version of the original WordsEye system [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ],
which was the rst system to use a large library of 3D objects to depict scenes in a
free-form manner using natural language. The current system contains 2,200 3D
objects and 10,000 images and a lexicon of approximately 15,000 nouns. It
supports language-based control of objects, spatial relations, and surface properties
(e.g., textures and colors); and it handles simple coreference resolution,
allowing for a variety of ways of referring to objects. The original WordsEye system
handled 200 verbs in an ad hoc manner with no systematic semantic modeling
of verb alternations and argument combinations. In the current system, we are
instead adding frame semantics to support verbs more robustly. To do this, we
are utilizing our own lexical knowledge-base, called the SBLR (Scenario-Based
Lexical Resource) [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. The SBLR consists of an ontology and lexical semantic
information extracted from WordNet [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] and FrameNet [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] which we are
augmenting to include the ner-grained relations and properties on entities needed
for depicting scenes as well as capturing the di erent senses of prepositions
related to those properties and relations.
      </p>
      <p>
        The system works by rst parsing each input sentence into a dependency
structure. These dependency structures are then processed to resolve anaphora
and other coreferences. The lexical items and dependency links are then
converted to semantic nodes and roles drawing on lexical valence patterns and other
information in the SBLR. The resulting semantic relations are then converted
to a nal set of graphical constraints representing the position, orientation, size,
color, texture, and poses of objects in the scene. Finally, the scene is composed
from these constraints and rendered in OpenGL (http://www.opengl.org) and
optionally ray-traced in Radiance [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. The user can then provide a title and
caption and save the scene in our online gallery where others can comment and
create their own pictures in response. See Figure 1.
4
      </p>
    </sec>
    <sec id="sec-3">
      <title>Spatial Relations</title>
      <p>WordsEye uses spatial tags and other spatial and functional properties on
objects to resolve the meaning of spatial relations. We focus here on the
interpretation of NPs containing spatial prepositions of the form \X-preposition-Y",
where we will refer to X as the figure and Y as the ground. For example, in
snow is on the roof, snow is the figure and roof is ground. The
interpretation of the spatial relation often depends upon the types of the arguments to
the preposition. There can be more than one interpretation of a spatial relation
for a given preposition The geometric and semantic information associated with
those objects will, however, help narrow down the possibilities.</p>
      <p>Base
Stem</p>
      <p>Canopy</p>
      <p>Top surface</p>
      <p>Side surface
Cup</p>
      <p>Enclosure</p>
      <p>Touch-point handle
{ Is-a: The lexical category to which the given object belongs.
{ Spatial tags identifying the following regions: (See Figure 2)</p>
      <p>Canopy: A canopy-like area \under" an object (e.g., under a tree).
Cup: A hollow area, open above, that forms the interior of an object.
Enclosure: An interior region, bounded on all sides (holes allowed).
Top/side/bottom/front/back: For both inner and outer surfaces.
Named-part: For example, the hood on car.</p>
      <p>Stem: A long thin, vertical base.</p>
      <p>Opening: An opening to an object's interior (e.g., doorway to a room).
Hole-through: A hole through an object. For example, a ring or donut.
Touch-point: Handles and other functional parts on the object. For
example, in John opened the door, the doorknob would be marked as a
handle, allowing the hand to grasp at that location.</p>
      <p>Base: The region of an object where it supports itself.
{ Overall shape: A dominant overall shape used in resolving various spatial
relations. For example, sheet, block, ribbon, cup, tube, disk, rod.
{ Forward/Upright direction: The object's default orientation.
{ Size: The default real-world size of the object. This is also used in
spatial relations where the figure and ground size must be compatible. For
example, ring on a stick versus *life-preserver on a pencil.
{ Length axis: The axis for lengthening an object.
{ Segmented/stretchable: Some objects don't change size in all
dimensions proportionally. For example, a fence can be extended inde nitely in
length without a corresponding change in height.
{ Embeddable: Some objects, in their normal function, are embedded in
others. For example, replaces are embedded in walls, and boats in water.
{ Wall-item and Ceiling-item: Some objects are commonly attached to
walls or ceilings or other non-upward surfaces. Some (e.g., pictures) do this
by virtue of their overall shape, while for others (e.g., sconces) the
orientation of the object's base is used to properly position the object.
{ Flexible: Flexible objects such as cloth and paper allow an object to hang
or wrap. For example, towel over a chair.
{ Surface element: Any object that can be part of a at surface or layer.</p>
      <p>For example, a crack, smudge, decal, or texture.
{ Semantic properties such as Path, Seat, Airborne for object function.</p>
      <p>
        Some of these features were used in earlier versions of our system [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Features
we have added to the current version include: surface element, embeddable,
overall shape, length axis, segmented/stretchable. Other features,
including (flexible, opening, hole-through and various semantic features)
are in the development stage. The implemented tagset supports the generation
of scenes such as Figure 3.
      </p>
      <p>
        In order to resolve a spatial relation, we nd the spatial tags and other
features of the figure and ground objects that are applicable for the given
preposition. For example, if the preposition is under, a canopy region for the
ground object is relevant, but not a top surface. Various other factors, such
as size, must also be considered. With enclosed-in, the figure must fully
t in the ground. For embedded-in, only part need t. For other relations
(e.g., next-to), the objects can be any size, but the figure location might
vary. For example, The mosquito is next to the horse and The dog is next to the
horse position the figure in di erent places, either in the air or on the ground,
Spatial Relation Example Partial Conditions
on-top-surface vase on table Ground is upward-surface
on-vertical-surface postcard on fridge Ground is vertical-surface
on-downward-surface fan on ceiling Ground is downward-surface
on-outward-surface pimple on nose Ground is surface
pattern/coating-on plaid pattern on shirt Figure is texture or layer
t-on-custom train on track special base pairing
ring-on-pole bracelet on wrist Figure=ring-shape,
Ground=pole-shape
on-vehicle man on bus Ground=public-transportation
on-region on the left side of... ground=region-designator
hang-on towel on rod figure is hangable
embedded-in pole in ground Ground is mass
embedded-in boat in water Figure is embeddable
buried-in treasure in ground Ground is terrain
enclosed-in-volume bird in cage Ground has enclosure
enclosed-in-area tree in yard Ground is area
in-2D-representation man in the photo Ground is 2D representation
in-cup cherries in bowl ground has cup
in-horiz-opening in doorway ground has opening
stem-in-cup ower in vase figure has stem, ground has cup
wrapped-in chicken in the foil ground is flexible/sheet
member-of-arrangement plate in stack ground is arrangement
in-mixture dust in air Figure/Ground=substance
in-entanglement bird in tree Ground has entanglement
tted-in hand in glove Figure/Ground=fit
in-grip pencil in hand Ground=gripper
depending on whether the given object is commonly airborne or not. We also
note that the figure is normally the smaller object while the ground functions
as a landmark. So it's normal to say The ower bed is next to the house, but
unnatural to say *The house is next to the owerbed. This is discussed in [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
See Table 1 for some mappings we make from prepositions to spatial relations.
      </p>
      <p>In order to use the object features described above to resolve the spatial
meaning of prepositions, linguistically referenced subregions must also be
considered. Spatial relations often express regions relative to an object (e.g., left side
of in The chair is on the left side of the room). The same subregion designation
can yield di erent interpretations, depending on the features of the objects.
{ external-vertical-surface: shutters on the left side of the house
{ interior-vertical-surface: picture on the left side of the room
{ region-of-horiz-surface: vase on the left side of the room
{ neighboring-area: car on the left side of the house
These regions (when present) are combined with the other constraints on spatial
relations to form the nal interpretation.</p>
      <p>Input text: A large magenta ower is
in a small vase. The vase is under an
umbrella. The umbrella is on the right
side of a table. A picture of a woman is
on the left side of a 16 foot long wall. A
brick texture is on the wall. The wall is
2 feet behind the table. A small brown
horse is in the ground. It is a foot to
the left of the table. A red chicken is in
a birdcage. The cage is to the right of
the table. A huge apple is on the wall. It
is to the left of the picture. A large rug
is under the table. A small blue chicken
is in a large ower cereal bowl. A pink
mouse is on a small chair. The chair is
5 inches to the left of the bowl. The bowl
is in front of the table. The red chicken
is facing the blue chicken. . .</p>
    </sec>
    <sec id="sec-4">
      <title>Conclusions and Ongoing and Future Work</title>
      <p>In order to represent spatial relations more robustly, much remains to be done
at the language, graphical, and application levels.</p>
      <p>
        We are augmenting the system to resolve verbs to semantic frames using
information in our SBLR, and mapping those in turn to corresponding poses
and spatial relations [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Figure 4 illustrates this process, which currently is
supported for a limited set of verbs and their arguments. This enhanced capability
also requires contextual information about actions and locations that we are
acquiring using human annotations obtained via Amazon's Mechanical Turk and
by extracting information from corpora using automatic methods [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. We will
be evaluating our software in partnership with a non-pro t after-school program
in New York City.
      </p>
      <p>Acknowledgments
This work was supported in part by the NSF IIS- 0904361. Any opinions, ndings
and conclusions or recommendations expressed in this material are the authors'
and do not necessarily re ect those of the sponsors.</p>
      <p>The truck chased the man down the road...</p>
      <p>The man ran across the sidewalk...</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Adorni</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Di</surname>
            <given-names>Manzo</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Giunchiglia</surname>
          </string-name>
          ,
          <string-name>
            <surname>F.</surname>
          </string-name>
          :
          <article-title>Natural language driven image generation</article-title>
          .
          <source>COLING</source>
          pp.
          <volume>495</volume>
          {
          <issue>500</issue>
          (
          <year>1984</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Badler</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bindiganavale</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bourne</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Palmer</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shi</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schule</surname>
            ,
            <given-names>W.:</given-names>
          </string-name>
          <article-title>A parameterized action representation for virtual human agents</article-title>
          .
          <source>Workshop on Embodied Conversational Characters</source>
          , Lake
          <string-name>
            <surname>Tahoe</surname>
          </string-name>
          (
          <year>1998</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Baker</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fillmore</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lowe</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>The Berkeley FrameNet Project</article-title>
          .
          <article-title>COLING-ACL (</article-title>
          <year>1998</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Clay</surname>
            ,
            <given-names>S.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wilhelms</surname>
          </string-name>
          , J.: Put:
          <article-title>Language-based interactive manipulation of objects</article-title>
          .
          <source>IEEE Computer Graphics and Applications</source>
          pp.
          <volume>31</volume>
          {
          <issue>39</issue>
          (
          <year>1996</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Coyne</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rambow</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hirschberg</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sproat</surname>
          </string-name>
          , R.:
          <article-title>Frame semantics in text-toscene generation</article-title>
          .
          <source>14th International Conference on Knowledge-Based and Intelligent Information &amp; Engineering Systems</source>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Coyne</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sproat</surname>
          </string-name>
          , R.:
          <article-title>WordsEye: An automatic text-to-scene conversion system</article-title>
          .
          <source>SIGGRAPH</source>
          , Computer Graphics Proceedings pp.
          <volume>487</volume>
          {
          <issue>496</issue>
          (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Dupuy</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Egges</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Legendre</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nugues</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Generating a 3d simulation of a car accident from a written description in natural language: The carsim system</article-title>
          .
          <source>Proceedings of ACL Workshop on Temporal and Spatial Information</source>
          Processing pp.
          <volume>1</volume>
          {
          <issue>8</issue>
          (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Fellbaum</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>WordNet: an electronic lexical database</article-title>
          . MIT Press (
          <year>1998</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Herskovits</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Language and Spatial Cognition: an Interdisciplinary Study of the Prepositions in English</article-title>
          . Cambridge University Press, Cambridge, England (
          <year>1986</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Larson</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shakespeare</surname>
          </string-name>
          , R.:
          <article-title>Rendering with Radiance</article-title>
          . The Morgan Kaufmann Series in Computer Graphics (
          <year>1998</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Levinson</surname>
            ,
            <given-names>S.:</given-names>
          </string-name>
          <article-title>Space in Language and Cognition: Explorations in Cognitive Diversity</article-title>
          . Cambridge University Press, Cambridge (
          <year>2003</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Ma</surname>
          </string-name>
          , M.:
          <article-title>Automatic Conversion of Natural Language to 3D Animation</article-title>
          .
          <source>Ph.D. thesis</source>
          , University of Ulster (
          <year>2006</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Simmons</surname>
            ,
            <given-names>R.:</given-names>
          </string-name>
          <article-title>The clowns microworld</article-title>
          .
          <source>Proceedings of TINLAP</source>
          pp.
          <volume>17</volume>
          {
          <issue>19</issue>
          (
          <year>1998</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Sproat</surname>
          </string-name>
          , R.:
          <article-title>Inferring the environment in a text-to-scene conversion system</article-title>
          .
          <source>First International Conference on Knowledge Capture</source>
          , Victoria,
          <string-name>
            <surname>BC</surname>
          </string-name>
          (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>