<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Approximate subgraph matching for detection of topic variations</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Mitja Trampuš</string-name>
          <email>mitja.trampus@ijs.si</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Dunja Mladenic´</string-name>
          <email>dunja.mladenic@ijs.si</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Jozef Stefan Institute</institution>
          ,
          <addr-line>Jamova 39, Ljubljana</addr-line>
          ,
          <country country="SI">Slovenia</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2011</year>
      </pub-date>
      <abstract>
        <p>The paper presents an approach to detection of topic variations based on approximate graph matching. Text items are represented as semantic graphs and approximately matched based on a taxonomy of node and edge labels. Best-matching subgraphs are used as a template against which to align and compare the articles. The proposed approach is applied on news stories using WordNet as the prede ned taxonomy. Illustrative experiments on real-world data show that the approach is promising.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. INTRODUCTION</title>
      <p>One of the classic goals of text mining is to structure
natural language text { for obvious reasons: the amount of
information we can extract from the data using shallow
approaches like bag-of-words is limited. By enhancing text
with structure, we can start to observe information that is
encoded in more than one word or sentence. Also,
structure enables us to bring the additional power of semantic
methods and background knowledge into the play.</p>
      <p>While reasonably reliable methods have been developed
for structuring text by annotating and identifying a speci c
subset of information, mostly named entities, little work has
been done on semantically capturing the macro-level aspects
of the text. In this article, we present some early work on
constructing domain templates, a generic \summary" that
ts many pieces of text on a speci c theme (e.g. news stories
about bombings) at the same time.</p>
      <p>The genericness of the template provides for data
exploration in two ways:
1. By automatically mapping speci c facts and entities
in an article to the more general ones in a template,
2.</p>
      <p>
        Because it aligns articles to a common template, our method
has much in common with other information extraction
mechanisms. Automatic construction of information extraction
templates is already relatively well-researched. Most
methods aim for attribute extraction, where the goal is to extract
a single prede ned type of information, e.g. the title of a
book. Each separate type of information requires a separate
classi er and training data. Examples of such approaches
are [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ].
      </p>
      <p>
        More recently, a generalized problem of relation extraction
has received considerable attention. The goal is to nd pairs
of items related by a prede ned relation. As an example,
Probst et al. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] mine product descriptions to
simultaneously identify product attributes and their values. Relation
extraction is particularly popular in biomedicine where pairs
of proteins in a certain relation (e.g. one inhibits the other)
are often of interest.
      </p>
      <p>
        The task in this article is more generalized still; we
attempt to decide both what information is interesting to
extract as well as perform the extraction. This is known as
domain template extraction. To our knowledge, little work has
been done in the area so far. The most closely related work
is by Filatova et al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], who nd templates by mining
frequent parse subtrees. Also closely related is work by Li et al.
[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]; similarly to Filatova, they mine frequent parse subtrees
but then cluster them into \aspects" with a novel graphical
model. Both approaches produce syntax-level patterns.
Unlike ours, neither of the two approaches exploits background
knowledge. Also belonging to this group is our previous
work [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] which mostly shares the goal and data
representation ideas with this article, but uses di erent methods apart
from preprocessing.
      </p>
      <p>
        Graph-based templates are also used in [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] in a context
similar to ours, though the semantics are shallower. Also,
the authors focus on information extraction and do not
attempt to generalize the templates.
      </p>
      <p>Templates somewhat similar to those we aim to construct
automatically and with no knowledge of the domain have
already been created manually by domain experts. FrameNet
[?] is a collection of templates for the events like "disclosing
a secret", "speaking", "killing", "arresting" etc. They focus
mostly on low-level events, of which typically many can be
found in a single document, be it a news article or not. The
project does not concern itself with the creation of the
templates, other than from the methodological point of view.
There is little support for automatic annotation of natural
language with the FrameNet frames.</p>
    </sec>
    <sec id="sec-2">
      <title>METHOD OVERVIEW</title>
      <p>This section describes the various stages in our data
processing pipeline. The assumed input data is, as discussed
above, a collection of text items on the same topic. The
goal is to identify a pattern which semantically matches a
substantial number of the input texts.</p>
      <p>The key idea is rather simple: we rst represent our input
data as semantic graphs, i.e. graphs of ontology-aligned
entities and relations. A pattern is then de ned as a (smaller)
graph such that, by specializing some of its entities, a
subgraph of at least input graphs ( being a parameter). We
seek to identify all such patterns.</p>
      <p>We proceed to describe our approach to the construction
of semantic graphs and to the mining of approximate
subgraphs.
3.1</p>
    </sec>
    <sec id="sec-3">
      <title>Data Preprocessing</title>
      <p>
        Starting with plain text, we rst annotate it with some
basic semantic and linguistic information. Using the
ANNIE tool from the GATE framework, we rst detect named
entities and tag them as person, location or organization.
Following that, we use the Stanford parser [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] to extract
subject-verb-object triplets. We then use the web service
by Rusu [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] to perform coreference and pronoun resolutions
("Mr. Obama", "President Barack Obama" and "he" might
all refer to the same entity within an article).
      </p>
      <p>We acknowledge that the triplets acquired in this way do
not necessarily provide a proper semantic description of the
article data. The discrepancies go both ways:</p>
      <p>We include some triplets which do not make sense
semantically, e.g. \people..kill..Monday" coming from
\93 people were killed on Monday".</p>
      <p>We fail to create triplets for information not encoded
with (lexicogrammatically) transitive verbs. For
example, "President's visit to China ..." will not spawn
\president..visit..China". In our experiments, this
shortcoming is alleviated by using redundant
information - each story, e.g. president's visit to China, is
described by several articles which increases the
probability that at least one will convey this information in a
form we can detect. However, the problem is not
completely overcome this way { some information e.g. the
\93" in \93 people were killed on Monday" will never
appear as the object of a transitive verb.</p>
      <p>As a last step, we align all triplets to WordNet; that
is, for each subject, verb and object appearing in any of
the triplets, we try to nd the corresponding concept (or
"synset", as they are called) in WordNet. We rst remove
in ection from the words using python NLTK (Natural
Language Toolkit), then align it to the corresponding synset. If
more than one synset matches, we choose the most common
sense which is a well-tried and surprisingly good strategy.
For words not found in WordNet, we create a new synset
on the y. If the new word (e.g. \Obama") was previously
tagged by ANNIE (with e.g. \person"), the new synset's
hypernym is set accordingly.
3.2</p>
    </sec>
    <sec id="sec-4">
      <title>Semantic Graph Construction</title>
      <p>From a collection of triplets, we proceed to construct the
semantic graph. Here, we rely rather heavily on the fact
that news articles tend to be focused in scope: we do not
disambiguate entities other than by name (not necessarily
a proper name; e.g. \book" is also a name). As an
example, if an article mentions two buildings, one of which burns
down and the second of which has a green roof, our method
detects a single \building" and assigns both properties to it.
In the newswire domain, we have not found this to be a
signi cant issue: entities which do need to be disambiguated
are presented with more unique names (\France" instead of
\country" etc.). This rationale would have to be revised if
one wanted to apply the approach to longer texts.</p>
      <p>This assumption greatly simpli es the construction of the
semantic graph: we start by treating each triplet as a
2node component of a single very fragmented graph and then
collapse the nodes with the same labels.</p>
      <p>Dataset specifics.</p>
      <p>In our experiments, each input \document" in the sense
described here was in fact a concatenation of actual
documents, all of which were reporting on the exact same news
event. Section 4 contains the details and rationale.</p>
    </sec>
    <sec id="sec-5">
      <title>Approximate Pattern Detection</title>
      <p>Given a collection of labeled graphs, we now wish to
identify frequent \approximate subgraphs", i.e. patterns as
described at the beginning of Section 3.</p>
      <p>Formal task de nition: Given a set of labeled graphs
S = fG1; : : : ; Gng, a transitive antisymmetric relation on
graph labels genl( ; ) (with genl(A; B) interpreted as \label
A is a generalization of label B") and a number , we wish
to construct all maximal graphs H that are approximate
subgraphs of at least graphs from S. A graph H is said to
be an approximate subgraph of G i there is a mapping f of
V (H) onto a subset of V (G) such that genl(v; f (v)) holds
for all v 2 V (H).</p>
      <p>
        This is not an easy task. Mining frequent subgraphs is in
itself computationally demanding because of isomorphisms;
satisfactorily fast algorithms for this seemingly basic
problem are relatively recent [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. By further requiring that the
frequent subgraph only match the input graphs in a soft
way implied by a taxonomy (here WordNet hyperymy), the
complexity becomes insurmountable. We compensate by
introducing two assumptions.
      </p>
      <p>1. The hierarchy imposed by genl has a tree-like form, it
is not a general DAG. This is true of WordNet: every
synset has at most one hypernym de ned.
2. Very generic patterns are not interesting and can (or
even should) be skipped. This too is a safe
assumption in our scenario: a pattern in which every node is
labeled with the most generic label entity is hardly
informative regardless of its graph structure.</p>
      <p>We can now employ a simple but e ective three-stage
search. The stages are illustrated in 1 with the minimal
example of two two-node graphs.</p>
      <p>1. Generalize all the labels of input graphs to the
maximum extent permissible. Under the rst assumption,
\generalizing a label" is a well-de ned operation. The
exact meaning of \maximum extent permissible" is
governed by the second assumption; no label should be
generalized so much as to fall in the uninteresting
category. In our experience with WordNet, the following
simple rule worked very well: generalize verbs as much
as possible and generalize nouns to two levels below
the hierarchy root. See steps 1 to 2 in Fig. 1.
2. Mine -frequent maximal subgraphs with support of
the generalized input graphs. This step cannot be
shown in Fig. 1 as the graphs are too small.
3. Formally, the resulting subgraphs already satisfy our
demands. However, to make them as descriptive as
possible, we try to specialize the pattern's labels,
taking care not to perform a specialization that would
reduce the pattern's support below . Specialization,
unlike generalization, is not a uniquely de ned
operation (a synset can have many hyponyms), but with
some we can a ord to recursively explore the whole
space of possible specializations. We use the sum of
labels' depth in the WordNet hierarchy as a measure
of pattern descriptiveness that we optimize. See steps
2 to 3 in Fig. 1.</p>
      <p>
        For frequent subgraph mining, we developed our own
algorithm, inspired by the current state-of-art[
        <xref ref-type="bibr" rid="ref11 ref3">11, 3</xref>
        ]. We
included some improvements pertaining to maximality of
out2)
3)
 
1) assasin-blow_up→president
robber-murder→officer
person-kill→person
      </p>
      <p>person-kill→person
criminal-kill→person
put graphs and to scalability { all existing open-source
software crashed on our full input data.
4.</p>
    </sec>
    <sec id="sec-6">
      <title>PRELIMINARY EXPERIMENTS</title>
    </sec>
    <sec id="sec-7">
      <title>AND RESULTS</title>
      <p>As a preliminary, let us de ne some terminology suitable
for our experiment domain. An article is a single web page
which is assumed to report on a single story. A story is
an event that is covered by one or more articles. Each story
may t some domain template (also event template or simply
template) describing a certain type of event.</p>
      <p>We obtained a month's worth of articles from Google
News by crawling. Each article was cleaned of all HTML
markup, advertisements, navigation and similar. Articles
were grouped into stories according to Google News.</p>
      <p>For each story, a semantic graph was constructed. The
reason to use an aggregate story graph rather than
individual article graphs was twofold. First, by representing each
story as a single graph, all stories were represented
equivalently (as opposed to the case where each article contributed
a graph, resulting in stories being weighted proportionally
to the number of their articles). Second, the method for
extracting triplets has relatively low precision and recall; it
therefore makes sense to employ the redundancy inherent
in the collection of articles reporting on the same event. To
construct the aggregate story graph, we simply concatenated
the plain text of individual articles; aggregation at this early
stage has the added bene t of providing cross-article entity
resolution. Finally, the collection of semantic graphs from
stories on a single topic was input to the pattern mining
algorithm.</p>
      <p>We de ned ve topics on which to observe the behavior
of the method: bomb attacks, award ceremonies, worker
layo s, political visits and court sentencings. For each, we
identi ed about 10 stories of interest. Note that each story
further comprises about 100 articles, clustering courtesy of
Google News; in total, about 5000 articles were therefore
processed.</p>
      <p>As semantic graphs were constructed on the level of stories
rather than articles, their structure was relatively rich. They
had about 1000 nodes each and an average node degree of
roughly 2.5. The 20% most connected nodes, which are also
the ones likely to appear in the patterns, had an average
degree of about 20.</p>
      <p>For each topic, graphs of all its stories were input to the
algorithm. The minimal pattern support was set at 30%
for all the topics. The algorithm output several patterns for
each topic; the sizes of the outputs along with the interesting
patterns are presented in Figure 2.</p>
      <p>For instance, the last person in the \visits" domain shows
that in at least 30% of the stories, there was a male person
(\he", e.g. Obama) who traveled to France (a coincidence),
and that same person met a \leader" (a president in some of
the stories, a minister in other).</p>
      <p>Bombing attacks (8 patterns in total)
weekday ←kill- person -kill→ attack -take→ place
himself ←have- suicide bomber -explode→ device
himself ←have- suicide bomber -blow→ building
Court sentencings (7 patterns in total)
correctional institution ←be- person -face→ year ←be- sentence
innocent ←be- person -face→ year ←be- sentence
Award ceremonies (2 patterns in total)
period of time ←have- person -be→ feeling
Political visits (3 patterns in total)
summit ←attend- he -- hold→ talk
||`-be→ leader
|`--tell→ communicator
`---express → feeling
need ←stress - he - hold→ talk
|`-attend → summit
`--be→ leader
leader ←meet- he -travel→ France</p>
      <p>Worker layoffs (0 patterns in total)</p>
    </sec>
    <sec id="sec-8">
      <title>DISCUSSION AND FUTURE WORK</title>
      <p>The preliminary results seem sound. The mappings of
individual stories onto the patterns (not given here) also
provide a semantically correct alignment. We can observe
how each story ts the template with slightly di erent
entities. Sometimes, the variations are almost imperceptible {
\correctional facility" from the \court" domain, for example,
appears as either \jail" or \prison", which for some reason
are two distinct concepts in WordNet.</p>
      <p>In other cases, the distinctions are signi cant and express
the subtopical diversity we were looking for. For example,
the groundings for \leader" in the \visits" domain varied even
in our small dataset over president, minister, instigator or
simply leader. In the same domain, \feeling" was either
sorrow, disappointment or satisfaction. The \building" in the
\bombings" domain was generalized from mosque,
restaurant, hotel and building. It might be interesting to
investigate this further and use the amount of variation between
pattern groundings as a measure of pattern interestingness.</p>
      <p>Unexpectedly, diversity can occasionally be found in the
natural clustering that the patterns provide. Observe the
two patterns in the \court" domain: in both, the defendant
is facing a sentence of (one or more) years, but is found
innocent in one cluster and sent to(?) the jail in the other.</p>
      <p>While the current experiments are too small to draw any
conclusive evidence, we can make some speculations about
precision and recall. While the rst is low but usable (a
data analyst should not mind going through e.g. 5 patterns
to identify a useful one), the latter seems a bigger issue. We
hope to improve the results signi cantly by developing a
better triplet extractor1; the previously discussed de ciencies of
current triplets appear to hit performance most.</p>
      <p>The tests also indicate that the method is not equally
suitable for all domains. The \layo s" domain, for example,
had no single pattern which would occur in 30% of the
stories. (A threshold of 25% produces a single rather
nonsensical pattern \it|cut !job lose|people"). The \awards;;
domain does not fare much better. Most probably, these
two topics are too broad, causing stories to have only little
overlap.
1But this is a new project in itself.</p>
      <p>Note that in current implementation, all nal patterns
with less than three nodes (e.g. worker..lose..job for
the \layo s" topic) were discarded. Partly this is because
we are, in perspective, interested in (dis)proving that
structured patterns can provide more information than
sentencelevel patterns found in related work2. Partly, however, it is
also because including two-node patterns would introduce
additional noise in the output. Even now, the precision is
relatively low; it would therefore be interesting to
investigate measures of interestingness of patterns other than raw
frequency.
6.</p>
    </sec>
    <sec id="sec-9">
      <title>ACKNOWLEDGMENTS</title>
      <p>This work was supported by the Slovenian Research Agency
and the IST Programme of the EC under PASCAL2
(ISTNoE-216886), ACTIVE (IST-2008-215040) and RENDER
(FP7-257790).
7.
2The \visits" domain is a nice indication that this may be
true.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Arasu</surname>
          </string-name>
          and
          <string-name>
            <given-names>H.</given-names>
            <surname>Garcia-Molina</surname>
          </string-name>
          .
          <article-title>Extracting structured data from web pages</article-title>
          .
          <source>pages</source>
          <volume>337</volume>
          {
          <fpage>348</fpage>
          ,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>S.</given-names>
            <surname>Brin</surname>
          </string-name>
          .
          <article-title>Extracting patterns and relations from the world wide web</article-title>
          .
          <source>Lecture Notes in Computer Science</source>
          , pages
          <volume>172</volume>
          {
          <fpage>183</fpage>
          ,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Nijssen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Muntz</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Kok</surname>
          </string-name>
          .
          <article-title>Frequent subtree mining-an overview</article-title>
          .
          <source>Fundamenta Informaticae</source>
          ,
          <volume>66</volume>
          (
          <issue>1</issue>
          ):
          <volume>161</volume>
          {
          <fpage>198</fpage>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>E.</given-names>
            <surname>Filatova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Hatzivassiloglou</surname>
          </string-name>
          , and
          <string-name>
            <given-names>K.</given-names>
            <surname>McKeown</surname>
          </string-name>
          .
          <article-title>Automatic creation of domain templates</article-title>
          .
          <source>In Proceedings of COLING/ACL</source>
          <year>2006</year>
          , pages
          <fpage>207</fpage>
          {
          <fpage>214</fpage>
          ,
          <string-name>
            <surname>Morristown</surname>
          </string-name>
          , NJ, USA,
          <year>2006</year>
          .
          <article-title>Association for Computational Linguistics</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>D.</given-names>
            <surname>Klein</surname>
          </string-name>
          and
          <string-name>
            <given-names>C.</given-names>
            <surname>Manning</surname>
          </string-name>
          .
          <article-title>Accurate unlexicalized parsing</article-title>
          .
          <source>In Proceedings of the 41st Annual Meeting on Association for Computational Linguistics</source>
          , pages
          <volume>423</volume>
          {
          <fpage>430</fpage>
          . Association for Computational Linguistics,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>P.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Jiang</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          .
          <article-title>Generating templates of entity summaries with an entity-aspect model and pattern mining</article-title>
          .
          <source>Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics</source>
          , pages
          <volume>640</volume>
          {
          <fpage>649</fpage>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>K.</given-names>
            <surname>Probst</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ghani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Krema</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Fano</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          .
          <article-title>Semi-supervised learning of attribute-value pairs from product descriptions</article-title>
          .
          <source>pages 2838{2843</source>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>D.</given-names>
            <surname>Rusu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Fortuna</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Grobelnik</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D.</given-names>
            <surname>Mladenic</surname>
          </string-name>
          .
          <article-title>Semantic Graphs Derived From Triplets With Application In Document Summarization</article-title>
          .
          <source>Informatica Journal</source>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>H.</given-names>
            <surname>Tanev</surname>
          </string-name>
          and
          <string-name>
            <given-names>B.</given-names>
            <surname>Magnini</surname>
          </string-name>
          .
          <article-title>Weakly supervised approaches for ontology population</article-title>
          .
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>M.</given-names>
            <surname>Trampus</surname>
          </string-name>
          and
          <string-name>
            <given-names>D.</given-names>
            <surname>Mladenic</surname>
          </string-name>
          .
          <article-title>Learning event templated from news articles</article-title>
          .
          <source>In Proceedings of SiKDD09</source>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>X.</given-names>
            <surname>Yan</surname>
          </string-name>
          and J. Han. gSpan:
          <article-title>Graph-based substructure pattern mining</article-title>
          .
          <source>page 721</source>
          ,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>