<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>LIMSI-CNRS@CLEF 2014: Invalidating Answers for Multiple Choice Question Answering</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Martin Gleize</string-name>
          <email>gleize@limsi.fr</email>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Anne-Laure Ligozat</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Brigitte Grau</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>ENSIIE</institution>
          ,
          <addr-line>Evry</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>LIMSI-CNRS</institution>
          ,
          <addr-line>Rue John von Neumann, 91405 Orsay CEDEX</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Universite Paris-Sud</institution>
          ,
          <addr-line>Orsay</addr-line>
        </aff>
      </contrib-group>
      <fpage>1386</fpage>
      <lpage>1394</lpage>
      <abstract>
        <p>This paper describes our participation to the Entrance Exams Task of CLEF 2014's Question Answering Track. The goal is to answer multiple-choice questions on short texts. Our system rst retrieves passages relevant to the question, through lexical expansion involving a structured use of the Simple English Wiktionary and WordNet. Then it extracts predicate-argument structures (PAS) from each answer choice and aligns them to PAS found in the passages retrieved in the rst step. Finally, manually crafted rules are applied to those alignments to try to invalidate answer choices. If enough answer choices are thus invalidated, we make a decision on the remaining answer choices based on their alignment scores with the passages. We submitted several runs in the task, only one of which reached the random baseline (c@1 of 0.25). In the last section, we provide an analysis of the di erences between our relatively good results obtained on trial data and the poor performance of our test run.</p>
      </abstract>
      <kwd-group>
        <kwd>Question Answering</kwd>
        <kwd>Passage Retrieval</kwd>
        <kwd>Textual Entailment</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        The task focuses on the reading of single documents and identi cation of the
correct answer to a question from a set of possible answer options. The
identi cation of the correct answer requires various kinds of inference and the
consideration of previously acquired background knowledge. Japanese University
Entrance Exams include questions formulated at various levels of complexity
and test a wide range of capabilities. The challenge of "Entrance Exams" aims
at evaluating systems under the same conditions humans are evaluated to enter
the University. Previously the evaluation campaign Question Answering For
Machine Reading Evaluation (QA4MRE at CLEF) [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] focused on multiple-choice
questions designed to evaluate computer systems, but this new task takes on
challenges typically o ered to humans. It naturally translates into more
complex inference phenomena to solve [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], and thus usually lower performance of
systems: QA4MRE 2013's best run [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] on the Main task outperformed by a
large margin its counterpart on the Entrance Exams pilot task (c@1 of 0.59 on
5-choice questions, compared to 0.42 on 4-choice questions) [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
    </sec>
    <sec id="sec-2">
      <title>System Architecture</title>
      <p>The overarching goal of our system is to essentially invalidate as many
incorrect answer choices as possible. Although we also introduce some elements of
validation of the correct answer, our study of the trial question set revealed
that we might have a better chance of nding the right answer by
elimination of the wrong candidates. The architecture of our multiple-choice
questionanswering system is described in Figure 1. Its pipeline is composed of mainly four
modules: preprocessing, passage retrieval, predicate-argument extractor and
validation/invalidation. The remaining of this section is dedicated to the detailed
description of those modules.</p>
      <p>Document
Questions
Answers</p>
      <p>Stanford
CoreNLP
TruthTeller</p>
      <p>Passage</p>
      <p>
        Retrieval
We use Stanford CoreNLP as our main Natural Language annotation tool. Each
sentence from the document, questions or answer choices is tagged with
PartOf-Speech [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] and syntactically parsed [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. In addition, we apply coreference
resolution [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] on the whole document. We did not use the Named Entity
Recognition component.
      </p>
      <p>
        We also use TruthTeller [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], a semantic annotator that assigns truth values to
predicate occurrences, on all the sentences. More details about the annotation
types are available in its presentation paper, but we generally only use the
Predicate Truth annotation, which is the nal value assigned by TruthTeller. It is
one of P (Positive), U (Uncertain) or N (Negative), and indicates whether the
predicate itself is entailed by its containing sentence, in the classical sense of
textual entailment.
2.2
      </p>
      <sec id="sec-2-1">
        <title>Passage Retrieval</title>
        <p>
          The passage retrieval module aims at ranking document snippets of 3 to 5
sentences by relevance to the question. Words of the question act as the query.
However, it is very rare that words of the question exactly appear in the
relevant passage of the document, so we have to use some form of query expansion.
We use a method similar to [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ]: we basically expand each query and document
word with the words of their de nition in the Simple English Wiktionary [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ].
We can do that recursively, expanding de nition words with their own de nition,
thus de ning a kind of word tree. We compute for each (query word, document
word) pair a weighted number of common words in their tree. We weigh the
words according to their depth in the de nition tree.
        </p>
        <p>Figure 2 shows a portion of the de nition trees for the words cat and wolf. The
word animal is found in the Simple English Wiktionary de nition of both words:
we say that it is found at depth 1 {depth 0 being the word itself. The word pet
on the other hand is found in the de nition of dog, a word of the de nition of
wolf : we say that it is of depth 2. We add the depth of common words in both
trees: animal is a common word of depth 2 and pet is a common word of depth 3
for the pair (cat, wolf ). This roughly captures the intuition that the word animal
is more accurate to describe common traits of a cat and a wolf rather than the
word pet.</p>
        <p>We enrich the de nition trees with coreference information and WordNet
synonyms and antonyms {to complete the coverage of the Simple Wiktionary,
and weigh all the words in the tree by the IDF score of the word in the document.
We ran TruthTeller on all the Simple English Wiktionary de nitions, so we can
propagate predicate truth values in the de nition tree to compute whether a
word in the tree is entailed by the root word. The propagation rules are simple:
if a word is of truthvalue P or U, each of its direct de nition words keeps the
truth value assigned by TruthTeller, if it is N, we take the inverse TruthTeller
annotation for the de nition word (inverse of P is N, inverse of N is P, and
inverse of U is U). We do not use predicate truth information for scoring, it will
be used in the validation/invalidation step.</p>
        <p>With the resulting word-to-word semantic relatedness scores, we compute a
1-to1 maximum sum alignment of the words using the Integer Linear Programming
solver lpsolve. We then rank the aligned passages by alignment scores. This
makes up our basic ranking system and it is also used in the predicate-argument
alignment described later in the paper.</p>
        <p>One interesting property of our data is that in general, the order of the passages
in the text preserve the order of the questions they're relevant to: if question 2
is after question 1 in the reading test, then its relevant passage in the document
is most likely after that of question 1. We use a simple dynamic programming
algorithm {omitted here for space{ to compute the best sequence of passages
conserving this property, and place the passages computed this way at the top
of their respective per-question list.
2.3</p>
      </sec>
      <sec id="sec-2-2">
        <title>Predicate-arguments Aligner</title>
        <p>Predicate-arguments Extractor Predicate-arguments structures (PAS) aim
at capturing essentially \Who does what to what?" in a shallow way. We use
the Stanford dependency graph from the parsed sentences and take the verbs
(Penn POS tag starting with V) as predicates and we classify their
dependencies between subject and arguments. We take as subject the nodes in a nsubj
or nsubjpass dependency with the verb, and consider the rest as arguments. We
enrich subjects and arguments with their own dependencies to not lose out on
adjectives, certain numerical determiners and prepositions. When considering
dependencies, we lter out determiners (with the DT tag).</p>
        <p>Alignment of answer PAS and passage PAS With the previously described
method, we extract PAS for the sentences of each answer choice. We also
extract PAS for the sentences in the relevant passages obtained in section 2.2. We
align answer PAS and passage PAS with the same alignment method using our
semantic relatedness method and the ILP formulation, and rank the alignments
by alignment score. Let us note that we align regardless of the function the word
plays in the PAS: we can align predicates with subjects and arguments, and
subjects with arguments.</p>
        <p>At this point, we summarize the information available to us in those alignments.
For each aligned word pair, we know:
{ the function that each word plays in its respective containing PAS: subject,
predicate or argument.
{ a semantic relatedness score.</p>
        <p>{ truth value annotations, as annotated by TruthTeller.
2.4</p>
      </sec>
      <sec id="sec-2-3">
        <title>Validation/Invalidation</title>
        <p>Algorithm Our goal in this last module is to eliminate as many answer choices
as possible without eliminating the right answer. Let K be the maximum number
of answer choices we are allowed to keep to take the nal decision. The following
algorithm computes whether we take a nal decision for the current question.
ALL_ANSWERS := all answer choices
AnswersChoices := all answers
Passages := all relevant passages
While (|AnswersChoices| &gt; K &amp;&amp; |Passages| &gt; 0) {</p>
        <p>Passage = Passages.pop()
Validated = {}
Invalidated = {}
Foreach (AnswerChoice in AnswerChoices) {</p>
        <p>PASAlignments = align(AnswerChoice, Passage)
Foreach (PASAlignment in PASAlignments) {
if (validate(PASAlignment))</p>
        <p>Validated.add(AnswerChoice)
if (invalidate(PASAlignment))</p>
        <p>Invalidated.add(AnswerChoice)
}
}
ToRemove := {}
if (|Invalidated| &lt; |ALL_ANSWERS|) {</p>
        <p>ToRemove := Invalidated
if (|Validated| &gt; 0) {</p>
        <p>ToRemove := ToRemove U (ALL_ANSWERS \ Validated)</p>
        <p>AnswersChoices := AnswersChoices \ ToRemove
}
return |AnswersChoices| &lt;= K
We provide a rough explanation of what goes on in this pre-decision process.
We explore all the ranked relevant passages as long as we have not eliminated
enough answers. When faced with a new passage, we PAS-align each answer
choice with the passage (section 2.3). The ranked alignments go through
validation and invalidation rules and the corresponding answer is invalidated if a PAS
alignment is found invalid.</p>
        <p>Final decision If we go through all the passages without invalidating enough
answers, we choose not to answer the question. If we have K or less answers
remaining, we pick the one with the strongest PAS alignment score found in the
passages.</p>
        <p>The rules We manually built 2 validation rules and 3 invalidation rules. They
operate on PAS alignments.</p>
        <p>In the current iteration of our system, the validation rules must be red
simultaneously to validate an answer choice, whereas only one invalidation rule red
is enough to invalidate an answer choice. We consider that it is indeed usually</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Data and results</title>
      <p>much more di cult, even for a human, to invalidate a wrong answer than it is
to validate a correct one with accuracy.</p>
      <p>In the description of the rules, polarity means the predicate truth value of the
common word found in the semantic relatedness score. Compatible polarities are
P with P, N with N, and U with P,N,U.</p>
      <p>Strong alignment means that the word-to-word alignment score is above a
threshold, manually set in this system. The validation rules are as follows:
{ Rule 1: Subject and Predicate are strongly aligned in both PAS, and all
polarities are compatible.
{ Rule 2: Predicate and one Argument are strongly aligned in both PAS, and
all polarities are compatible.</p>
      <p>The invalidation rules are as follows:
{ Rule 1: One polarity mismatch is found in a strong alignment.
{ Rule 2: Predicates are strongly aligned, but their Subjects are not aligned
at all.
{ Rule 3: The alignment is located at least 2 sentences before the best
(question, passage) alignment. We noticed on the trial corpus that the correct
answer was usually found after the mention of the question in the
document.</p>
      <sec id="sec-3-1">
        <title>CLEF 2014 QA Track: Entrance Exams data and evaluation</title>
        <p>Our data consist of the trial and test sets at CLEF 2014 Question Answering
Track, Task 3: Entrance Exams. Both trial and test sets feature the same format:
a series of 12 texts, and for each of them, 5 multiple-choice questions to answer: 60
questions in total. There are 4 answer choices possible for each of the questions.
This corpus has been extracted from the Tokyo University Entrance Exam in
English as a foreign language.</p>
        <p>Systems are evaluated according to their c@1, de ned in equation 1.</p>
        <p>C@1 = n1 (nR + nU nnR ) (1)
with n the total number of questions, nR the number of correctly answered
questions, nU the number of unanswered questions.
3.2</p>
      </sec>
      <sec id="sec-3-2">
        <title>Results</title>
        <p>In this section, we report the results of our best run on both trial and test
question sets. This system uses K = 2 as the maximum number of answer choices
allowed to make the decision. Passages are 5 sentences long and the algorithm
described in section 2.4 uses a maximum of 10 passages before reaching
nondecision. We tried other values of those parameters, but the corresponding runs
were performing worse on both trial and data sets, so we do not feel the need to
further expose results about them. We focus instead on the di erences on trial
and test results for the described run. Both datasets contain 60 questions. Table
1 reports global results of our system on both sets of questions. As we can see,
performance is quite satisfactory on the trial dataset, but really poor {at the
level of the random baseline{ on the test dataset.</p>
        <p>We also analyzed in a ne-grained way the accuracy and e ciency of our
validation and invalidation rules. Table 2 describes the error rate on questions
where invalidation rules ended up eliminating a correct answer choice. In each
cell, we nd the number of questions where the particular rule eliminated a
correct answer, and we also nd the number of questions it red on {where
it eliminated at least one answer choice. As we can see, all rules mis re more
on the test dataset, which seems to correlate well with the overall poor test
performance. Validation rules did not re as often as the invalidation rules (4
times for trial, and 5 times for test).</p>
        <p>It is possible to frame our rule system as a kind of information retrieval
system and evaluate it in term of validation precision and recall, and invalidation
precision and recall. For validation, relevant items are the correct answers, and
for invalidation, relevant items are the incorrect answer choices. We compare
our results in both trial and test datasets with a random baseline, which
basically has a 50% chance of validating and invalidating each answer choice and
stops under the same conditions as our system (when K or less answer choices
remain). Results in term of precision and recall are shown in table 3. As we can
see, surprisingly, validation/invalidation does not seem to perform signi cantly
di erently from random guesses on test data, while it is not the case at all on
trial data.</p>
        <p>We nally report the performance of our passage retrieval system compared
to several baselines. Prior to this participation, we had annotated the correct
passages on 45 of the 60 questions of the trial dataset, this allows us to compute
the MRR (Mean Reciprocal Rank) of our passage retrieval method. Results are
shown in table 4. Baseline 1 is a simple word-overlap-based measure. Baseline
2 is Baseline 2 with TF-IDF weighting and WordNet keyword expansion. We
then ran our system with and without taking into account the order preserving
property of the reading tests, mentioned at the end of section 2.2. As we can
see, our passage retrieval method vastly outperforms typical baselines, and the
assumption about order of questions being preserved in the order of passages
seems to hold and help quite a bit, at least on trial questions.
Our system has been developed with the invalidation of wrong answer
candidates in mind, speci cally to answer multiple-choice questions. In the CLEF
2014 evaluation campaign, Question Answering track, the submitted run
performed at the level of the random baseline in the Entrance exams task and thus
did not reproduce satisfactory trial performance.</p>
        <p>In further works, we plan to study the reasons behind the di erences in
performance of our system on trial and test results and to better integrate
predicatearguments structure selection with passage retrieval, the module of our system
which seemed to perform well. Other features considered include a character
name resolver addition to coreference resolution, and analyzing discourse
relations.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Banerjee</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bhaskar</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pakray</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bandyopadhyay</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gelbukh</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Multiple choice question (mcq) answering system for entrance examination</article-title>
          .
          <source>CLEF 2013 Working Notes</source>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Gleize</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grau</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>A hierarchical taxonomy for classifying hardness of inference tasks</article-title>
          . In: Chair),
          <string-name>
            <given-names>N.C.C.</given-names>
            ,
            <surname>Choukri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            ,
            <surname>Declerck</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            ,
            <surname>Loftsson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            ,
            <surname>Maegaard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            ,
            <surname>Mariani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Moreno</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Odijk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Piperidis</surname>
          </string-name>
          , S. (eds.)
          <source>Proceedings of the Ninth International Conference on Language Resources and Evaluation (LREC'14)</source>
          .
          <source>European Language Resources Association (ELRA)</source>
          , Reykjavik, Iceland (may
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Gleize</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grau</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ligozat</surname>
            ,
            <given-names>A.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pho</surname>
            ,
            <given-names>V.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Illouz</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Giannetti</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lahondes</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Selecting answers with structured lexical expansion and discourse relations</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Klein</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manning</surname>
          </string-name>
          , C.D.:
          <article-title>Accurate unlexicalized parsing</article-title>
          .
          <source>In: Proceedings of the 41st Annual Meeting on Association for Computational Linguistics-Volume</source>
          <volume>1</volume>
          . pp.
          <volume>423</volume>
          {
          <fpage>430</fpage>
          .
          <article-title>Association for Computational Linguistics (</article-title>
          <year>2003</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Lotan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stern</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dagan</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Truthteller: Annotating predicate truth</article-title>
          .
          <source>In: Proceedings of NAACL-HLT</source>
          . pp.
          <volume>752</volume>
          {
          <issue>757</issue>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6. Pen~as,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Miyao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            ,
            <surname>Hovy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            ,
            <surname>Forner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Kando</surname>
          </string-name>
          , N.:
          <article-title>Overview of qa4mre 2013 entrance exams task</article-title>
          .
          <source>Working Notes</source>
          , CLEF (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Recasens</surname>
          </string-name>
          , M., de Marne e, M.C.,
          <string-name>
            <surname>Potts</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>The life and death of discourse entities: Identifying singleton mentions</article-title>
          .
          <source>In: Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          . pp.
          <volume>627</volume>
          {
          <issue>633</issue>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Sutcli</surname>
            <given-names>e</given-names>
          </string-name>
          , R., Pen~as,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Hovy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            ,
            <surname>Forner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Rodrigo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Forascu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            ,
            <surname>Benajiba</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            ,
            <surname>Osenova</surname>
          </string-name>
          ,
          <string-name>
            <surname>P.</surname>
          </string-name>
          :
          <article-title>Overview of qa4mre main task at clef 2013</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Toutanova</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Klein</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manning</surname>
            ,
            <given-names>C.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Singer</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>Feature-rich part-of-speech tagging with a cyclic dependency network</article-title>
          .
          <source>In: Proceedings of the 2003 Conference of the North American Chapter of the Association for Computational Linguistics on Human Language Technology-Volume 1</source>
          . pp.
          <volume>173</volume>
          {
          <fpage>180</fpage>
          .
          <article-title>Association for Computational Linguistics (</article-title>
          <year>2003</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Wikimedia</surname>
          </string-name>
          <article-title>Foundation: The simple english wiktionary</article-title>
          , http://http://simple. wiktionary.org/
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>