<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Identifying Problems and Solutions in Scienti c Text</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Cambridge Computer Laboratory</institution>
          ,
          <addr-line>15 JJ Thomson Avenue, Cambridge CB3 0FD</addr-line>
          ,
          <country country="UK">UK</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Research is often described as a problem-solving activity, and as a result, descriptions of problems and solutions are an essential part of the scienti c discourse used to describe research activity. We present an automatic classi er that, given a phrase that may or may not be a description of a scienti c problem or a solution, makes a binary decision about problemhood and solutionhood of that phrase. We recast the problem as a supervised machine learning problem, and de ne a set of 8 features correlated with the target categories, and use several machine learning algorithms on this task. We also create our own corpus of 2000 positive and negative examples of problems and solutions. We nd that we can distinguish problems from non-problems with an accuracy of 82.4%, and solutions from non-solutions with an accuracy of 81.5%. Our two most helpful features for the task are syntactic information (POS tags) and document and word embeddings.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Kevin He ernan and Simone Teufel
Problem solving is generally regarded as the most important cognitive activity in
everyday and professional contexts [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Many studies on formalising the cognitive
process behind problem-solving exist, for instance [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] argues that we all share
knowledge of the thought/action problem-solution process involved in real life,
and so our writings will often re ect this order. There is general agreement
amongst theorists that state that the nature of the research process can be
viewed as a problem-solving activity [4{7].
      </p>
      <p>
        One of the best-documented problem-solving patterns was established by
Winter [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Winter analysed thousands of examples of technical texts, and noted
that these texts can largely be described in terms of a four-part pattern
consisting of Situation, Problem, Solution and Evaluation. This is very similar to
the pattern described by [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], which consists of Introduction-Theory,
ProblemExperiment-Comment and Conclusion. The di erence is that in Winter's view,
a solution only becomes a solution after it has been evaluated positively. Hoey
changes Winter's pattern by introducing the concept of Response in place of
Solution [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. This seems to describe the situation in science better, where
evaluation is mandatory for research solutions to be accepted by the community. In
Hoey's pattern, the Situation (which is generally treated as optional) provides
background information; the Problem describes an issue which requires
attention; the Response provides a way to deal with the issue, and the Evaluation
assesses how e ective the response is.
      </p>
      <p>An example of this pattern in the context of the Goldilocks story can be seen
in Figure 1. In this text, there is a preamble providing the setting of the story (i.e.
Goldilocks is lost in the woods), which is called the Situation in Hoey's system.
A Problem in encountered when Goldilocks becomes hungry. Her rst Response
is to try the porridge in big bear's bowl, but she gives this a negative Evaluation
(\too hot!") and so the pattern returns to the Problem. This continues in a cyclic
fashion until the Problem is nally resolved by Goldilocks giving a particular
Response a positive Evaluation of baby bear's porridge (\it's just right").</p>
      <p>It would be attractive to detect problem and solution statements
automatically in text. This holds true both from a theoretical and a practical viewpoint.
Theoretically, we know that sentiment detection is related to problem-solving
activity, because of the perception that \bad" situations are transformed into
\better" ones via problem-solving. The exact mechanism of how this can be
detected would advance the state of the art in text understanding. In terms of
linguistic realisation, problem and solution statements come in many variants
and reformulations, often in the form of positive or negated statements about
the conditions, results and causes of problem{solution pairs. Detecting and
interpreting those would give us a reasonably objective manner to test a system's
understanding capacity. Practically, being able to detect any mention of a
problem is a rst step towards detecting a paper's speci c research goal. Being able
to do this has been a goal for scienti c information retrieval for some time, and
if successful, it would improve the e ectiveness of scienti c search immensely.
Detecting problem and solution statements of papers would also enable us to
compare similar papers and eventually even lead to automatic generation of
review articles in a eld.</p>
      <p>
        There has been some computational e ort on the task of identifying
problemsolving patterns in text. However, most of the prior work has not gone beyond
the usage of keyword analysis and some simple contextual examination of the
pattern. [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] presents a corpus-based analysis of lexio-grammatical patterns for
problem and solution clauses using articles from professional and student reports.
Problem and solution keywords were used to search their corpora, and each
occurrence was analysed to determine grammatical usage of the keyword. More
interestingly, the causal category associated with each keyword in their context
was also analysed. For example, Reason-Result or Means-Purpose were common
causal categories found to be associated with problem keywords.
      </p>
      <p>
        The goal of the work by [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] was to determine words which are semantically
similar to problem and solution, and to determine how these words are used to
signal problem-solution patterns. However, their corpus-based analysis used
articles from the Guardian newspaper. Since the domain of newspaper text is very
di erent from that of scienti c text, we decided not to consider those keywords
associated with problem-solving patterns for use in our work.
      </p>
      <p>
        Instead of a keyword-based approach, [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] used discourse markers to
examine how the problem-solution pattern was signalled in text. In particular, they
examined how adverbials associated with a result such as \thus, therefore, then,
hence" are used to signal a problem-solving pattern. Problem solving also has
been studied in the framework of discourse theories such as Rhetorical Structure
Theory [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] and Argumentative Zoning [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. RST uses a solutionhood relation as
one of the 23 relations that can hold between elementary discourse units.
However, the de nition of problemhood used in RST di ers too much from ours to be
of direct use here. Argumentative Zoning [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] contains zones such as Gap/Weak
which have a close relation to our de nition of problemhood, and so knowledge
from this particular zone may prove bene cial in the future, although we do not
study this e ect in the current paper.
      </p>
      <p>
        In this work, we approach the task of identifying problem-solving patterns
in scienti c text. We choose to use the model of problem-solving described by
Hoey [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. This pattern comprises four parts: Situation, Problem, Response and
Evaluation. The Situation element is considered optional to the pattern, and so
our focus centres on the core pattern elements.
2
      </p>
    </sec>
    <sec id="sec-2">
      <title>Goal statement and task</title>
      <p>Many surface features in the text o er themselves up as potential signals for
detecting problem-solving patterns in text. However, since Situation is an optional
element, we decided to focus on either Problem or Response and Evaluation as
signals of the pattern. Moreover, we decide to look for each type in isolation.
Our reasons for this are as follows: It is quite rare for an author to introduce a
problem without resolving it using some sort of response, and so this is a good
starting point in identifying the pattern. There are exceptions to this, as authors
will sometimes introduce a problem and then leave it to future work, but overall
there should be enough signal in the Problem element to make our method of
looking for it in isolation worthwhile. The second signal we look for is the use of
Response and Evaluation within the same sentence. Similar to Problem elements,
we hypothesise that this formulation is well enough signalled externally to help
us in detecting the pattern. For example, consider the following Response and
Evaluation: \One solution is to use smoothing." In this statement, the author
is explicitly stating that smoothing is a solution to a problem which must have
been mentioned in a prior statement. In scienti c text, we often observe that
solutions implicitly contain both Response and Evaluation (positive) elements.
Therefore, due to these reasons there should be su cient external signals for the
two pattern elements we concentrate on here.</p>
      <p>When attempting to nd Problem elements in text, we run into the issue
that the word \problem" actually has at least two word senses that need to be
distinguished. There is a word sense of \problem" that means something which
must be undertaken (i.e. task), while another sense is the core sense of the word,
something that is problematic and negative. Only the latter sense is aligned
with our sense of problemhood. This is because the simple description of a task
does not predispose problemhood, just a wish to perform some act. Consider the
following examples, where the non-desired word sense is being used:
{ \Das and Petrov (2011) also consider the problem of unsupervised bilingual</p>
      <p>
        POS induction." [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ].
{ \In this paper, we describe advances on the problem of NER in Arabic
Wikipedia." [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ].
      </p>
      <p>
        Here, the author explicitly states that the phrases in orange are problems,
they align with our de nition of research tasks and not with what we call here
`problematic problems'. We will now give some examples from our corpus for the
desired, core word sense:
{ \The major limitation of supervised approaches is that they require
annotations for example sentences." [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ].
{ \To solve the problem of high dimensionality we use clustering to group the
words present in the corpus into much smaller number of clusters." [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ].
      </p>
      <p>When creating our corpus of positive and negative examples, we took care to
select only problem strings that satisfy our de nition of problemhood; section 3
will explain how we did that.</p>
    </sec>
    <sec id="sec-3">
      <title>Corpus creation</title>
      <p>
        Our new corpus is a subset of the latest version of the ACL anthology released
in March, 20161 which contains 22,878 articles in the form of PDFs and OCRed
text.2 The 2016 version was also parsed using ParsCit [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]. ParsCit recognises
not only document structure, but also bibliography lists as well as references
within running text. A random subset of 2,500 papers was collected covering
the entire ACL timeline. In order to disregard non-article publications such as
introductions to conference proceedings or letters to the editor, only documents
containing abstracts were considered. The corpus was preprocessed using
tokenisation, lemmatisation and dependency parsing with the Rasp Parser [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ].
      </p>
      <p>Our goal was to de ne a ground truth for problem and solution strings, while
covering as wide a range as possible of syntactic variations in which such strings
naturally occur. However, to simplify the task, we only consider here problem
and solution descriptions that are at most one sentence long. In reality, of course,
many problem descriptions and solution descriptions go beyond single sentence,
and require for instance an entire paragraph. However, we also know that short
summaries of problems and solutions are very prevalent in science, and also that
these tend to occur in the most prominent places in a paper. This is because
scientists are trained to express their contribution and the obstacles possibly
hindering their success, in an informative, succinct manner. That is the reason
why we can a ord to only look for shorter problem and solution descriptions,
ignoring those that cross sentence boundaries.</p>
      <p>
        To de ne our ground truth, we examined the parsed dependencies and looked
for a target word (\problem/solution") in subject position, and then chose its
syntactic argument as our candidate problem or solution phrase. To increase the
variation, i.e., to nd as many di erent-worded problem and solution descriptions
as possible, we additionally used semantically similar words (near-synonyms) of
the target words \problem" or \solution" for the search. Semantic similarity
1 http://acl-arc.comp.nus.edu.sg/
2 The corpus comprises 3,391,198 sentences, 71,149,169 words and 451,996,332
characters.
was de ned as cosine in a deep learning distributional vector space, trained
using Word2Vec [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] on 18,753,472 sentences from a biomedical corpus based on
all full-text Pubmed articles [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ]. From the 200 words which were semantically
closest to \problem", we manually selected 28 clear synonyms. From the 200
semantically closest words to \solution" we similarly chose 19. Of the sentences
matching our dependency search, a subset of problem and solution candidate
sentences were randomly selected. An example of this is shown in Figure 2.
Here, the target word \drawback" is in subject position (highlighted in red),
and its clausal argument (ccomp) is \(that) it achieves low performance"
(highlighted in purple). Examples of other arguments we searched for included copula
constructions and direct/indirect objects.
      </p>
      <p>If more than one candidate was found in a sentence, one was chosen at
random. Non-grammatical sentences were excluded; these might appear in the
corpus as a result of its source being OCRed text.</p>
      <p>The potential phrases expressing problems and solutions, respectively, were
then independently checked for correctness by two annotators (the two authors
of this paper). Correctness was de ned by two criteria:
{ The sentence must unambiguously and clearly state the phrase's status as
either a problem or a solution. For problems, the guidelines state that the
phrase has to represent one of the following:
1. an unexplained phenomenon or a problematic state in science; or
2. a research question; or
3. an artifact that does not ful l its stated speci cation.</p>
      <p>For solutions, the phrase had to represent a response to a problem with a
positive evaluation. Implicit solutions were also allowed.
{ The phrase must not lexically give away its status as problem or solution
phrase.</p>
      <p>The second criterion saves us from machine learning cues that are too
obvious. If for instance, the phrase itself contained the words \lack of" or
\problematic" or \drawback", our manual check rejected it, because it would be too easy
for the machine learner to learn such cues, at the expense of many other, more
generally occurring cues.</p>
      <p>We next needed to nd negative examples for both cases. We wanted them
not to stand out on the surface as negative examples, so we chose them so as to
mimic the obvious characteristics of the positive examples as closely as possible.
We call the negative examples `non-problems' and `non-solutions' respectively.
We wanted the only di erences between problems and non-problems to be of
a semantic nature, nothing that could be read o on the surface. We therefore
sampled a population of phrases that obey the same statistical distribution as
our problem and solution strings while making sure they really are negative
examples. We started from sentences not containing any problem/solution words
(i.e. those used as target words). From each such sentence, we at random
selected one syntactic subtree contained in it. From these, we randomly selected a
subset of negative examples of problems and solutions that satisfy the following
conditions:
{ The distribution of the head POS tags of the negative strings should
perfectly match the head POS tags3 of the positive strings. This has the purpose
of achieving the same proportion of surface syntactic constructions as
observed in the positive cases.
{ The average lengths of the negative strings must be within a tolerance of the
average length of their respective positive candidates e.g., non-solutions must
have an average length very similar (i.e. +/- small tolerance) to solutions.
We chose a tolerance value of 3 characters.</p>
      <p>Again, a human quality check was performed on non-problems and
nonsolutions. For each candidate non-problem statement, the candidate was
accepted if it did not contain a phenomenon, a problematic state, a research
question or a non-functioning artefact. If the string expressed a research task, without
explicit statement that there was anything problematic about it (i.e., the `wrong'
sense of \problem", as described above), it was allowed as a non-problem. A
clause was con rmed as a non-solution if the string did not represent both a
response and positive evaluation.</p>
      <p>If the annotator found that the sentence had been slightly mis-parsed, but did
contain a candidate, they were allowed to move the boundaries for the candidate
clause. This resulted in cleaner text, e.g., in the frequent case of coordination,
when non-relevant constituents could be removed.</p>
      <p>From the set of sentences which passed the quality-test for both
independent assessors, 500 instances of positive and negative problems/solutions were
randomly chosen (i.e. 2000 instances in total).
4
4.1</p>
    </sec>
    <sec id="sec-4">
      <title>Method</title>
      <sec id="sec-4-1">
        <title>Experimental design</title>
        <p>
          In our experiments, we used three classi ers, namely Nave Bayes, Logistic
Regression and a Support Vector Machine. For all classi ers an implementation
from the WEKA machine learning library [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ] was chosen. Given that our dataset
is small, 10-fold cross-validation was used instead of a held out test set. All
signi cance tests were conducted using the (two-tailed) Sign Test [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ].
3 The head POS tags were found using a modi cation of the Collins'
Head Finder. This modi ed algorithm addresses some of the limitations
of the head nding heuristics described by [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ] and can be found here:
http://nlp.stanford.edu/nlp/javadoc/javanlp/edu/stanford/nlp/trees/
ModCollinsHeadFinder.html.
4.2
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>Linguistic correlates of problem- and solution-hood</title>
        <p>We rst de ne a set of features without taking the phrase's context into account.
This will tell us about the disambiguation ability of the problem/solution
description's semantics alone. In particular, we cut out the rest of the sentence
other than the phrase and never use it for classi cation. This is done for similar
reasons to excluding certain `give-away' phrases inside the phrases themselves
(as explained above). As the phrases were found using templates, we know that
the machine learner would simply pick up on the semantics of the template,
which always contains a synonym of \problem" or \solution", thus drowning
out the more hidden features hopefully inherent in the semantics of the phrases
themselves. If we allowed the machine learner to use these stronger features, it
would su er in its ability to generalise to the real task.</p>
        <p>
          Of these, bags of words are traditionally successfully used for classi cation
tasks in NLP, so we included bags of words (lemmas) within the candidate
phrases as one of our features (and treat it as a baseline later on). Our second
feature concerns the polarity of selected words, by determining the head of each
candidate phrase and performing word sense disambiguation of each head using
the Lesk algorithm [
          <xref ref-type="bibr" rid="ref26">26</xref>
          ]. The polarity of the resulting synset in SentiWordNet
[
          <xref ref-type="bibr" rid="ref27">27</xref>
          ] was then looked up and used as a feature. Next, a set of syntactic features
were de ned by using the POS tags present in each candidate. We were careful
not to base the model directly on the head POS tag and the length of each
candidate phrase, as these are de ning characteristics used for determining the
non-problem and non-solution candidate set. Not all words are assigned a sense
by the Lesk algorithm, so we need to take care when that happens to a phrasal
head. In those cases, the distributional semantic similarity of the phrasal head
is compared to two words with a known polarity, namely \poor" and
\excellent".These particular words have traditionally been consistently good
indicators of polarity status in many studies [
          <xref ref-type="bibr" rid="ref28 ref29">28, 29</xref>
          ]. Semantic similarity was de ned
as cosine similarity on the embeddings of the Word2Vec model (cf. Section 3).
        </p>
        <p>
          Given that solutions often involve an activity (e.g. a task), we also model the
subcategorisation properties of the verbs involved. Our intuition was that since
problematic situations are often described as non-actions, then these are more
likely to be intransitive. Conversely solutions are often actions and are likely to
have at least one argument. This feature was calculated by running the C&amp;C
parser [
          <xref ref-type="bibr" rid="ref30">30</xref>
          ] on each sentence. C&amp;C is a supertagger and parser that has access
to subcategorisation information.
        </p>
        <p>
          We also wanted to add more information using word embeddings. This was
done in two di erent ways. Firstly, we created a Doc2Vec model [
          <xref ref-type="bibr" rid="ref31">31</xref>
          ], which was
trained on 19 million sentences from scienti c text (no overlap with our data
set). An embedding was created for each candidate sentence. Secondly, word
embeddings were calculated using the Word2Vec model (cf. Section 3). For each
candidate head, the full word embedding was included as a feature.
Modality Responses to problems in scienti c writing often express possibility
and necessity, and so have a close connection with modality. Modality can be
broken into three main categories, as described by [
          <xref ref-type="bibr" rid="ref32">32</xref>
          ], namely epistemic
(possibility), deontic (permission / request / wish) and dynamic (expressing ability).
        </p>
        <p>
          Problems have a strong relationship to modality within scienti c writing.
Often, this is due to a tactic called \hedging" [
          <xref ref-type="bibr" rid="ref33">33</xref>
          ] where the author uses speculative
language in an attempt to make either noncommital or vague statements. This
has the e ect of allowing the author to distance themselves from the statement,
and is often employed when discussing negative or problematic topics.
        </p>
        <p>
          To take this linguistic correlate into account as a feature, we replicated a
modality classi er as described by Ruppenhofer et al. [
          <xref ref-type="bibr" rid="ref34">34</xref>
          ]. More sophisticated
modality classi ers have been recently introduced, for instance using a wide
range of features and convolutional neural networks, e.g, [
          <xref ref-type="bibr" rid="ref35 ref36">35, 36</xref>
          ]. However, we
wanted to check the e ect of a simpler method of modality classi cation on the
nal outcome rst before investing heavily into their implementation. We trained
three classi ers using the subset of features which Ruppenhofer et al. reported
as performing best, and evaluated them on the gold standard dataset provided
by the authors4. The dataset contains annotations of English modal verbs on
the 535 documents of the rst MPQA corpus release [
          <xref ref-type="bibr" rid="ref37">37</xref>
          ].
        </p>
        <p>
          The modality classi er was then retrained on the entirety of the dataset used
by [
          <xref ref-type="bibr" rid="ref34">34</xref>
          ] using the best performing model from training (Logistic Regression).
This new model was then used in the upcoming experiment to predict modality
labels for each instance in our dataset.
5
5.1
        </p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Results</title>
      <sec id="sec-5-1">
        <title>Problems</title>
        <p>4 https://www.uni-hildesheim.de/ruppenhofer/data/modalia release1.0.tgz</p>
        <p>IG Features
0.018 single
0.014 limit, experiment
0.010 datum, information
0.009 require, generate, error, many
0.008 explosion</p>
        <p>As can be seen from Figure 3, we are able to achieve good results for
distinguishing a problematic statement from non-problematic one. The bag of words
baseline achieves a reasonable performance of 68.1% for the SVM classi er,
showing that there is enough signal in the candidate phrases alone to distinguish them
much better than random chance. Taking a look at Figure 4, which shows the
information gain for the top lemmas, we can see that the top lemmas are indeed
indicative of problemhood (e.g. \limit",\explosion"). The transitivity feature
provided some improvement over the baseline but was unable to achieve the
marked improvement we were expecting. Upon taking a closer look at our data,
we saw that our hypothesis that intransitive verbs are commonly used in
problematic statements was true, with over 30% of our problems (153) using them.
However, due to our sampling method for the negative cases we also picked up
many intransitive verbs (163). This explains the small improvement given that
the distribution of intransitive verbs amongst the positive and negative
candidates was almost even.</p>
        <p>
          The modality feature was the most expensive to produce, and managed to
increase performance in both the Bayesian and SVM classi ers but degraded
performance in the Logistic Regression. This surprising result may be partly
due to a data sparsity issue where only a small portion (67) of our instances
contained modal verbs. If the accumulation of additional data was possible, we
think that this feature may have the potential to be much more valuable in
determining problemhood. Additionally, modality has also shown to be helpful
in determining contextual polarity [
          <xref ref-type="bibr" rid="ref38">38</xref>
          ] and argumentation [
          <xref ref-type="bibr" rid="ref39">39</xref>
          ], so using the
output from this modality classi er may also prove useful for further feature
engineering taking this into account in future work.
        </p>
        <p>
          Polarity also didn't achieve the performance we were expecting, improving
only the SVM, but this feature also su ers from a sparsity issue resulting from
cases where the Lesk algorithm [
          <xref ref-type="bibr" rid="ref26">26</xref>
          ] is not able to resolve the synset of the
syntactic head. Knowledge of syntax provides a big improvement with a signi cant
increase in results from two of the classi ers. Examining this in greater detail,
POS tags with high information gain included WH- tags and VB- tags.
WHtags encode a problematic nature, that of being unsure / awaiting resolution
and so this may be one of the reasons for the marked increase.
        </p>
        <p>The embeddings from Doc2Vec allowed us to obtain the most signi cant
increase in performance (80.8 with Nave Bayes) and Word2Vec provided the
best overall result (82.4 with Nave Bayes). The addition of these vectors may
be seen as a form of smoothing in cases where previous linguistic features had a
sparsity issue i.e., instead of a NULL entry, the embeddings provide some sort
of value for each candidate. Particularly wrt. the polarity feature, cases where
Lesk was unable to resolve a synset meant that a ZERO entry was added to the
vector supplied to the machine learner. However, using the word embeddings for
the head in addition to the head's polarity meant that even if Lesk was unable
to resolve the synset, the embedding were able to provide some sort of signal.
5.2</p>
      </sec>
      <sec id="sec-5-2">
        <title>Solutions</title>
        <p>Feature Sets</p>
        <p>Classi cation Acc.</p>
        <p>NB LR SVM
Baselinebow 66.8 63.4 69.5
+Transitivity 66.1 63.3 69.1
+Polarity 68.0 66.3 71.5
+Syntax 70.4 69.5 73.9
+Doc2Vec 75.5 75.5 78.8
+Word2Vec 75.0 81.5 79.8
+Word2Vecsmoothed 74.6 81.3 80.1
Fig. 6: Information gain (IG) in bits
of top lemmas from the best
performing model in Figure 5 (Word2Vec
with LR).</p>
        <p>The results for disambiguation of solutions from non-solutions can be seen
in Figure 5. The bag of words baseline performs much better than random, with
the performance being quite high with regard to the SVM (this result was also
higher than any of the baseline performances from the problem classi ers). As
shown in Figure 6, the top ranked lemmas from the best performing model (using
information gain) were \use" and \method". These lemmas are very indicative
of solutionhood and so give some insight into the high baseline returned from
the machine learners. Transitivity provided no improvement and actually
degraded performance in all three classi ers. However, this low performance is due
to the sampling of the non-solutions (the same reason for the low performance
of the problem transitivity feature). When tting the POS-tag distribution for
the negative samples, we noticed that over 80% of the head POS-tags were verbs
(much higher than the problem heads). The most frequent verb type being the
in nite form. This is not surprising given that a very common formulation to
describe a solution is to use the in nitive \TO" since it often describes a task
e.g., \One solution is to nd the singletons and remove them." Therefore, since
the head POS tags of the non-solutions had to match this high distribution of
in nitive verbs present in the solution, the transitivity feature is not particularly
discriminatory. Polarity and syntactic features were slightly more discriminate,
improving results in all three classi ers. However, similar to the problem
experiment, the embeddings from Word2Vec and Doc2Vec provide the highest result
(81.5% from LR).</p>
        <p>There was no signi cant increase in performance as each feature set was
added. However, the best performing models for each classi er (Doc2Vec, Word2Vec
and Word2Vecsmoothed) were all signi cant with regard to the baseline (P &lt; 0.01).
6</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Discussion</title>
      <p>In this work, we have presented new supervised classi ers for the task of
identifying problem and solution statements in scienti c text. We have also introduced a
new corpus for this task and used it for evaluating our classi ers. Great care was
taken in constructing the corpus by ensuring that the negative and positive
samples were closely matched in terms of syntactic shape. If we had simply selected
random subtrees for negative samples without regard for any syntactic similarity
with our positive samples, the machine learner may have found easy signals such
as sentence length. Additionally, since we did not allow the machine learner to
see the surroundings of the candidate string within the sentence, this made our
task even harder. Our performance on the corpus shows promise for this task,
and proves that there are strong signals for determining both the problem and
solution parts of the problem-solving pattern independently.</p>
      <p>With regard to classifying problems from non-problems, features such as the
POS tag and document and word embeddings provide the best results, with the
Word2Vec embeddings providing the highest performance (82.4%). Classifying
solutions from non-solutions also performs well using these features, with the
best result coming from the embeddings (81.5%).</p>
      <p>In future work, we plan to link problem and solution statements which were
found independently during our corpus creation. Given that our classi ers were
trained on data solely from the ACL anthology, we also hope to investigate
the domain speci city of our classi ers and see how well they can generalise
to domains other than ACL (e.g. bioinformatics). Since we took great care at
removing the knowledge our classi ers have of the explicit statements of problem
and solution (i.e. the classi ers were trained only on the syntactic argument of the
explicit statement of problem-/solution-hood), our classi ers should in principle
be in a good position to generalise, i.e., nd implicit statements too. In future
work, we will measure to which degree this is the case.</p>
      <p>To facilitate further research on this topic, all code and data used in our
experiments can be found here: www.cl.cam.ac.uk/ kh562/ps.html</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>David H Jonassen</surname>
          </string-name>
          .
          <article-title>Toward a design theory of problem solving</article-title>
          .
          <source>Educational Technology Research and Development</source>
          ,
          <volume>48</volume>
          (
          <issue>4</issue>
          ):
          <volume>63</volume>
          {
          <fpage>85</fpage>
          ,
          <year>2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>Balakrishnan</given-names>
            <surname>Chandrasekaran</surname>
          </string-name>
          .
          <article-title>Towards a taxonomy of problem solving types</article-title>
          .
          <source>AI Magazine</source>
          ,
          <volume>4</volume>
          (
          <issue>1</issue>
          ):
          <fpage>9</fpage>
          ,
          <year>1983</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Michael P Jordan</surname>
          </string-name>
          .
          <article-title>Short texts to explain problem-solution structures-and vice versa</article-title>
          .
          <source>Instructional Science</source>
          ,
          <volume>9</volume>
          (
          <issue>3</issue>
          ):
          <volume>221</volume>
          {
          <fpage>252</fpage>
          ,
          <year>1980</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Jo</surname>
          </string-name>
          <article-title>rg Strubing. Research as pragmatic problem-solving: The pragmatist roots of empirically-grounded theorizing</article-title>
          .
          <source>The Sage Handbook of Grounded Theory</source>
          , pages
          <volume>580</volume>
          {
          <fpage>602</fpage>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>Teun</given-names>
            <surname>Adrianus Van Dijk</surname>
          </string-name>
          .
          <article-title>Text and context explorations in the semantics and pragmatics of discourse</article-title>
          .
          <source>Longman Linguistics Library</source>
          ,
          <year>1980</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>John</given-names>
            <surname>Hutchins</surname>
          </string-name>
          .
          <article-title>On the structure of scienti c texts</article-title>
          .
          <source>UEA Papers in Linguistics</source>
          ,
          <volume>5</volume>
          (
          <issue>3</issue>
          ):
          <volume>18</volume>
          {
          <fpage>39</fpage>
          ,
          <year>1977</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <given-names>Joseph</given-names>
            <surname>Evans</surname>
          </string-name>
          <article-title>Grimes. The thread of discourse</article-title>
          , volume
          <volume>207</volume>
          . Walter de Gruyter,
          <year>1975</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <given-names>EO</given-names>
            <surname>Winter</surname>
          </string-name>
          .
          <article-title>Some aspects of cohesion</article-title>
          . University of London,
          <year>1968</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <given-names>Michael</given-names>
            <surname>Hoey</surname>
          </string-name>
          .
          <article-title>Textual interaction: An introduction to written discourse analysis</article-title>
          . Psychology Press,
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <given-names>Lynne</given-names>
            <surname>Flowerdew</surname>
          </string-name>
          .
          <article-title>Corpus-based analyses of the problem-solution pattern: a phraseological approach</article-title>
          , volume
          <volume>29</volume>
          . John Benjamins Publishing,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <given-names>Mike</given-names>
            <surname>Scott</surname>
          </string-name>
          .
          <article-title>Mapping key words to problem and solution</article-title>
          . John Benjamins Publishing,
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <given-names>Maggie</given-names>
            <surname>Charles</surname>
          </string-name>
          .
          <article-title>Adverbials of result: Phraseology and functions in the problem{ solution pattern</article-title>
          .
          <source>Journal of English for Academic Purposes</source>
          ,
          <volume>10</volume>
          (
          <issue>1</issue>
          ):
          <volume>47</volume>
          {
          <fpage>60</fpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13. William C Mann and
          <article-title>Sandra A Thompson. Rhetorical structure theory: Toward a functional theory of text organization</article-title>
          .
          <source>Text-Interdisciplinary Journal for the Study of Discourse</source>
          ,
          <volume>8</volume>
          (
          <issue>3</issue>
          ):
          <volume>243</volume>
          {
          <fpage>281</fpage>
          ,
          <year>1988</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <given-names>Simone</given-names>
            <surname>Teufel</surname>
          </string-name>
          .
          <article-title>Argumentative zoning: Information extraction from scienti c text</article-title>
          .
          <source>PhD thesis</source>
          , University of Edinburgh,
          <year>2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Desai</surname>
            <given-names>Chen</given-names>
          </string-name>
          , Chris Dyer, Shay B Cohen,
          <article-title>and Noah A Smith. Unsupervised bilingual pos tagging with markov random elds</article-title>
          .
          <source>In Proceedings of the First Workshop on Unsupervised Learning in NLP</source>
          , pages
          <volume>64</volume>
          {
          <fpage>71</fpage>
          . Association for Computational Linguistics,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Behrang</surname>
            <given-names>Mohit</given-names>
          </string-name>
          , Nathan Schneider, Rishav Bhowmick,
          <article-title>Kemal O azer, and Noah A Smith. Recall-oriented learning of named entities in arabic wikipedia</article-title>
          .
          <source>In Proceedings of the 13th Conference of the European Chapter of the Association for Computational Linguistics</source>
          , pages
          <volume>162</volume>
          {
          <fpage>173</fpage>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <given-names>Hoifung</given-names>
            <surname>Poon</surname>
          </string-name>
          and
          <string-name>
            <given-names>Pedro</given-names>
            <surname>Domingos</surname>
          </string-name>
          .
          <article-title>Unsupervised semantic parsing</article-title>
          .
          <source>In Proceedings of the 8th Conference on Empirical Methods in Natural Language Processing</source>
          , pages
          <volume>1</volume>
          {
          <fpage>10</fpage>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Sujan Kumar</surname>
            <given-names>Saha</given-names>
          </string-name>
          , Pabitra Mitra, and
          <string-name>
            <given-names>Sudeshna</given-names>
            <surname>Sarkar</surname>
          </string-name>
          .
          <article-title>Word clustering and word selection based feature reduction for maxent based hindi ner</article-title>
          .
          <source>In Proceedings of the Conference on Human Language Translations</source>
          , pages
          <volume>488</volume>
          {
          <fpage>495</fpage>
          . Association for Computational Linguistics,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Isaac G Councill</surname>
            ,
            <given-names>C Lee</given-names>
          </string-name>
          <string-name>
            <surname>Giles</surname>
          </string-name>
          , and
          <string-name>
            <surname>Min-Yen Kan</surname>
          </string-name>
          .
          <article-title>ParsCit: an open-source CRF reference string parsing package</article-title>
          .
          <source>In Proceedings of The 6th International Conference on Language Resources and Evaluation</source>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Ted</surname>
            <given-names>Briscoe</given-names>
          </string-name>
          , John Carroll, and
          <string-name>
            <given-names>Rebecca</given-names>
            <surname>Watson</surname>
          </string-name>
          .
          <article-title>The second release of the rasp system</article-title>
          .
          <source>In Proceedings of the Joint Conference of the International Committee on Computational Linguistics and the Association for Computational Linguistics</source>
          , pages
          <volume>77</volume>
          {
          <fpage>80</fpage>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Tomas</surname>
            <given-names>Mikolov</given-names>
          </string-name>
          , Ilya Sutskever, Kai Chen, Greg S Corrado, and
          <string-name>
            <given-names>Je</given-names>
            <surname>Dean</surname>
          </string-name>
          .
          <article-title>Distributed representations of words and phrases and their compositionality</article-title>
          .
          <source>In Proceedings of Advances in Neural Information Processing Systems</source>
          <volume>26</volume>
          , pages
          <fpage>3111</fpage>
          {
          <fpage>3119</fpage>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Kathy</surname>
            <given-names>McKeown</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Hal</given-names>
            <surname>Daume</surname>
          </string-name>
          , Snigdha Chaturvedi, John Paparrizos, Kapil Thadani, Pablo Barrio, Or Biran, Suvarna Bothe, Michael Collins,
          <string-name>
            <surname>Kenneth R Fleischmann</surname>
          </string-name>
          , et al.
          <article-title>Predicting the impact of scienti c concepts using full-text features</article-title>
          .
          <source>Journal of the Association for Information Science and Technology</source>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <given-names>Michael</given-names>
            <surname>Collins</surname>
          </string-name>
          .
          <article-title>Head-driven statistical models for natural language parsing</article-title>
          .
          <source>Computational linguistics</source>
          ,
          <volume>29</volume>
          (
          <issue>4</issue>
          ):
          <volume>589</volume>
          {
          <fpage>637</fpage>
          ,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24. Mark Hall, Eibe Frank, Geo rey Holmes, Bernhard Pfahringer,
          <string-name>
            <given-names>Peter</given-names>
            <surname>Reutemann</surname>
          </string-name>
          , and
          <string-name>
            <surname>Ian H Witten</surname>
          </string-name>
          .
          <article-title>The weka data mining software: an update</article-title>
          .
          <source>Explorations Newsletter from The Association for Computing Machinery's Special Interest Group on Knowledge Discovery and Data Mining</source>
          ,
          <volume>11</volume>
          (
          <issue>1</issue>
          ):
          <volume>10</volume>
          {
          <fpage>18</fpage>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <string-name>
            <given-names>Sidney</given-names>
            <surname>Siegel</surname>
          </string-name>
          .
          <article-title>Nonparametric statistics for the behavioral sciences</article-title>
          .
          <source>McGraw-Hill</source>
          ,
          <year>1956</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          26.
          <string-name>
            <given-names>Michael</given-names>
            <surname>Lesk</surname>
          </string-name>
          .
          <article-title>Automatic sense disambiguation using machine readable dictionaries: how to tell a pine cone from an ice cream cone</article-title>
          .
          <source>In Proceedings of The 5th Annual International Conference on Systems Documentation</source>
          , pages
          <volume>24</volume>
          {
          <fpage>26</fpage>
          .
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery,
          <year>1986</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          27.
          <string-name>
            <surname>Stefano</surname>
            <given-names>Baccianella</given-names>
          </string-name>
          ,
          <source>Andrea Esuli, and Fabrizio Sebastiani. Sentiwordnet 3</source>
          .
          <article-title>0: An enhanced lexical resource for sentiment analysis and opinion mining</article-title>
          .
          <source>In In Proceedings of The 7th International Conference on Language Resources and Evaluation</source>
          , volume
          <volume>10</volume>
          , pages
          <fpage>2200</fpage>
          {
          <fpage>2204</fpage>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          28.
          <string-name>
            <surname>Peter D Turney</surname>
          </string-name>
          .
          <article-title>Thumbs up or thumbs down?: semantic orientation applied to unsupervised classi cation of reviews</article-title>
          .
          <source>In Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics</source>
          , pages
          <volume>417</volume>
          {
          <fpage>424</fpage>
          ,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          29.
          <string-name>
            <given-names>Tony</given-names>
            <surname>Mullen</surname>
          </string-name>
          and
          <string-name>
            <given-names>Nigel</given-names>
            <surname>Collier</surname>
          </string-name>
          .
          <article-title>Sentiment analysis using support vector machines with diverse information sources</article-title>
          .
          <source>In Proceedings of the Conference on Empirical Methods in Natural Language Processing</source>
          , volume
          <volume>4</volume>
          , pages
          <fpage>412</fpage>
          {
          <fpage>418</fpage>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          30.
          <string-name>
            <surname>James R Curran</surname>
            , Stephen Clark, and
            <given-names>Johan</given-names>
          </string-name>
          <string-name>
            <surname>Bos</surname>
          </string-name>
          .
          <article-title>Linguistically motivated largescale nlp with c&amp;c and boxer</article-title>
          .
          <source>In Proceedings of the 45th Annual Meeting of the Association for Computational Linguistics on Interactive Poster and Demonstration Sessions</source>
          , pages
          <volume>33</volume>
          {
          <fpage>36</fpage>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          31.
          <string-name>
            <surname>Quoc</surname>
            <given-names>V Le</given-names>
          </string-name>
          and
          <string-name>
            <given-names>Tomas</given-names>
            <surname>Mikolov</surname>
          </string-name>
          .
          <article-title>Distributed representations of sentences and documents</article-title>
          .
          <source>In Proceedings of the International Conference on Machine Learning</source>
          , volume
          <volume>14</volume>
          , pages
          <fpage>1188</fpage>
          {
          <fpage>1196</fpage>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          32.
          <string-name>
            <given-names>Angelika</given-names>
            <surname>Kratzer</surname>
          </string-name>
          . Modality. In von Stechow &amp; Wunderlich (eds.),
          <source>Semantics: An international handbook of contemporary research. de Gruyter</source>
          ,
          <year>1991</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          33.
          <string-name>
            <given-names>Ben</given-names>
            <surname>Medlock</surname>
          </string-name>
          and
          <string-name>
            <given-names>Ted</given-names>
            <surname>Briscoe</surname>
          </string-name>
          .
          <article-title>Weakly supervised learning for hedge classi cation in scienti c literature</article-title>
          .
          <source>In Proceedings of the Association for Computational Linguistics</source>
          , volume
          <volume>2007</volume>
          , pages
          <fpage>992</fpage>
          {
          <fpage>999</fpage>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          34.
          <string-name>
            <given-names>Josef</given-names>
            <surname>Ruppenhofer</surname>
          </string-name>
          and
          <string-name>
            <given-names>Ines</given-names>
            <surname>Rehbein</surname>
          </string-name>
          .
          <article-title>Yes we can!? annotating the senses of english modal verbs</article-title>
          .
          <source>In Proceedings of the 8th International Conference on Language Resources and Evaluation</source>
          , pages
          <volume>24</volume>
          {
          <fpage>26</fpage>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          35.
          <string-name>
            <surname>Mengfei</surname>
            <given-names>Zhou</given-names>
          </string-name>
          , Anette Frank, Annemarie Friedrich, and
          <string-name>
            <given-names>Alexis</given-names>
            <surname>Palmer</surname>
          </string-name>
          .
          <article-title>Semantically enriched models for modal sense classi cation</article-title>
          .
          <source>In Workshop on Linking Models of Lexical</source>
          , Sentential and Discourse-level
          <string-name>
            <surname>Semantics</surname>
          </string-name>
          (LSDSem),
          <source>page 44</source>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          36.
          <string-name>
            <given-names>Ana</given-names>
            <surname>Marasovic</surname>
          </string-name>
          and
          <string-name>
            <given-names>Anette</given-names>
            <surname>Frank</surname>
          </string-name>
          .
          <article-title>Multilingual modal sense classi cation using a convolutional neural network</article-title>
          .
          <source>In Proceedings of the 1st Workshop on Representation Learning for NLP</source>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          37.
          <string-name>
            <surname>Janyce</surname>
            <given-names>Wiebe</given-names>
          </string-name>
          , Theresa Wilson, and
          <string-name>
            <given-names>Claire</given-names>
            <surname>Cardie</surname>
          </string-name>
          .
          <article-title>Annotating expressions of opinions and emotions in language. Language Resources and Evaluation (formerly Computers and the Humanities</article-title>
          ),
          <volume>39</volume>
          (
          <issue>2</issue>
          ):
          <volume>165</volume>
          {
          <fpage>210</fpage>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          38.
          <string-name>
            <surname>Theresa</surname>
            <given-names>Wilson</given-names>
          </string-name>
          , Janyce Wiebe, and Paul Ho mann.
          <article-title>Recognizing contextual polarity in phrase-level sentiment analysis</article-title>
          .
          <source>In Proceedings of the Conference on Human Language Technology and Empirical Methods in Natural Language Processing</source>
          , pages
          <volume>347</volume>
          {
          <fpage>354</fpage>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          39.
          <string-name>
            <surname>Maria</surname>
            <given-names>Becker</given-names>
          </string-name>
          , Alexis Palmer, and Anette Frank.
          <article-title>Clause types and modality in argumentative microtexts</article-title>
          .
          <source>In Workshop on Foundations of the Language of Argumentation (in conjunction with COMMA)</source>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>