<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Towards Discourse Parsing-inspired Semantic Storytelling</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Georg Rehm</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Karolina Zaczynska</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Julian Moreno-Schneider</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Malte Ostendor</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Peter Bourgonje</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Maria Berger</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jens Rauenbusch</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andre Schmidt</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mikka Wild</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Corresponding Author: Georg Rehm</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>georg.rehm@dfki.de</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>3pc GmbH Neue Kommunikation</institution>
          ,
          <addr-line>Prinzessinnenstra e 1, 10969 Berlin</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>DFKI GmbH, Alt-Moabit 91c</institution>
          ,
          <addr-line>10559 Berlin</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Previous work of ours on Semantic Storytelling uses text analytics procedures including Named Entity Recognition and Event Detection. In this paper, we outline our longer-term vision on Semantic Storytelling and describe the current conceptual and technical approach. In the project that drives our research we develop AI-based technologies that are veri ed by partners from industry. One long-term goal is the development of an approach for Semantic Storytelling that has broad coverage and that is, furthermore, robust. We provide rst results on experiments that involve discourse parsing, applied to a concrete use case, \Explore the Neighbourhood!", which is based on a semi-automatically collected data set with documents about noteworthy people in one of Berlin's districts. Though automatically obtaining annotations for coherence relations from plain text is a non-trivial challenge, our preliminary results are promising. We envision our approach to be combined with additional features (NER, coreference resolution, knowledge graphs).</p>
      </abstract>
      <kwd-group>
        <kwd>Semantic Storytelling</kwd>
        <kwd>Natural Language Processing</kwd>
        <kwd>Dis- course Parsing</kwd>
        <kwd>Rhetorical Structure Theory</kwd>
        <kwd>Penn Discourse TreeBank</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Cultural institutions such as museums, archives or libraries often rely on public
funding and therefore need to communicate their value to the public constantly.
One successful way to achieve this goal is to employ storytelling, which can be
de ned as creating emotional, interactive narratives in a digital format.
Storytelling enables cultural institutions to make use of their digitized collections,
demonstrating their relevance and reaching out to new audiences. Due to the
extremely large amounts of available digital content, the curation of stories is
typically performed by human knowledge workers. This calls for automated
procedures. Such procedures should 1) label the content for several types of
metadata semi-automatically, allowing for relevant categorisation. And 2) process
the individual content pieces to present the information contained in them to
a knowledge worker in an intuitive way. Since cultural organisations are often
unlikely to be able to face this challenge on their own, we develop a platform
supporting this use case in the the technology transfer project QURATOR. Our
goal are semi-automatic technologies that keep the human in the loop and allow
for fast, e cient and intuitive exploration of large and highly domain-speci c
data sets. Relating events into a schematic structure, i. e., storytelling, and
ordering them, e. g., in terms of topic, locality or causal or temporal relationships,
aid humans in nding meaningful patterns in data [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
      </p>
      <p>
        In earlier work, we described approaches to Semantic Storytelling making
use of Named Entity Recognition (NER) and Event Detection [
        <xref ref-type="bibr" rid="ref16 ref23 ref24">16, 24, 23</xref>
        ]. In
this article, we explore ways to present a knowledge worker the semantic
structure between text segments in an incoming text collection, making it possible
to nd interesting and surprising connections and information inside texts
regarding a prede ned topic. We focus on means of relating text segments to each
other by borrowing from frameworks for the processing of coherence relations.
From Rhetorical Structure Theory (RST) [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] we borrow the idea that larger
sequences of texts (i. e., non-elementary discourse units) are related, moving
beyond the shallow parsing of individual coherence relations. From the Penn
Discourse TreeBank (PDTB) [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ] we use the sense inventory and perform a
series of experiments, relating text segments according to the four top-level classes
of the PDTB sense hierarchy. The experiments are centered around the use case
\Explore the Neighbourhood!". This tool, currently in development, is an urban
exploration app that makes uses of documents on the Berlin district of Moabit.
It allows users to follow stories, created by an editor semi-automatically, while
exploring the district both physically and digitally.
      </p>
      <p>The remainder of this paper is structured as follows. Section 2 reviews
relevant work, in particular, approaches using discourse relations in text. Section 3
explains the use case in more detail. Section 4 provides a technical de nition,
while Section 5 outlines the experiments on the data set we created. Finally,
Section 6 provides a summary and suggests directions for future work.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>
        The act of storytelling and the resulting stories, can be seen as a strategy to
uncover meaningful patterns in the world around us [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. At the core of research
on classical narratology, essential to storytelling, is the uncovering of the rules
that underlie this strategy, or at least the ways to best achieve the goal. Early
work on narratology is described in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], de ning a narrative as a discourse
following a plot structure that has a chronological and logical event order. More
recently, [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] applied this de nition of plot structure to (chrono)logically ordered
events. Another line of work on narratology is represented by the work of [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ],
who analyzes the basic, irreducible, structural elements of Russian folk tales.
More recently, Propp's work was used by [
        <xref ref-type="bibr" rid="ref33">33</xref>
        ] for their story detection and
generation systems. The same authors, in [
        <xref ref-type="bibr" rid="ref32">32</xref>
        ], make use of another eld of research
related to text coherence, namely that of the processing of coherence relations.
They apply the work of [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ] on hierarchical discourse relations to work out how
paragraphs behave when being used as discourse-structural units in news
articles, with the ultimate goal of understanding the importance and temporal
order of story items. Our work follows a similar approach, but uses PDTB sense
hierarchy labels. The PDTB [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ] is an (English) corpus of Wall Street Journal
articles (a subsection of the Penn TreeBank [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]) annotated for individual
discourse relations. We adopt the PDTB sense hierarchy, because it is the single
largest corpus annotated for coherence relations and therefore the corpus best
facilitating machine-learning based approaches. Due to the shallow nature of
the PDTB framework (it only annotates individual relations, without making
commitment to larger text structure, or mutual importance or relevance), we
additionally source from RST [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], particularly the notion of nuclearity. In RST,
a text is divided into Elementary Discourse Units, which are joined together,
forming either a mono-nuclear relation (with one unit being the more
prominent, important or relevant nucleus and the other, less prominent unit being the
satellite) or a multi-nuclear relation. It is this notion of prominence, or relative
importance to the storyline at hand, that we adopt from RST.
      </p>
      <p>
        With regard to application-driven approaches, much work has been done
on the nal, surface realisation aspect of text generation [
        <xref ref-type="bibr" rid="ref7 ref8">7, 8</xref>
        ]. An approach
resembling more closely ours is described by [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], who use dependency parsing
in combination with discourse relations to determine sentence relations. In our
approach, however, in addition to nding relevant articles for the user, we want
to classify the type of relation the articles in question have to each other.
      </p>
      <p>
        In our own previous work we described tools supporting the processing and
generation of digital content with a strong industry focus, as is equally the
case in the current context of the QURATOR project. The functionality of the
curation technology platform is explained in [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ]. [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ] presents an example of
this platform applied to the use case of a personal communication archive, i. e.,
a collection of approx. 2,800 letters exchanged between the German architect
Erich Mendelsohn and his wife Luise between 1910 and 1953. From this, we
extracted, i. a., named entities, temporal expressions and events, combined these
and used them to track and visualise the movement (across the globe) of Erich
and Luise. Additional prototypes are presented in [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] and [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ].
3
      </p>
      <p>Industry Needs and Applications: The \Explore the
Neighbourhood!" Use Case
\Explore the Neighbourhood!" is a concept for a mobile app, which engages
urban explorers in semi-automatically created stories, making use of digitized
cultural collections. Moabit is a district in Berlin and was chosen due to its rich
history and lively present. Such an app could be made available by museums,
cities or municipalities, tourist information o ces or local marketing campaigns.
End users might be tourists, pupils studying in or visiting the neighborhood,
or residents. Value is created for all parties by entertaining and educating users
whilst communicating the district's or cultural institution's relevance. The app
o ers both curated and generated stories. While in a nal concept of \Explore
the Neighbourhood!" these di erences might not be noticed by the end user, in
the following we will present each approach separately to describe the concept
more precisely. We plan to fully integrate the approach described in Section 4.
3.1</p>
      <sec id="sec-2-1">
        <title>Curated Stories</title>
        <p>Upon launching the app a set of interactive stories is o ered to the end user who
can in uence the story's direction, depth, and pace. Nevertheless, it still contains
signi cant plot elements curated by an editor. The curation process requires the
editor to de ne several storylines in a customised tool, which contains search
capabilities and a recommendation system (Figure 1), both of which help surface
relevant content for each step along a story path. Such a tool is made possible
due to rich metadata which allow queries such as \poems describing Berlin in a
praising tone" (text classi cation and analysis detecting locations and sentiment)
or \photos showing Kurt Tucholsky next to a church" (image classi cation and
analysis detecting people and objects, in this case churches). Figure 1 shows the
user interface of such a tool.</p>
        <p>Curated stories can be published to the app (Figure 2a). Stories may contain
geographical points of interest within Moabit which are connected through an
overall story arch, such as a biography. The exemplary stories depicted in this
article follow the biography of Kurt Tucholsky (Figure 2b), a German-Jewish
journalist and writer born in Moabit in 1890. The stories contain locations,
historic photos and maps, scanned original works and editorial content.
(a) Description of Kommune 1</p>
        <p>(b) Kurt Tucholsky's biography
The existence of several storylines within a story, as well as several stories
in parallel, allows for connections to be forged. These connections can be based
on common topics, locations, or other parameters that support a consistent
and emotional narrative. Users can follow one path through a story, choose to
dive deeper into certain aspects of it, e. g., Kurt Tucholsky (Figure 2b), change
their perspective onto a topic by exploring alternative stories, or switch to a
completely di erent, yet connected story. The consumable stories are linked in
a network and limited only by the amount of pieces of information and the size
of the network created by the editor, who can extend it continuously.
3.2</p>
      </sec>
      <sec id="sec-2-2">
        <title>Generated Stories</title>
        <p>Unlike curated ones, generated stories are created entirely by a storytelling
engine. This is made possible due to a set of well-chosen parameters which in uence
the automatic selection and connection of content. These parameters are de ned
by several factors:
{ a chosen topic (initiated through a keyword or phrase)
{ the type of story being told (such as biography or travel guide),
{ users' preferences (such as available time, current sentiment, preferred mode
of travel),
{ users' behavior (such as current location, walking speed, orientation).</p>
        <p>Based upon the factors listed above, \Explore the Neighbourhood!"
automatically generates a story by selecting the right content based on its rich metadata.
The end result, which is the story consumed by the users, may not look so di
erent from editor-curated stories. Nevertheless, since generating a story happens
in real-time, it constantly adapts to users' choices, which creates a more personal
and more interactive experience.
4</p>
        <p>Semantic Storytelling: Technical Description
One of the goals of our Semantic Storytelling system is to aid knowledge workers
in selecting relevant pieces of content, e. g., the app editor who wants to curate
stories for the app. Following the prototype of the \Explore the Neighbourhood!"
app (Section 3), this section describes the technical details of the back-end.</p>
        <p>Let us assume the following situation. A user is visiting a city and wants
information about a topic T regarding the neighbourhood. The goal of the
curation prototype is, then, to identify and to suggest new content for the app that
can be included in the user's tour. To do so, we rst have to initialise the topic
T , for example, as a sentence, keyword or named entity. Next up, the tool has to
identify if, for example, a document in a collection or a web page is relevant for
topic T , and, if so, if it is important for T . Finally, we identify the semantic
relation between incoming texts and the provided topic T , which could be, among
others, background, cause, contrast, example etc. In the following, we describe
these steps in more detail (Figure 3).</p>
        <p>Possible instantiations of T
• Complete document
• Summary
• Claim or fact
• Event
• Named entity</p>
        <p>Incoming Content</p>
        <p>Web content</p>
        <p>Self-contained
document collection</p>
        <p>Wikipedia
Topic</p>
        <p>T</p>
        <p>User
generating
Stories</p>
        <sec id="sec-2-2-1">
          <title>1 Determine the relevance of a segment for T</title>
          <p>a Document relevance
b Segment relevance
A Sentence 1
B Sentence 5
C Sentence 4
Ranked list of
text segments</p>
        </sec>
        <sec id="sec-2-2-2">
          <title>2 Determine importance</title>
          <p>of a segment</p>
          <p>A isLessImportantThan</p>
          <p>B</p>
          <p>C isMoreImportantThanT isMoreImportantThan
3 sDeisgcmoeunrsteanredlatotiopnicbetween A</p>
          <p>Comparison T</p>
          <p>Comparison B
Expansion</p>
          <p>C
“Explore The Neighbourhood!”</p>
          <p>GUI
Fig. 3: Architecture of the Semantic Storytelling approach</p>
        </sec>
      </sec>
      <sec id="sec-2-3">
        <title>Step 1: Determine the Relevance of a Segment for a Topic The approach</title>
        <p>starts with a topic T , instantiated through a text segment such as a complete
document, a headline or a named entity. To identify content pieces relevant for T ,
we process incoming textual content, like a self-contained document collection,
a systematically compiled corpus or a knowledge base.</p>
        <p>
          For each piece of content, we need to decide whether its topic is relevant for
T , which can be computed in various ways. We can employ topic modeling (LDA,
LSA) or, without explicitly modeling topics, we can also perform pair-wise
comparisons of document similarity. Document pairs with a high similarity score are
assumed to cover the same topic, therefore, we start with the seed document
ds of which we know that it represents T and measure its similarity to other
candidates. To compute semantic similarity, documents are represented as
numerical vectors. Classical methods like bag-of-words or tf-idf encode documents
as sparse vectors [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ], while neural methods (word2vec, sent2vec, doc2vec, see
e. g., [
          <xref ref-type="bibr" rid="ref14 ref19 ref27">14, 19, 27</xref>
          ]) produce dense representations. In both cases, cosine similarity
can be used to compute the similarity of the document vectors.
        </p>
        <p>
          Step 2: Determine the Importance of a Segment If we have determined
all documents d which are related to T , we need to determine the importance
of d (or its segments or sentences) with regard to T . There is no o -the-shelf
approach to determine the importance of a segment with regard to a topic, but
various cues and indicators can potentially be exploited. One way of doing this
is to borrow from RST, especially the notion of nuclearity. Constructing an RST
tree involves decisions with regard to the status of text segments including their
discourse relation to other segments and also regarding their role as a nucleus
(the important core part of a relation) or satellite (the contributing part of a
relation) in the context of a speci c discourse relation. Two segments are assigned
either a satellite-nucleus (S-N), nucleus-satellite (N-S) or a nucleus-nucleus
(NN) structure. This sub-task can be done in isolation [
          <xref ref-type="bibr" rid="ref28 ref9">9, 28</xref>
          ], or in conjunction
with the relation classi cation task [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ]. When performed iteratively, this
pairwise classi cation can result in a set of most important segments regarding T .
Another way of determining topical importance is to treat it as a
segmentlevel question answering task. Given a document d consisting of text segments
(t1; t2; : : : tn), the aim is to nd the segment ti that contains the answer to
the input question (i. e., topic T ). Transformer language models have achieved
state-of-the-art results for question answering [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ], suggesting that those model
architectures would be bene cial for storytelling.
        </p>
      </sec>
      <sec id="sec-2-4">
        <title>Step 3: Semantic or Discourse Relation between two Segments After</title>
        <p>
          having established the relevance and relative importance, we proceed with
determining the semantic or discourse relation that exists between the text segments
and topic T . Our initial experiments are based on the PDTB due to its
considerably larger size with more than 1.1 million tokens over the RST-Discourse
TreeBank [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] with approx. 200k tokens. We adopt the PDTB's sense hierarchy,
which comprises four top-level classes, 16 types on the second level and 23
subtypes on the third. For now, our experiments are based on the top-level senses,
Temporal, Contingency, Comparison, Expansions, and an additional label, None.
5
        </p>
        <p>Experiment for \Explore the Neighbourhood!"
In this section, we describe our rst experiments, which aim to explore the
suitability of the approach and helps us gain an understanding of what we can
achieve in the long run. We concentrate on step 3, therefore, we created a data
set of crawled web documents about the Berlin district Moabit, and implemented
initial experiments to classify discourse relations between text segments inside
the data set. We would like to show a comparison with similar tools, but to the
best of our knowledge, there are no similar tools that are extracting semantic
relations through intra-document text segments (using PDTB).
5.1</p>
      </sec>
      <sec id="sec-2-5">
        <title>Data Set</title>
        <p>The data set is composed of documents containing information and stories
connected to the district of Moabit in Berlin. We are in the rst stages of developing
this data set. In the long term, the idea is to put together a much larger
collection of documents focused on Moabit so that it can be used for the Semantic
Storytelling prototype. We used the focused crawler Spidey3, which returns a
list of URLs from websites which are based on a set of prede ned query terms.
We manually de ned 28 queries about interesting places, buildings, or persons
connected to Moabit. Some of these terms are Moabit, Moabit gentri cation,
Kleiner Tiergarten, Kulturfabrik Moabit, Berlin Central Station and Kurt
Tucholsky. After obtaining the website URLs, we crawl and boilerplate the content
of the pages and their metadata4. The resulting data set is composed of slightly
more than 100 documents that have been ltered manually in a second step.
5.2</p>
      </sec>
      <sec id="sec-2-6">
        <title>Classi ers for Discourse Relation between Text Segments</title>
        <p>
          Our aim is to extract discourse relations from texts and so, being able to extract
relevant content from a text collection and, in the longer run, to nd new
storylines composed of semantically related parts of di erent text segments taken
from the collection. We train a relation sense classi er on PDTB2 [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ] and apply
it on two pieces of content. For training, we use the two arguments of a
relation, but at a later point we deploy it using individual sentences. We argue that
the sentence-level is the most appropriate level to use as input for our classi er
(as opposed to the shorter token or phrase level, or the longer paragraph level)
and that the discrepancy between argument shapes and typical sentence lengths
(itself very much dependent on the domain) is tolerable.
3 https://github.com/vikrambajaj22/Spidey-Focused-Web-Crawler
4 We use Newspaper3k, see https://github.com/codelucas/newspaper
Classi er Model Classifying the discourse relation between sentence pairs
requires a semantic understanding of the sentences. We encode the text as deep
contextual representations with a language model based on the Transformer
architecture [
          <xref ref-type="bibr" rid="ref30">30</xref>
          ]. To be precise, the pre-trained language model from DistilBERT
[
          <xref ref-type="bibr" rid="ref26">26</xref>
          ], a distilled version of Bidirectional Encoder Representations from
Transformers [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] is used5. BERT performs well for document classi cation tasks [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ].
        </p>
        <p>
          To classify the relation between two texts, we employ a Siamese architecture
[
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. In contrast to a classical Siamese model, in which a binary classi er is
employed on the output of the two identical sub-networks, we feed the sub-network
output into a multi-label classi er, as illustrated in Figure 4.
        </p>
        <p>Concatenation
MLP
Classification Layer
d1
BERT
d2</p>
        <p>BERT
ŷ = SemRel(*, , )</p>
        <p>Fig. 4: The architecture of the Siamese
BERT model for the classi cation of
discourse relations between two text
segments d1 and d2. The output of the
classi cation layer y^ holds the predicted
semantic relation according to the
toplevel PDTB2 senses.</p>
        <p>Text snippets d1 and d2 are inputs to the classi er. BERT's architecture
consists of six hidden layers, each layer consists of 768 units (66M parameters;
DistilBERT). BERT is used in a Siamese fashion such that hi = BERT(di) is
the encoded representation of text di where hi is the last hidden state of the last
BERT layer. The nal feature vector xf is a combined concatenation of the text
representations:</p>
        <p>
          xf = [h1; h2; jh1 h2j; h1 h2; h1 +2 h2 ] (1)
On top of the concatenation, we implement a Multi-Layer Perceptron (MLP).
The MLP consists of two fully-connected layers, Ff ( ) and Fg( ), where each layer
has 100 units and ReLU( ) is the activation function. The discourse relation y^ is
classi ed on the basis of the feature vector xf as follows:
y^ = (Ff (ReLU(Fd(xf ))))
(2)
The logistic softmax function ( ) generates probabilistic multi-label classi
cations. The dimension of y^ corresponds to the number of classi cation labels,
which are the four top-level PDTB2 senses (Temporal, Contingency,
Comparison, Expansions ) and one additional dimension (None).
5 We use the PyTorch implementation by HuggingFace [
          <xref ref-type="bibr" rid="ref31">31</xref>
          ].
        </p>
        <p>
          Transfer Learning Our target corpus of texts for \Explore the
Neighbourhood!" does not include any kind of annotated training data. Thus, we cannot
use the data set to train the classi er. Instead, we rely on the PDTB2 data
set. Training is performed with batch size b = 16, dropout probability d = 0:1,
learning rate = 2 5 (Adam optimizer) and 5 training epochs. These
hyperparameters are the ones proposed by [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] for BERT ne-tuning.
        </p>
        <p>PDTB Relation</p>
        <p>Precision</p>
        <p>Recall</p>
        <p>F1-score</p>
        <p>Support
Comparison
Contingency
Expansion
Temporal
None
Micro avg.</p>
        <p>Macro avg.</p>
        <p>The results that are derived from a 80-20 train-test-split are shown in Table 1.
For evaluation, we use the multi class metric F1-micro average, which calculates
the metrics globally by counting the total true positives, false negatives and false
positives to compute the average metric. In a multi-class classi cation setup,
micro-average is preferable if you suspect there might be class imbalance. In
the end, we achieve 0.55 micro average F1. Due to the fact that we have not
implemented features relating to the connective, our classi cation performs lower
than current state-of-the-art approaches.
5.3</p>
      </sec>
      <sec id="sec-2-7">
        <title>First Experiment on Use Case Data Set and Discussion</title>
        <p>Given the PTDB2-based classi er, we continue to nd discourse relations within
the corpus containing documents for the \Explore the Neighbourhood!" use case.
As a preprocessing step, we rst exclude all non-English documents and group
documents by topic based on the query terms for the focused crawler. Next, we
nd document pairs among the topic groups (only semantically similar document
pairs are considered). More precisely, documents are represented as tf-idf vectors
and the cosine similarity of a document pair da and db must be above a xed
threshold (cosine(da; db) &gt; 0:15). Our classi er is trained to detect sentence-level
relations, thus, we also split the documents into sentences6. After excluding all
sentences with less than ve words, we end up with 96,796 sentence pairs that
are passed to the classi er.
6 We use pySBD, see https://github.com/nipunsadvilkar/pySBD
Documents</p>
        <p>Discourse relations
Topic</p>
        <p>Segment A</p>
        <p>Segment B</p>
        <p>S. Co. Ct. E. T. N.
1 Farin</p>
        <p>Urlaub
Moabit</p>
        <p>In April 2012 they re- At the age of 16, Vet- .51 .01 .01 .04 .93 .1
leased another album ter went on a school trip
\auch" (\also") to London, and returned
home as a punk with dyed
blonde hair.
3 AEG
turbine
factory
4 Kurt
cholsky</p>
        <p>Moabit
5
Schultheiss
Brewery</p>
        <p>It is an in uential and However, when it came .32 .62 .16 .15 .01 .06
well-known example of in- to AEG's public image
dustrial architecture. and public perception, the
focus remained on
Peter Behrens: the famous
artist-cum-architect
overshadowed the engineer.</p>
        <p>Tu- Admittedly, Tucholsky is He saw himself as a left- .25 .45 .28 .19 .01 .08
seldom recognized as a wing democrat and paci st
philosopher. and warned against
antidemocratic tendencies {
above all in politics, the
military and justice { and
the threat of National
Socialism.
"Good people drink good Schultheiss is currently .16 .32 .19 .42 .01 .06
beer," Hunter S Thompson brewing far less beer
once said, writing about than at the time of
a beverage that is consid- re-uni cation.
ered to be typically
German and is, of course, also
popular in Berlin.</p>
        <p>To get a rst impression on the applicability of our approach, and to
motivate our next steps, we manually select ve example sentence pairs to evaluate
them qualitatively. The rst line of Table 2 shows an example where the
classi er correctly labels the discourse relation as Temporal, most likely because of
the temporal markers included. In the second line, the approach correctly
identi es the discourse relation as an Expansion, i. e., segment B can be seen as an
extension of the biography described in segment A. Nevertheless, in other
examples, the approach is often unable to handle coreference. The classi er is often
not detecting a discourse relation between two segments, even if those segments
reference the same entity, while one segment uses a pronoun for the entity. By
implementing a preprocessing step with rudimentary coreference resolution we
expect the classi cation to improve signi cantly. The classi er predicts the label
Comparison often when speci c lexical markers, such as however, but or while,
appear in segment B, like in example 3. Example 4 is an exception, where the
classi er predicts the relation Comparison correctly without needing a lexical
marker, but, generally we observe that this dependency on lexical features leads
to wrong predictions. We see one reason in the fact that the sentences are taken
from di erent sources, and the lexical markers for the discourse relation are
therefore often missing, also even if semantically it can be seen as a Comparison.
This is the case in example 5, which is wrongly predicted as an Extension while
we interpret it as a Comparison (paraphrased as 'Even if he is recognized as a
philosopher, he saw himself as a political activist'). On the other hand, in other
examples, the lexical markers cause false positives errors. Hence, future work will
extent the number of preprocessing steps to better group text segments which
have the same content and talk about the same entities, events or topics.
6</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Conclusions</title>
      <p>
        We describe rst experiments in order to apply our Semantic Storytelling
approach to an industrial use case. This use case, \Explore the Neighbourhood!",
makes it possible to interactively create a city guide with adjusting interesting
stories about a particular district built upon user-dependent parameters, such
as prede ned topics, keywords, etc. The basic idea is to automate storytelling by
detecting discourse relations between texts segments of di erent sources on the
same topic, which makes it possible to be able to detect and create new storylines
extracted from a document collection. We describe the di erent steps in order to
create a corresponding processing framework. In the experiment presented here,
we focus on the third step of our approach, the classi cation of discourse
relations between segments. By focusing more on steps one and two as described in
Section 4, we will be able to improve the results in further experiments. For
example, we expect the classi cation to improve signi cantly by using coreference
resolution during preprocessing. One way of improving the coreference resolution
would be to pretrain the classi er on the coreference task rst [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. As data sets
are still limited, we will expand the data set for our needs and create, in the
longer run, annotations to develop a gold standard.
      </p>
    </sec>
    <sec id="sec-4">
      <title>Acknowledgements</title>
      <p>The research presented in this article is funded by the German Federal
Ministry of Education and Research (BMBF) through the project QURATOR
(Unternehmen Region, Wachstumskern, no. 03WKDA1A). http://qurator.ai</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Bal</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          : Narratology: Introduction to the
          <source>Theory of Narrative</source>
          .
          <year>1985</year>
          . Trans. by Christine van Boheemen. Toronto: University of Toronto Press (
          <year>1985</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Bromley</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bentz</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bottou</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Guyon</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lecun</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moore</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Eduard</surname>
            <given-names>Sackinger</given-names>
          </string-name>
          , Shah, R.:
          <article-title>Signature Veri cation using a Siamese Time Delay Neural Network</article-title>
          .
          <source>International Journal of Pattern Recognition and Arti cial Intelligence</source>
          <volume>7</volume>
          (
          <issue>4</issue>
          ) (
          <year>1993</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Bruner</surname>
          </string-name>
          , J.:
          <source>The Narrative Construction of Reality. Critical Inquiry</source>
          <volume>18</volume>
          (
          <issue>1</issue>
          ),
          <volume>1</volume>
          {
          <fpage>21</fpage>
          (
          <year>1991</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Carlson</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Marcu</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Okurowski</surname>
            ,
            <given-names>M.E.: RST</given-names>
          </string-name>
          <string-name>
            <surname>Discourse Treebank</surname>
          </string-name>
          ,
          <year>ldc2002t07</year>
          (
          <year>2002</year>
          ), https://catalog.ldc.upenn.edu/LDC2002T07
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Caselli</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vossen</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>The Event StoryLine Corpus: A New Benchmark for Causal and Temporal Relation Extraction</article-title>
          .
          <source>In: Proceedings of the Events and Stories in the News Workshop</source>
          . pp.
          <volume>77</volume>
          {
          <fpage>86</fpage>
          .
          <string-name>
            <surname>ACL</surname>
          </string-name>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Devlin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>M.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Toutanova</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          : BERT:
          <article-title>Pre-training of Deep Bidirectional Transformers for Language Understanding</article-title>
          .
          <source>In: NAACL-HLT</source>
          . pp.
          <volume>4171</volume>
          {
          <issue>4186</issue>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Fan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lewis</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dauphin</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>Hierarchical Neural Story Generation</article-title>
          . In:
          <article-title>Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)</article-title>
          . pp.
          <volume>889</volume>
          {
          <issue>898</issue>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Fan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lewis</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dauphin</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>Strategies for Structuring Story Generation</article-title>
          . arXiv preprint arXiv:
          <year>1902</year>
          .
          <volume>01109</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Hernault</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Prendinger</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          , duVerle,
          <string-name>
            <given-names>D.A.</given-names>
            ,
            <surname>Ishizuka</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.:</surname>
          </string-name>
          <article-title>HILDA: A Discourse Parser Using Support Vector Machine Classi cation</article-title>
          .
          <source>Dialogue &amp; Discourse</source>
          <volume>1</volume>
          (
          <issue>3</issue>
          ),
          <volume>1</volume>
          {
          <fpage>33</fpage>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Joshi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Levy</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zettlemoyer</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weld</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>BERT for Coreference Resolution: Baselines and Analysis</article-title>
          .
          <source>In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)</source>
          . pp.
          <volume>5807</volume>
          {
          <fpage>5812</fpage>
          . Association for Computational Linguistics, Hong Kong,
          <source>China (Nov</source>
          <year>2019</year>
          ). https://doi.org/10.18653/v1/
          <fpage>D19</fpage>
          -1588, https://www.aclweb.org/ anthology/D19-1588
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Joty</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Carenini</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ng</surname>
          </string-name>
          , R.T.:
          <article-title>CODRA: A Novel Discriminative Framework for Rhetorical Analysis 41-3 (</article-title>
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Mann</surname>
            ,
            <given-names>W.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Thompson</surname>
            ,
            <given-names>S.A.</given-names>
          </string-name>
          :
          <article-title>Rhetorical Structure Theory: Toward a Functional Theory of Text Organization</article-title>
          .
          <source>Text</source>
          <volume>8</volume>
          ,
          <issue>243</issue>
          {
          <fpage>281</fpage>
          (
          <year>1988</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Marcus</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Marcinkiewicz</surname>
            ,
            <given-names>M.A.</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>MacIntyre</given-names>
            , R.,
            <surname>Bies</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Ferguson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Katz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            ,
            <surname>Schasberger</surname>
          </string-name>
          ,
          <string-name>
            <surname>B.</surname>
          </string-name>
          :
          <article-title>The Penn Treebank: Annotating Predicate Argument Structure</article-title>
          .
          <source>In: Proceedings of the Workshop on Human Language Technology</source>
          . pp.
          <volume>114</volume>
          {
          <fpage>119</fpage>
          . HLT '94,
          <string-name>
            <surname>ACL</surname>
          </string-name>
          (
          <year>1994</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Corrado</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dean</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>E cient Estimation of Word Representations in Vector Space</article-title>
          .
          <source>In: 1st International Conference on Learning Representations</source>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Moreno-Schneider</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bourgonje</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rehm</surname>
          </string-name>
          , G.:
          <article-title>Towards User Interfaces for Semantic Storytelling</article-title>
          . In: Yamamoto,
          <string-name>
            <surname>S</surname>
          </string-name>
          . (ed.)
          <article-title>Human Interface and the Management of Information: Information, Knowledge and Interaction Design</article-title>
          , 19th International Conference,
          <source>HCI International</source>
          <year>2017</year>
          (Vancouver, Canada). pp.
          <volume>403</volume>
          {
          <fpage>421</fpage>
          . Lecture Notes in Computer Science (LNCS), Springer (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Moreno-Schneider</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Srivastava</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bourgonje</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wabnitz</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rehm</surname>
          </string-name>
          , G.:
          <article-title>Semantic Storytelling, Cross-lingual Event Detection and Other Semantic Services for a Newsroom Content Curation Dashboard</article-title>
          .
          <source>In: Proceedings of the 2017 EMNLP Workshop: Natural Language Processing meets Journalism</source>
          . pp.
          <volume>68</volume>
          {
          <fpage>73</fpage>
          .
          <string-name>
            <surname>ACL</surname>
          </string-name>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Nie</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bennett</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goodman</surname>
          </string-name>
          , N.:
          <article-title>DisSent: Learning Sentence Representations from Explicit Discourse Relations</article-title>
          .
          <source>In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics</source>
          . pp.
          <volume>4497</volume>
          {
          <fpage>4510</fpage>
          .
          <string-name>
            <surname>ACL</surname>
          </string-name>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Ostendor</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bourgonje</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Berger</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moreno-Schneider</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rehm</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gipp</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Enriching BERT with Knowledge Graph Embeddings for Document Classi - cation</article-title>
          .
          <source>In: Proceedings of the GermEval 2019 Workshop</source>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Pagliardini</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gupta</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jaggi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Unsupervised Learning of Sentence Embeddings Using Compositional n-Gram Features</article-title>
          .
          <source>In: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          . pp.
          <volume>528</volume>
          {
          <issue>540</issue>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Prasad</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dinesh</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Miltsakaki</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Robaldo</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Joshi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Webber</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>The Penn Discourse Treebank 2.0</article-title>
          . In: In Proceedings of LREC (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Propp</surname>
          </string-name>
          , V.Y.:
          <article-title>Morphology of the Folktale</article-title>
          . Publication ... of the Indiana University Research Center in Anthropology, Folklore, and Linguistics, University of Texas Press (
          <year>1968</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Rehm</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>He</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schneider</surname>
            ,
            <given-names>J.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nehring</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Quantz</surname>
          </string-name>
          , J.:
          <article-title>Designing User Interfaces for Curation Technologies</article-title>
          . In: Yamamoto,
          <string-name>
            <surname>S</surname>
          </string-name>
          . (ed.)
          <article-title>Human Interface and the Management of Information: Information, Knowledge and Interaction Design</article-title>
          , 19th International Conference,
          <source>HCI International</source>
          <year>2017</year>
          (Vancouver, Canada). pp.
          <volume>388</volume>
          {
          <fpage>406</fpage>
          . Lecture Notes in Computer Science (LNCS), Springer (
          <year>2017</year>
          ), part I
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Rehm</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moreno-Schneider</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bourgonje</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Srivastava</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fricke</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Thomsen</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>He</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Quantz</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Berger</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , Konig, L., Rauchle,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Gerth</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Wabnitz</surname>
          </string-name>
          ,
          <string-name>
            <surname>D.</surname>
          </string-name>
          :
          <article-title>Di erent Types of Automated and Semi-Automated Semantic Storytelling: Curation Technologies for Di erent Sectors</article-title>
          . In: Rehm,
          <string-name>
            <given-names>G.</given-names>
            ,
            <surname>Declerck</surname>
          </string-name>
          , T. (eds.)
          <article-title>Language Technologies for the Challenges of the Digital Age: 27th International Conference</article-title>
          ,
          <source>GSCL 2017</source>
          , Berlin, Germany,
          <source>September 13-14</source>
          ,
          <year>2017</year>
          , Proceedings. pp.
          <volume>232</volume>
          {
          <fpage>247</fpage>
          . Lecture Notes in Arti cial Intelligence (LNAI),
          <article-title>Gesellschaft fur Sprachtechnologie und Computerlinguistik e</article-title>
          .V., Springer (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24.
          <string-name>
            <surname>Rehm</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moreno-Schneider</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bourgonje</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Srivastava</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nehring</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Berger</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , Konig, L., Rauchle,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Gerth</surname>
          </string-name>
          , J.:
          <article-title>Event Detection and Semantic Storytelling: Generating a Travelogue from a large Collection of Personal Letters</article-title>
          . In: Caselli,
          <string-name>
            <given-names>T.</given-names>
            ,
            <surname>Miller</surname>
          </string-name>
          ,
          <string-name>
            <surname>B.</surname>
          </string-name>
          , van Erp,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Vossen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Palmer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Hovy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            ,
            <surname>Mitamura</surname>
          </string-name>
          , T. (eds.)
          <source>Proceedings of the Events and Stories in the News Workshop</source>
          . pp.
          <volume>42</volume>
          {
          <fpage>51</fpage>
          .
          <string-name>
            <surname>ACL</surname>
          </string-name>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <string-name>
            <surname>Salton</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McGill</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Introduction to Modern Information Retrieval</article-title>
          . Computer Series,
          <string-name>
            <surname>McGraw-Hill</surname>
          </string-name>
          , New York (
          <year>1983</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          26.
          <string-name>
            <surname>Sanh</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Debut</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chaumond</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wolf</surname>
          </string-name>
          , T.:
          <article-title>DistilBERT, a Distilled Version of BERT: Smaller, Faster, Cheaper and Lighter</article-title>
          . CoRR pp.
          <volume>1</volume>
          {
          <issue>5</issue>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          27.
          <string-name>
            <surname>Selivanov</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          : text2vec:
          <article-title>Modern Text Mining Framework for R</article-title>
          .
          <source>Computer software manual](R package version 0.4. 0)</source>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          28.
          <string-name>
            <surname>Soricut</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Marcu</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Sentence Level Discourse Parsing using Syntactic and Lexical Information</article-title>
          .
          <source>In: Proceedings of the 2003 Human Language Technology Conference of the North American Chapter of the Association for Computational Linguistics</source>
          . pp.
          <volume>228</volume>
          {
          <issue>235</issue>
          (
          <year>2003</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          29. Van Dijk,
          <string-name>
            <surname>T.A.</surname>
          </string-name>
          :
          <article-title>News as Discourse</article-title>
          .
          <source>Routledge</source>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          30.
          <string-name>
            <surname>Vaswani</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shazeer</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Parmar</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Uszkoreit</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jones</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gomez</surname>
            ,
            <given-names>A.N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kaiser</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Polosukhin</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Attention Is All You Need</article-title>
          .
          <source>Advances in Neural Information Processing Systems</source>
          <volume>30</volume>
          (Nips),
          <volume>5998</volume>
          {6008 (jun
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          31.
          <string-name>
            <surname>Wolf</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Debut</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sanh</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chaumond</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Delangue</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cistac</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rault</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Louf</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Funtowicz</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brew</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>HuggingFace's Transformers: Stateof-the-art Natural Language Processing</article-title>
          (oct
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          32.
          <string-name>
            <surname>Yarlott</surname>
            ,
            <given-names>W.V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cornelio</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gao</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Finlayson</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Identifying the Discourse Function of News Article Paragraphs</article-title>
          .
          <source>In: Proceedings of the Workshop Events and Stories in the News</source>
          <year>2018</year>
          . pp.
          <volume>25</volume>
          {
          <issue>33</issue>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          33.
          <string-name>
            <surname>Yarlott</surname>
            ,
            <given-names>W.V.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Finlayson</surname>
            ,
            <given-names>M.A.</given-names>
          </string-name>
          :
          <article-title>ProppML: A Complete Annotation Scheme for Proppian Morphologies</article-title>
          .
          <source>In: CMN. OASICS</source>
          , vol.
          <volume>53</volume>
          , pp.
          <volume>8</volume>
          :
          <issue>1</issue>
          {8:
          <fpage>19</fpage>
          .
          <string-name>
            <surname>Schloss Dagstuhl - Leibniz-Zentrum fuer Informatik</surname>
          </string-name>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>