<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Identification of Relations between Text Segments for Semantic Storytelling</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Georg Rehm</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Malte Ostendorf</string-name>
          <email>malte.ostendorf@dfki.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Rémi Calizzano</string-name>
          <email>remi.calizzano@dfki.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Karolina Zaczynska</string-name>
          <email>karolina.zaczynska@dfki.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Julián Moreno Schneider</string-name>
          <email>julian.moreno_schneider@dfki.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>DFKI GmbH</institution>
          ,
          <addr-line>Berlin</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Semantic Storytelling is the (semi)-automatic generation of new content (storylines) based on information extracted from document collections presented with helpful visualisation techniques. This paper summarises our previous work and describes the Semantic Storytelling vision and technical approach. We describe experiments that focus on the identification of relations between text segments extracted from documents written by diferent authors; discourse relations have, so far, primarily been researched within single documents only. The results confirm our intuition that discourse parsing is dificult to apply to the inter-text level, but they are encouraging as a first step. Similarly, the identification of inter-text relations using pairwise document classification yields promising results. Lastly, we show the efectiveness of paragraph ordering for coherent story generation.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        With the ever-increasing amount of digital content, users face the challenge of coping with
enormous quantities of information. This challenge is especially true for digital content
curators, i. e., knowledge workers such as, among others, journalists [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], television producers [2],
designers [3], librarians [4], or academics [5]. These and other professional profiles have in
common that they monitor and process existing or incoming content to produce new content. In
several projects [6, 5], we have been developing technical approaches to support knowledge
workers in their day-to-day jobs, curating large amounts of mostly textual content more
eficiently and efectively. One of our focus areas is the generation of storylines. This includes
the (semi)automatic creation of new content as well as helpful presentation and visualisation
techniques. We call the approach Semantic Storytelling [
        <xref ref-type="bibr" rid="ref1">7, 8, 1, 9, 2, 10, 11</xref>
        ].
      </p>
      <p>Recently, storytelling has mostly been interpreted as a language generation task [see, e. g.,
12, 13, 14, 15], where the goal is to generate texts. We interpret the concept diferently by
concentrating on the extraction and presentation of stories and their parts, contained in content
streams, e. g., document collections or social media feeds. We see storylines as sets of building
blocks, which, depending on their combination (temporal, geographical, causal etc.), can be
(J. M. Schneider)
© 2022 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
assembled into a story in various ways. Our goal is the recognition of atomic pieces of
information (e. g., facts, propositions, events) in document collections and the identification of semantic
relations between these pieces. Applications can be conceptualised, e. g., as information systems
(for the retrieval of existing content) or recommender systems (for creating new content).</p>
    </sec>
    <sec id="sec-2">
      <title>2. Operationalising Storytelling</title>
      <p>We define Semantic Storytelling as the (semi-)automatic generation of storylines based on
information extracted from documents or social media streams which are processed, classified,
annotated and visualised, typically in an interactive way. This section describes the existing
components and utilized previous work, followed by an architecture description conceptualized
to meet the requirements of our Semantic Storytelling approach.</p>
      <sec id="sec-2-1">
        <title>2.1. Semantic Storytelling: Existing Components</title>
        <p>Three groups of text analytics components are the foundation of our architecture [see, e. g.,
7, 16, 8, 2]: (1) services that analyse documents or document collections to provide
documentlevel metadata, (2) services that extract, annotate and enrich specific parts of the incoming
content, and (3) services that transform the content, e. g., via summarization or translation.</p>
        <p>The tools are in diferent stages of technical maturity. They are orchestrated using a workflow
manager [17]. Named entity recognition and linking as well as time expression analysis are
performed to identify named entities of various types and classes. The integration of time
expression analysis allows reasoning over temporal expressions and anchoring entities and
events to a timeline. We use topic detection to assign abstract topics to individual sentences,
paragraphs, chapters, and documents. For the annotations, we use NLP Interchange Format
[NIF, 18], which allows the exploitation of the Linked Data paradigm and Linked Open Data
resources. We distinguish between diferent classes or genres of documents, i. e., we experiment
with diferent approaches for identifying document structures [ 10] and, on top of this, document
genres [19]. An ontology to represent a heterogeneous set of document characteristics, tying
together the diferent parts of annotations mentioned above, is currently under development.</p>
        <p>Incoming Content</p>
        <p>UC1</p>
        <p>Self-contained
document collection</p>
        <p>UC2
Web content</p>
        <p>UC2
RSS feeds</p>
        <p>UC3
Wikipedia
Possible instantiations of T
••••• SECCNvulaoaemmmimnmpetldoeartreeyfnadtciottycument
Prototype GUIs</p>
        <sec id="sec-2-1-1">
          <title>1 Determine the relevance of a segment for T</title>
          <p>Topic</p>
          <p>T
a Document relevance
b Segment relevance
CBA SSSeeennnttteeennnccceee 451 tReaxtnskeegd mlisetnotsf
2 Determine importance
of a segment
A isLessImportantThan
B</p>
        </sec>
        <sec id="sec-2-1-2">
          <title>C isMoreImportantThanT isMoreImportantThan</title>
          <p>3 Discourse relation between A
segment and topic
Comparison T</p>
        </sec>
        <sec id="sec-2-1-3">
          <title>Comparison B</title>
          <p>Expansion C
UC1</p>
          <p>UC2</p>
          <p>UC3</p>
          <p>We implemented a number of experimental prototypes and user interfaces [7, 9, 8, 2]. On top
of the semantic analysis of documents, we map the extracted information, whenever possible, to
Linked Open Data and visualise the result [20, 3]. By providing feedback to the output of certain
semantic services, content curators have control over the workflow. The storytelling UIs involve
the dynamic and interactive recomposition and visualisation of extracted information. This
involves arranging content elements (documents, paragraphs, sentences, events) on a dynamic
timeline or as a graph. Figure 1 shows an example that visualises a story as a graph in which
individual story units (content pieces) are represented as nodes. Diferent content types such as
topics and text segments are displayed in diferent colors. Edges represent specific relations.</p>
        </sec>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Architecture</title>
        <p>At the core of the Semantic Storytelling approach are three processing steps. Figure 2 and the
following paragraphs illustrate these steps in more detail.
2.2.1. Step 1: Determine the Relevance
We start with a topic  , instantiated through a text segment, e. g., a named entity, headline or
document. To identify content pieces relevant for  , we process an incoming content stream
and decide for each piece whether it is relevant for  . Relevance can be computed in various
ways, e. g., text similarity measures, or we can compute the overlap in terms of named entities.
The accuracy depends on the length of the segments. For example, we can perform pairwise
comparisons of document similarity starting with the seed document   of which we know that
it represents topic  and measure its similarity to other candidate documents. Document pairs
with a high similarity score are assumed to cover the same topic. If an incoming segment, e. g.,
a news article, is relevant to  or its seed document   , the next steps involve identifying the
important atomic segments (i. e., sentences or paragraphs) and determining the relations that
hold between these atomic segments and  .
2.2.2. Step 2: Determine the Importance
Given a document  related to  , we need to determine the importance of  (and its segments)
with regard to  . Various cues and indicators can be exploited, e. g., an incoming news piece
on  that was published only seconds ago and that includes the cue “BREAKING” in its title.
One way of determining the topical importance of an individual segment is to treat it as a
segment-level question answering task. Given a document  that consists of a sequence of
segments ( 1,  2, …   ), the aim is to find the segment   that contains the answer to the input
question. In our context, the input would be the topic  instead of a question. A weighting
schema could be applied such that, e. g., novel news pieces are preferred over old ones.
2.2.3. Step 3: Determine the Relation
Relations can exist between two segments on diferent levels. While we are primarily looking
at discourse or coherence relations, relations can also relate to more simple levels such as
segment order. The modeling of semantic, discourse, or coherence relations in textual content,
where propositions, statements or events are the individual units, is at the core of discourse
parsing frameworks. These analyse a text for, typically, intra-textual but not inter -sentential,
relations. We borrow from discourse parsing and experiment with PDTB2 annotations [21] (see
Section 3.1), but there are several added challenges. Crucially, our system needs to be able to
robustly find (discourse) relations of short segments extracted from diferent texts while we
have ample evidence from step 1 that the two texts are relevant to each other.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Identification of Relations between Text Segments</title>
      <p>
        While steps 1 and 2 can be considered common NLP tasks, step 3 is less standardized with only
little related work in the literature, which is why we elaborate on the identification of relations
between text segments in more detail. We summarise three separately published experiments
that represent diferent approaches to the task: PDTB2 discourse classification [ 22], Wikipedia
article relations [
        <xref ref-type="bibr" rid="ref2">23</xref>
        ], and text segment ordering [
        <xref ref-type="bibr" rid="ref3">24</xref>
        ].
      </p>
      <sec id="sec-3-1">
        <title>3.1. Experiment 1: PDTB2-based Discourse Relation Classification</title>
        <p>
          The goal of this experiment [22] is to get preliminary results regarding the (ambitious) task of
identifying discourse relations between two arbitrary text segments that are relevant to each
other and that both relate to the same topic. The experiments are based on PDTB2 [21], which
contains approx. 40,000 annotated relations. We establish a baseline training a classifier for the
four top level PDTB2 senses (Temporal, Contingency, Comparison, Expansion) or None if none of
the classes apply. We use DistilBERT to obtain contextual vector representations of the text
segments [
          <xref ref-type="bibr" rid="ref4">25</xref>
          ]. For the relation classification, we use a Siamese architecture.
        </p>
        <p>
          Table 1 shows the results of the classifier which performs best for Expansion, by far the most
frequent class. The performance is lower than state-of-the-art approaches, but a comparison is
not straightforward. First, our classification performs considerably lower than other approaches
because we have not implemented features relating to the connective. Second, we only use the
four PDTB2 top level classes. Most other approaches use a more fine-grained set, resulting in
many more classes and lower performance. See for example [
          <xref ref-type="bibr" rid="ref5">26</xref>
          ], who report an F1-score of
40.70 for implicit relations. For our experiments, however, we argue that a coarse classification,
with more training examples and higher accuracy, is better suited.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Experiment 2: Semantic Relations of Wikipedia Articles</title>
        <p>
          In this experiment [
          <xref ref-type="bibr" rid="ref2">23</xref>
          ], we transfer the relation classification from a sentence-level to a
document-level and from PDTB relations to more generic ones. Given a seed document   ,
we are interested in finding a target document   that shares the semantic relation   with   .
We model the task of finding the relation  of a document pair (  ,   ) as a pairwise multi-class
document classification problem. Wikipedia articles are utilized as documents and Wikidata
properties [
          <xref ref-type="bibr" rid="ref11">32</xref>
          ] define their semantic relations. For example, the Wikipedia article on Albert
Einstein and its Wikidata item are connected to German Empire through the property country of
citizenship. The Wikidata property acts as the relation between the article pair and the class
label in the training data for this pair of documents. We selected nine relations ranging from
country of citizenship to facet of to opposite of. Besides the number of available Wikipedia
article pairs, diversity was also a criterion in our selection with respect to the diferent semantic
meanings of properties. We evaluate six diferent methods: Document vectors from average
GloVe word vectors [
          <xref ref-type="bibr" rid="ref6">27</xref>
          ], Paragraph Vectors [
          <xref ref-type="bibr" rid="ref7">28</xref>
          ], BERT [
          <xref ref-type="bibr" rid="ref8">29</xref>
          ], XLNet [
          <xref ref-type="bibr" rid="ref10">31</xref>
          ], and Siamese variations
of BERT and XLNet [
          <xref ref-type="bibr" rid="ref12 ref9">33, 30</xref>
          ].
        </p>
        <p>
          Section 2 presents the empirical results. BERT yields the best micro average F1-score with
0.933, followed by XLNet with 0.926 F1. The vanilla Transformers, BERT and XLNet, generally
outperform their Siamese counterparts. The shared contextual information during the encoding
of document pairs most likely yields the better performance for vanilla Transformers. Even
abstract relations, like facet of, yield a considerable high F1-score (0.91 for BERT). Siamese BERT
(0.870 F1) and Siamese XLNet (0.870 F1) are even outperformed by AvgGloVe (0.875 F1) despite
AvgGloVe requiring only a fraction of the computing resources compared to the Transformer
models. A qualitative evaluation and detailed analysis is presented in [
          <xref ref-type="bibr" rid="ref2">23</xref>
          ]. Our results suggest
that pairwise classification is suitable for classifying semantic relations between documents. In
another study [
          <xref ref-type="bibr" rid="ref13">34</xref>
          ], we confirm this finding also for the domain of research papers.
        </p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Experiment 3: Text Segment Ordering</title>
        <p>
          In the last experiment, we study the relation of coherence between text segments [
          <xref ref-type="bibr" rid="ref3">24</xref>
          ]. When
segments are related to the same topic, we want to determine which of the segments (sentences
or paragraphs) should precede the other in order to maximise discourse coherence. In the best
case, the predicted order would correspond to a complete and coherent story line although
being composed of individual segments. As opposed to pairwise ordering (similar to pairwise
classification, Section 3.2), we apply direct ordering of all segments using a Pointer network
[
          <xref ref-type="bibr" rid="ref14">35</xref>
          ] combined with the pre-trained encoder-decoder model BART [
          <xref ref-type="bibr" rid="ref15">36</xref>
          ]. As baseline, we rely on
Hierarchical Attention Networks (HAN) inspired by [
          <xref ref-type="bibr" rid="ref14">35</xref>
          ] that uses a Pointer network,
MultiHead Attention and LSTMs for sequence representations. We evaluate the BART-Pointer and
the HAN baseline on two new paragraph ordering datasets tailored to Semantic Storytelling.
As opposed to common segment ordering datasets using only single sentences as segments,
for Semantic Storytelling, we are also interested in paragraphs with more than one sentence.
Hence, we construct two new datasets for paragraph ordering based on the CNN DailyMail
(CNN-DM) dataset and on Wikipedia.
        </p>
        <p>Our model outperforms the HAN baseline on both datasets. With a Perfect Match Ratio (PMR)
of 0.3699 for Wikipedia and 0.0171 for CNN-DM, the BART-Pointer combination is significantly
better than the baseline, which yields 0.2100 PMR for Wikipedia and 0.0049 PMR for CNN-DM.
An evaluation with Kendall’s Tau metric ( ) shows that for 36.99% of the test samples our model
is able to perfectly order the shufled paragraphs from the introduction of Wikipedia articles
while only 1.71% of the CNN-DM articles can be ordered perfectly. CNN-DM seems to be a
greater challenge to the model by the number of paragraphs with an average of 14.5 sequences
to order, whereas the Wikipedia dataset has an average of 6.29 paragraphs. We assume that the
introductions in Wikipedia articles is often more consistent than those in CNN-DM, thus, the
order is easier to learn. To sum up, we evaluate the BART-Pointer combination as suitable for
ordering paragraphs to create a coherent text from an unordered collection of segments.</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Discussion</title>
        <p>The (semi)automatic identification and generation of storylines from text segments is still in its
infancy. In this paper, we focus on one crucial step of our Semantic Storytelling approach, i. e.,
the identification of the relation between text segments. We approach the task from diferent
angles, i.e., the notion of relation is defined as PDTB2 discourse relation, Wikidata properties,
or order relation. The results of the PDTB2 experiments (Section 3.1) reveal a below state-of-art
performance of the Siamese BERT approach. We attribute the low performance to the inability
of the Siamese network to encode relational information of the two text segments.</p>
        <p>This assumption is confirmed by the method evaluation of the second experiment (Section 3.2)
carried out as a pairwise multi-class document classification task with more training data and
on a variety of non-discourse relations. The methods from the experiment yield substantially
higher accuracy scores compared to experiment 1. On a methodological level, vanilla
Transformer models turn out to be more suitable for the relation classification task as their Siamese
counterparts. We find that also presumable dificult relations like facet of achieve promising
results that would be suitable for our use cases. However, the pairwise document classification
approach has one drawback: the approach only classifies a single pair of segments while stories
consists of multiple segments.</p>
        <p>To address this, experiment 3 explores relation identification as a paragraph ordering task
(Section 3.3). The order relation ensures a coherent story generation, i. e., more than two
segments are arranged in a meaningful manner. In our experiment, we demonstrate that a
Pointer network in combination with an encoder-decoder model like BART, is capable of not
only ordering sentences but also text segments of paragraph length. Given the dificulty of this
task, our results are very encouraging.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Related Work</title>
      <p>This brief overview of related work refers to several areas including narratology, discourse
theory, as well as applied work in computational linguistics and language technology.</p>
      <p>Several approaches grounded in narratology address storytelling as a way of automatising the
detection of instances of story grammars [37], especially events, in texts. Caselli and Vossen [38]
present a data set for the detection of temporal and causal relations and use a plot structure [39]
to order events found in narratives or text documents, chronologically and logically. According
to Bal [39], narratives follow a plot structure that consists of ordered events, told by an agent
or author and caused or experienced by actors. Yarlott and Finlayson [40] use Propp’s (1968)
morphology of Russian hero tales for story detection and generation systems. His book, first
published in 1928, Propp analyzes the structural elements of Russian folk tales, which always
occur in a fixed, consecutive order. Yan et al. [42] describe a system that learns “functional
story schemas” as sets of functional structures (e. g., character introduction, conflict setup, etc.)
in social media narratives. They extract patterns of functional structures. Afterwards, their
formation in a story is analyzed across all stories to find schematic structures. Vice versa,
Gordon et al. [43] use stories from blog articles to perform automated causal reasoning. Bois
et al. [44] recommend articles based on simple lexical similarity. They link news articles in the
form of a graph and label links to inform users on the nature of the relation between two news
pieces. Ribeiro et al. [45] cluster news articles based on identified event instances and word
alignment. They attempt to form clusters of online articles that deal with a certain event type.
Nie et al. [46] use dependency parsing and discourse relations to determine sentence relations
by learning vector representations. Yarlott et al. [47] apply the discourse theory by Dijk [48] to
examine how paragraphs behave when used as discourse structure units in news articles.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusions and Future Work</title>
      <p>
        After laying the groundwork for Semantic Storytelling through various experiments, we are
now approaching the final phase, in which we attempt to combine the components described
in Section 2.1, including text analytics and enrichment services that operate on documents,
document collections or text segments [10], with the emerging set of technologies described
in Section 2. To this end, we identified the key building blocks for Semantic Storytelling
technologies. While step 1 can be implemented using one of several known approaches, steps 2
and 3 are much more challenging (Section 2). Our approach is grounded in the assumption that
diferent texts that deal with the same topic but that are from diferent authors and diferent sources
can be interconnected in a meaningful way through relations, which we attempt to extract
automatically. We want to support content curators and make use of these relations holding
between two segments by exposing them explicitly and exploiting them in the construction of
storylines in a semiautomatic or fully automatic way. While our experiments are promising
[
        <xref ref-type="bibr" rid="ref2 ref3">49, 24, 23</xref>
        ] they also show that additional research is needed before we can integrate the
technologies into prototypes. Data sets annotated for rhetorical or discourse structure are still
rather limited both in availability and in size. Our future work will focus on expanding our
setup, especially with regard to the analysis and classification of discourse relations and more
sophisticated processing of connectives. We will integrate a more flexible approach with regard
to the processing of single documents by concentrating on larger parts of a document including
longer summaries and paraphrased variants to increase coverage. Taking into account explicit
ontological knowledge to identify semantic relations between texts will also be an important
next step towards the completion of the envisaged Semantic Storytelling prototype [10].
      </p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>The work presented in this paper has received funding from the German Federal Ministry of
Education and Research (BMBF) through the projects QURATOR (Wachstumskern no. 03WKDA1A)
and PANQURA (no. 03COV03E).
[2] G. Rehm, J. Moreno Schneider, P. Bourgonje, A. Srivastava, R. Fricke, J. Thomsen, J. He,
J. Quantz, A. Berger, L. König, S. Räuchle, J. Gerth, D. Wabnitz, Diferent Types of
Automated and Semi-Automated Semantic Storytelling: Curation Technologies for Diferent
Sectors, in: G. Rehm, T. Declerck (Eds.), Language Technologies for the Challenges of
the Digital Age: 27th Int. Conf., GSCL 2017, Berlin, Germany, September 13-14, 2017,
Proceedings, number 10713 in Lecture Notes in Artificial Intelligence (LNAI), Springer,
2018, pp. 232–247. 13/14 September 2017.
[3] G. Rehm, J. He, J. Moreno-Schneider, J. Nehring, J. Quantz, Designing User Interfaces for
Curation Technologies, in: S. Yamamoto (Ed.), Human Interface and the Management of
Information: Information, Knowledge and Interaction Design, 19th Int. Conf., HCI Int.
2017 (Vancouver, Canada), number 10273 in Lecture Notes in Computer Science (LNCS),
Springer, 2017, pp. 388–406.
[4] C. Neudecker, G. Rehm, Digitale Kuratierungstechnologien für Bibliotheken, Zeitschrift
für Bibliothekskultur 027.7 4 (2016).
[5] G. Rehm, M. Lee, J. M. Schneider, P. Bourgonje, Curation Technologies for a Cultural
Heritage Archive: Analysing and transforming a heterogeneous data set into an interactive
curation workbench, in: A. Antonacopoulos, M. Büchler (Eds.), DATeCH 2019: Proceedings
of the 3rd Int. Conf. on Digital Access to Textual Cultural Heritage, Brussels, Belgium,
2019, pp. 117–122. 8-10 May 2019.
[6] G. Rehm, F. Sasaki, Digitale Kuratierungstechnologien – Verfahren für die efiziente
Verarbeitung, Erstellung und Verteilung qualitativ hochwertiger Medieninhalte, in:
Proceedings der Frühjahrstagung der Gesellschaft für Sprachtechnologie und Computerlinguistik
(GSCL 2015), Duisburg, 2015, pp. 138–139.
[7] P. Bourgonje, J. Moreno Schneider, G. Rehm, F. Sasaki, Processing Document Collections to
Automatically Extract Linked Data: Semantic Storytelling Technologies for Smart Curation
Workflows, in: A. Gangemi, C. Gardent (Eds.), Proceedings of the 2nd Int. Workshop
on Natural Language Generation and the Semantic Web (WebNLG 2016), The Assoc. for
Computational Linguistics, Edinburgh, UK, 2016, pp. 13–16.
[8] J. Moreno Schneider, P. Bourgonje, J. Nehring, G. Rehm, F. Sasaki, A. Srivastava, Towards
Semantic Story Telling with Digital Curation Technologies, in: L. Birnbaum, O. Popescu,
C. Strapparava (Eds.), Proceedings of Natural Language Processing meets Journalism –
IJCAI-16 Workshop (NLPMJ 2016), New York, 2016.
[9] G. Rehm, J. Moreno Schneider, P. Bourgonje, A. Srivastava, J. Nehring, A. Berger, L. König,
S. Räuchle, J. Gerth, Event Detection and Semantic Storytelling: Generating a Travelogue
from a large Collection of Personal Letters, in: T. Caselli, B. Miller, M. van Erp, P. Vossen,
M. Palmer, E. Hovy, T. Mitamura (Eds.), Proceedings of the Events and Stories in the News
Workshop, ACL, 2017, pp. 42–51.
[10] G. Rehm, K. Zaczynska, J. M. Schneider, Semantic Storytelling: Towards Identifying
Storylines in Large Amounts of Text Content, in: A. Jorge, R. Campos, A. Jatowt, S. Bhatia
(Eds.), Proceedings of Text2Story – Second Workshop on Narrative Extraction From Texts
co-located with 41th European Conf. on Information Retrieval (ECIR 2019), Cologne,
Germany, 2019, pp. 63–70. 14 April 2019.
[11] M. Raring, M. Ostendorf, G. Rehm, Semantic Relations between Text Segments for
Semantic Storytelling: Annotation Tool – Dataset – Evaluation, in: N. Calzolari, F. Béchet,
P. Blache, C. Cieri, K. Choukri, T. Declerck, H. Isahara, B. Maegaard, J. Mariani, J. Odijk,
S. Piperidis (Eds.), Proceedings of the 13th Language Resources and Evaluation Conference
(LREC 2022), European Language Resources Association (ELRA), Marseille, France, 2022,
pp. 4923–4932. June 20-25, 2022.
[12] A. Fan, M. Lewis, Y. Dauphin, Hierarchical neural story generation, in: Proc. of the 56th
Annual Meeting of the Assoc. for Computational Linguistics (Volume 1: Long Papers),
2018, pp. 889–898.
[13] A. Fan, M. Lewis, Y. Dauphin, Strategies for structuring story generation, arXiv preprint
arXiv:1902.01109, 2019.
[14] P. Xu, M. Patwary, M. Shoeybi, R. Puri, P. Fung, A. Anandkumar, B. Catanzaro,
MEGATRONCNTRL: Controllable story generation with external knowledge using large-scale language
models, in: Proceedings of the 2020 Conference on Empirical Methods in Natural
Language Processing (EMNLP), Association for Computational Linguistics, Online, 2020, pp.
2831–2845. doi:10.18653/v1/2020.emnlp-main.226.
[15] X. Hua, A. Sreevatsa, L. Wang, DYPLOC: Dynamic planning of content using mixed
language models for text generation, in: Proceedings of the 59th Annual Meeting of the
Association for Computational Linguistics and the 11th International Joint Conference on
Natural Language Processing (Volume 1: Long Papers), Association for Computational
Linguistics, Online, 2021, pp. 6408–6423. doi:10.18653/v1/2021.acl-long.501.
[16] P. Bourgonje, J. Moreno Schneider, J. Nehring, G. Rehm, F. Sasaki, A. Srivastava, Towards
a Platform for Curation Technologies: Enriching Text Collections with a Semantic-Web
Layer, in: H. Sack, G. Rizzo, N. Steinmetz, D. Mladenic, S. Auer, C. Lange (Eds.), The
Semantic Web, number 9989 in Lecture Notes in Computer Science, Springer, 2016, pp.
65–68. ESWC 2016 Satellite Events. Heraklion, Crete, Greece, May 29 – June 2.
[17] J. Moreno-Schneider, P. Bourgonje, F. Kintzel, G. Rehm, A Workflow Manager for Complex
NLP and Content Curation Pipelines, in: G. Rehm, K. Bontcheva, K. Choukri, J. Hajic,
S. Piperidis, A. Vasiljevs (Eds.), Proceedings of the 1st Int. Workshop on Language
Technology Platforms (IWLTP 2020, co-located with LREC 2020), Marseille, France, 2020, pp.
73–80. 16 May 2020.
[18] S. Hellmann, J. Lehmann, S. Auer, M. Brümmer, Integrating NLP using Linked Data, in:</p>
      <p>Proc. of the 12th Int. Semantic Web Conf., 2013. 21-25 October 2013.
[19] G. Rehm, Hypertextsorten: Definition – Struktur – Klassifikation, Books on Demand,
Norderstedt, 2007. PhD thesis in Applied and Computational Linguistics,
Justus-LiebigUniversität Giessen, 2005.
[20] J. Moreno Schneider, P. Bourgonje, G. Rehm, Towards User Interfaces for Semantic
Storytelling, in: S. Yamamoto (Ed.), Human Interface and the Management of Information:
Information, Knowledge and Interaction Design, 19th Int. Conf., HCI Int. 2017 (Vancouver,
Canada), number 10274 in Lecture Notes in Computer Science (LNCS), Springer, Cham,
Switzerland, 2017, pp. 403–421. Part II.
[21] R. Prasad, N. Dinesh, A. Lee, E. Miltsakaki, L. Robaldo, A. K. Joshi, B. L. Webber, The Penn</p>
      <p>Discourse TreeBank 2.0, in: Proc. of LREC 2008, ELRA, 2008.
[22] G. Rehm, K. Zaczynska, P. Bourgonje, M. Ostendorf, J. Moreno-Schneider, M. Berger,
J. Rauenbusch, A. Schmidt, M. Wild, J. Böttger, J. Quantz, J. Thomsen, R. Fricke, Semantic
Storytelling: From Experiments and Prototypes to a Technical Solution, in: T. Caselli,
[37] D. E. Rumelhart, Notes on a Schema for Stories, in: D. G. Bobrow, A. Collins (Eds.),
Representation and Understanding – Studies in Cognitive Science, Academic Press, New
York, San Francisco, London, 1975, pp. 211–236.
[38] T. Caselli, P. Vossen, The event storyline corpus: A new benchmark for causal and temporal
relation extraction, in: Proc. of the Events and Stories in the News Workshop, ACL, 2017,
pp. 77–86.
[39] M. Bal, Narratology: Introduction to the theory of narrative, University of Toronto Press
(1985). Trans. by Christine van Boheemen.
[40] W. V. H. Yarlott, M. A. Finlayson, ProppML: A Complete Annotation Scheme for Proppian
Morphologies, in: B. Miller, A. Lieto, R. Ronfard, S. G. Ware, M. A. Finlayson (Eds.), 7th
Workshop on Computational Models of Narrative (CMN 2016), volume 53 of OpenAccess
Series in Informatics (OASIcs), Schloss Dagstuhl–Leibniz-Zentrum für Informatik, Dagstuhl,
Germany, 2016, pp. 8:1–8:19.
[41] V. Y. Propp, Morphology of the Folktale, Originally published in 1928; Translation 1968,</p>
      <p>The American Folklore Society and Indiana University, University of Texas Press, 1968.
[42] X. Yan, A. Naik, Y. Jo, C. Rose, Using functional schemas to understand social media
narratives, in: Proc. of the Second Workshop on Storytelling, ACL, 2019, pp. 22–33.
[43] A. S. Gordon, C. A. Bejan, K. Sagae, Commonsense causal reasoning using millions of
personal stories, in: Proc. of the Twenty-Fifth AAAI Conf. on Artificial Intelligence, 2011.
[44] R. Bois, G. Gravier, E. Jamet, M. Robert, E. Morin, P. Sébillot, M. Robert, Language-based
construction of explorable news graphs for journalists, in: Proc. of the 2017 EMNLP
Workshop on Natural Language Processing meets Journalism, ACL, 2017, pp. 31–36.
[45] S. Ribeiro, O. Ferret, X. Tannier, Unsupervised event clustering and aggregation from
newswire and web articles, in: Proc. of the 2017 EMNLP Workshop: Natural Language
Processing meets Journalism, Assoc. for Computational Linguistics, Copenhagen, Denmark,
2017, pp. 62–67.
[46] A. Nie, E. Bennett, N. Goodman, Dissent: Learning sentence representations from explicit
discourse relations, in: Proc. of the 57th Annual Meeting of the Assoc. for Computational
Linguistics, ACL, 2019, pp. 4497–4510.
[47] W. V. Yarlott, C. Cornelio, T. Gao, M. Finlayson, Identifying the discourse function of news
article paragraphs, in: Proc. of the Workshop Events and Stories in the News 2018, Assoc.
for Computational Linguistics, Santa Fe, New Mexico, USA, 2018, pp. 25–33.
[48] T. v. Dijk, News as discourse, Communication Series, L. Erlbaum Associates, 1988.
[49] G. Rehm, K. Zaczynska, J. M. Schneider, M. Ostendorf, P. Bourgonje, M. Berger, J.
Rauenbusch, A. Schmidt, M. Wild, Towards Discourse Parsing-inspired Semantic Storytelling,
in: A. Paschke, C. Neudecker, G. Rehm, J. A. Qundus, L. Pintscher (Eds.), Proceedings
of QURATOR 2020 – The Conf. for intelligent content solutions, Berlin, Germany, 2020.
CEUR Workshop Proceedings, Volume 2535. 20/21 January 2020.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J. Moreno</given-names>
            <surname>Schneider</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Srivastava</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bourgonje</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Wabnitz</surname>
          </string-name>
          , G. Rehm,
          <article-title>Semantic storytelling, cross-lingual event detection and other semantic services for a newsroom content curation dashboard</article-title>
          ,
          <source>in: Proceedings of the 2017 EMNLP Workshop: Natural Language Processing meets Journalism, ACL</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>68</fpage>
          -
          <lpage>73</lpage>
          . E. Hovy,
          <string-name>
            <given-names>M.</given-names>
            <surname>Palmer</surname>
          </string-name>
          , P. Vossen (Eds.),
          <source>Computational Analysis of Storylines: Making Sense of Events, Studies in Natural Language Processing</source>
          , Cambridge University Press, Cambridge,
          <year>2021</year>
          , pp.
          <fpage>240</fpage>
          -
          <lpage>259</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ostendorf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Ruas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Schubotz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Rehm</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Gipp</surname>
          </string-name>
          ,
          <article-title>Pairwise Multi-Class Document Classification for Semantic Relations between Wikipedia Articles</article-title>
          ,
          <source>in: Proc. of the 2020 ACM/IEEE Joint Conf. on Digital Libraries (JCDL'20)</source>
          ,
          <year>2020</year>
          . arXiv:
          <year>2003</year>
          .09881.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>R.</given-names>
            <surname>Calizzano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ostendorf</surname>
          </string-name>
          , G. Rehm,
          <article-title>Ordering Sentences and Paragraphs with Pretrained Encoder-Decoder Transformers and Pointer Ensembles</article-title>
          ,
          <source>in: Proceedings of the 21st ACM Symposium on Document Engineering (DocEng</source>
          <year>2021</year>
          ),
          <article-title>Association for Computing Machinery</article-title>
          , Limerick, Ireland,
          <year>2021</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>9</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>V.</given-names>
            <surname>Sanh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Debut</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chaumond</surname>
          </string-name>
          , T. Wolf,
          <article-title>DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter</article-title>
          ,
          <source>CoRR</source>
          (
          <year>2019</year>
          )
          <fpage>1</fpage>
          -
          <lpage>5</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Dai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Taneja</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <article-title>Fine-grained structure-based news genre categorization</article-title>
          ,
          <source>in: Proc.of the Workshop Events and Stories in the News</source>
          <year>2018</year>
          , ACL,
          <year>2018</year>
          , pp.
          <fpage>61</fpage>
          -
          <lpage>67</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>J.</given-names>
            <surname>Pennington</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Socher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Manning</surname>
          </string-name>
          , Glove:
          <article-title>Global Vectors for Word Representation</article-title>
          ,
          <source>in: Proc. of the 2014 Conf. on Empirical Methods in Natural Language Processing (EMNLP)</source>
          ,
          <year>2014</year>
          , pp.
          <fpage>1532</fpage>
          -
          <lpage>1543</lpage>
          . doi:
          <volume>10</volume>
          .3115/v1/
          <fpage>D14</fpage>
          - 1162.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>Q. V.</given-names>
            <surname>Le</surname>
          </string-name>
          , T. Mikolov,
          <article-title>Distributed Representations of Sentences and Documents</article-title>
          ,
          <source>Proc. of the 31st Int. Conf. on Machine Learning</source>
          <volume>32</volume>
          (
          <year>2014</year>
          )
          <fpage>1188</fpage>
          -
          <lpage>1196</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , BERT:
          <article-title>Pre-training of Deep Bidirectional Transformers for Language Understanding</article-title>
          ,
          <source>in: Proc. of the 2019 Conf. of the North American Chapter of the Assoc. for Computational Linguistics</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>4171</fpage>
          -
          <lpage>4186</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>N19</fpage>
          - 1423.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>N.</given-names>
            <surname>Reimers</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gurevych</surname>
          </string-name>
          , Sentence-BERT:
          <article-title>Sentence Embeddings using Siamese BERTNetworks</article-title>
          ,
          <source>in: The 2019 Conf. on Empirical Methods in Natural Language Processing (EMNLP</source>
          <year>2019</year>
          ),
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Dai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yang</surname>
          </string-name>
          , J. Carbonell, R. Salakhutdinov,
          <string-name>
            <given-names>Q. V.</given-names>
            <surname>Le</surname>
          </string-name>
          ,
          <article-title>XLNet: Generalized Autoregressive Pretraining for Language Understanding</article-title>
          ,
          <source>in: Advances in Neural Inf. Processing Syst</source>
          .
          <volume>32</volume>
          ,
          <year>2019</year>
          , pp.
          <fpage>5754</fpage>
          -
          <lpage>5764</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>D.</given-names>
            <surname>Vrandecic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Krötzsch</surname>
          </string-name>
          ,
          <article-title>Wikidata: a free collaborative knowledgebase</article-title>
          ,
          <source>Commun. ACM</source>
          <volume>57</volume>
          (
          <year>2014</year>
          )
          <fpage>78</fpage>
          -
          <lpage>85</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [33]
          <string-name>
            <given-names>J.</given-names>
            <surname>Bromley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bentz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Bottou</surname>
          </string-name>
          , I. Guyon,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Lecun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Moore</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Sackinger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Shah</surname>
          </string-name>
          ,
          <article-title>Signature verification using a Siamese time delay neural network</article-title>
          ,
          <source>Int. J. of Pattern Recognition and Artificial Intelligence</source>
          <volume>7</volume>
          (
          <year>1993</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [34]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ostendorf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Ruas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Blume</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Gipp</surname>
          </string-name>
          , G. Rehm,
          <article-title>Aspect-based document similarity for research papers</article-title>
          ,
          <source>in: Proceedings of the 28th International Conference on Computational Linguistics</source>
          ,
          <source>International Committee on Computational Linguistics</source>
          , Barcelona,
          <source>Spain (Online)</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>6194</fpage>
          -
          <lpage>6206</lpage>
          . URL: https://aclanthology.org/
          <year>2020</year>
          .coling-main.
          <volume>545</volume>
          . doi:
          <volume>10</volume>
          . 18653/v1/
          <year>2020</year>
          .coling- main.545.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [35]
          <string-name>
            <given-names>T.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wan</surname>
          </string-name>
          ,
          <article-title>Hierarchical attention networks for sentence ordering</article-title>
          ,
          <source>in: Proc. of the AAAI Conf. on Artificial Intelligence</source>
          , volume
          <volume>33</volume>
          ,
          <year>2019</year>
          , pp.
          <fpage>7184</fpage>
          -
          <lpage>7191</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [36]
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ghazvininejad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mohamed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          , L. Zettlemoyer, Bart:
          <article-title>Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension</article-title>
          , arXiv (
          <year>2019</year>
          ) arXiv-
          <fpage>1910</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>