<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Method for Coherent Storyline Extraction via Maximum Capacity Path Optimization</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Fausto German</string-name>
          <email>fgermanj@vt.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Brian Keith</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Chris North</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Narrative Extraction, Coherence Graph, Information Extraction, Information Retrieval, Sensemaking</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>In: R. Campos, A. Jorge, A. Jatowt, S. Bhatia, M. Litvak (eds.): Proceedings of the Text2Story'25 Workshop</institution>
          ,
          <addr-line>Lucca</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Universidad Católica del Norte</institution>
          ,
          <addr-line>Av. Angamos 0610, Antofagasta, 1270709</addr-line>
          ,
          <country country="CL">Chile</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Virginia Tech</institution>
          ,
          <addr-line>Blacksburg, Virginia 24061</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>Traditional information retrieval is primarily concerned with finding relevant information from large datasets without imposing a structure within the retrieved pieces of data. However, structuring information in the form of narratives-ordered sets of documents that form coherent storylines-allows us to identify, interpret, and share insights about the connections and relationships between the ideas presented in the data. Despite their significance, current approaches for algorithmically extracting storylines from data are scarce, with existing methods primarily relying on intricate word-based heuristics and auxiliary document structures. Moreover, many of these methods are dificult to scale to large datasets and general contexts, as they are designed to extract storylines for narrow tasks. In this paper, we propose Narrative Trails, an eficient, general-purpose method for extracting coherent storylines in large text corpora. Specifically, our method uses the semantic-level information embedded in the latent space of deep learning models to build a sparse coherence graph and extract narratives that maximize the minimum coherence of the storylines. By quantitatively evaluating our proposed methods on two distinct narrative extraction tasks, we show the generalizability and scalability of Narrative Trails in multiple contexts while also simplifying the extraction pipeline. The code for our algorithm, evaluations, and examples are available at https://github.com/faustogerman/narrative-trails ∗Corresponding author.</p>
      </abstract>
      <kwd-group>
        <kwd>Optimization</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        In the last couple of decades, the fields of data science and data analytics have seen significant growth,
helping people make sense of large, complex, and often interwoven data. A common task in the
sensemaking process is to structure data in a format that aids analysis and information retrieval for
downstream tasks [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. For example, structuring information in the form of narratives can help scientists
communicate advanced ideas to the general public [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ] and can aid with finding information in
collaborative settings [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. That is, narratives serve as tools for structuring complex datasets into
coherent, manageable units that facilitate more efective communication and understanding of the
information, ultimately reducing the cognitive load needed to make sense of information [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. By
organizing disparate data points into narrative structures, we enable people to identify underlying
patterns, connections, and themes that might not be immediately evident by the data. For instance,
placing documents in a sequential storyline may help a student researching the relationship between
“computer vision” (CV) and “natural language processing” (NLP) to discover that “image captioning”
and “visual question answering” bridge the concepts of CV and NLP.
      </p>
      <p>
        However, despite the importance of narrative extraction from data, eficient algorithmic approaches
are scarce, with current methods primarily relying on intricate word-based heuristics [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] and linear
programming formulations [
        <xref ref-type="bibr" rid="ref5 ref7">5, 7</xref>
        ] with limited scalability and versatility. To solve some of the scalability
and availability issues of current methods, in this work we propose Narrative Trails, an algorithm for
extracting coherent storylines from text documents. This helps to relate potentially disconnected ideas
https://faustogerman.com (F. German)
      </p>
      <p>CEUR</p>
      <p>
        ceur-ws.org
and extract information from large datasets. To achieve this, we approach the narrative extraction
problem as a maximum capacity path optimization problem. Specifically, we utilize Dijkstra’s algorithm
with a MaxiMin objective to identify  distinct paths that maximize the minimum coherence between
documents in a sparse coherence graph representation extracted from the dataset. Following the
metaphor of Narrative Maps [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], the Narrative Trails algorithm is analogous to route maps for hiking
trails, which often provide a multiple adjacent path between a starting point and a destination.
      </p>
      <p>To assess the performance of our proposed method, we present a quantitative comparison of the
Narrative Trails algorithm against a random sampling, shortest simple path, and a simplified version of
the Narrative Maps extraction method on two distinct tasks across four datasets. Our analyses show
that the proposed approach has a lower computational cost and better performance in terms of narrative
coherence.</p>
      <p>In this paper, we make the following contributions to computational narrative extraction: (1)
Approach: We provide a more abstractive approach to narrative extraction, focused primarily on the
abstract semantic relationships between documents in a dataset; (2) Algorithm: We describe an eficient
algorithm that merges dimensionality reduction and path optimization for coherent storyline extraction
from large datasets; and (3) Extensibility: We provide a repository with details of our algorithm that
can be used to reproduce our results or to extend our methods to other contexts beyond text.</p>
      <p>In the next section, we review related work on narrative extraction. Section 3 formalizes the Narrative
Trails extraction method and its core components. Sections 4 and 5 present and analyze our evaluation
results across multiple extraction tasks. Finally, we discuss the limitations of our work and potential
lines of future research in Section 6, followed by conclusions in Section 7.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        Computational narrative extraction lies at the intersection of artificial intelligence, natural language
processing (NLP), and combinatorial optimization. The interdisciplinary nature of computational
narratives makes them useful for knowledge discovery and information synthesis from large sets of
data, as they allow the extraction of structured content from unstructured content. Moreover, recent
developments in NLP and deep learning models have driven computational narrative extraction to a
notable area of scholarly discussion [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>
        Many frameworks, models, and algorithms have been developed for computational narrative
extraction, including linear sequences [
        <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
        ] or timelines [
        <xref ref-type="bibr" rid="ref10 ref9">9, 10</xref>
        ], parallel stories [
        <xref ref-type="bibr" rid="ref11 ref12">11, 12</xref>
        ], and directed
acyclic graphs [
        <xref ref-type="bibr" rid="ref13 ref14 ref7">7, 13, 14</xref>
        ]. Each of these revolves around the idea of storylines that weave narrative
elements—sentences, documents, or clusters of documents—into coherent sequences of events. For
instance, in the Narrative Maps [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] and Metro Maps [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] approaches, the authors build directed acyclic
graphs that interconnect multiple storylines into single narratives with potentially many starting
and ending events around particular subjects. These approaches aim to extract the underlying graph
structure of documents through optimization techniques that prioritize the coherence of the storylines
while adhering to the global structural constraints of the narratives.
      </p>
      <p>
        Similarly, the work of Xu et al. [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] aims to build multiple timelines or storylines parallel to one
another that share similarities or revolve around a complex topic. On the other hand, The newsLens
algorithm [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] builds multiple parallel timelines across an entire dataset that may or may not share
common themes. These methods underscore the importance of narrative frameworks in providing a
multifaceted view of complex topics, allowing for a richer, more nuanced understanding of events and
their interconnections while acknowledging the separation in ideas between each storyline.
      </p>
      <p>
        Finally, in the Connect the Dots approach [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], the authors focus on extracting singular sequences of
events that together form a narrative chain between two fixed endpoints. This method only considers
one storyline at a time, displaying a linear narrative progression of the most relevant documents from
a defined start to an end. They achieve this by optimizing a set of word-based linear programming
constraints that maximize the coherence of the chain.
      </p>
      <p>In our work, we also focus on extracting singular narrative chains. However, we focus on an
       RAG
       Computer Vision</p>
      <p>User-Selected Endpoints
e
c
a
p
S
t
n
e
t
a
L</p>
      <p>Angular Similarity
Projection Space
UMAP + HDBSCAN
y
itr
a
li
m
i
S
c
i
p
o
T
h
h
p
a
r
G
e
c
n
e
r
e
o
C</p>
      <p>Computer Vision</p>
      <p>Natural Language Processing
Path Optimization
Redundancy Reduction</p>
      <p>Narrative Trail</p>
      <p>SStotoryrlyinline e 
Main</p>
      <p>Storyline
          C ComompuptuetreVrVisiisoinon
       Computer Vision
       Object Detection
       Multimodal Learning
         NLP
         NVLisPual QA
       Text Generation
       NLP
that connect them with maximum capacity for coherence.
abstractive approach based on the semantics of a piece of text rather than the activations of individual
words within the text and their exact appearance across the narrative. This allows us to extract chains
between documents that may not use the exact wording but nevertheless share similar themes. Moreover,
given enough data, our abstractive approach can find smooth transitions between unrelated source and
target documents.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology: Narrative Trails</title>
      <p>In this section, we introduce the formal definition of a Narrative Trail, which serves as the foundation
for the structure of our narrative extraction methods.</p>
      <p>Definition 1 (Narrative Trail). Let  = ( , ,</p>
      <p>) be a weighted, directed graph built from a set of
documents  = {
(,  ) ∈ 
1,  2, ...,   }, with each node  ∈ 
having an associated document   ∈ 
and edges
weighted through some coherence function   (,  ) =  (
 ,   ) ≥ 0. A Narrative Trail ℳ is

a collection of storylines { 1,  2, ...,   } in  , where each storyline   = {  ,  1

,  2 , ..., 
sequence of documents that maximize the minimum edge weight   (,  )


user-selected endpoints ,  ∈  . Documents  1</p>
      <p>,  2 , ..., 
 are unique to their storylines.</p>
      <p>,   } is an ordered
for all (,  ) ∈ 
to connect</p>
      <p>
        Our algorithm is a simplification of Narrative Maps [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] for the single-pair narrative extraction case
with up to  distinct storylines between the source and target documents. However, unlike previous
work, Narrative Trails does not rely on linear programming to extract storylines. Instead, the algorithm
focuses on the latent spaces learned by deep learning models for path optimization, with the coherence
graph structure and its maximum spanning tree being natural representations of the connection between
documents in this space. More concretely, we divide the algorithm into projection space construction,
coherence graph representation, and path optimization. Figure 1 provides an overview of the extraction
pipeline and the following subsections provide more details about each step.
      </p>
      <sec id="sec-3-1">
        <title>3.1. Projection Space Construction</title>
        <p>
          Our narrative extraction algorithm relies on a low-dimensional data representation constructed from
the latent space learned by a deep learning model to discover topics for each document, which we
achieve through a neural topic modeling method similar to BERTopic [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ]. More specifically, we use
the text-embedding-3-small embedding model from OpenAI’s API service. While OpenAI’s
secondgeneration model, text-embedding-ada-002, had previously shown great performance in various
NLP tasks [
          <xref ref-type="bibr" rid="ref16 ref17">16, 17</xref>
          ], their latest embedding models outperform the previous models in benchmarks for
English tasks and multi-language retrieval [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ]. We also note that other embedding models can produce
similar results. For instance, the all-mpnet-base-v2 model from the Sentence Transformers library
[
          <xref ref-type="bibr" rid="ref19">19</xref>
          ] has shown state-of-the-art performance in multiple embedding tasks such as semantic similarity
and retrieval [
          <xref ref-type="bibr" rid="ref20 ref21">20, 21</xref>
          ], providing an excellent balance between speed and performance for embedding
extraction. However, it is limited to a smaller context length of 384 tokens, restricting the amount of
information that can be captured within the embedding representations. OpenAI’s models do not have
this limitation, as they allow API calls with thousands of tokens at a time.
        </p>
        <p>
          We then use a combination of the Uniform Manifold Approximation and Projection (UMAP) algorithm
[
          <xref ref-type="bibr" rid="ref22">22</xref>
          ] and HDBSCAN [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ] to project the data into two dimensions and assign clusters to each document.
UMAP has been shown to outperform other dimensionality reduction techniques such as t-SNE [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ] and
PCA by preserving more of the local and global structure of the embeddings [
          <xref ref-type="bibr" rid="ref25 ref26">25, 26</xref>
          ]. Since HDBSCAN
is a density-based clustering algorithm, UMAP also serves as a complementary technique by creating
low-dimensional projections where documents are more densely grouped, thereby mitigating the efects
of the curse of dimensionality [
          <xref ref-type="bibr" rid="ref27">27</xref>
          ]. Additionally, HDBSCAN’s soft-clustering feature allows us to obtain
cluster probability distribution vectors, which capture the probability of each document belonging
to each of the discovered clusters. Comparing the topic distributions of two documents then allows
us to quantify their topic similarity [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ], which is an important component for ensuring the extracted
storylines follow smooth topic transitions.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Coherence Graph Representation</title>
        <p>
          Narrative Trails builds storylines by connecting documents based on the content similarity encoded
in their embeddings. A naive approach to constructing the storylines is to let some storyline   be
the shortest Euclidean path in the latent space between the source and target documents. However,
using spatial information from the latent space alone does not provide enough structure for narrative
extraction. This is because as the size of the dataset approaches infinity, finding a path between two
embeddings using only their Euclidean distance would equate to finding a path of shortest distance in
ℝ , which would resemble a straight line between the source and target points [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ]. However, deep
learning models do not learn globally linear embedding spaces. Therefore, a straight line through
the latent space may not capture the semantics of the documents well enough to define a coherent
narrative. Instead, we need to define a quantity that measures how plausible it is for two documents to
be connected in a storyline based on the semantics encoded in the latent space. That is, we need to
define a measure for document pairwise coherence.
        </p>
        <sec id="sec-3-2-1">
          <title>3.2.1. Base Coherence</title>
          <p>
            Building on the premise that documents within a narrative should share content and context similarity
[
            <xref ref-type="bibr" rid="ref29 ref30 ref5">5, 29, 30</xref>
            ], prior work [
            <xref ref-type="bibr" rid="ref7">7</xref>
            ] defines the coherence (  ,   ) between two documents   and   as the
geometric mean of their angular similarity in the high-dimensional embedding space and their topic
similarity in the low-dimensional projection space. Formally, the coherence is defined as:
(  ,   ) = √(  ,   ) (  ̂ ,  ̂ ) = √(1 − arccos(cos_sim(  ,   ))/ ) (1 − JSD( ̂ ,  ̂ ))
(1)
where   ,   ∈ ℝ are the high-dimensional embeddings in the latent space for documents   and   ,
and  ̂ ,  ̂ ∈ ℝ are their low-dimensional projections (with  ≪  ). The angular similarity (  ,   )
maps the angle between the embeddings to a similarity measure in the interval [
            <xref ref-type="bibr" rid="ref1">0, 1</xref>
            ], and the topic
similarity  (  ̂ ,  ̂ ) is defined using the Jensen-Shannon divergence (JSD) between the cluster membership
distributions of the documents in the low-dimensional space. We obtain the cluster membership
distribution vectors from HDBSCAN during the projection space construction step. Thus, two documents
are considered highly coherent with respect to one another if they exhibit both high content similarity
from (  ,   ) and high topic similarity from  (  ̂ ,  ̂ ).
          </p>
        </sec>
        <sec id="sec-3-2-2">
          <title>3.2.2. Sparse Coherence</title>
          <p>
            We note that the base coherence results in a complete, undirected graph  = ( , ,   ), where the
weights   are defined by the function (  ,   ). However, we can induce sparsity into  by leveraging
properties of its maximum spanning tree MaxST , which is the inverse of the minimum spanning tree
[
            <xref ref-type="bibr" rid="ref31">31</xref>
            ]. Specifically, we observe that in an undirected graph, any  - path in the maximum spanning tree
is also a path that maximizes the minimum edge weight between  and  in the original graph  [32].
This follows from the fact that MaxST is constructed by selecting the highest-weight edges while
maintaining connectivity [
            <xref ref-type="bibr" rid="ref31">31</xref>
            ], which ensures that for any two nodes, the weakest edge along their path
is as strong as possible.
          </p>
          <p>Given the equivalence between the storylines in the original graph and those in its maximum spanning
tree, a tree-based approach serves as an additional optimization strategy by reducing the search space
from the source node to the target during the storyline extraction phase. However, while it is possible to
use MaxST directly to identify the storylines   , trees inherently provide only a single path between any
two vertices. This limitation conflicts with the requirement to extract multiple storylines, as outlined in
Definition 1. Consequently, this structure may be too restrictive in scenarios where multiple storylines
are necessary. For example, extracting the top- distinct storylines can aid in sensemaking, as each
storyline may reveal a diferent perspective on the relationship between the endpoints. To address this,
we extend the base coherence function by introducing sparse coherence, which is defined as follows to
balance sparsity with the number of possible storylines between two documents:
 (  ,   ) =
[ ≠  and (  ,   ) ≥   ] (  ,   )
(2)</p>
          <p>
            Where  is the bottleneck edge weight of MaxST . The parameter  ≥ 0 scales the bottleneck weight
to set a hard cutof on the minimum coherence between any two documents. It follows from Equation 2
and Kruskal’s algorithm [
            <xref ref-type="bibr" rid="ref31">31</xref>
            ] for maximum spanning trees that if  &gt; 1 , MaxST becomes disconnected
and therefore the resulting sparse graph may also become disconnected. This is because a value of
 &gt; 1 has the efect of raising  past the bottleneck weight that holds together MaxST . In contrast, if
 ≤ 1 , the tree remains connected, and we can construct at least one storyline between two nodes in  .
          </p>
        </sec>
        <sec id="sec-3-2-3">
          <title>3.2.3. Incorporating Task-Specific Information</title>
          <p>While so far we have only discussed undirected graphs, Equation 2 allows us to model complex
constraints through task-specific information that induce explicit directionality to the storylines. For
instance, time dependencies between documents can be enforced by refining the sparse coherence
definition to include an additional condition  (  ,   ) within the indicator function. The function  (  ,   )
returns true if and only if the date of document   is later than that of document   , ensuring that the
extracted storylines follow chronological order. To that end, the final step in the Coherence Graph
Representation is to construct a weighted, directed coherence graph   = ( , ,   ), where nodes  ∈ 
represent the documents in our dataset and edges (,  ) ∈  with weights   (,  ) are formed based on
the sparse coherence  (  ,   ) between documents. Specifically, an edge from node  to node  exists if
and only if   (,  ) &gt; 0 .</p>
        </sec>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Path Optimization</title>
        <p>Recall that our objective is to extract a collection of  distinct storylines that maximize the minimum
edge weight to connect a source document   to a target document   in a weighted directed graph  . By
letting  equal the sparse coherence graph   constructed in section 3.2, the extracted storylines then
aim to optimize the minimum edge value or, equivalently, maximize the minimum coherence required
to connect a predefined source document   to a target document   .</p>
        <p>The problem outlined mirrors the Maximum Capacity Path problem [32], aiming to maximize the
minimum edge weight within a graph. While linear-time algorithms exist for solving this problem [33],
they require additional constraints or assumptions about the graph, such as undirected graphs with
distinct edges or weights in ℕ [34]. Our coherence graph does not meet either of those requirements
since it is a directed graph with real-valued edge weights from the sparse coherence between nodes.
Given these constraints, we repurposed Dijkstra’s algorithm [35] with a single-pair path-finding task
and a maximin objective for storyline extraction. In particular, this version of Dijkstra’s algorithm
updates the tentative score of a node by taking the minimum between the current path’s minimum
edge weight and the edge weight leading to the node.</p>
        <sec id="sec-3-3-1">
          <title>3.3.1. Extracting  Distinct Chains</title>
          <p>To identify the top  distinct storylines from the source node  to the target node  , we modify Dijkstra’s
algorithm to exclude nodes that have already been included in previously discovered paths. Specifically,
after each execution of the algorithm, we record the set of nodes   = ⋃=1 (  ∖ {  ,   }) representing the
nodes contained in each previously discovered path   and update the graph to ignore these nodes in
subsequent extractions. That is, we iteratively run the modified algorithm  times, each time operating
on the updated graph  ′ = ( ∖   ,  ′), where  ′ includes only edges between the remaining nodes.
By efectively “hiding” these nodes in subsequent extractions, we prevent them from being part of any

new storylines.</p>
        </sec>
        <sec id="sec-3-3-2">
          <title>3.3.2. Redundancy Reduction</title>
          <p>
            An inherent property of spanning trees is that the paths between vertices may not always be the most
direct, often resulting in longer routes through the graph [
            <xref ref-type="bibr" rid="ref31">31</xref>
            ]. This occurs because eliminating cycles
reduces the number of possible shortcuts between nodes. In our pipeline, this means that the extracted
storylines can sometimes be excessively long or include redundant documents. To address this issue,
we implement a fast post-processing step that removes redundant documents from the storylines.
          </p>
          <p>For each extracted storyline   , we examine consecutive triplets of documents (, , )
and calculate
the sparse coherence values between them. We let  be the geometric mean of the base coherence
values  (, )
and  (, )</p>
          <p>. Since  is, by definition, greater than or equal to the minimum edge weight
  in the storyline, we only check whether  (, ) ≥  − 
if the edge (, )
exists in the sparse
coherence graph, where  is a redundancy threshold parameter. If this is true, we consider document 
to be redundant and remove it from the storyline. This process creates a shortcut from  to  without
significantly compromising the overall coherence, resulting in potentially more concise storylines.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments &amp; Evaluations</title>
      <p>In this section, we implement our proposed Narrative Trails pipeline and evaluate its performance
on two distinct narrative extraction tasks to address the following questions: (RQ1) How well does
Narrative Trails align with human-derived shortest semantic paths? and (RQ2) How do the storylines
extracted by Narrative Trails compare to those extracted by the current state-of-the-art method? Our
goal with these evaluations is to demonstrate the generalizability of Narrative Trails across multiple
domains and tasks, as well as its adaptability to various sensemaking scenarios. Additionally, we
illustrate how the flexibility of our sparse coherence definition allows us to incorporate task-specific
information and constraints into the extraction pipeline.</p>
      <sec id="sec-4-1">
        <title>4.1. Evaluation Metrics</title>
        <p>To evaluate the intrinsic quality of the storylines, we (1) measure the minimum coherence within each
storyline to verify how well our method maximizes the weakest link in the chain and (2) calculate the
geometric mean of the coherence values of the edges in the storylines. We call this the “reliability” of
the storyline as it indicates the likelihood that the documents collaboratively form a coherent storyline.</p>
        <p>To evaluate the similarity between two storylines of potentially diferent lengths, we first identify
the Dynamic Time Warping (DTW) path [36] between them using the Euclidean distance between the
low-dimensional embeddings of their documents as the matching metric and normalize by the number
of matches in the path (referred to hereafter as nDTW Distance), then compute the average pairwise
cosine similarities between those embeddings along the resulting DTW path (referred to hereafter as
DTW Similarity). Dynamic Time Warping is commonly used in time series search [37] and as a metric
for curve similarity [38]. In this context, the DTW metrics measure the semantic alignment between
storylines, which we represent as curves in the low-dimensional projections. This allows us to quantify
the similarity between storylines of diferent lengths that share related but distinct documents.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Experimental Setup</title>
        <p>We used four datasets and two tasks to evaluate our proposed methods. To answer RQ1, we use a
subset of human-derived paths over the Wikipedia network through the Wikispeedia game [39]. In this
game, users are tasked with finding a path between two Wikipedia pages in as few clicks as possible.
Although the objective of the game—to find a shortest path—is slightly diferent than the goal of our
proposed methods—to find a path of maximum semantic capacity—the paths extracted by humans in
the Wikispeedia game have been shown to encode context about how humans perceive and explore
information, especially in relation to the semantics of the network and the assigned goal page [40]. To
that end, we select a subset of 10,607 finished paths from the dataset (covering 3,928 Wikipedia pages)
with lengths between 7 and 20 pages per path and no back links to use as a ground truth dataset.</p>
        <p>To answer RQ2, we use a collection of 540 news articles related to the COVID-19 pandemic and the
2021 Cuban protests [41], 840 randomly sampled research articles from the VisPub dataset [42], and
1,140 randomly sampled research articles related to machine learning and AI from the AMiner dataset
[43]. These datasets feature various subtopics under a single main topic, allowing for a more focused
evaluation of our experiments. In addition, we selected the AMiner and VisPub subsets at random to
minimize any bias in the subtopic distributions that could provide an advantage to any of the algorithms.
We implemented the Narrative Maps algorithm as a baseline by removing the coverage constraint from
its linear programming formulation and setting the expected length for the main storyline extraction to
12 documents, ensuring a fair comparison with Narrative Trails.</p>
        <p>Since OpenAI’s text-embedding-3-small model provides 1536-dimensional embeddings, we project
them to a 48 dimensions using UMAP before clustering with HDBSCAN. In all experiments, we
implemented the proposed Narrative Trails algorithm and used the default values of  = 1 and  = 0.2
for the sparse coherence formulation and redundancy reduction, respectively. The choice of  = 1
follows from Equation 2, ensuring that the graph remains connected by the critical edge weight  .
Additionally, our empirical analysis shows an average critical coherence value of 0.58 ± 0.104 across
datasets. Based on this, we set  = 0.2 to balance coherence and flexibility, preventing the storylines
from becoming overly incoherent while allowing some variation in the documents considered redundant.
In the Wikispeedia experiments, we incorporated the directed edges of the Wikipedia network as an
additional task-specific constraint within the sparse coherence formulation. When comparing against
Narrative Maps, we used the publication dates of the articles to enforce directionality on the edges, as
detailed in section 3.2.3. Additionally, we benchmarked Narrative Trails against a random sampling
method and a shortest simple path algorithm across all experimental setups.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Results</title>
      <sec id="sec-5-1">
        <title>5.1. Alignment with Human-Derived Paths</title>
        <p>For our Wikispeedia evaluations, we extracted the top  = 3 distinct paths between the source and
target documents in each of the ground truth human-extracted storylines. We average the scores of
the top- storylines to provide a sense of the cumulative extraction quality. Table 1 summarizes the
results of this experiment. In most cases, Narrative Trails outperforms the random sampling and simple
shortest path algorithms. However, in the case of nDTW Distance, the shortest simple paths outperform
our methods since it more closely models the underlying task of the Wikispeedia game by node count.</p>
        <p>Evaluations of how humans navigate information networks using a directed  -task demonstrate that
users typically visit hub nodes—Wikipedia pages with many incoming and outgoing links—before zeroing
in on the target document [40]. Building on this observation, we investigated whether Narrative Trails
could emulate similar behavior by multiplying each node’s base coherence by their closeness centrality
Comparison of absolute coherence and reliability, along with DTW similarity and distance for the top- extracted
storylines between Narrative Trails, Narrative Trails with closeness centrality (CC), and shortest simple path
using the human-derived paths from the Wikispeedia dataset as ground truth.
[44] score along outgoing edges. As illustrated in Table 1, Narrative Trails with closeness centrality
becomes the second-best performer in most cases of DTW Similarity (underlined), demonstrating the
lfexibility of our algorithm to approximate human sensemaking.</p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Comparison with Narrative Maps</title>
        <p>For our comparison with Narrative Maps, we extracted the top storyline with Narrative Trails and
the equivalent main storyline with Narrative Maps for 50 randomly sampled source-target pairs from
each dataset. Table 2 summarizes the results of our experiments with Narrative Trails, using Narrative
Maps as a baseline algorithm. In all datasets, Narrative Trails extracts storylines with higher coherence
and reliability. Moreover, when compared against the random sampling and shortest simple paths
methods with the results form Narrative Maps as ground truth, Narrative Trails extracts storylines that
semantically align better with the state-of-the-art Narrative Maps algorithm.</p>
        <p>Similar to our coherence evaluation, we measured the execution time of Narrative Trails and Narrative
Maps on all datasets. Since each dataset had a diferent number of documents, we could get a sense for
how well each of the algorithms scales during the storyline extraction phase. Specifically, we measured
the time it takes the algorithms to extract storylines and disregarded the time required to construct the
projection space and coherence graphs. These excluded steps are considered preprocessing tasks in
both algorithms, involving non-critical and cacheable operations from the end user’s perspective.</p>
        <p>Our simplified narrative extraction pipeline substantially speeds up our algorithm’s performance
compared to Narrative Maps. Figure 2 shows the average execution time per extracted storyline for both
algorithms on the News Articles, VisPub, and AMiner datasets. We also note that since Narrative Maps
is required to extract many storylines within the same narrative, we have divided the total extraction
time by the number of storylines extracted (orange chart) to maintain a fair evaluation. However, we
also show the execution times for the full extraction of each narrative map (red chart). These results
demonstrate the eficiency of Narrative Trails when compared to Narrative Maps, especially for datasets
with thousands of documents.</p>
        <p>Lastly, Figure 3 showcases example storylines extracted by both algorithms to connect the first
reported death from the SARS-CoV-2 virus to airlines worldwide canceling flights to China. While both
storylines track the development and international response to the coronavirus outbreak, Narrative
Trails emphasizes the specific developments within the health crisis that prompted an immediate
response from airlines, making the storyline focused on the cause (the spread of the virus) and the
efect (the suspension of flights).</p>
      </sec>
      <sec id="sec-5-3">
        <title>5.3. Statistical Analyses</title>
        <p>In most cases, our base Narrative Trails algorithm, without redundancy reduction, shows statistically
significant diferences across all metrics compared to other methods. However, in some instances—
specifically when compared to shortest paths—the results for DTW Similarity and Distance for the
redundancy reduced Narrative Trails are not statistically significant. In Tables 1 and 2, we mark such
cases with †. This suggests that our redundancy reduction algorithm causes the maximum capacity
path to approximate the shortest simple path between the source and target documents. Additional
details on the statistical analyses are available in the linked GitHub repository.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Limitations and Future Work</title>
      <p>Evaluation Limitations We note that the evaluations with the user-extracted Wikispeedia paths as
a ground-truth dataset may be dificult to interpret, as the task of the game difers slightly from the
objective of narrative extraction. Thus, performing a user study or an expert-based analysis of our
proposed algorithm could help to verify and contextualize its broader usability and alignment to human
perception of storyline coherence.</p>
      <p>
        Similarly, we do not include methods such as Connect the Dots [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] and newsLens [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] in our
evaluations. This is because the highly focused nature of these algorithms and their code availability
make it dificult to incorporate them into our generalized evaluation pipeline. Additionally, our focus on
single storylines in the evaluations against Narrative Maps, as opposed to the more complex narrative
structures possible with other algorithms, points to future work to develop general narrative evaluation
metrics that can take into account multiple interconnected storylines.
      </p>
      <p>Lastly, while our evaluations report coherence—which our algorithm optimizes—as a key metric,
the additional DTW metrics used in our evaluations do not rely on our optimized coherence function.
Instead, DTW measures sequence similarity based on structural alignment, providing an independent
assessment of storyline similarity against the Wikispeedia and Narrative Maps baselines.
Comparison Issues with Narrative Maps For a fair comparison against our methods, we removed
the coverage constraint from the linear programming formulation of the Narrative Maps algorithm.
The reason for this removal was twofold: (1) Narrative Trails does not include coverage constraints in
its formulation, focusing solely on coherence; and (2) adding the coverage constraints is a bottleneck of
the Narrative Maps extraction algorithm when the number of clusters is high, as in our evaluations,
which would inflate the execution times unfairly. In practical terms, removing these constraints is
equivalent to requiring a minimum average coverage of 0% in the Narrative Maps extraction algorithm.
In this context, we note that removing the coverage constraints can lead Narrative Maps to focus on a
single topical cluster as it no longer needs to address the diversity requirements in topical coverage,
which reduces its overall performance.</p>
      <p>Another relevant point of comparison is that, in Narrative Maps, the extraction process can be guided
by the user through semantic interactions that can directly alter its linear programming constraints.
This approach provides a direct way to guide the narrative extraction and generate relevant narratives
for the user. In contrast, Narrative Trails requires constraints and task-specific information to be
encoded during pre-processing. Future research could explore semantic interaction models that enable
users to dynamically modify and update the sparse coherence graph during extraction. For instance,
deep-learning-based search agents could encode human-controllable constraints directly into its learned
parameters, ofering a balance between speed, accuracy, and user-defined narrative requirements as
part of the semantic interaction pipeline.</p>
      <p>Potential Extensions to Image Data Lastly, we note that the Narrative Trails algorithm is data and
model-agnostic and can be extended to any context where embeddings are available. Beyond text, a
trivial extension of this algorithm could extract “concept narratives” from image data that transition
the concepts in a source image to the concepts of a target image. Applications of this extension
include guided storyboarding and multi-modal narrative generation. However, these scenarios require
considering the semantics of the data, as well as the representation power of the deep learning model
used for embedding extraction, information retrieval, and similarity search in the context of computer
vision and multi-modal learning.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Conclusions</title>
      <p>In this paper, we presented Narrative Trails, a method for storyline extraction from large datasets with
task-specific graph structures. This method addresses the limitations of current algorithms by providing
an abstractive approach to narrative extraction. More specifically, we leverage the representational
power of deep learning models to capture the semantics of data to form storylines. Our main insight
stems from the parallels between our definition of coherent storylines and the maximum capacity path
problem. The results from our experiments on human-derived paths from the Wikispeedia dataset and
the comparative study with Narrative Maps demonstrate the ability of our presented algorithm to not
only generalize to task-specific domains, but also to extract storylines with high coherence. Moreover,
our simplified graph construction and extraction pipeline improves the overall time complexity over
current methods, opening the door to future methods where extraction time is critical.</p>
      <p>Our narrative extraction algorithm opens new avenues for future research, including contextualized
global coherence and semantic interaction. One key avenue lies in improving our definition of coherence
with global metrics that better guide the extraction process and regulate storyline length. Additionally,
integrating deep-learning-based search agents with human-controllable constraints could improve the
balance between speed, accuracy, and user-defined requirements. These improvements could further
generalize our approach, broadening its applicability to context-specific domains and tasks.</p>
      <p>Despite its limitations, our approach marks a significant contribution to computational narrative
extraction, ofering a more general method that simplifies the storyline extraction pipeline. Additionally,
the data and model-agnostic nature of Narrative Trails opens the doors to extracting narratives from
diverse datasets, including image data, thus expanding the method’s applicability and impact.</p>
    </sec>
    <sec id="sec-8">
      <title>Acknowledgments</title>
      <p>Brian Keith is supported by Project 202311010033-VRIDT-UCN.
[32] M. Pollack, Letter to the Editor—The Maximum Capacity Through a Network, Operations Research
8 (1960) 733–736. doi:10.1287/opre.8.5.733, publisher: INFORMS.
[33] A. P. Punnen, A linear time algorithm for the maximum capacity path problem, European Journal
of Operational Research 53 (1991) 402–404. doi:10.1016/0377-2217(91)90073-5.
[34] V. Kaibel, M. Peinhardt, On the Bottleneck Shortest Path Problem, Technical Report 06-22, ZIB,</p>
      <p>Takustr. 7, 14195 Berlin, 2006.
[35] E. W. Dijkstra, A note on two problems in connexion with graphs, Numerische Mathematik 1
(1959) 269–271. doi:10.1007/BF01386390.
[36] M. Müller, Dynamic time warping, Information retrieval for music and motion (2007) 69–84.
[37] T. Rakthanmanon, B. Campana, A. Mueen, G. Batista, B. Westover, Q. Zhu, J. Zakaria, E. Keogh,
Searching and mining trillions of time series subsequences under dynamic time warping, in:
Proceedings of the 18th ACM SIGKDD International Conference on Knowledge Discovery and
Data Mining, KDD ’12, Association for Computing Machinery, New York, NY, USA, 2012, p.
262–270. doi:10.1145/2339530.2339576.
[38] K. Wang, T. Gasser, Alignment of curves by dynamic time warping, The Annals of Statistics 25
(1997) 1251–1276.
[39] R. West, J. Pineau, D. Precup, Wikispeedia: an online game for inferring semantic distances between
concepts, in: Proceedings of the 21st International Joint Conference on Artificial Intelligence,
IJCAI’09, Morgan Kaufmann Publishers Inc., San Francisco, CA, USA, 2009, p. 1598–1603. URL:
https://dl.acm.org/doi/10.5555/1661445.1661702.
[40] R. West, J. Leskovec, Human wayfinding in information networks, in: Proceedings of the 21st
International Conference on World Wide Web, WWW ’12, Association for Computing Machinery,
New York, NY, USA, 2012, p. 619–628. doi:10.1145/2187836.2187920.
[41] B. F. Keith Norambuena, T. Mitra, C. North, Mixed multi-model semantic interaction for
graphbased narrative visualizations, in: Proceedings of the 28th International Conference on Intelligent
User Interfaces, IUI ’23, Association for Computing Machinery, New York, NY, USA, 2023, p.
866–888. doi:10.1145/3581641.3584076.
[42] P. Isenberg, F. Heimerl, S. Koch, T. Isenberg, P. Xu, C. Stolper, M. Sedlmair, J. Chen, T. Möller,
J. Stasko, vispubdata.org: A metadata collection about IEEE visualization (VIS) publications, IEEE
Transactions on Visualization and Computer Graphics 23 (2017) 2199–2206. doi:10.1109/TVCG.
2016.2615308.
[43] J. Tang, J. Zhang, L. Yao, J. Li, L. Zhang, Z. Su, Arnetminer: extraction and mining of academic
social networks, in: Proceedings of the 14th ACM SIGKDD International Conference on Knowledge
Discovery and Data Mining, KDD ’08, Association for Computing Machinery, New York, NY, USA,
2008, p. 990–998. doi:10.1145/1401890.1402008.
[44] L. C. Freeman, Centrality in social networks conceptual clarification, Social Networks 1 (1978)
215–239. doi:https://doi.org/10.1016/0378-8733(78)90021-7.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>P.</given-names>
            <surname>Pirolli</surname>
          </string-name>
          ,
          <string-name>
            <surname>S. Card,</surname>
          </string-name>
          <article-title>The sensemaking process and leverage points for analyst technology as identified through cognitive task analysis</article-title>
          ,
          <source>in: Proceedings of International Conference on Intelligence Analysis</source>
          , volume
          <volume>5</volume>
          ,
          <string-name>
            <surname>McLean</surname>
            ,
            <given-names>VA</given-names>
          </string-name>
          , USA,
          <year>2005</year>
          , pp.
          <fpage>2</fpage>
          -
          <lpage>4</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. E. Hobbs,</surname>
          </string-name>
          <article-title>The power of stories: Narratives and information framing efects in science communication</article-title>
          ,
          <source>American Journal of Agricultural Economics</source>
          <volume>102</volume>
          (
          <year>2020</year>
          )
          <fpage>1271</fpage>
          -
          <lpage>1296</lpage>
          . doi:https: //doi.org/10.1002/ajae.12078.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>M. W.</given-names>
            <surname>Kreuter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. C.</given-names>
            <surname>Green</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. N.</given-names>
            <surname>Cappella</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. D.</given-names>
            <surname>Slater</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. E.</given-names>
            <surname>Wise</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Storey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. M.</given-names>
            <surname>Clark</surname>
          </string-name>
          ,
          <string-name>
            <surname>D. J. O'Keefe</surname>
            ,
            <given-names>D. O.</given-names>
          </string-name>
          <string-name>
            <surname>Erwin</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Holmes</surname>
            ,
            <given-names>L. J.</given-names>
          </string-name>
          <string-name>
            <surname>Hinyard</surname>
          </string-name>
          , T. Houston, S. Woolley,
          <article-title>Narrative communication in cancer prevention and control: a framework to guide research and application</article-title>
          ,
          <source>Annals of Behavioral Medicine: A Publication of the Society of Behavioral Medicine</source>
          <volume>33</volume>
          (
          <year>2007</year>
          )
          <fpage>221</fpage>
          -
          <lpage>235</lpage>
          . doi:
          <volume>10</volume>
          .1007/BF02879904.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A.</given-names>
            <surname>Karunakaran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Reddy</surname>
          </string-name>
          ,
          <article-title>The role of narratives in collaborative information seeking</article-title>
          ,
          <source>in: Proceedings of the 2012 ACM International Conference on Supporting Group Work</source>
          , GROUP '12,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2012</year>
          , p.
          <fpage>273</fpage>
          -
          <lpage>276</lpage>
          . doi:
          <volume>10</volume>
          .1145/ 2389176.2389217.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>D.</given-names>
            <surname>Shahaf</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Guestrin, Connecting the dots between news articles</article-title>
          ,
          <source>in: Proceedings of the 16th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>
          , KDD '10,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2010</year>
          , p.
          <fpage>623</fpage>
          -
          <lpage>632</lpage>
          . doi:
          <volume>10</volume>
          .1145/1835804.1835884.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>D.</given-names>
            <surname>Shahaf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Guestrin</surname>
          </string-name>
          ,
          <article-title>Connecting two (or less) dots: Discovering structure in news articles</article-title>
          ,
          <source>ACM Trans. Knowl. Discov. Data</source>
          <volume>5</volume>
          (
          <year>2012</year>
          ). doi:
          <volume>10</volume>
          .1145/2086737.2086744.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>B. F.</given-names>
            <surname>Keith Norambuena</surname>
          </string-name>
          , T. Mitra,
          <article-title>Narrative maps: An algorithmic approach to represent and extract information narratives</article-title>
          ,
          <source>Proc. ACM Hum.-Comput. Interact</source>
          .
          <volume>4</volume>
          (
          <year>2021</year>
          ). doi:
          <volume>10</volume>
          .1145/3432927.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>R.</given-names>
            <surname>Campos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Jorge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Jatowt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bhatia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Litvak</surname>
          </string-name>
          ,
          <source>The 6th international workshop on narrative extraction from texts: Text2story</source>
          <year>2023</year>
          , in: J.
          <string-name>
            <surname>Kamps</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Goeuriot</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Crestani</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Maistro</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Joho</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Davis</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Gurrin</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          <string-name>
            <surname>Kruschwitz</surname>
            ,
            <given-names>A</given-names>
          </string-name>
          . Caputo (Eds.),
          <source>Advances in Information Retrieval</source>
          , Springer Nature Switzerland, Cham,
          <year>2023</year>
          , pp.
          <fpage>377</fpage>
          -
          <lpage>383</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Evolutionary hierarchical Dirichlet process for timeline summarization</article-title>
          , in: H.
          <string-name>
            <surname>Schuetze</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Fung</surname>
          </string-name>
          , M. Poesio (Eds.),
          <source>Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (Volume</source>
          <volume>2</volume>
          :
          <string-name>
            <surname>Short</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <source>Association for Computational Linguistics</source>
          , Sofia, Bulgaria,
          <year>2013</year>
          , pp.
          <fpage>556</fpage>
          -
          <lpage>560</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>F.</given-names>
            <surname>ren Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.-H.</given-names>
            <surname>Liang</surname>
          </string-name>
          ,
          <article-title>Storyline-based summarization for news topic retrospection, Decision Support Systems 45 (</article-title>
          <year>2008</year>
          )
          <fpage>473</fpage>
          -
          <lpage>490</lpage>
          . doi:https://doi.org/10.1016/j.dss.
          <year>2007</year>
          .
          <volume>06</volume>
          .009, special Issue Clusters.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>S.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <surname>Y. Zhang,</surname>
          </string-name>
          <article-title>Summarizing complex events: a cross-modal solution of storylines extraction and reconstruction</article-title>
          , in: D.
          <string-name>
            <surname>Yarowsky</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Baldwin</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Korhonen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Livescu</surname>
          </string-name>
          , S. Bethard (Eds.),
          <source>Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing</source>
          , Association for Computational Linguistics, Seattle, Washington, USA,
          <year>2013</year>
          , pp.
          <fpage>1281</fpage>
          -
          <lpage>1291</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>P.</given-names>
            <surname>Laban</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Hearst, newsLens: building and visualizing long-ranging news stories</article-title>
          , in: T.
          <string-name>
            <surname>Caselli</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Miller</surname>
            ,
            <given-names>M. van Erp</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Vossen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Palmer</surname>
          </string-name>
          , E. Hovy,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mitamura</surname>
          </string-name>
          , D. Caswell (Eds.),
          <source>Proceedings of the Events and Stories in the News Workshop</source>
          , Association for Computational Linguistics, Vancouver, Canada,
          <year>2017</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>9</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>W17</fpage>
          -2701.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>D.</given-names>
            <surname>Shahaf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Guestrin</surname>
          </string-name>
          , E. Horvitz,
          <article-title>Trains of thought: generating information maps</article-title>
          ,
          <source>in: Proceedings of the 21st International Conference on World Wide Web, WWW '12</source>
          ,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2012</year>
          , p.
          <fpage>899</fpage>
          -
          <lpage>908</lpage>
          . doi:
          <volume>10</volume>
          .1145/2187836.2187957.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>D.</given-names>
            <surname>Shahaf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Guestrin</surname>
          </string-name>
          , E. Horvitz,
          <article-title>Metro maps of science</article-title>
          ,
          <source>in: Proceedings of the 18th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>
          , KDD '12,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2012</year>
          , p.
          <fpage>1122</fpage>
          -
          <lpage>1130</lpage>
          . doi:
          <volume>10</volume>
          .1145/2339530.2339706.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>M.</given-names>
            <surname>Grootendorst</surname>
          </string-name>
          , Bertopic:
          <article-title>Neural topic modeling with a class-based tf-idf procedure</article-title>
          ,
          <year>2022</year>
          . arXiv:
          <volume>2203</volume>
          .
          <fpage>05794</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>A.</given-names>
            <surname>Neelakantan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Puri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Radford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Han</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J</given-names>
            .
            <surname>Tworek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Yuan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Tezak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. W.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Hallacy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Heidecke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Shyam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Power</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. E.</given-names>
            <surname>Nekoul</surname>
          </string-name>
          , G. Sastry,
          <string-name>
            <given-names>G.</given-names>
            <surname>Krueger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Schnurr</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. P.</given-names>
            <surname>Such</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Hsu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Thompson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Khan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Sherbakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Jang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Welinder</surname>
          </string-name>
          , L. Weng,
          <article-title>Text and code embeddings by contrastive pre-training</article-title>
          ,
          <year>2022</year>
          . arXiv:
          <volume>2201</volume>
          .
          <fpage>10005</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>E.</given-names>
            <surname>Kamalloo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Ogundepo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Thakur</surname>
          </string-name>
          ,
          <string-name>
            <surname>D.</surname>
            Alfonso-hermelo, M. Rezagholizadeh,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Lin</surname>
          </string-name>
          ,
          <article-title>Evaluating embedding APIs for information retrieval</article-title>
          ,
          <year>2023</year>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          . acl-industry.
          <volume>50</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>OpenAI</surname>
          </string-name>
          ,
          <article-title>New embedding models</article-title>
          and
          <source>API updates</source>
          ,
          <year>2024</year>
          . URL: https://openai.com/index/ new-embedding
          <article-title>-models-and-api-updates/.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>N.</given-names>
            <surname>Reimers</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gurevych</surname>
          </string-name>
          , Sentence-BERT:
          <article-title>Sentence embeddings using Siamese BERT-networks</article-title>
          , in: K. Inui,
          <string-name>
            <given-names>J.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Ng</surname>
          </string-name>
          ,
          <string-name>
            <surname>X.</surname>
          </string-name>
          Wan (Eds.),
          <source>Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)</source>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Hong Kong, China,
          <year>2019</year>
          , pp.
          <fpage>3982</fpage>
          -
          <lpage>3992</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>D19</fpage>
          -1410.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>N.</given-names>
            <surname>Muennighof</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Tazi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Magne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Reimers</surname>
          </string-name>
          , MTEB:
          <article-title>Massive text embedding benchmark</article-title>
          , in: A.
          <string-name>
            <surname>Vlachos</surname>
          </string-name>
          , I. Augenstein (Eds.),
          <source>Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics</source>
          , Association for Computational Linguistics, Dubrovnik, Croatia,
          <year>2023</year>
          , pp.
          <fpage>2014</fpage>
          -
          <lpage>2037</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          .eacl-main.
          <volume>148</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <surname>D. M. Mistry</surname>
            ,
            <given-names>A. A.</given-names>
          </string-name>
          <string-name>
            <surname>Minai</surname>
          </string-name>
          ,
          <article-title>A comparative study of sentence embedding models for assessing semantic variation</article-title>
          ,
          <source>in: Artificial Neural Networks and Machine Learning - ICANN 2023: 32nd International Conference on Artificial Neural Networks</source>
          , Heraklion, Crete, Greece,
          <source>September 26-29</source>
          ,
          <year>2023</year>
          , Proceedings,
          <string-name>
            <surname>Part</surname>
            <given-names>X</given-names>
          </string-name>
          , Springer-Verlag, Berlin, Heidelberg,
          <year>2023</year>
          , p.
          <fpage>1</fpage>
          -
          <lpage>12</lpage>
          . doi:
          <volume>10</volume>
          .1007/ 978-3-
          <fpage>031</fpage>
          -44204-
          <issue>9</issue>
          _
          <fpage>1</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>L.</given-names>
            <surname>McInnes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Healy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Melville</surname>
          </string-name>
          , Umap:
          <article-title>Uniform manifold approximation and projection for dimension reduction</article-title>
          ,
          <year>2020</year>
          . arXiv:
          <year>1802</year>
          .03426.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>L.</given-names>
            <surname>McInnes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Healy</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.</surname>
          </string-name>
          <article-title>Astels, hdbscan: Hierarchical density based clustering</article-title>
          ,
          <source>Journal of Open Source Software</source>
          <volume>2</volume>
          (
          <year>2017</year>
          )
          <article-title>205</article-title>
          . doi:
          <volume>10</volume>
          .21105/joss.00205.
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>L. van der</given-names>
            <surname>Maaten</surname>
          </string-name>
          , G. Hinton,
          <article-title>Viualizing data using t-sne</article-title>
          ,
          <source>Journal of Machine Learning Research</source>
          <volume>9</volume>
          (
          <year>2008</year>
          )
          <fpage>2579</fpage>
          -
          <lpage>2605</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>A.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Nashaat</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Miller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Quader</surname>
          </string-name>
          ,
          <article-title>Context-based evaluation of dimensionality reduction algorithms-experiments and statistical significance analysis</article-title>
          ,
          <source>ACM Trans. Knowl. Discov. Data</source>
          <volume>15</volume>
          (
          <year>2021</year>
          ). doi:
          <volume>10</volume>
          .1145/3428077.
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>M.</given-names>
            <surname>Sánchez-Rico</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Hoertel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Alvarado</surname>
          </string-name>
          ,
          <article-title>Combination of cluster analysis with dimensionality reduction techniques for pattern recognition studies in healthcare data: Comparing pca, t-sne and umap (</article-title>
          <year>2023</year>
          ). doi:
          <volume>10</volume>
          .31234/osf.io/zxvf2.
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>M.</given-names>
            <surname>Allaoui</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. L.</given-names>
            <surname>Kherfi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Cheriet</surname>
          </string-name>
          ,
          <article-title>Considerably Improving Clustering Algorithms Using UMAP Dimensionality Reduction Technique: A Comparative Study</article-title>
          ,
          <source>Image and Signal Processing</source>
          <volume>12119</volume>
          (
          <year>2020</year>
          )
          <fpage>317</fpage>
          -
          <lpage>325</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>030</fpage>
          -51935-3_
          <fpage>34</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>S. J.</given-names>
            <surname>Hwang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. B.</given-names>
            <surname>Damelin</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. O. H. III</surname>
          </string-name>
          ,
          <article-title>Shortest path through random points</article-title>
          ,
          <source>The Annals of Applied Probability</source>
          <volume>26</volume>
          (
          <year>2016</year>
          )
          <fpage>2791</fpage>
          -
          <lpage>2823</lpage>
          . doi:
          <volume>10</volume>
          .1214/15-
          <fpage>AAP1162</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>D.</given-names>
            <surname>Shahaf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Suen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Jacobs</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Leskovec</surname>
          </string-name>
          ,
          <article-title>Information cartography: creating zoomable, large-scale maps of information</article-title>
          ,
          <source>in: Proceedings of the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>
          , KDD '13,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2013</year>
          , p.
          <fpage>1097</fpage>
          -
          <lpage>1105</lpage>
          . doi:
          <volume>10</volume>
          .1145/2487575.2487690.
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>P.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Cao</surname>
          </string-name>
          ,
          <article-title>Emmbtt: A novel event evolution model based on tfxief and tdc in tracking news streams</article-title>
          ,
          <source>in: 2017 IEEE Second International Conference on Data Science in Cyberspace (DSC)</source>
          , Shenzhen, China,
          <year>2017</year>
          , pp.
          <fpage>102</fpage>
          -
          <lpage>107</lpage>
          . doi:
          <volume>10</volume>
          .1109/DSC.
          <year>2017</year>
          .
          <volume>53</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>J.</given-names>
            <surname>Kleinberg</surname>
          </string-name>
          , E. Tardos, Algorithm Design, Pearson,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>