<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Paris, France
£ nicolas.ruth@studserv.uni-leipzig.d(Ne. Ruth); bernhard.liebl@uni-leipzig.d(Be. Liebl);
burghardt@informatik.uni-leipzig.d(Me. Burghardt)
ç https://ch.uni-leipzig.de/(N. Ruth); https://ch.uni-leipzig.de/(B. Liebl);https://ch.uni-leipzig.de/
(M. Burghardt)
ȉ</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>From Clusters to Graphs - Toward a Scalable Viewing of News Videos</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>NicolasRuth</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>BernhardLiebl</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>ManuelBurghardt</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Computational Humanities Group, Institute for Computer Science, Leipzig University</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0002</lpage>
      <abstract>
        <p>In this paper, we present a novel approach that combines density-based clustering and graph modeling to create a scalable viewing application for the exploration of similarity patterns in news videos. Unlike most existing video analysis tools that focus on individual videos, our approach allows for an overview of a larger collection of videos, which can be further examined based on their connections or communities. By utilizing scalable reading, speci昀椀c subgraphs can be selected from the overview and their respective clusters can be explored in more detail on the video frame level.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;scalable viewing</kwd>
        <kwd>hdbscan clustering</kwd>
        <kwd>community detection</kwd>
        <kwd>graph visualization</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction: Video Analytics and Scalable Viewing</title>
      <p>
        analysis perspectives for video material, termed “scalable viewin3g,”1[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. However, for video
material, a signi昀椀cant challenge lies in providing a comprehensive overview from a distant
perspective in a dynamic medium.
      </p>
      <p>
        In this paper, we propose a novel approach to address this challenge by visualizing
relationships and patterns of similarity within a video collection. The basis for our research stems from
the ongoingFakeNarratives1 project, in which we investigate the use of narrative strategies for
the purpose of disinformation in German news video2s1[]. We build on theZoetrope prototype
[
        <xref ref-type="bibr" rid="ref10 ref11">11, 10</xref>
        ], a tool developed in the project’s early phase, initially designed for the analysis of
individual news videos. Zoetrope relies on a complex multimodal information extraction pipeline
from which we have acquired CLIP embeddings1[
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] for a sample of 117 German news videos
from “Tagesschau”, including all the videos from 01.01.2022 to 14.03.2022.
      </p>
      <p>
        In the following, we present an approach that allows researchers to interactively navigate
extensive video collections and to explore underlying patterns that are latent in the CLIP
embeddings. Our approach opens up novel opportunities for scalable viewing of visual media in
computational humanities research. The suggested approach enhances the existing landscape
of video analysis tools such as thDeistant Viewing Toolkit [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] or VIAN [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], which are mostly
focused on the analysis of single video2s.In a sense our interactive visualization can be
compared to recent cultural analytics tools, suchPaixsPlot3 or theCollection Space Navigator [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ],
which visualize patterns in large collections of images. However, we extend these existing
tools as we propose a two-fold approach that combines initial clustering of similar images and
a 3D network visualization of similar image clusters. Also, our approach is optimized for being
used with video frames rather than other types of imagery. Please note that we describe a
novel analytical work昀氀ow rather than a ready-to-use too4lW.hile the use case we present is
based on the analysis of news videos, the work昀氀ow can also be adapted for other collections
of video material.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. From Clusters to Graphs – Conceptual Approach</title>
      <p>
        The employed methodology aims to summarize the visual content of news formats and to
construct an exploratory visualization to investigate their connections and interrelationships.
In the following sections, we describe our conceptual approach in some more detail and also
provide information about the utilized clustering and graph algorithms and there
parameterization.
1Project website and further informationh:ttps://fakenarratives.github.io/
2For a comprehensive overview of similar video analysis tools see the survey paper by Pustu-Iren, Sittel, Mauer,
Bulgakowa, and Ewerth 1[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
3Yale DH Lab – PixPlot:https://github.com/YaleDHLab/pix-plot
4The full code is available vihattps://github.com/Nicolas-le/from-clusters-to-gra p.hDsue to copyright issues, we
currently cannot share a demo visualization of the news video use case. However, to get a basic idea of what the
interactive visualization actually looks like, you will also 昀椀nd a short demo video (2:05 min) in the above repository.
      </p>
      <sec id="sec-2-1">
        <title>2.1. Feature Extraction via CLIP</title>
        <p>
          First, we are concerned with the extraction of semantic elements from the video materi2a]l [
by means of CLIP embeddings, which were calculated for every frame every 0.5 seconds for
each of the 117 news videos in the dataset, resulting in a total of 300,029 embedding vectors.
Contrastive Language-Image Pre-Training (CLIP) is a neural network that was trained on 400
million image-text pairs with the primary objective to combine the textual description of an
image and the image itself within a common vector space. As a result, CLIP embeddings have
the ability to represent the content description of an image in a high-dimensional vector space.
Therefore, embedding vectors with high similarity indicate a resemblance in the semantic
content of the corresponding images1[
          <xref ref-type="bibr" rid="ref8">8</xref>
          ].
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Dimensionality Reduction and Density-Based Clustering</title>
        <p>
          Next we aggregate similar content across videos via a clustering of the embedding vectors. To
reduce the dimensionality of the embeddings, the data was standardized using the scikit-learn
standard scaler and then followed by a Principal Component Analysis (PCA), to retain only
30 principal components5. The parameterization of the PCA was the result of a mutual
adjustment with the parameters of the clustering algorithm with the focus on the interpretability of
the visualization and had mainly the goal to reduce the computation time of the clustering. In
subsequent extensions of the procedure, this part can be extended by systematic determination
of the principal components by, for example, a threshold of the explained variance and by
tests with further dimension reduction procedures. In the early stages of the project, we
experimented with KMeans clustering. However, this approach had certain limitations,
including the clustering of noise and the requirement to de昀椀ne a 昀椀xed number of clusters.
KMeans also showed some problems in dealing with outliers. Following that, our transition to
a density-based clustering approach led us to using the HDBScan algorithm. THhierarchical
Density-Based Spatial Clustering of Applications with Noise (HDBScan) extends DBScan with a
complete clustering hierarchy composed of all possible density-based clusters. It thus inherits
the advantages of density-based clustering, among others the recognition of noise and the
non-static number of clusters, and improves them by the ability to detect clusters of varying
densities [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. The algorithm can be parameterized in numerous manners. For instance, the
Euclidean distance has been designated as the distance metric. However, a critical parameter
to consider pertains to the minimum number of data pointmsin_cluster_size required
to constitute a cluster. There is typically no upper limit, so the algorithm is not likely to
split bigger clusters into smaller sub-clusters, which can lead to clusters with signi昀椀cantly
varying sizes. The optimization ofmin_cluster_size poses challenges due to a trade-o昀: On
one hand, lower parameter values enable the algorithm to detect 昀椀ne-grained clean clusters,
o昀琀en caused by consecutive video frames. On the other hand, higher values may lead to the
complete loss of smaller clusters. Another trade-o昀 is introduced by the second parameter,
the minimum number of neighbors for a point to be a core point. Lowering the parameter
values can result in the formation of larger clusters, which, in turn, ampli昀椀es the issue of
5Scikit-learn standard scalerh:ttps://scikit-learn.org/stable/modules/generatedkl/searn.preprocessing.StandardSc
aler.html
semanticallyimpure clusters, which refers to clusters in which the topics exhibited within
the frames of the cluster manifest considerable heterogeneity. Semanticalpluyre clusters are
favored by higher values, but they increase the detection of noise and encourage the focus
on overly dense clusters created by successive frames or clusters with frequently occurring
items, like recurring images. A昀琀er conducting extensive experiments to compare the results
and to assess usability in the visualization process, thmein_cluster_size was set to 100, to
detect 昀椀ne-grained clusters. The minimum threshold for a core point was set to default, that is
the same as the min_cluster_size. However, the 昀椀ne-tuning of the parameters needs some
more consideration in future work. In our current approach, the detection of small clusters
was given greater weight, which resulted in the detection of altogether 231 clusters of varying
size. The deliberate emphasis on smaller, semantically more distinctive clusters also favors
the incidental classi昀椀cation of substantial quantities of data points as noise. To counteract the
focus on too similar data points, the points previously clustered as noise were added again to
the clustering, using a function in the HDBScan implementation o1f3[] to assign new data
points into the computed clustering, which returns the probability of the new point belonging
to the individual clusters. The points were added to the cluster with the highest probability if
it exceeded a threshold of 10%. Out of a total of 300,029 embeddings, 198,520 were classi昀椀ed as
noise, while 10,685 embeddings were reassigned to clusters, following the described process.
However, it became apparent that the application of a uniform static threshold across all
clusters leads to the contamination of smaller as well as visually very complex clusters, such
as cluster 162 (also see Sec. 3.1), which represents images in medical contexts. The cluster
manifests an array of distinct scenarios, all semantically interlinked through the presence of
medical themes. However, due to the heightened diversity of represented scenes, the cluster’s
demarcation from other topics becomes less pronounced and elevating the threshold leads to
a reduction in its semantic distinctiveness. Please note that clusters that only occur in a single
news video will not be displayed in the later graph visualization.
        </p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. Bidirectional Graph Modeling</title>
        <p>The next step toward a scalable visualization relates the clusters to each other by mapping them
into a bidirectional graph structure. The nodes of the graph represent the previously
generated clusters. For the creation of the edges, we experimented with two di昀erent approaches,
resulting in two di昀erent graphs.</p>
        <p>1. Co-appearance frequency: The 昀椀rst approach inserts an edge between clusters, if their
constituent elements appear together in news videos and weights the edges based on the
normalized amount of co-appearance. Frames belonging to a cluster must appear at
least 昀椀ve times in a news video for the cluster to be counted as present in the video, and
clusters must appear at least three times together for the edge to exist.
2. Pearson correlation: The second approach draws an edge if the clusters appear at
least two times together and are weighted based on the Pearson correlation coe昀케cient
between the clusters. Edges with a weight under 0.3 were deleted.</p>
        <p>The graph based on the number of co-appearances, aims towards an understanding of the
levels of centrality of motifs in the Tagesschau, so they can be experienced intuitively. F1ig.
shows that the graph takes a circular form and concentrates around frequent topics or core
motifs in central nodes. The periphery shows the less prominent clusters. A core topic could
be the weather forecast and a core motif the intro sequence of each broadcast. The edges of this
graph exhibit a high degree of intuitive comprehensibility. However, the mere co-appearance
does not directly indicate a correlation between clusters.</p>
        <p>Therefore, the second graph aims towards the exploration of the correlation explicitly. In
doing so, the visualization of this graph reduces the in昀氀uence of the central topics of the
Tagesschau. It creates a network that also highlights edges between clusters that are less frequent
in the dataset, but still have high correlations. These clusters may then represent a speci昀椀c
interest, because they involve concrete issues and are less likely to be core motifs. This graph
can be seen in Fig. 2.</p>
        <p>Even though the two-fold graph structure is mainly used to create an interactive
visualization, it also o昀ers the possibility to use graph algorithms to further enrich the visualization
with information that can be included as needed. Thus, communities were calculated with
the Clauset-Newman-Moore greedy modularity maximization, including consideration of
the weights. In this application, the community algorithm can be interpreted as a second
clustering algorithm on the generated graph. In the later visualization, this will provide initial
support for the identi昀椀cation of linked subgraphs. The communities, i.e. subgraphs, thus
represent networks of frequently occurring motifs in daily broadcasts, or: networks of strong
correlations. The resolution parameter, that if it is less than 1, favors larger communities and
if it is greater than 1 smaller communities, was chosen to be 1.2.</p>
      </sec>
      <sec id="sec-2-4">
        <title>2.4. Visualization and Interactivity</title>
        <p>
          The created data structure is visualized in an interactive three-dimensional approach. The
frontend of the web app is based on a Javascript framework built on ThreeJS/WebGL
(https://github.com/vasturiano/3d-force-graph) and the force engine works with D3. The
visualization is guided on a theoretical level by Arnold’s and Tilton’s third demand for distant
viewing [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], with a pronounced emphasis on thescalability of the observational granularity.
Additionally, it follows the principles by Shneiderman: “Overview 昀椀rst, zoom and 昀椀lter, then
details-on-demand” 1[
          <xref ref-type="bibr" rid="ref9">9</xref>
          ].
        </p>
        <p>Zooming is possible via scrolling, allowing for an individually adjustable view magni昀椀cation
of the interrelationships. The cluster nodes are represented by a randomly selected frame of
the cluster. The diameter of the edges depends on the edge weight and the graph structure
represented in the 3-D space based on gravitational force in昀氀uenced by the weight of the edge
(Fig. 1 and Fig. 2). By clicking in the space and moving the cursor, the perspective of the camera
can be changed, which allows the focus to be on di昀erent parts of the graph and, together with
the zoom, provides a certain immersive ‘昀氀ight e昀ect’. Individual nodes can be moved via drag
and drop and the graph reorients itself depending on the force. By hovering over the node, the
ID of the cluster can be explored. For a more detailed insight into the semantics of the cluster,
a scrollable overlay can be opened by right-clicking on a cluster (F3)i,g.in which up to 80
random frames of the cluster are displayed. This feature enables a detailed examination and
analysis of single clusters.</p>
        <p>To reduce the complexity of the cluster and to set a focus, it is possible to display only directly
connected clusters by le昀琀-clicking on a cluster. Using the GUI, nodes can also be displayed
that are connected via another edge. Another possibility to view sub graphs is the possibility
to display one of the calculated communities in the graph via the GUI (Fi4g).</p>
        <p>All of these features aim to create an interactive experience for the user that encourages
continuous scaling of the viewing level, breaking down the barrier between micro- and
macroanalytical viewing. Therefore, the developed visualization facilitates a comprehensive
examination of large-scale patterns, pertaining to motifs and topics present in the data source, in
this case speci昀椀cally news videos. It serves as a crucial entry point for diving deeper into the
detailed analysis of extensive volumes of video data, which would otherwise be impracticable
to achieve manually.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Examples Analyses</title>
      <p>The following examples are meant to demonstrate the exploratory analysis of video data with
our methodology. Please note that all of these examples are using the correlation coe昀케cient
graph.
3.1. Topic: Covid-19
During the immersive exploration of the motifs of the Tagessschau in the visualization, cluster
162 stood out prominently. A昀琀er scrolling through numerous images within the cluster and
exploring their semantics via right-click, the majority of the images were con昀椀rmed to
pertain to medical topics, predominantly related to Covid-19. Next, the focus was shi昀琀ed toward
clusters that showed direct linkage, which were accessed by le昀琀-clicking on cluster 162. As
a result, four associated clusters emerged, each displaying distinct characteristics: the 昀椀rst (1)
cluster depicted images from thBeundespressekonferenz, another (2) exhibited an anchor with
a headline discussing Boris Johnson’Psartygate scandal, there was an (3) interview cluster, and
most notably, a cluster (4) presenting multiple instances of tMheeinung-format, which
translates toopinion. This exploration revealed a potential object of further analysis: the interaction
between topics related to Covid-19 and the utilization of a speci昀椀c format that allows
individuals to express subjective opinions. Fig5. shows the directly connected nodes of the medical
cluster 162.</p>
      <sec id="sec-3-1">
        <title>3.2. Topic: Ukraine War</title>
        <p>The second example starts by analyzing the communities in the graph. Community 2 exhibits
a sub-graph encompassing topics related to the war in Ukraine and another subgraph featuring
a core element of the Tagesschau, namely, the weather forecast (Fig6.).</p>
        <p>Both sub-graphs are interconnected through a few edges, thereby forming a cohesive
community together. Within the ‘Ukraine-sub-graph,’ three distinct clusters have been identi昀椀ed.
The 昀椀rst cluster contains images of Volodymyr Zelenskyy, the current president of Ukraine.
The second cluster consists of maps of Ukraine, and the last cluster showcases calls for
donations to support Ukraine. At the periphery of the graph, two clusters featuring war reporters
can be observed. Moreover, there is another link connecting to a cluster containing an
anchorman and background reporting on war criminals, and another cluster showcasing Robert
Habeck, the Federal Minister for Economic A昀airs and Climate Action. The discovered
connection of motifs appears to be more closely related to the contents of the Tagesschau. Hence, the
visualization enables interactive exploration and presentation of complex themes that co-occur
across multiple videos, allowing for the enrichment of these themes with additional
information as required.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Discussion</title>
      <p>The force simulation of the graph proved particularly bene昀椀cial in gaining an intuitive
understanding of signi昀椀cant clusters. For instance, in the circular graph depicting co-appearance,
it becomes evident that the force simulation places the core topics of the Tagesschau at the
center. Indeed, prominent subjects within the dataset, such as the introductions, weather
forecasts, charts displaying results of the German football league, and the previously described
medical cluster 162, all converge at the center of the circular graph due to the force simulation.
The analysis of the two graph variants should be considered in combination, as their
complementary focus allows us to gain valuable insights into the motif network of the Tagesschau.
Together, they o昀er a comprehensive perspective on the relationships and patterns within the
dataset.</p>
      <p>In the context of exploratory visualization, another crucial aspect to consider is that the
ability to freely focus on speci昀椀c elements while having abundant information might lead
individuals to seek con昀椀rmation of their own hypotheses rather than openly discovering new
connections between clusters. Similarly, the projection of semantics onto a cluster needs
careful re昀氀ection and must be balanced and negotiated with one’s own con昀椀rmation bias.
Furthermore, the co-appearance in a news video as a unifying element and resulting correlations must
always be considered in connection with certain basic conditions of the respective broadcast.
This refers, for instance, to external factors that in昀氀uence the program, such as unpredictable
current events. However, this does not make the interaction of special topics, set focuses,
motifs and special broadcast elements any less important. Finally, we want to highlight that
during the creation of the graph it became evident that the parameters have a strong in昀氀uence
on the actual visualization. Their adjustment must therefore be critically examined to avoid
the con昀椀rmation bias of a desired hypothesis.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion and Future Work</title>
      <p>In this short paper, we present a novel approach that combines density-based clustering and
graph modeling to create a scalable viewing application for the exploration of similarity
patterns in news videos. Unlike most existing video analysis tools that focus on individual videos,
our approach allows for an overview of a larger collection of videos, which can be further
examined based on their connections or communities. By utilizing scalable viewing, speci昀椀c
subgraphs can be selected from the overview, and their respective clusters can be explored in
more detail on the video frame level. The next step involves connecting the aforementioned
prototype for analyzing individual videos with the current scalable viewing prototype, enabling
analyses of news videos at various levels of granularity.</p>
      <p>Currently, the similarity function between frame clusters relies solely on CLIP embedding
vectors. However, other information, such as written and spoken language, object and face
recognition, color information, etc. can also be easily extracted from videos and will serve as
additional information sources for future visualizations. While our current experiments focus
on news videos, the scalable viewing approach can be applied to other video formats as well.
Depending on the type and genre of the video, some clustering parameters may need to be
adjusted since they are currently optimized for news videos.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Acknowledgements</title>
      <p>This research was made possible through funding provided by the GermFaenderal Ministry of
Education and Research as part of the ”FakeNarratives” project (16KIS1516).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>T.</given-names>
            <surname>Arnold</surname>
          </string-name>
          and
          <string-name>
            <given-names>L.</given-names>
            <surname>Tilton</surname>
          </string-name>
          . “
          <article-title>Distant viewing toolkit: A python package for the analysis of visual culture”</article-title>
          .
          <source>InJ:ournal of Open Source So昀琀ware 5</source>
          .45 (
          <year>2020</year>
          ), p.
          <year>1800</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>T.</given-names>
            <surname>Arnold</surname>
          </string-name>
          and
          <string-name>
            <given-names>L.</given-names>
            <surname>Tilton</surname>
          </string-name>
          . “
          <article-title>Distant viewing: analyzing large visual corpora”D.Iingi:tal Scholarship in the Humanities 34</article-title>
          .Supplement_
          <volume>1</volume>
          (
          <issue>2019</issue>
          ), pp.
          <fpage>i3</fpage>
          -
          <lpage>i16</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>M.</given-names>
            <surname>Burghardt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kao</surname>
          </string-name>
          , and
          <string-name>
            <given-names>N.-O.</given-names>
            <surname>Walkowski</surname>
          </string-name>
          . “
          <article-title>Scalable MovieBarcodes-An Exploratory Interface for the Analysis of Movies”</article-title>
          .
          <source>InI:EEE VIS Workshop on Visualization for the Digital Humanities</source>
          . Vol.
          <volume>2</volume>
          .
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>R. J. G. B.</given-names>
            <surname>Campello</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Moulavi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Zimek</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Sander</surname>
          </string-name>
          . “
          <article-title>Hierarchical Density Estimates for Data Clustering, Visualization, and Outlier Detection”</article-title>
          .
          <source>AInC:M Trans. Knowl. Discov. Data 10.1</source>
          (
          <year>2015</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>G.</given-names>
            <surname>Halter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ballester-Ripoll</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Flueckiger</surname>
          </string-name>
          , and
          <string-name>
            <given-names>R.</given-names>
            <surname>Pajarola</surname>
          </string-name>
          . “
          <article-title>VIAN: A visual annotation tool for 昀椀lm analysis”</article-title>
          .
          <source>In:Computer Graphics Forum</source>
          . Vol.
          <volume>38</volume>
          . 3. Wiley Online Library.
          <year>2019</year>
          , pp.
          <fpage>119</fpage>
          -
          <lpage>129</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S.</given-names>
            <surname>Jänicke</surname>
          </string-name>
          , G. Franzini,
          <string-name>
            <given-names>M. F.</given-names>
            <surname>Cheema</surname>
          </string-name>
          , and
          <string-name>
            <surname>G. Scheuermann.</surname>
          </string-name>
          “
          <article-title>On Close and Distant Reading in Digital Humanities: A Survey and Future Challenges</article-title>
          .” IEnu: roVis (STARs)
          <year>2015</year>
          (
          <year>2015</year>
          ), pp.
          <fpage>83</fpage>
          -
          <lpage>103</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>S.</given-names>
            <surname>Jänicke</surname>
          </string-name>
          , G. Franzini,
          <string-name>
            <given-names>M. F.</given-names>
            <surname>Cheema</surname>
          </string-name>
          , and
          <string-name>
            <surname>G. Scheuermann.</surname>
          </string-name>
          “
          <article-title>Visual text analysis in digital humanities”</article-title>
          .
          <source>In:Computer Graphics Forum</source>
          . Vol.
          <volume>36</volume>
          . 6. Wiley Online Library.
          <year>2017</year>
          , pp.
          <fpage>226</fpage>
          -
          <lpage>250</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>D.</given-names>
            <surname>Keim</surname>
          </string-name>
          , G. Andrienko,
          <string-name>
            <surname>J.-D. Fekete</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Görg</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Kohlhammer</surname>
            , and
            <given-names>G.</given-names>
          </string-name>
          <article-title>MelançoVni.sual analytics: De昀椀nition, process</article-title>
          , and challenges . Springer,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>K.</given-names>
            <surname>Kurzhals</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>John</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Heimerl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Kuznecov</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D.</given-names>
            <surname>Weiskopf</surname>
          </string-name>
          . “
          <article-title>Visual movie analytics”</article-title>
          .
          <source>In: IEEE Transactions on Multimedia 18.11</source>
          (
          <year>2016</year>
          ), pp.
          <fpage>2149</fpage>
          -
          <lpage>2160</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>B.</given-names>
            <surname>Liebl</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Burghardt</surname>
          </string-name>
          . “
          <article-title>Designing a Prototype for Visual Exploration of Narrative Patterns in News Videos”</article-title>
          .
          <source>In:INFORMATIK 2023 - Lecture Note in Informatics (LNI)</source>
          . Ed. by
          <string-name>
            <given-names>M.</given-names>
            <surname>Klein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Krupka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Winter</surname>
          </string-name>
          , and
          <string-name>
            <given-names>V.</given-names>
            <surname>Wohlgemuth</surname>
          </string-name>
          .
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>B.</given-names>
            <surname>Liebl</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Burghardt</surname>
          </string-name>
          . “
          <article-title>Zoetrope - Interactive Feature Exploration in News Videos”</article-title>
          .
          <source>In: Annual International Conference of the Alliance of Digital Humanities Organizations, DH</source>
          <year>2022</year>
          , Graz, Austria,
          <source>July 10-14</source>
          ,
          <year>2023</year>
          ,
          <string-name>
            <given-names>Conference</given-names>
            <surname>Abstracts</surname>
          </string-name>
          . Ed. by
          <string-name>
            <given-names>A.</given-names>
            <surname>Baillot</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Tasovac</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Scholger</surname>
          </string-name>
          , and
          <string-name>
            <given-names>G.</given-names>
            <surname>Vogeler</surname>
          </string-name>
          .
          <year>2023</year>
          . doi1:
          <fpage>0</fpage>
          .5281/zenodo.8107770. url: https://doi.org/10 .5281/zenodo.8107770.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>L.</given-names>
            <surname>Manovich</surname>
          </string-name>
          . Cultural analytics. Mit Press,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>L.</given-names>
            <surname>McInnes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Healy</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Astels</surname>
          </string-name>
          .
          <article-title>“hdbscan: Hierarchical density based clustering”</article-title>
          .
          <source>In: The Journal of Open Source So昀琀ware 2</source>
          .11 (
          <year>2017</year>
          ), p.
          <fpage>205</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>F.</given-names>
            <surname>Moretti</surname>
          </string-name>
          .
          <article-title>Graphs, maps, trees: abstract models for a literary history</article-title>
          .
          <source>Verso</source>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>T.</given-names>
            <surname>Ohm</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. C.</given-names>
            <surname>Solà</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Karjus</surname>
          </string-name>
          , and
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>SchichC.ollection Space Navigator: An Interactive Visualization Interface for Multidimensional Datasets</article-title>
          .
          <year>2023</year>
          . arXiv:
          <volume>2305</volume>
          .06809 [cs.CV].
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>J.</given-names>
            <surname>Pause</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Burghardt</surname>
          </string-name>
          , and
          <string-name>
            <given-names>N.-O.</given-names>
            <surname>Walkowski</surname>
          </string-name>
          . “
          <article-title>Scalable Viewing in den Filmwissenscha昀琀en”</article-title>
          . In: DHd 2019:
          <article-title>Digital Humanities: multimedial &amp; multimodal</article-title>
          . Konferenzabstracts. Ed. by
          <string-name>
            <given-names>P.</given-names>
            <surname>Sahle</surname>
          </string-name>
          .
          <source>Frankfurt &amp; Mainz</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>201</fpage>
          -
          <lpage>204</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>K.</given-names>
            <surname>Pustu-Iren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Sittel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Mauer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Bulgakowa</surname>
          </string-name>
          , and
          <string-name>
            <given-names>R.</given-names>
            <surname>Ewerth</surname>
          </string-name>
          . “
          <source>Automated Visual Content Analysis for Film Studies: Current Status and Challenges.”DInH:Q: Digital Humanities Quarterly 14.4</source>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>A.</given-names>
            <surname>Radford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. W.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Hallacy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ramesh</surname>
          </string-name>
          , G. Goh,
          <string-name>
            <given-names>S.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Sastry</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Askell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Mishkin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Clark</surname>
          </string-name>
          , G. Krueger, and I.
          <source>SutskeveLr.earning Transferable Visual Models From Natural Language Supervision</source>
          .
          <year>2021</year>
          . arXiv:
          <volume>2103</volume>
          .00020 [cs.CV].
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>B.</given-names>
            <surname>Shneiderman</surname>
          </string-name>
          . “
          <article-title>The eyes have it: a task by data type taxonomy for information visualizations”</article-title>
          .
          <source>In: Proceedings 1996 IEEE Symposium on Visual Languages</source>
          .
          <year>1996</year>
          , pp.
          <fpage>336</fpage>
          -
          <lpage>343</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>B.</given-names>
            <surname>Shneiderman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Plaisant</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. S.</given-names>
            <surname>Cohen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Jacobs</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Elmqvist</surname>
          </string-name>
          , and
          <string-name>
            <given-names>N.</given-names>
            <surname>Diakopoulos</surname>
          </string-name>
          .
          <article-title>Designing the user interface: strategies for e昀ective human-computer interaction</article-title>
          .
          <source>Pearson</source>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>C.</given-names>
            <surname>Tseng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Liebl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Burghardt</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Bateman</surname>
          </string-name>
          . “
          <article-title>FakeNarratives - First Forays in Understanding Narratives of Disinformation in Public and Alternative News Videos”</article-title>
          .
          <source>In: 9. Tagung des Verbands Digital Humanities im deutschsprachigen Raum</source>
          ,
          <source>DHd</source>
          <year>2023</year>
          , Belval, Luxembourg and Trier, Germany, March 13 - 17,
          <year>2022</year>
          . Ed. by
          <string-name>
            <given-names>A.</given-names>
            <surname>Busch</surname>
          </string-name>
          and
          <string-name>
            <given-names>P.</given-names>
            <surname>Trilcke</surname>
          </string-name>
          .
          <year>2023</year>
          , p.
          <fpage>138</fpage>
          . doi:
          <volume>10</volume>
          .5281/zenodo.7715277. url: https://doi.org/10.5281/zenodo.771527.7
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>C.</given-names>
            <surname>Ware</surname>
          </string-name>
          .
          <article-title>Information visualization: perception for design</article-title>
          . Morgan Kaufmann,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>T.</given-names>
            <surname>Weitin</surname>
          </string-name>
          . “Scalable reading”.
          <source>InZ:eitschri昀琀 für literaturwissenscha昀琀 und linguistik</source>
          (
          <year>2017</year>
          ), pp.
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>