<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Narratives About the Energy Transition in Germany</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jonas Rieger</string-name>
          <email>rieger@statistik.tu-dortmund.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Lars Grönberg</string-name>
          <email>lars.groenberg@tu-dortmund.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Carmen Loschke</string-name>
          <email>c.loschke@oeko.de</email>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sibylle Braungardt</string-name>
          <email>s.braungardt@oeko.de</email>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Statistics, TU Dortmund University</institution>
          ,
          <addr-line>44221 Dortmund</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>In: R. Campos, A. Jorge, A. Jatowt, S. Bhatia, M. Litvak (eds.): Proceedings of the Text2Story'25 Workshop</institution>
          ,
          <addr-line>Lucca</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Oeko-Institut</institution>
          ,
          <addr-line>Merzhauser Str. 173, 79100 Freiburg</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>The energy transition as a global challenge is widely discussed in Germany and shaped by diferent narratives. We investigate how newspapers, comments and social media posts difer between sources, platforms, and entities to visualize and monitor respective indicators of the debate over time. For this purpose, we propose a framework of an unsupervised and automated topic model-based approach, the results of which are visualized in a dashboard in conjunction with results from a reliable and supervised classification model using parameter-eficient fine-tuning methods and language models. The analysis of the classification results is refined by integrating the information on annotator disagreement which is captured in a new human-coder labeled sample of our text corpus on the German energy transition. We place a particular emphasis on highlighting the issue of social (in)justice in the debate on energy transition. narrative, energy transition, topic model, parameter-eficient fine-tuning, annotation The energy transition is a critical global issue, addressing environmental, economic, and social challenges that afect and involve everyone. As a transformative process aimed at shifting from fossil fuels to renewable energy, it has become a central topic in (social) media and public debates. The respective coverage not only reflects public opinion but also plays a significant role in shaping it, influencing how society perceives and engages with the transition. Understanding and tracking these discussions is therefore essential to support informed, inclusive, and efective energy policy development.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>CEUR</p>
      <p>
        ceur-ws.org
more equitable and inclusive policy development in the energy transition. As a complementary and
expanding approach to Loschke et al. [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], which is limited to data from the platform X on the topic of
German energy saving in the context of the Russian invasion of Ukraine, we cover the entire media
debate on the German energy transition, including a focus on its socially (un)just implementation.
      </p>
      <p>For the automated representation of the debate we will make use of continuous, i.e. updateable
topic models, while for the snapshot analysis we will utilize parameter-eficient fine-tuning (PEFT) in
combination with few-shot methods on (large) language models (LLMs). Our text corpus will primarily
comprises national and regional newspapers, to which we selectively add online comments and social
media content.</p>
      <p>In the following, we present our methodological workflow, which is based on the findings of Schofield
et al. (2015), that “suggest not automating these workflows but adding transparency and easy-to-use
tools to support practitioners’ context-specific judgments and interventions” [ 3]. This paper aims to
provide a concrete description of a best practice case study for the continuous and reliable
extraction and presentation of narrative elements from real world application scenarios
— and to open it up for further discussion and improvement. To provide a first glimpse of
the possibilities of our project, we present preliminary results of the our continuous topic models
with integrated change detection as an example analysis. Further elements of the (later) presented
framework are currently still under development. The database and the parameters used for the
presented results are preliminary as well. However, their heuristic choice illustrates the potential of
our methods for an even more sophisticated parameter choice. We provide the programming code,
analyses and (permitted) data of the project in the GitHub repository https://github.com/LarsG321/
Dissemination-Monitoring-narratives-about-the-energy-transition-in-Germany. We welcome any type
of questions, suggestions or criticisms via the issue tracker of the respective repository.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Methods</title>
      <p>Latent Dirichlet allocation (LDA) [4], estimated with a Gibbs sampler [5], is supposedly (still) the
most frequently used topic model variant. In particular in the social sciences [6] it still enjoys great
popularity years after the publication of the structural topic model (STM) [ 7], probably due to the
simplicity of application, the modest model assumptions and the meaningful performance, which cannot
be outperformed consistently even by neural topic models [8, 9].</p>
      <p>After the publication of the seminal paper of the attention algorithm [ 10] and the resulting
performance boost for the corresponding transformer-based language models, BERTopic [11] as a neural
topic model (NTM) based on language models (LMs) has also become quite popular. That is likely
because it is particularly convenient to use, in fact BERTopic is even more convenient than LDA as it
estimates the number of topics data-driven by default. However, it has already been shown in a few
analyses that BERTopic performs inferior to Gibbs LDA in terms of several (in particular those based on
human judgments) performance measures [8, 9]. In combination with the ongoing scrutiny of existing
automated evaluation methods for topic models [12, 13, 14], even if the contextualized topic model
[15] works better for well-chosen parameters, overall LDA serves as a robust, stable, reliable universal
tool — and many NTMs perform worse in comparison. In our project, since we want to update the
resulting topic models continuously, we evaluate which of the two methods RollingLDA [16], including
a dedicated monitoring tool to analyze the topic evolutions [17], or BERTrend [18], that is based on
BERTopic, is better suited for our use case and select the more suitable model for application.</p>
      <p>The research fields of parameter-eficient fine-tuning (PEFT) and mixture of experts (MoE) are very
volatile and at the same time strongly overlapping in both aims and architectures, so that a number of
publications also explicitly consider combinations. While PEFT in its origin (and name) aims at eficiency
and implements this by keeping a large part of the pre-trained parameters fixed, MoE (according to
its name) primarily targets performance improvement through input-dependent (de-)activation or
weighting of parameters. However, implicitly — or explicitly — the respective other objective is usually
achieved as well. In recent years, many innovations, mostly with incremental improvements in certain
domains, have been published. The supposedly most well-known method in this area is LoRA [19],
which has subsequently undergone various further developments, such as PRILoRA [20], VeRA [21],
DoRA [22], RoSA [23] and the quantized — or quantization-aware — versions QLoRA [24] and QA-LoRA
[25]. Moreover, several studies were able to show that the combination of PEFT methods in MoE
architectures proves efective. These include mixture of vectors (MoV) and mixture of LoRA (MoLoRA)
[26] or parameter-eficient routed fine-tuning (PERFT) [ 27]. Furthermore, it was shown that in MoE
PEFT architectures it is advantageous to activate only the PEFT modules that are most relevant for the
task, while freezing the other experts for fine-tuning [ 28].</p>
      <p>
        For our use case, we rely on the well-established PEFT variant LoRA, which has been proven to
perform well across diverse tasks. In fact, we compare LoRA to its quantized version QLoRA and
combine both with the language modeling head PETapter [29]. Thus, we unite the advantages of PEFT,
i.e. eficiency, robustness, modularity, ease of use and sharing, with the additional performance gain
by using a few-shot method that integrates the task description into the input utilizing the masked
language modeling objective in a PET-alike manner [30]. For the pretrained base (L)LM, we are focusing
on German or multilingual models since the corpus is in German language. Options include the models
German BERT [31], LeoLM [
        <xref ref-type="bibr" rid="ref3">32</xref>
        ], Llama [
        <xref ref-type="bibr" rid="ref4">33</xref>
        ], Mixtral/Ministral [
        <xref ref-type="bibr" rid="ref5">34</xref>
        ], Gemma [
        <xref ref-type="bibr" rid="ref6">35</xref>
        ], Qwen [
        <xref ref-type="bibr" rid="ref7">36</xref>
        ], which we
want to compare in an appropriate size in an evaluation study and use the best-performing model for
the application.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Framework</title>
      <p>
        For the design of our framework, we follow a human-in-the-loop best practice [
        <xref ref-type="bibr" rid="ref8">37</xref>
        ] with consideration
of active-learning strategies for incremental improvement of the classification results. We conduct the
evaluation process in a task-based manner [
        <xref ref-type="bibr" rid="ref9">9, 12, 13, 38</xref>
        ], so that we do not solely rely on existing NLP
benchmarks for the selection of the (L)LM, as their reliability is limited by the inherent potential for
data contamination [
        <xref ref-type="bibr" rid="ref10">39</xref>
        ].
      </p>
      <p>
        In Figure 1, our project’s methodological workflow is illustrated. As a database, we focus on national
and regional German newspapers. We provide a list of example newspapers that we have access to and
are considering to include in Appendix A. The resulting corpus is complemented with commentary
content from the websites of, e.g., Bild, Spiegel, Süddeutsche Zeitung, Welt, taz, and Zeit, for which we
need to evaluate the full public availability of the respective community comments. Further, we collect
social media posts utilizing promising software packages for the platforms X [
        <xref ref-type="bibr" rid="ref11">40</xref>
        ] and Telegram [
        <xref ref-type="bibr" rid="ref12">41</xref>
        ].
      </p>
      <p>
        Next, we define a filter that primarily aims for a recall, i.e. relevance with regard to the energy
transition, close to 100% while maintaining the highest possible precision. In Section 4, we provide
an example of a filter for the newspapers Bild, Süddeutsche Zeitung, and Welt. The filtered and
preprocessed corpora are modeled using RollingLDA or BERTrend, respectively. Every week, the
procedure is repeated with the newly added data. After postprocessing, the results of the modeling
are made publicly available in our dashboard on an ongoing weekly basis. For this, we make use of
automated topic labeling [
        <xref ref-type="bibr" rid="ref13">42</xref>
        ] and highlight hot/cold topics, popular topics per month [cf. 43], as well as
a monitoring of the topics’ evolution [17] with tailored graphics [
        <xref ref-type="bibr" rid="ref15">44</xref>
        ], all of which is implemented in an
unsupervised manner. In addition, we present the most important entities (e.g., politicians, companies,
countries) in the respective corpus, put them in context using network analysis methods, and report
the overall sentiment over time as well as top articles per topic and month. Furthermore, as far as the
data (availability) allows, we intend to not only highlight temporal changes but also spatial diferences,
e.g., among diferent federal states or regions of Germany [cf. 45]. Besides, we explore the use of LLMs
for writing short summaries of topics for the latest analyzed week.
      </p>
      <p>The interpretation of the results is used for the development of the codebook [cf. 46] for supervised
classification, which we use for a restrospective snapshot analysis. In the definition of the codebook,
we will (need to) focus on the operationalizable definition of what social justice means in order to avoid
too much disagreement in the labels due to insuficiently precise definitions [cf. 47, 48, 49]). We aim
for the resulting labeled data set to provide evidence for (at least) the following hypotheses:</p>
      <sec id="sec-3-1">
        <title>National Regional</title>
      </sec>
      <sec id="sec-3-2">
        <title>Newspapers Newspapers</title>
      </sec>
      <sec id="sec-3-3">
        <title>Comments Social Media</title>
      </sec>
      <sec id="sec-3-4">
        <title>RollingLDA / BERTrend</title>
      </sec>
      <sec id="sec-3-5">
        <title>Dashboard</title>
        <p>s
g
n
i
d
n
i
F</p>
      </sec>
      <sec id="sec-3-6">
        <title>Aggregation</title>
        <p>(Gold Standard)</p>
      </sec>
      <sec id="sec-3-7">
        <title>Annotator</title>
      </sec>
      <sec id="sec-3-8">
        <title>Disagreement</title>
      </sec>
      <sec id="sec-3-9">
        <title>QLoRA &amp; PETapter</title>
      </sec>
      <sec id="sec-3-10">
        <title>Evaluation: Performance</title>
      </sec>
      <sec id="sec-3-11">
        <title>Scores</title>
      </sec>
      <sec id="sec-3-12">
        <title>Analysis: Popularity, Subjectivity, Polarization</title>
        <p>
          • The (media-perceptible) acceptance for the German energy transition dropped in the last two
years (2023, 2024).
• The negative aspects regarding the energy transition are disseminated disproportionately often
by certain media.
• The argument of social (in)justice is more often used as a counter-argument for the German
energy transition as a whole than as an argument for a more socially just implementation of it.
The annotation is done by five independent coders in order to be able to provide not only a (reliable) gold
standard label but also topic-based uncertainty estimates through (reasonable) annotator disagreement
for our evaluations. In addition to the evaluation of the annotator disagreement [
          <xref ref-type="bibr" rid="ref18 ref19">47, 48</xref>
          ], we make use of
majority vote gold standard labels for fine-tuning [
          <xref ref-type="bibr" rid="ref20">49</xref>
          ]. We implement an active learning approach [
          <xref ref-type="bibr" rid="ref8">37</xref>
          ]
to keep track of and be able to incrementally improve in terms of the model’s performance measures.
Our architecture uses QLoRA and PETapter, the utility of which has already been proven in real-world
social sciences applications [
          <xref ref-type="bibr" rid="ref17">29, 46</xref>
          ].
        </p>
        <p>Supposing a suficiently high annotator agreement, we employ the fine-tuned model based on the
codebook-labeled partial dataset to predict and analyze a variety of descriptive characteristics, e.g.,
polarization and subjectivity scores for topics, arguments, newspapers and media types, respectively,
over time. These findings are in turn fed back into our dashboard to provide an comprehensive overview
of the German news media debate on the energy transition and its social (in)justice.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. First Insights</title>
      <p>This first analysis is concerned with the left half of our framework in Figure 1 up to the development of
the dashboard and presents first results for the RollingLDA methodology.</p>
      <p>In order to investigate the German energy transition and its accompanying narratives, we retrieve
articles from three major German newspapers Bild, Süddeutsche Zeitung (SZ), and Welt from 2014
to 2024. These outlets are selected to capture a broad ideological and stylistic range: SZ is generally
considered left-leaning, Welt is viewed as conservative, and Bild is known for its more opinion-based and
sensationalist approach. We apply a pre-filtering step to retrieve only relevant articles, i.e. containing
terms explicitly related to the German energy transition. For this, we make use of the keywords
• Bild: (wende | gesetz | politik | habeck | hammer) &amp; (waerme | energie | klima | heiz)
• Süddeutsche Zeitung and Welt: (wende | gesetz | politik ) &amp; (waerme | energie | klima | heizung)
In addition, for Bild, we exclude articles belonging to the Sports section according to meta data as well
as all articles containing the pattern leserbriefe or seite in the title since these articles revealed to be
noisy and seem to be falsely extracted by the data provider. These steps result in a dataset of 21,046
articles from SZ, 21,107 from Welt, and 2,263 from Bild.</p>
      <sec id="sec-4-1">
        <title>4.1. Topic Modeling</title>
        <p>
          To explore the thematic structure of the filtered articles, we applied RollingLDA [ 16] for various topic
numbers in combination with a change detection procedure [17] to identify changes in topics and the
most influential words for these respective changes. The topic models were trained independently
on each newspaper corpus. In Appendix B, we provide details about the preprocessing and heuristic
hyperparameter choices, i.e. a more sophisticated choice of parameters could lead to even clearer (and
better interpretable) results.. As explained for our framework in Figure 1, we employed automated
topic labeling leveraging LLMs [
          <xref ref-type="bibr" rid="ref13">42</xref>
          ] in combination with subsequent human verification and refinement
(where necessary). Among the tested configurations,  = 10 appeared most promising in terms of
interpretability and thematic clarity across the three corpora.
        </p>
        <p>The basic topic model results, i.e. top ten explanatory words per topic and the monthly topic
frequencies, can bee accessed in Appendix C. Within the SZ corpus, the topics covered both the German
energy transition itself and international political afairs, with frequent mentions of Ukraine and Russia.
Additional thematic clusters included climate protection and protest (Topic 4) and climate change in
relation to agriculture (Topic 2), indicating that the filtering process efectively isolated content directly
relevant to the German energy transition. Notably, topics related to sustainability and renewable
energies were found consistently across multiple topic number configurations. A similarly broad thematic
distribution emerged within the Welt corpus. The identified topics ranged from Russia-EU relations in
the context of the Ukraine conflict (Topic 8) and German energy policy (e.g., topics 1, 4, 6, 7) to real estate
and infrastructure (e.g., topics 5, 10). As in the SZ corpus, themes tied to sustainability, transportation,
and the German energy transition were prevalent and showed consistent patterns across diferent topic
number configurations. Furthermore, comparative analyses between the two newspapers revealed that
certain topic clusters — particularly those pertaining to the Ukraine conflict, sustainable energy, and
transportation — were present in both corpora, with occasional co-moving patterns in topic prevalence
over time. In contrast, the Bild corpus exhibited a more pronounced focus on national politics and
prominent public figures. The most frequently identified topics included German domestic politics in the
context of the Ukraine conflict (Topic 1), Thuringian state-level politics (Topic 2) and German politicians
and public figures (nearly all topics). This orientation aligns with Bild’s reputation for spotlighting
highprofile personalities and emphasizing sensationalist content. Despite these diferences, Bild’s articles
also contained relevant discussions on the German energy transition, demonstrating the efectiveness
of the pre-filtering approach in capturing pertinent material, even within a more personality-driven
news environment.</p>
        <p>Overall, the RollingLDA analysis successfully extracted thematically coherent topics that illuminate
the debate on the German energy transition. Süddeutsche Zeitung and Welt both covered a broad
range of energy- and sustainability-related issues alongside international political afairs, while Bild’s
coverage skewed more heavily toward national politics and individual political figures. These findings
ofer an important foundation for understanding how diferent newspaper orientations shape narratives
surrounding the German energy transition, laying the groundwork for deeper investigations into public
debate and policy implications in this domain.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Change Detection</title>
        <p>Moving on to the change detection, several meaningful changes were identified. Figure 2 displays
some results for the change detection. Blue curves indicate intra-topic similarities over time, while
the red curves represent statistically motivated bootstrapped thresholds [17]. That is, vertical lines
mark detected changes in time (top row). For illustration, the plots start in 2018 displaying five distinct
changes for topic 5 “German Energy Transition” of SZ. We select the first of those five changes for
a more comprehensive analysis, focusing on the most influential words associated with this change
(bottom row). Accordingly, it was triggered by an increased usage of terms related to fossil fuels
(e.g., gas, oel) and the word russland. This aligns with the start of the Russian invasion of Ukraine in
February 2022, which significantly impacted the fossil fuel crisis in Germany. For Welt, the third topic
emerged as the most relevant to the energy transition. Here, the most pertinent change occurred in
July 2021, attributed to the catastrophic flooding (180 deaths) in the German Ahrtal, which received
extensive media attention. The increased frequency of terms such as klimawandel (climate change) and
katastrophenschutz (disaster prevention) supports this interpretation. For Bild, it is noteworthy that the
intra-topic cosine similarity are rather low, indicating that topic 5 for Bild, due to the smaller number
of articles, demonstrates less stability over time compared to the other two corpora. This volatility
leads to a reduced cosine similarity, suggesting a less persistent thematic presence within this corpus.
Nonetheless, the method identified the beginning of the Russia-Ukraine conflict as a relevant change,
leveraged by an increased usage of the terms Putin, Ukraine, and Russischen.</p>
        <p>To summarize the preliminary findings, it becomes evident that the Süddeutsche Zeitung and Welt
appear to contain several relevant topics related to the German energy transition, whereas the analysis
of Bild reveals challenges due to a relative lack of energy-related topics. In addition, we observed the
cross theme of social injustice to be barely represented in the current corpus, indicating areas for further
refinement and investigation for future analyses.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Challenges</title>
      <p>Despite our concise approach, there are still several challenges that we will briefly address here. It
is likely to be easily possible to classify thematic aspects, to depict sentiment over time, as well as to
present key actors in the debate. However, it is still challenging to render entire narratives (over time)
beyond individual elements [cf., e.g., 50, 51]. Narratives are complex to extract (even if, by nature, they
rely on content simplification) and have several components, whereby there are already many diferent
concepts in the (theoretical) definition and hence also in the (operationalized) extraction.</p>
      <p>
        Regarding data, it should be noted that social media data availability (e.g., X as most prominent
example) is increasingly burdening analyses utilizing such data. Furthermore, Instagram, TikTok, but
also X contain ever-increasing video and image content, posing the challenge to appropriately combine
these diferent data types for modeling (we limit ourselves to text data). This is already evident for the
articles of the Bild newspaper, which, due to their heavy use of visual content, but also due to their use
of neologisms (especially in connection with the 2023 Gebäudeenergiegesetz GEG, “Buildings Energy
Act”, which coined the term Heizungshammer, “heating hammer”), require special treatment when it
comes to preprocessing and keyword searches (cf. Section 4). Further, it is unclear how to adequately
take into account diferent, e.g., age-related, media consumption habits when evaluating perception. In
case we get access to regional newspaper data, moreover, we will investigate the regional diferences in
reporting on the German energy transition, similar to Ozgun and Broekel [
        <xref ref-type="bibr" rid="ref16">45</xref>
        ], who were able to prove
a strong link between regional characteristics and the amount of news on innovation.
      </p>
      <sec id="sec-5-1">
        <title>5.1. Next Steps</title>
        <p>In our project “Diskurs Energiewende”, additionally to the presented framework, we aim to contribute to
the development of NLP methods. For RollingLDA, we plan to implement an extension of emerging and
fading topics, for which we will evaluate the positive and negative aspects of how BERTrend handles
them. This might be especially helpful for the RollingLDA results presented above, since the method
ifxates the topics in the first iteration. This leads to the efect that non-present topics in later iterations
are mapped to topics of the first iteration. Thus, if the narrative of social injustice gets important in the
year 2020, for example, and was not present in the first iteration in year 2018 it does not receive its own
topic. Further, we intend to develop advancements in the overlapping area of narrative detection using
PEFT and the use of advanced learning methods making special use of annotator disagreement. These
would allow us to analyze the dissemination and timeline of narratives through diferent media sources.
In turn, it will be challenging to develop a fair performance score to compare models fine-tuned with
classical gold standard labels and those incorporating annotator disagreement [cf. 47, 48, 49]).</p>
        <p>While the current preliminary analysis is capable of detecting topics with respect to the energy
transition, certain limitations with regard to the narrative detection of social injustice apply. To
overcome this problem, we will extend the RollingLDA analyses and compare them with other topic
model results. We will extend the keyword search to obtain a larger corpus, i.e. a higher recall. In
addition, a focused sub-corpus analysis of social injustice related articles within the corpus of the
energy transition articles will be conducted. Further, as incremental improvements we will incorporate
lemmatization and improve our preprocessing function to exclude non-alphanumeric letters (e.g. “€”),
which further normalizes the text corpus and might improve the narrative extraction. Based on the
improved corpus, we will evaluate the usefulness of RollingLDA, BERTrend [18] and BERTopic [11] to
identify narratives in our data and to implement it in our final publicly available dashboard.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>This work was funded by the German Federal Ministry for Economic Afairs and Climate Action
(Bundesministerium für Wirtschaft und Klimaschutz, BMWK) as part of the joint project Social (in)justice
in the energy transition - from the digital debate to the living world (project no. 03EI5267B).
[3] A. Schofield, S. Wu, T. Bayard de Volo, T. Kuze, A. Gomez, S. Sultana, ”My very subjective human
interpretation”: Domain expert perspectives on navigating the text analysis loop for topic models,
Proceedings of the ACM on Human-Computer Interaction 9 (2025). doi:10.1145/3701201.
[4] D. M. Blei, A. Y. Ng, M. I. Jordan, Latent Dirichlet allocation, Journal of Machine Learning Research
3 (2003) 993–1022. doi:10.1162/jmlr.2003.3.4-5.993.
[5] T. L. Grifiths, M. Steyvers, Finding scientific topics, Proceedings of the National Academy of</p>
      <p>Sciences 101 (2004) 5228–5235. doi:10.1073/pnas.0307752101.
[6] D. Ramage, E. Rosen, J. Chuang, C. D. Manning, D. A. McFarland, Topic modeling for the social
sciences, in: Workshop on Applications for Topic Models, NIPS, 2009. URL: http://vis.stanford.
edu/papers/topic-modeling-social-sciences.
[7] M. E. Roberts, B. M. Stewart, D. Tingley, E. M. Airoldi, The structural topic model and applied
social science, in: NIPS-Workshop on Topic Models: Computation, Application, and Evaluation,
2013. URL: https://www.wcfia.harvard.edu/files/wcfia/files/stmnips2013.pdf.
[8] A. M. Hoyle, R. Sarkar, P. Goel, P. Resnik, Are neural topic models broken?, in: Findings of the
Association for Computational Linguistics: EMNLP 2022, Association for Computational Linguistics,
Abu Dhabi, United Arab Emirates, 2022, pp. 5321–5344. doi:10.18653/v1/2022.findings-emnlp.
390.
[9] Z. Li, A. Mao, D. Stephens, P. Goel, E. Walpole, A. Dima, J. Fung, J. Boyd-Graber, Improving the
TENOR of labeling: Re-evaluating topic models for content analysis, in: Proceedings of the 18th
Conference of the European Chapter of the Association for Computational Linguistics (Volume 1:
Long Papers), Association for Computational Linguistics, St. Julian’s, Malta, 2024, pp. 840–859.</p>
      <p>URL: https://aclanthology.org/2024.eacl-long.51/.
[10] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. u. Kaiser, I. Polosukhin,
Attention is all you need, in: I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S.
Vishwanathan, R. Garnett (Eds.), Advances in Neural Information Processing Systems, volume 30,
Curran Associates, Inc., 2017. URL: https://proceedings.neurips.cc/paper_files/paper/2017/file/
3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf.
[11] M. Grootendorst, BERTopic: Neural topic modeling with a class-based tf-idf procedure, 2022.</p>
      <p>arXiv:2203.05794.
[12] C. Doogan, W. Buntine, Topic model or topic twaddle? Re-evaluating semantic interpretability
measures, in: Proceedings of the 2021 Conference of the North American Chapter of the Association
for Computational Linguistics: Human Language Technologies, Association for Computational
Linguistics, Online, 2021, pp. 3824–3848. doi:10.18653/v1/2021.naacl-main.300.
[13] A. Hoyle, P. Goel, A. Hian-Cheong, D. Peskov, J. Boyd-Graber, P. Resnik, Is automated topic model
evaluation broken? The incoherence of coherence, in: Advances in Neural Information Processing
Systems, volume 34, Curran Associates, Inc., 2021, pp. 2018–2033. URL: https://proceedings.neurips.
cc/paper_files/paper/2021/file/0f83556a305d789b1d71815e8ea4f4b0-Paper.pdf.
[14] D. Stammbach, V. Zouhar, A. Hoyle, M. Sachan, E. Ash, Revisiting automated topic model evaluation
with large language models, in: Proceedings of the 2023 Conference on Empirical Methods in
Natural Language Processing, Association for Computational Linguistics, Singapore, 2023, pp.
9348–9357. doi:10.18653/v1/2023.emnlp-main.581.
[15] F. Bianchi, S. Terragni, D. Hovy, Pre-training is a hot topic: Contextualized document embeddings
improve topic coherence, in: Proceedings of the 59th Annual Meeting of the Association for
Computational Linguistics and the 11th International Joint Conference on Natural Language
Processing (Volume 2: Short Papers), Association for Computational Linguistics, Online, 2021, pp.
759–766. doi:10.18653/v1/2021.acl-short.96.
[16] J. Rieger, C. Jentsch, J. Rahnenführer, RollingLDA: An update algorithm of Latent Dirichlet
Allocation to construct consistent time series from textual data, in: Findings of the Association
for Computational Linguistics: EMNLP 2021, Association for Computational Linguistics, Punta
Cana, Dominican Republic, 2021, pp. 2337–2347. doi:10.18653/v1/2021.findings-emnlp.201.
[17] J. Rieger, K.-R. Lange, J. Flossdorf, C. Jentsch, Dynamic change detection in topics based on rolling
LDAs, in: Proceedings of the Text2Story’22 Workshop, CEUR-WS 3117, 2022, pp. 5–13. URL:
https://ceur-ws.org/Vol-3117/paper1.pdf.
[18] A. Boutaleb, J. Picault, G. Grosjean, BERTrend: Neural topic modeling for emerging trends
detection, in: Proceedings of the Workshop on the Future of Event Detection (FuturED),
Association for Computational Linguistics, Miami, Florida, USA, 2024, pp. 1–17. doi:10.18653/v1/2024.
futured-1.1.
[19] E. J. Hu, Y. Shen, P. Wallis, Z. Allen-Zhu, Y. Li, S. Wang, L. Wang, W. Chen, LoRA: Low-rank
adaptation of large language models, in: International Conference on Learning Representations,
2022. URL: https://openreview.net/forum?id=nZeVKeeFYf9.
[20] N. Benedek, L. Wolf, PRILoRA: Pruned and rank-increasing low-rank adaptation, in: Y.
Graham, M. Purver (Eds.), Findings of the Association for Computational Linguistics: EACL 2024,
Association for Computational Linguistics, St. Julian’s, Malta, 2024, pp. 252–263. URL: https:
//aclanthology.org/2024.findings-eacl.18/.
[21] D. J. Kopiczko, T. Blankevoort, Y. M. Asano, VeRA: Vector-based random matrix adaptation, in:
The Twelfth International Conference on Learning Representations, 2024. URL: https://openreview.
net/forum?id=NjNfLdxr3A.
[22] S.-Y. Liu, C.-Y. Wang, H. Yin, P. Molchanov, Y.-C. F. Wang, K.-T. Cheng, M.-H. Chen, DoRA:
Weight-decomposed low-rank adaptation, in: R. Salakhutdinov, Z. Kolter, K. Heller, A. Weller,
N. Oliver, J. Scarlett, F. Berkenkamp (Eds.), Proceedings of the 41st International Conference on
Machine Learning, volume 235, PMLR, 2024, pp. 32100–32121. URL: https://proceedings.mlr.press/
v235/liu24bn.html.
[23] M. Nikdan, S. Tabesh, E. Crnčević, D. Alistarh, RoSA: Accurate parameter-eficient fine-tuning via
robust adaptation, in: Proceedings of the 41st International Conference on Machine Learning,
ICML’24, JMLR.org, 2024. URL: https://dl.acm.org/doi/10.5555/3692070.3693618.
[24] T. Dettmers, A. Pagnoni, A. Holtzman, L. Zettlemoyer, QLoRA: Eficient finetuning of
quantized LLMs, in: Advances in Neural Information Processing Systems, volume 36, Curran
Associates, Inc., 2023, pp. 10088–10115. URL: https://proceedings.neurips.cc/paper_files/paper/2023/
file/1feb87871436031bdc0f2beaa62a049b-Paper-Conference.pdf.
[25] Y. Xu, L. Xie, X. Gu, X. Chen, H. Chang, H. Zhang, Z. Chen, X. Zhang, Q. Tian, QA-LoRA:
Quantization-aware low-rank adaptation of large language models, in: The Twelfth
International Conference on Learning Representations, 2024. URL: https://openreview.net/forum?id=
WvFoJccpo8.
[26] T. Zadouri, A. Üstün, A. Ahmadian, B. Ermis, A. Locatelli, S. Hooker, Pushing mixture of experts
to the limit: Extremely parameter eficient MoE for instruction tuning, in: The Twelfth
International Conference on Learning Representations, 2024. URL: https://openreview.net/forum?id=
EvDeiLv7qc.
[27] Anonymous, PERFT: Parameter-eficient routed fine-tuning for mixture-of-expert model, 2025.</p>
      <p>URL: https://openreview.net/forum?id=PPjpGTPG5K.
[28] Z. Wang, D. Chen, D. Dai, R. Xu, Z. Li, Y. Wu, Let the expert stick to his last: Expert-specialized
ifne-tuning for sparse architectural large language models, in: Y. Al-Onaizan, M. Bansal, Y.-N.
Chen (Eds.), Proceedings of the 2024 Conference on Empirical Methods in Natural Language
Processing, Association for Computational Linguistics, Miami, Florida, USA, 2024, pp. 784–801.
doi:10.18653/v1/2024.emnlp-main.46.
[29] J. Rieger, M. Ruckdeschel, G. Wiedemann, PETapter: Leveraging PET-style classification heads for
modular few-shot parameter-eficient fine-tuning, 2024. arXiv:2412.04975.
[30] T. Schick, H. Schütze, Exploiting cloze-questions for few-shot text classification and natural
language inference, in: P. Merlo, J. Tiedemann, R. Tsarfaty (Eds.), Proceedings of the 16th
Conference of the European Chapter of the Association for Computational Linguistics: Main
Volume, Association for Computational Linguistics, Online, 2021, pp. 255–269. doi:10.18653/v1/
2021.eacl-main.20.
[31] B. Chan, S. Schweter, T. Möller, German‘s next language model, in: D. Scott, N. Bel, C. Zong (Eds.),
Proceedings of the 28th International Conference on Computational Linguistics, International
Committee on Computational Linguistics, Barcelona, Spain (Online), 2020, pp. 6788–6796. doi:10.</p>
      <p>A. Potential national and regional newspapers under study
The following newspapers are considered to be taken into account in our analyses:
• national newspapers: Bild, Der Tagesspiegel, Die Welt, Süddeutsche Zeitung, taz — die</p>
      <p>Tageszeitung
• magazines/weekly newspapers: Der Spiegel, Stern
• regional newspapers: Reutlinger Nachrichten, Schwarzwälder Bote, Stuttgarter Nachrichten,</p>
      <p>Stuttgarter Zeitung, Südwest Presse
• thematic outlets: EID Energie Informationsdienst, Energie &amp; Management/Powernews.org,
Handelsblatt, KI — Kälte Luft Klimatechnik,</p>
    </sec>
    <sec id="sec-7">
      <title>B. Preprocessing and hyperparameters</title>
      <p>The programming code is publicly available via our GitHub project repository https://github.com/
LarsG321/Dissemination-Monitoring-narratives-about-the-energy-transition-in-Germany. Due to
license restrictions, we are unable to share the raw data. The following preprocessing steps are
conducted for all corpora:
• Filter for texts between 2014-01-01 and 2024-03-31
• Remove duplicated texts
• Remove umlauts, HTML and XML nodes
• Remove special characters (apparently we missed removing the € sign)
• Replaced line breaks with spaces
• Lowercase the text
• Remove German stopwords (leveraging a common and rather conservative list)
• Remove punctuation
• Remove numbers
LDA hyperparameters used for all corpora:</p>
      <p>•  = 5, 6, 7, 8, 9, 10, 15, 20 ; alpha = eta = 1/K
RollingLDA hyperparameters used for all corpora (if not specified, we use the default parameters from
the function RollingLDA in the software package rollinglda [16]):
• init = "2014-02-01" (first month of data)
• chunks = "month" (monthly updates)
• memory = "month" (one month of memory)
• memory.fallback = 0, 1, 2 (number of documents used as memory if there are no documents
in the “original” memory time period)
Change detection [17] hyperparameters used for all corpora:
•  = {1, 2, 3, 4} (maximum number of months in the reference period)
•  = {0.80, 0.81, 0.82, 0.83, 0.84, 0.85, 0.86, 0.87, 0.88, 0.89, 0.90} (quantile to determine threshold)
Final parameters for RollingLDA and change detection:
• Bild: memory.fallback=0,  = 10,  = 1,  = 0.85
• SZ: memory.fallback=0,  = 10,  = 1,  = 0.85
• Welt: memory.fallback=2,  = 10,  = 1,  = 0.80</p>
    </sec>
    <sec id="sec-8">
      <title>C. Topic model results</title>
      <p>BILD − Topic 1 to Topic 5
BILD − Topic 6 to Topic 10
1500
:
1
ic m
p re
o</p>
      <p>T G
afd
trump
roy
putin
rechts
land
lindner
putins
s
t
c
i
l
f
n
o
d
n
a
s
c
i
t
i
l
o
P
n
a
:
3
ic m
p re
o</p>
      <p>T G
luetzerath
graichen
krim
hitler
steinmeier
flick
s
c
i
t
i
l
o
P
e
t
a
t
S
n
a
i
: g
2 n
ic ir
p u
o h</p>
      <p>T T
euro
kemmerich
manta
tina
haut
schweiger
goetz
inflation
syrien
adresse
frodeno
s
e
i
t
i
l
a
n
o
s
r
e
P
d
n
a
s
c
i
t
i
l
o
P
n
a
:
4
ic m
p re
o</p>
      <p>T G
scholz
kohl
laschet
habeck
robert
heynckes
olaf
jupp
eier
y
c
i
l
o
P
y
m
o
n
o
c
E
n
e
e
r
G
n
a
:
5
ic m
p re
o</p>
      <p>T G
habeck
gruene
robert
trump
minister
waermepumpe
gasheizungen
akw
heizhammer
2014
2016
2018</p>
      <p>2020
Date
2022
2024
2014
2016
2018</p>
      <p>2020
Date
2022
2024
cdu
neubauer
ukraine
luisa
druckerei
klamroth
newsprint
erdogan
putin
tipp
olaf
kueblboeck
h
t
l
a
e
H
d
n
a
,
s
b
: o
9 J
c ,
i o
p r
o u</p>
      <p>T E
jobgeld
gesundheit
tipp
ps
geld
zaman
y
m
o
n
o
c
E
d
n
a
s
c
i
t
i
l
o
P
n
a
:
0
1
ic m
p re
o</p>
      <p>T G
tagestrend
kohl
jobgeld
charles
neuinfektionen
gorbatschow
euro
rasen
muskeln
baerbock
euro
€
tagestrend helmut
10
r
e
c
c
o
S
h
c
i
n
u
:ic1opT reaynBM
spd
gruenen
partei
cdu
fdp
scholz
csu
koalition
ki
afd
fridays
m
s
i
v
i
t
c</p>
      <p>A
:4 te
c a
ip im
oT lC
stadt
schueler
meter
wasser
for
greta
future
thunberg
autos
n
o
i
t
i
s
n
a
r
T
y
g
r
e
n</p>
      <p>E
:5 an
icopT reGm
strom
euro
wasserstof
gas
kosten
deutschland
bundesregierung
pro
energien
s
e
u
s
s
I
l
a
i
c
o
S
d
n
:ic6p liitsca
oT oP
trump
biden
polizei
aktivisten
generation
israel
hamas
e
c
n
a
n
i
F
d
n
a
y
m
o
n
o
c</p>
      <p>E
:7 an
icopT reGm
unternehmen
euro
inflation
milliarden
firmen
preise
wirtschaft
demokratie geld
men- ezb
schen
gesellschaft banken
putin
china
trump
usa
praesident
europa
russischen
russland
museum
kinder
e
r
u
t
l
u
:ic9 adnC
p t
oT rA
kunst
buch
menschen
the
welt
leben
kuenstler
e
f
i
L
d
:0 an
1 ly
c i
iopT aFm
frau
leben
mutter
eltern
mann
leute
familie
geschichte frauen
gesellschaft kind
s
t
r
o
p
S
d
n
a
s
c
i
t
i
l
o</p>
      <p>P
: n
1 a
ic m
op re</p>
      <p>T G
seite
lesen
fc
dax
innenpolitik
sport
polizei
angaben
t
n
e
m
n
o
r
i
v
n
E
d
n
a
e
g
n
a
h</p>
      <p>C
:3 te
c a
iop lim</p>
      <p>T C
erde
co
wissenschaftler
meter
klimawandel
holz
y
l
i
m
a</p>
      <p>F
: d
2 n
c a
i
op ife</p>
      <p>T L
leben
mann
buch
frau
eltern
mutter
menschen
stadt
geschichte grad
s
e
i
t
r
a
P
l
a
c
i
t
i
l
o</p>
      <p>P
: n
4 a
ic m
op re</p>
      <p>T G
spd
cdu
afd
fdp
union
laschet
scholz
ezb
zinsen
s
c
i
t
i
l
o
P
d
n
a
y
g
r
e
n</p>
      <p>E
: n
6 a
c i
i s
p s
o u</p>
      <p>T R
trump
euro
putin
strom
Date
forscher
partei
s
c
i
t
i
l
o</p>
      <p>P
: n
7 a
ic m
op re</p>
      <p>T G
spd
partei
cdu
csu
gruenen
merkel
afd
fdp
seehofer
usa
russische
geht
trump
politik
freiheit
jungen
men- eu
schen
gesellschaft
milliarden
leben euro
gibt
unternehmen
kinder dollar
e
d
a
r
T
d
n
a
y
m
:0 on
1 o
ic cE
p
o U</p>
      <p>T E
europaeischen
europa
china</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.</given-names>
            <surname>Braungardt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Keimeyer</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Loschke, Is the “heating hammer” hitting energy eficiency policy? Learnings from the debate around the German Buildings Energy Act</article-title>
          ,
          <source>in: Proceedings of the eceee summer study</source>
          ,
          <year>2024</year>
          . URL: https://www.oeko.de/fileadmin/oekodoc/3-028-24_Braungardt.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>C.</given-names>
            <surname>Loschke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Braungardt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Rieger</surname>
          </string-name>
          ,
          <article-title>What motivates and demotivates energy savings in times of crisis? - An argument mining analysis using X/Twitter data</article-title>
          ,
          <source>Energy Eficiency</source>
          <volume>18</volume>
          (
          <year>2025</year>
          ).
          <source>doi:10.1007/s12053-024-10283-0</source>
          . 18653/v1/
          <year>2020</year>
          .coling-main.
          <volume>598</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>B.</given-names>
            <surname>Plüster</surname>
          </string-name>
          ,
          <source>LeoLM: Igniting German-language LLM research</source>
          ,
          <year>2023</year>
          . URL: https://laion.ai/blog/ leo-lm/.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [33]
          <string-name>
            <surname>Llama</surname>
            <given-names>Team</given-names>
          </string-name>
          ,
          <source>The Llama 3 herd of models</source>
          ,
          <year>2024</year>
          . arXiv:
          <volume>2407</volume>
          .
          <fpage>21783</fpage>
          ,
          <article-title>(for full author list see arXiv)</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [34]
          <string-name>
            <surname>Mixtral</surname>
            <given-names>Team</given-names>
          </string-name>
          , Mixtral of experts,
          <year>2024</year>
          . arXiv:
          <volume>2401</volume>
          .
          <fpage>04088</fpage>
          ,
          <article-title>(for full author list see arXiv)</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [35]
          <string-name>
            <surname>Gemma</surname>
            <given-names>Team</given-names>
          </string-name>
          ,
          <article-title>Gemma 2: Improving open language models at a practical size</article-title>
          ,
          <year>2024</year>
          . arXiv:
          <volume>2408</volume>
          .
          <fpage>00118</fpage>
          ,
          <article-title>(for full author list see arXiv)</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [36]
          <string-name>
            <surname>Qwen</surname>
            <given-names>Team</given-names>
          </string-name>
          ,
          <source>Qwen2 technical report</source>
          ,
          <year>2024</year>
          . arXiv:
          <volume>2407</volume>
          .
          <fpage>10671</fpage>
          ,
          <article-title>(for full author list see arXiv)</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [37]
          <string-name>
            <given-names>J. F.</given-names>
            <surname>Fung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. K.</given-names>
            <surname>Stephens</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Goel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Walpole</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Dima</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. L.</given-names>
            <surname>Boyd-Graber</surname>
          </string-name>
          ,
          <article-title>Humanin-the-loop Technical Document Annotation</article-title>
          ,
          <source>Technical Report, National Institute of Standards and Technology</source>
          , Gaithersburg,
          <string-name>
            <surname>MD</surname>
          </string-name>
          ,
          <year>2024</year>
          . doi:
          <volume>10</volume>
          .6028/NIST.TN.2287,
          <string-name>
            <surname>NIST</surname>
            <given-names>TN</given-names>
          </string-name>
          2287.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [38]
          <string-name>
            <given-names>Q. V.</given-names>
            <surname>Liao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Xiao</surname>
          </string-name>
          ,
          <article-title>Rethinking model evaluation as narrowing the socio-technical gap</article-title>
          ,
          <year>2023</year>
          . arXiv:
          <volume>2306</volume>
          .
          <fpage>03100</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [39]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Dong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Jin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Gu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Generalization or memorization: Data contamination and trustworthy evaluation for large language models</article-title>
          , in: L.
          <string-name>
            <surname>-W. Ku</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Martins</surname>
          </string-name>
          , V. Srikumar (Eds.),
          <source>Findings of the Association for Computational Linguistics: ACL</source>
          <year>2024</year>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Bangkok, Thailand,
          <year>2024</year>
          , pp.
          <fpage>12039</fpage>
          -
          <lpage>12050</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2024</year>
          .findings-acl.
          <volume>716</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [40]
          <string-name>
            <given-names>S.</given-names>
            <surname>Peeters</surname>
          </string-name>
          , Zeeschuimer,
          <year>2023</year>
          . URL: https://github.com/digitalmethodsinitiative/zeeschuimer. doi:
          <volume>10</volume>
          .5281/zenodo.7525702.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [41]
          <string-name>
            <given-names>P.</given-names>
            <surname>Kessling</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. V.</given-names>
            <surname>Münch</surname>
          </string-name>
          ,
          <article-title>tegracli: A convenience wrapper around Telethon and the Telegram API for research purposes</article-title>
          ,
          <year>2023</year>
          . URL: https://github.com/Leibniz-HBI/tegracli. doi:
          <volume>10</volume>
          .5281/zenodo. 8043362.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [42]
          <string-name>
            <given-names>J.</given-names>
            <surname>Rieger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Peters</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Fischer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Lauer</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . Bittermann, topiclabels:
          <source>Automated Topic Labeling with Language Models</source>
          ,
          <year>2024</year>
          . URL: https://github.com/PetersFritz/topiclabels. doi:
          <volume>10</volume>
          .32614/CRAN. package.topiclabels,
          <source>R package version 0.2.0.</source>
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [43]
          <string-name>
            <given-names>A.</given-names>
            <surname>Bittermann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Rieger</surname>
          </string-name>
          ,
          <article-title>Finding scientific topics in continuously growing text corpora</article-title>
          ,
          <source>in: Proceedings of the Third Workshop on Scholarly Document Processing</source>
          , Association for Computational Linguistics, Gyeongju, Republic of Korea,
          <year>2022</year>
          , pp.
          <fpage>7</fpage>
          -
          <lpage>18</lpage>
          . URL: https://aclanthology.org/
          <year>2022</year>
          .sdp-
          <volume>1</volume>
          .2/.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [44]
          <string-name>
            <given-names>C.</given-names>
            <surname>Krause</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Rieger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Flossdorf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Jentsch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Beck</surname>
          </string-name>
          ,
          <article-title>Visually analyzing topic change points in temporal text collections</article-title>
          ,
          <source>in: Vision</source>
          , Modeling, and
          <string-name>
            <surname>Visualization</surname>
          </string-name>
          , The Eurographics Association,
          <year>2023</year>
          . doi:
          <volume>10</volume>
          .2312/vmv.20231231.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [45]
          <string-name>
            <given-names>B.</given-names>
            <surname>Ozgun</surname>
          </string-name>
          , T. Broekel,
          <article-title>The geography of innovation and technology news - an empirical study of the German news media</article-title>
          ,
          <source>Technological Forecasting and Social Change</source>
          <volume>167</volume>
          (
          <year>2021</year>
          )
          <article-title>120692</article-title>
          . doi:https://doi.org/10.1016/j.techfore.
          <year>2021</year>
          .
          <volume>120692</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [46]
          <string-name>
            <given-names>J.</given-names>
            <surname>Rieger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Yanchenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ruckdeschel</surname>
          </string-name>
          , G. von Nordheim,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kleinen-von Königslöw</surname>
          </string-name>
          , G. Wiedemann,
          <article-title>Few-shot learning for automated content analysis: Eficient coding of arguments and claims in the debate on arms deliveries to Ukraine, Studies in Communication and Media 13 (</article-title>
          <year>2024</year>
          )
          <fpage>72</fpage>
          -
          <lpage>100</lpage>
          . doi:
          <volume>10</volume>
          .5771/
          <fpage>2192</fpage>
          -4007-2024-1-72.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [47]
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Plank</surname>
          </string-name>
          , ACTOR:
          <article-title>Active learning with annotator-specific classification heads to embrace human label variation</article-title>
          ,
          <source>in: Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing</source>
          , Association for Computational Linguistics, Singapore,
          <year>2023</year>
          , pp.
          <fpage>2046</fpage>
          -
          <lpage>2052</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          .emnlp-main.
          <volume>126</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [48]
          <string-name>
            <given-names>N.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , S. Liu,
          <string-name>
            <given-names>W.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Mihalcea</surname>
          </string-name>
          ,
          <article-title>You are what you annotate: Towards better models through annotator representations</article-title>
          ,
          <source>in: Findings of the Association for Computational Linguistics: EMNLP</source>
          <year>2023</year>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Singapore,
          <year>2023</year>
          , pp.
          <fpage>12475</fpage>
          -
          <lpage>12498</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          .findings-emnlp.
          <volume>832</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [49]
          <string-name>
            <given-names>E.</given-names>
            <surname>Fleisig</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Abebe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Klein</surname>
          </string-name>
          ,
          <article-title>When the majority is wrong: Modeling annotator disagreement for subjective tasks</article-title>
          ,
          <source>in: Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing</source>
          , Association for Computational Linguistics, Singapore,
          <year>2023</year>
          , pp.
          <fpage>6715</fpage>
          -
          <lpage>6726</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          .emnlp-main.
          <volume>415</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [50]
          <string-name>
            <given-names>A.</given-names>
            <surname>Piper</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. J.</given-names>
            <surname>So</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Bamman</surname>
          </string-name>
          ,
          <article-title>Narrative theory for computational narrative understanding</article-title>
          , in: M.
          <article-title>-</article-title>
          <string-name>
            <surname>F. Moens</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Specia</surname>
          </string-name>
          , S. W.-t. Yih (Eds.),
          <source>Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing</source>
          , Association for Computational Linguistics, Online and
          <string-name>
            <given-names>Punta</given-names>
            <surname>Cana</surname>
          </string-name>
          , Dominican Republic,
          <year>2021</year>
          , pp.
          <fpage>298</fpage>
          -
          <lpage>311</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2021</year>
          .emnlp-main.
          <volume>26</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [51]
          <string-name>
            <given-names>E.</given-names>
            <surname>Ash</surname>
          </string-name>
          , G. Gauthier, P. Widmer, Relatio:
          <article-title>Text semantics capture political and economic narratives</article-title>
          ,
          <source>Political Analysis</source>
          <volume>32</volume>
          (
          <year>2024</year>
          )
          <fpage>115</fpage>
          -
          <lpage>132</lpage>
          . doi:
          <volume>10</volume>
          .1017/pan.
          <year>2023</year>
          .
          <volume>8</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          <source>WELT − Topic 1 to Topic 5</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>