<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>July</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>A Benchmark for Text Classification in News Recommendations</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Xinyi Li</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Edward C. Malthouse</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Northwestern University</institution>
          ,
          <addr-line>Evanston, Illinois</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2022</year>
      </pub-date>
      <volume>15</volume>
      <issue>2022</issue>
      <abstract>
        <p>Text classification is an important task in natural language processing. In the current era, people mainly obtain information from online news resources. It is then important to have an automatic and accurate news classifier to categorize every day's news stories such that readers can find articles of interested more easily. We use news story data from the McClatchy organization to establish benchmarks on how accurately stories can be classified by multiple existing deep learning classifiers. Among the models we evaluated, Bidirectional Encoder Representations from Transformers (BERT) provides the best accuracy, macro-averaging precision, micro-averaging precision, macro-averaging recall and micro-averaging recall. Diferent from many other benchmark news data set, McClatchy provides both headline and full-text for each news story. We compare the performance of every deep learning-based classifier using headlines versus full-texts-the top three predicted categories include the labeled value 95% of the time with full-texts training and 92% with headlines only. Furthermore, the defined topics in McClatchy are not mutually exclusive. Some predictions identified as inaccurate are in fact classified into reasonable topics. We further provide a visualization of stories from various defined topics. The predicted results and the visualization of news stories illustrate the untrustworthiness of labeled classes and the intrinsic dificulty of categorizing news stories.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;News recommendations</kwd>
        <kwd>news taxonomy</kwd>
        <kwd>news topic categorization</kwd>
        <kwd>text classification</kwd>
        <kwd>deep learning</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        The newspaper industry has seen a steady and steep decline over the past decade in part because
traditional, ad-supported revenue models are no longer viable. There have been widespread
layofs and closures, resulting in ‘ghost newspapers’ and ‘news deserts’, where almost 200 out
of 3,143 counties in the U.S have been left with no daily newspaper and 1,540 counties with only
one weekly newspaper [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. The demise of local newspapers is not only a business problem, but
also a public-good and societal problem. Communities without news organizations have seen
an increase in government spending due to a lack of accountability [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Citizens who consume
less news are unable to evaluate elected oficials and less likely to vote. Reading news is one
way for people to gain knowledge and to become more open-minded. It is therefore important
to help local news organizations find a viable revenue model, which relies on a larger portion
of subscriber revenue [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Retaining a subscriber depends on developing a reading habit [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ],
and recommmender systems (RS) can play a vital role in helping readers find stories of interest
and creating the habit.
      </p>
      <p>
        News RS are diferent from other RS in eCommerce [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] because only focusing on accuracy
might lead to adverse social efects [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. If a news RS keeps recommending items based on a
user’s preference, readers might have narrow opinions, lack certain information because they
are in a filter bubble [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], and be in an echo chamber [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. The spread of information online is
rapid and wide [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Once an echo chamber is generated, the reader can become a spreader of
online fake news [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], which will cause serious social issues and misinformation. The collapse
of local journalism also exposes the public to the risk of receiving and spreading misinformation
[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. Raza stated that improving the diversity of recommendations can solve these problems to
some extent [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>
        Diversity can be measured from the content perspective[
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. In general, improving the
diversity of recommended items can prevent users from falling into a similarity hole [
        <xref ref-type="bibr" rid="ref13 ref14">13, 14</xref>
        ].
First, in the domain of news, improving the diversity can relieve echo chambers. Beam et al. [15]
and Hannak et al. [16] found that news diversity plays an important role in shaping people’s
political and public opinion. Lee et al. [17] showed that diversity in news recommendations
would increase users’ satisfaction. Therefore, it is necessary to have a taxonomy to understand
a text. A second reason to automatically classify news topics is to help newsrooms allocate
resources. News organizations can count the number of stories of diferent types read by
subscribers and use the counts in churn models to understand what types of stories are associated
with retaining a subscriber versus driving one to churn [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. These insights can help news
managers assign reporters to cover stories that will engage and retain paying customers, as
opposed to stimulating page views to generate dwindling ad revenue, often with sensational
content. Such churn models require a reliable content taxonomy, and therefore it is necessary to
develop an automatic news classifier [ 18]. Third, a well-performed taxonomy is also expected to
distinguish the fake and real news stories such that some social issues caused by the spreading
of fake news can be alleviated. Lee et al. [17] proposed the trustworthy of recommended
news would also increase users’ satisfaction. Therefore, it is necessary to have an automatic
text classifier to filter out fake articles. Finally, newspaper websites often make some
nonpersonalized news recommendations [19], which also requires a good news taxonomy.
      </p>
      <p>This paper studies news articles from 33 categories provided by the McClatchy news
organization. Diferent from most existing benchmark news datasets, the topics are not mutually
exclusive. For example, news articles belonging to the topics ‘localOpin’ and ‘localGovt’ also
belong to the topic ‘local’. Our contribution is to establish a benchmark for how accurately
news stories can be classified by existing deep learning (DL) techniques. Furthermore, most
news organizations can provide a text headline for each story, but many have great dificulty in
providing the full text of stories. It is therefore of interest to know how accurately stories can
be classified only from headlines versus with full texts, which we evaluate. Lastly, we provide a
visualization of news articles to illustrate the intrinsic dificulty of classifying news stories.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Literature Review</title>
      <p>Many DL techniques have been applied to news classification, but most studies only evaluated
binary outcomes. Jang [20] evaluated Convolutional Neural Network (CNN) based models
using Word2Vec to classify news articles a either relevant or irrelevant. Detecting fake news
is getting more attention and Liu et al. [21] proposed a hybrid model of Long Short-Term
Memory (LSTM) and CNN for detecting it. Jadhav et al. [22] proposed a hybrid of Recurrent
Neural Network (RNN) and Deep Structured Semantic models to identify important features
of fake news. Qasim et al. [23] studied nine BERT models such as BERT-base, BERT-large,
RoBERTa-base, and RoBERTa-large, etc. on detecting COVID-19 fake news. In this paper, we
focus on multi-class news topic categorization.</p>
      <p>There is some extant research on multi-class news categorization. Zhang [24] studied
multiclass news categorization and proposed a customized DCLSTM-MLP model. Diferent from
the existing CLSTM-MLP model, this model uses a discrete vector to represent the probability
variance of each word and considers the contribution of each word in a classification. Due to
the innovated discrete vector representation of each word, Zhang’s model had better prediction
accuracy than CNN-MLP and CLSTM [24]. However, their news text sample is small. Huang and
Jiang [25] studied the classification of Chinese news articles with machine learning (ML) and
DL techniques including LSTM, Bi-LSTM, CNN Naïve Bayes, TFIDF-SVM, and Word2Vec-SVM.
They showed that ML and DL methods have similar accuracy if the texts are pre-processed
properly. Deng et al. [18] proposed a model that performs better than CNN-BiLSTM, attention
CNN, attention BiLSTM, and BERT-DCNN-BiGRU to classify long Chinese texts. Pre-processing
Chinese texts is diferent from English ones because Chinese characters can either be morphemes
or words and there are no spaces to mark word boundaries. Deng’s work did not evaluate
short texts (e.g., headline only) or languages other than Chinese. We evaluate several the most
popular and well-known DL techniques for English news topic categorization.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <p>Minaee et al. [26] proposed a comprehensive review of DL techniques for multiple text
classification tasks. In this paper, we establish a benchmark of how DL classifiers perform on McClatchy
news topic classification.</p>
      <sec id="sec-3-1">
        <title>3.1. McClatchy dataset</title>
        <p>The McClatchy news dataset consists of 74,672 news items from 25 geographically dispersed
US markets. It provides headline and full-text for each story. The average headline length is
around 13 words, the the average story length is around 163 words. The training, validation,
and testing data set is split into an 8:1:1 ratio. The dataset is imbalanced, with the most common
topic ‘health’ (around 24%) and the least common ‘NHL’. The ratio between the number of
‘health’ and ‘NHL’ is around 65:1, which is acceptable. McClatchy news has 33 topics (shown in
Figure 1); however, diferent from other benchmark news datasets such as AG News [27] and
20 Newsgroups [28], the boundary between some defined topics is blurred. For example, it is
reasonable and accurate to classify ‘stateGovt’ articles into ‘state’.</p>
        <p>Defined topics of McClathy articles.</p>
        <p>Visualization of article embedding studied by
BERT.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Evaluated models</title>
        <p>We investigate the performance of a spectrum of deep models applied to the McClathy news
classification task. These models include feed-forward networks (FFN), convolutional neural
networks (CNN), recurrent neural networks (RNN), attention-based deep models, and their
interpolated variants.</p>
        <p>Feed-forward networks such as FastText [29] have been a workhorse for industry-level
NLP applications owning to their computational eficiency in producing quality sentence
representations from character-level features; yet their feed-forward nature thwarts them from
explicitly capturing sequential dependencies. The RNNs leverage a recurrent bias to alleviate this
issue. We primarily study LSTM [30], an RNN variant that address a typical vanishing/exploding
gradient bottleneck of vanilla recurrent architectures. Empirical studies demonstrate that LSTMs
and its bi-directional variant have achieved state-of-the-art performance in many natural
language applications [31].</p>
        <p>CNNs are mainly used in the area of computer vision [32] but have recently been adopted
for natural language understanding [33, 34, 35]. While RNNs excel in capturing sequential
structure, CNNs demonstrate better capacity in detecting local and position-invariant patterns
to detect key phrases in a sentence [18]. When applying a CNN for text classification, we obtain
the embedding with a fixed dimension for each word, and then a context can be represented
as a matrix (i.e. each row is a wording embedding). In our experiments, we adopt textCNN
[36], which represent the context as a matrix and then applies 1D convolution along the time
dimension (i.e. temporal convolution) to obtain a hidden representation. We then apply a
classification head to this hidden representation for news classification. Recent works have
also studied causal convolution, a technique that prevents information leakage from future to
past, and achieves competitive results on sequence modeling benchmarks against recurrent
counterparts [37]. In experiments we do not find this variant to be beneficial, and defer further
investigation of such causal variants to future work.</p>
        <p>The attention mechanisms was first proposed as an improvement for RNNs [ 38] in machine
translation tasks; subsequently full attention-based models [39] were introduced. Intuitively,
the attention mechanism assigns weights to each word or sentence based on its importance in a
context [39]. In this paper, we evaluate the BiLSTM-attention model proposed by Zhou et al.
[40]—an attention layer is added such that the final context vector is a weighted sum of feature
vectors studied by BiLSTM.</p>
        <p>RNNs and CNNs have their own issues. RNNs easily forget the information from the beginning,
while textCNN lacks interpretability, ignores dependencies among local features, and is dificult
to weight the importance of each feature. Therefore, hybrid models that combine LSTM and
CNN architectures have been proposed. We will evaluate Zhu et al.’s C-LSTM [41]. The key
phrases studied by CNN are fed in Bi-LSTM to obtain the sentence representation. We further
evaluate a hybrid model Bi-LSTM-CNN-attention [42] that combines Bi-LSTM, CNN and the
attention mechanism.</p>
        <p>Lastly, we evaluate Google’s BERT [43]. Diferent from RNNs and CNNs, which require
sequential text inputs and embedding for each word, BERT is a transformer model pre-trained
on tasks masked language modeling and next sentence prediction. It is trained by randomly
masking 15% words and predicting the masked token based on the other independent tokens,
and relying on the self-attention mechanism, BERT understands each word by connecting it to
every other words [43].</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experimental Results</title>
      <p>Tuning parameters such as the batch size, dropout rate, embedding dimension, etc. are chosen
based on the model’s performance on the validation dataset. For BERT models we apply its
pre-trained tokenizer and for other models we apply the ‘Spacy’ tokenizer and ‘glove.6B.100d’
word representations. The 33 topics are imbalanced and the news topic categorization belongs
to the multi-class classification task. Therefore, besides the accuracy metric we also apply
macro-averaging precision (Macro-P), micro-averaging precision (Micro-P), macro-averaging
recall (Macro-R) and micro-averaging recall (Micro-R) for the model evaluation [44].
Macroaveraging of a measure (i.e., precision, recall) is the average of the same measure calculated for
each class, and micro-averaging is computed by first summing up counts of true positives, true
negatives, false positives and false negatives for all classes [44]. For each model, we evaluate
these metrics with headlines only and separately using full texts.</p>
      <p>Tables 1 and 2 show the numerical performances of selected DL techniques mentioned in
Section 3. To make better comparisons, models’ performances with respect to each metric
are visualized in Figure 2. In general, models trained with full-text outperform those trained
with headlines only. Furthermore, BERT has much better performances than all the other
models models. However, compared to BERT performance on the DBpedia [45] data set with 14
categories reported in [26], BERT has worse accuracy on McClatchy.</p>
      <p>As mentioned in Section 3.1, the categories defined by McClatchy are not as mutually exclusive
as the other benchmark datasets. Therefore, we first take a look at some predictions made by our
best model BERT. Among the 191 news articles belonging to the defined category ‘localGovt’,
112 are predicted correctly, and then 12 are predicted to be ‘local’. From Figure 1, we believe that
it is reasonable and accurate to categorize ‘localGovt’ articles into ‘local’. These observations
imply that if we apply broader topics, then the models’ performances should be improved.
Second, instead of classifying news into the topic with the highest probability, we decide to
evaluate models’ performances by making top-3 predictions. When the defined class is in the
top-3 probabilities, we count it as a ‘correct’ prediction, and then we compute top-3 accuracy,
macro-precision, micro-precision, macro-recall and micro-recall for each model. Tables 1 and 2
show that performances are very good using top-3 prediction—the top three predicted categories
include the labeled value 95% of the time with full-text training and 92% with headline only.
Figure 2 also shows that if top-3 prediction is applied, BERT using only headlines performs
better than all the other models using full-texts. These computational results imply that models
understand texts well but the boundaries among defined categories must be more clear.</p>
      <p>We lastly project the vector representations for some headlines from some selected topics to
2D using t-SNE and visualize the spread of these news stories (shown in Figure 1). We observe
that the defined topics ‘local’, ‘localOpin’, ‘localGovt’ and ‘stateGovt’ are intrinsically hard
to be separated from each other, and the ‘local’ category has the widest spread, indicating
it is the most ambiguous category. The fact that some identified inaccurate predictions are
actually classified into reasonable topics, models’ performances are dramatically improved if
top-3 prediction is adopted, and the visualization of some stories illustrate the untrustworthy of
labeled classes and the intrinsic dificulty of categorizing McClatchy news stories.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Discussion</title>
      <p>Having a reliable news taxonomy can improve the context understanding, manage news
resources for news organizations, prevent the spread of fake news, and study the components
of non-personalized news RS. We use data from McClatchy to establish a benchmark of how
accurately stories from ambiguously defined topics can be classified by some popular DL news
classifiers. Among the models we evaluated, BERT performs the best. We further evaluate the
importance of having full-texts by comparing each model’s performances with headlines only
and full-texts. From our experimental results, full-texts provide more contextual information
for better classification performances. Our computational experiments also show that (1) some
predictions identified as being inaccurate are in fact reasonable; (2) models have much better
performances if we apply top-3 predictions, indicating that we could not completely rely on
labeled classes; and (3) our visualization of news stories from some selected topics further
confirm the intrinsic dificulty in classifying news items.</p>
      <p>There are some limitations in our study. First, in this paper, we just establish a benchmark
on how some selected DL news classifiers perform on McClatchy news stories whose defined
topics are not mutually exclusive. The techniques that we evaluated are not completed and
all of them have better variants. Therefore, we expect to provide a more comprehensive
model evaluations on McClatchy. Second, we have identified that the defined categories in
McClatchy are intrinsically hard to be classified. Some predicted classes are in fact accurate.
This observation motivates us to think of the necessity of having the current detailed topics
and want to identify a better way to categorize news articles by applying news RS algorithms.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>We would like to express our gratitude to Oliver Zhuoran Liu from Industrial Engineering and
Management Science department at Northwestern University for providing his valuable and
constructive advice during the course of this research. We thank the Northwestern Local News
Initiative for providing the data. We would also like to thank all the anonymous reviewers for
reviewing this work and providing insightful comments.
[15] M. A. Beam, G. M. Kosicki, Personalized news portals: Filtering systems and increased
news exposure, Journalism &amp; Mass Communication Quarterly 91 (2014) 59–77.
[16] A. Hannak, P. Sapiezynski, A. Molavi Kakhki, B. Krishnamurthy, D. Lazer, A. Mislove,
C. Wilson, Measuring personalization of web search, in: Proceedings of the 22nd
International Conference on World Wide Web, arXiv, 2013, pp. 527–538.
[17] S. Y. Lee, S. W. Lee, Normative or efective? the role of news diversity and trust in news
recommendation services, International Journal of Human–Computer Interaction (2022)
1–14.
[18] J. Deng, L. Cheng, Z. Wang, Attention-based bilstm fused cnn with gating mechanism
model for chinese long text classification, Computer Speech &amp; Language 68 (2021) 101182.
[19] A. Chakraborty, S. Ghosh, N. Ganguly, K. P. Gummadi, Optimizing the
recency-relevancediversity trade-ofs in non-personalized news recommendations, Information Retrieval
Journal 22 (2019) 447–475.
[20] B. Jang, I. Kim, J. W. Kim, Word2vec convolutional neural networks for classification of
news articles and tweets, PloS one 14 (2019) e0220976.
[21] Y. Liu, Y.-F. Wu, Early detection of fake news on social media through propagation path
classification with recurrent and convolutional networks, in: Proceedings of the AAAI
Conference on Artificial Intelligence, AAAI Press, 2018.
[22] S. S. Jadhav, S. D. Thepade, Fake news identification and classification using dssm and
improved recurrent neural network classifier, Applied Artificial Intelligence 33 (2019)
1058–1068.
[23] R. Qasim, W. H. Bangyal, M. A. Alqarni, A. Ali Almazroi, A fine-tuned bert-based transfer
learning approach for text classification, Journal of Healthcare Engineering 2022 (2022).
[24] M. Zhang, Applications of deep learning in news text classification, Scientific Programming
2021 (2021).
[25] C.-M. Huang, Y.-J. Jiang, An empirical study on the classification of chinese news articles
by machine learning and deep learning techniques, in: 2019 International Conference on
Machine Learning and Cybernetics (ICMLC), IEEE, 2019, pp. 1–6.
[26] S. Minaee, N. Kalchbrenner, E. Cambria, N. Nikzad, M. Chenaghlu, J. Gao, Deep learning–
based text classification: a comprehensive review, ACM Computing Surveys (CSUR) 54
(2021) 1–40.
[27] X. Zhang, J. Zhao, Y. LeCun, Character-level convolutional networks for text classification,</p>
      <p>Advances in neural information processing systems 28 (2015).
[28] K. Clark, M.-T. Luong, Q. V. Le, C. D. Manning, Electra: Pre-training text encoders as
discriminators rather than generators, arXiv preprint arXiv:2003.10555 (2020).
[29] A. Joulin, E. Grave, P. Bojanowski, M. Douze, H. Jégou, T. Mikolov, Fasttext.zip:
Compressing text classification models, arXiv preprint arXiv:1612.03651 (2016).
[30] F. A. Gers, E. Schmidhuber, Lstm recurrent networks learn simple context-free and
context-sensitive languages, IEEE Transactions on Neural Networks 12 (2001) 1333–1340.
[31] S. Merity, N. S. Keskar, R. Socher, Regularizing and optimizing lstm language models,
arXiv preprint arXiv:1708.02182 (2017).
[32] S. Khan, H. Rahmani, S. A. A. Shah, M. Bennamoun, A guide to convolutional neural
networks for computer vision, Synthesis Lectures on Computer Vision 8 (2018) 1–207.
[33] Y. Chen, Convolutional neural network for sentence classification, Master’s thesis,
University of Waterloo, 2015.
[34] J. P. A. Vieira, R. S. Moura, An analysis of convolutional neural networks for sentence
classification, in: 2017 XLIII Latin American Computer Conference (CLEI), IEEE, 2017, pp.
1–5.
[35] J. Du, L. Gui, Y. He, R. Xu, X. Wang, Convolution-based neural attention with applications
to sentiment classification, IEEE Access 7 (2019) 27983–27992.
[36] Y. Kim, Convolutional neural networks for sentence classification, CoRR abs/1408.5882
(2014). URL: http://arxiv.org/abs/1408.5882. arXiv:1408.5882.
[37] S. Bai, J. Z. Kolter, V. Koltun, An empirical evaluation of generic convolutional and
recurrent networks for sequence modeling, arXiv preprint arXiv:1803.01271 (2018).
[38] D. Bahdanau, K. Cho, Y. Bengio, Neural machine translation by jointly learning to align
and translate, arXiv preprint arXiv:1409.0473 (2014).
[39] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, I.
Polosukhin, Attention is all you need, Advances in neural information processing systems 30
(2017).
[40] P. Zhou, W. Shi, J. Tian, Z. Qi, B. Li, H. Hao, B. Xu, Attention-based bidirectional long
short-term memory networks for relation classification, in: Proceedings of the 54th annual
meeting of the association for computational linguistics (volume 2: Short papers), 2016,
pp. 207–212.
[41] C. Zhou, C. Sun, Z. Liu, F. Lau, A c-lstm neural network for text classification, arXiv
preprint arXiv:1511.08630 (2015).
[42] Y. Zhu, X. Gao, W. Zhang, S. Liu, Y. Zhang, A bi-directional lstm-cnn model with attention
for aspect-level text classification, Future Internet 10 (2018) 116.
[43] J. Devlin, M. Chang, K. Lee, K. Toutanova, BERT: pre-training of deep bidirectional
transformers for language understanding, CoRR abs/1810.04805 (2018). URL: http://arxiv.
org/abs/1810.04805. arXiv:1810.04805.
[44] M. Sokolova, G. Lapalme, A systematic analysis of performance measures for classification
tasks, Information Processing &amp; Management 45 (2009) 427–437.
[45] J. Lehmann, R. Isele, M. Jakob, A. Jentzsch, D. Kontokostas, P. N. Mendes, S. Hellmann,
M. Morsey, P. Van Kleef, S. Auer, et al., Dbpedia–a large-scale, multilingual knowledge
base extracted from wikipedia, Semantic web 6 (2015) 167–195.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>P. M.</given-names>
            <surname>Abernathy</surname>
          </string-name>
          ,
          <article-title>The expanding news desert, Center for Innovation and Sustainability in Local Media</article-title>
          ,
          <source>School of Media and Journalism</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>P.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Murphy</surname>
          </string-name>
          ,
          <article-title>Municipal borrowing costs and state policies for distressed municipalities</article-title>
          ,
          <source>Journal of Financial Economics</source>
          <volume>132</volume>
          (
          <year>2019</year>
          )
          <fpage>404</fpage>
          -
          <lpage>426</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S. J.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. C.</given-names>
            <surname>Malthouse</surname>
          </string-name>
          ,
          <string-name>
            <surname>Y.</surname>
          </string-name>
          <article-title>Kamyab Hessary, In search for an audiencesupported business model for local newspapers: Findings from clickstream and subscriber data, Digital Journalism (</article-title>
          <year>2021</year>
          )
          <fpage>1</fpage>
          -
          <lpage>21</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. J.</given-names>
            <surname>Calder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. C.</given-names>
            <surname>Malthouse</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y. K.</given-names>
            <surname>Hessary</surname>
          </string-name>
          ,
          <article-title>Not all clicks are equal: detecting engagement with digital content</article-title>
          ,
          <source>Journal of Media Business Studies</source>
          <volume>0</volume>
          (
          <year>2021</year>
          )
          <fpage>1</fpage>
          -
          <lpage>18</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>T.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Harrington</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <article-title>Bundle recommendation in ecommerce</article-title>
          ,
          <source>in: Proceedings of the 37th international ACM SIGIR conference on Research &amp; development in information retrieval</source>
          , Association for Computing Machinery, New York, NY, USA,
          <year>2014</year>
          , pp.
          <fpage>657</fpage>
          -
          <lpage>666</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S.</given-names>
            <surname>Raza</surname>
          </string-name>
          ,
          <article-title>A news recommender system considering temporal dynamics and diversity</article-title>
          ,
          <source>arXiv preprint arXiv:2103.12537</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>R.</given-names>
            <surname>Fletcher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. K.</given-names>
            <surname>Nielsen</surname>
          </string-name>
          , Automated serendipity:
          <article-title>The efect of using search engines on news repertoire balance and diversity</article-title>
          ,
          <source>Digital Journalism</source>
          <volume>6</volume>
          (
          <year>2018</year>
          )
          <fpage>976</fpage>
          -
          <lpage>989</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M.</given-names>
            <surname>Cinelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. D. F.</given-names>
            <surname>Morales</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Galeazzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Quattrociocchi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Starnini</surname>
          </string-name>
          ,
          <article-title>The echo chamber efect on social media</article-title>
          ,
          <source>Proceedings of the National Academy of Sciences</source>
          <volume>118</volume>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>O.</given-names>
            <surname>Stitini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kaloun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Bencharef</surname>
          </string-name>
          ,
          <article-title>Towards the detection of fake news on social networks contributing to the improvement of trust and transparency in recommendation systems: Trends and challenges</article-title>
          ,
          <source>Information</source>
          <volume>13</volume>
          (
          <year>2022</year>
          )
          <fpage>128</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. A.</given-names>
            <surname>Ghorbani</surname>
          </string-name>
          ,
          <article-title>An overview of online fake news: Characterization, detection, and discussion</article-title>
          ,
          <source>Information Processing &amp; Management</source>
          <volume>57</volume>
          (
          <year>2020</year>
          )
          <fpage>102025</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>R.</given-names>
            <surname>Picard</surname>
          </string-name>
          , Media and communications policy making, Springer,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ziegler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Iida</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Pfeifer</surname>
          </string-name>
          ,
          <article-title>Cheap” underwater locomotion: Morphological properties and behavioral diversity</article-title>
          , in: IROS05 Workshop on Morphology,
          <source>Control and Passive Dynamics</source>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>L.</given-names>
            <surname>Candillier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chevalier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dudognon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Mothe</surname>
          </string-name>
          ,
          <article-title>Diversity in recommender systems</article-title>
          ,
          <source>in: Proceedings: The Fourth International Conference on Advances in Human-oriented and Personalized Mechanisms</source>
          , Technologies, and
          <article-title>Services</article-title>
          . CENTRIC,
          <year>2011</year>
          , pp.
          <fpage>23</fpage>
          -
          <lpage>29</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>P.</given-names>
            <surname>Castells</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. J.</given-names>
            <surname>Hurley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Vargas</surname>
          </string-name>
          ,
          <article-title>Novelty and diversity in recommender systems</article-title>
          ,
          <source>in: Recommender systems handbook</source>
          , Springer,
          <year>2015</year>
          , pp.
          <fpage>881</fpage>
          -
          <lpage>918</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>