<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Detecting Fake News in Tweets from Text and Propagation Graph: IRISA&apos;s Participation to the FakeNews Task at MediaEval 2020</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Vincent</forename><surname>Claveau</surname></persName>
							<email>vincent.claveau@irisa.fr</email>
							<affiliation key="aff0">
								<orgName type="institution" key="instit1">CNRS</orgName>
								<orgName type="institution" key="instit2">IRISA</orgName>
								<orgName type="institution" key="instit3">Univ. Rennes</orgName>
								<address>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Detecting Fake News in Tweets from Text and Propagation Graph: IRISA&apos;s Participation to the FakeNews Task at MediaEval 2020</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">6BD7115D8EBF8E55EE5240E9C3C1DFF1</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T07:12+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This paper presents the participation of IRISA to the task of fake news detection from tweets, relying either on the text or on propagation information. For the text based detection, variants of BERTbased classification are proposed. In order to improve this standard approach, we investigate the interest of augmenting the dataset by creating tweets with fine-tuned generative models. For the graph based detection, we have proposed models characterizing the propagation of the news or the users' reputation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">TEXT-BASED APPROACHES 2.1 Pre-processing</head><p>From the tweets still online 1 , the text is extracted and pre-processed as follows. Emojis are transformed into texts <ref type="bibr" target="#b12">[13]</ref>. URLs are changed 1 At retrieval time, respectively 227, 128 and 80 tweets were no longer available for the class 'non', '5G', 'other' in the dev set.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">INTRODUCTION AND RELATED WORK</head><p>This paper describes the systems that we developed for the textbased and structure-based MediaEval 2020 Fake News detection challenge. These two subtasks and the datasets are detailed in <ref type="bibr" target="#b9">[10]</ref> and <ref type="bibr" target="#b11">[12]</ref>.</p><p>Text classification is a common NLP task <ref type="bibr" target="#b5">[6]</ref>. Although simple machine learning approaches have shown promising results for fake news detection <ref type="bibr" target="#b7">[8]</ref>, the recent transformer-based architectures, such as BeRT <ref type="bibr" target="#b1">[2]</ref>, have set new standards. Several large pre-trained transformer models are now available; they are known to yield stateof-the-art results on many NLP tasks including text classification [16, inter alia]. We rely on one of these pre-trained models to build our systems. In order to improve this standard approach, we have investigated the interest of augmenting the dataset artificially by generating tweets with fine-tuned generative models (one for each class). These approaches and results are detailed in Sec. 2.</p><p>Similarly, classification of data represented as a graph, and in particular node classification, is not new but the recent trend is to use deep learning <ref type="bibr" target="#b4">[5]</ref>. Yet, for the specific domain of fake news detection, other approaches are possible. In particular, it has been shown that the fake news are propagated differently (and faster) than legit news <ref type="bibr" target="#b14">[15]</ref>. The use of node reputation and link-based analysis, as it is done in the detection of spam web pages from the Web graph (such as TrustRank <ref type="bibr" target="#b3">[4]</ref>, an adaptation of PageRank <ref type="bibr" target="#b0">[1]</ref>) is another inspiration for our approaches. Our two approaches are further detailed in Sec. 3.</p><p>to the fixed string 'URL'. Twitter usernames are removed if they appear once, others are kept and the @ removed. The intuition is that some often cited users may be associated to a specific class. Hashtags are kept (with # removed), and decomposed when they contain a mix of capital and small letters (eg. #CovidHoax is changed in CovidHoax Covid Hoax).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Generating artificial examples</head><p>For this task we wanted to investigate the use of generative models in order to artificially augment and balance the datasets. Indeed, the performance of neural language models based on transformers <ref type="bibr" target="#b13">[14]</ref> makes this task realistic. To do so, we use GPT2 (Generative Pre-Trained Transformers), a model built from stacked transformers (precisely, decoders) trained on a large corpus by auto-regression <ref type="bibr" target="#b10">[11]</ref>. Three GPT2 models -one for each class -are fine-tuned (from the 355M-parameter pre-trained model) with the tweets of the dev set. The amount of tweets available is very small; we stopped the iterations when perplexity reached 0.5. The way this stopping criterion impacts the results would need further investigations, which were not possible due to the limited time of the challenge. For the generation, we randomly picked up tweets and kept the two first words to serve as bootstrap. The temperature, which controls the creativity of the model, was set at 0.7. Here again, we had no time to investigate the impact of this parameter. Approximately 20,000 tweets were generated for each class. Here are some tweets generated for the class '5G conspiracy':</p><p>Crude and unproductive! Turn off the 5G in your area and see if that helps. Covid19 is not funny. I hope that the Wuhan government puts an end to this immediately. "Immigrants are the cause of 5G towers, they're the cause of the coronavirus outbreak, they're the covid-19 victims, the 5G towers are the weapon which will eradicate the world population, 5G lays the microchips for the virus, i read somewhere that the 5G was debuting prior to the introduction of the COVID-19 virus to negate some of the hype around COVID-19  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3">Classification models</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.4">Results of text-based detection</head><p>The results of our models are given in Tab. 1. When available, in addition to the official score on the test set, we provide Matthews correlation coefficient (MCC), micro-F1 (accuracy) and macro-F1 on the dev data (80% for training, 20% for validation). Note that due to the cost of the artificial example generation and the small amount of data, the GPT2 models are fine-tuned on all the available dev data; we do not have reliable results for models 3 and 4 (generated tweets added to the training set can be very similar to those in the validation set).</p><p>From the results, we see that fine-tuning the representation (model 2 vs. model 1) is beneficial. Unfortunately, the artificially generated tweets (model 3 and 4) do not yield the expected improvement. From the confusion matrices, one can see that the class 'other conspiracy' has the poorest results, with tweets being equally labeled as '5G', 'non' or 'other'.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">GRAPH-BASED APPROACHES</head><p>For the second sub-task, we have proposed two models, based on two different sets of features. They are described in the following subsections, as well as the machine learning algorithms adopted and their results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Modeling the user's reputation</head><p>This set of features aims at taking into account if one of the users posting or propagating the news has already be seen. Each user is indeed associated with a score for each possible label, computed from the numbers of training samples of each class it was associated with. We also take into account the scores of the neighbors of this user, their own neighbors, and so on... In practice, this is implemented with the PageRank algorithm <ref type="bibr" target="#b0">[1]</ref> on the undirected graph with a dumping factor set to 0.8 (optimized by cross-validation). Finally, each sample ends up with one value for each class; these three scores are the features used by the classifier.</p><p>Several learning algorithms have been tested (logistic regression, random forests, SVM; as implemented in scikit learn <ref type="bibr" target="#b8">[9]</ref>). The optimal settings for their hyper-parameters are grid-searched using 20% of the dev set as validation set. The weight of each sample is adapted according to the inverse of its class proportion ('balanced' strategy). With their optimal settings, the different learning algorithms finally show little differences. For this set of features, the submitted run was produced with a random forest (1,000 trees with a maximal depth set to 5, Out-of-Bag weights used in the prediction).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Modeling the propagation</head><p>This set of features is built by considering how the tweet is propagated (without considering the users' reputations). These features can be used even if every involved user has never been seen before and is not connected any known user. The features include (with n 0 the first user tweeting the piece of news): number of nodes in the propagation graph; total number of friends and followers (for all nodes implied), as well as the median, 25% percentile, 75% percentile of followers; number of followers and friends of n 0 ; difference between the number of followers and friends of n 0 ; maximal, minimal, average, median, 25% percentile, 75% percentile of retweet time; times to reach at least 100, 1,000, 10,000 followers and so on up to 200,000 followers. With this set of features, a SVM has been used with the following parameters: standardized features (removed mean and scaled to unit variance), RBF kernel, C=0.9, gamma automatically set with the 'scale' heuristics.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Results of graph-based detection</head><p>The results of the systems are given in Tab. 1. The cross-validation and official results are consistent; they both show the advantage of the reputation-based approach, especially when considering micro-F1. The difference between cross-validation and official test score may be explained by a lower amount of already seen nodes in the test set, compared to what was generated by cross-validation. A system exploiting all the proposed features (propagation + reputation) was also tested but obtained no statistical difference with the reputation only features.</p><p>For both models, the 'other conspiracy' class is again the most error-prone (proportionally), with an equal amount of the its tweets being classified in the three classes. Overall, for both feature sets, many errors are caused by confusion between the 5G and non 5G conspiracy tweets.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">CONCLUSION AND FUTURE WORK</head><p>For the detection of fake news based on the text, we have adopted a state-of-the-art approach based on RoBerta. The scores obtained show that there exists a large margin for progress, especially when dealing with close classes (5G vs. other conspiracies). The idea of incorporating artificially generated examples did not result in better performance and still needs some work. First, we may find better ways to set the training and generation hyper-parameters. Secondly, we plan to investigate the use of generative model to expand the sample at inference time.</p><p>For the detection based on the structure, we have shown that simple approaches like reputation already offered promising results, even on small datasets with many unseen-before nodes. In addition to this type of approach, we want to explore more recent node representation techniques that make it possible to use deep learning, such as node2vec <ref type="bibr" target="#b2">[3]</ref> or subsequent variants.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Our 4</head><label>4</label><figDesc>classification variants are based on the RoBerta-large model<ref type="bibr" target="#b6">[7]</ref>. It was preferred over other transformer-based representations because its tokenizer is expected to be more suited for the tweet writing specifics. We have tested models with different classification layers (SVM, logistic regression), with or without fine tuning, and with or without artificial examples. Finally, the submitted runs are the following ones: model 1: tweet embedding from the Roberta model (not fine-tuned), and SVM (RGB kernel); model 2: Roberta model with a linear classification layer, fine-tuned on the task (3 epochs);</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 :</head><label>1</label><figDesc>Performance of the proposed systems for the textbased and graph-based detection; models are detailed in Sec. 2 and 3.</figDesc><table><row><cell></cell><cell cols="3">cross validation results</cell><cell>official</cell></row><row><cell>model</cell><cell cols="4">MCC micro-F1 macro-F1 MCC</cell></row><row><cell>model 1 (text)</cell><cell>0.4654</cell><cell>0.7460</cell><cell>0.5924</cell><cell>0.4680</cell></row><row><cell>model 2 (text)</cell><cell>0.5345</cell><cell>0.7945</cell><cell>0.6253</cell><cell>0.5571</cell></row><row><cell>model 3 (text)</cell><cell>-</cell><cell>-</cell><cell>-</cell><cell>0.4937</cell></row><row><cell>model 4 (text)</cell><cell>-</cell><cell>-</cell><cell>-</cell><cell>0.4888</cell></row><row><cell>reputation (graph)</cell><cell>0.4415</cell><cell>0.7274</cell><cell>0.5900</cell><cell>0.4093</cell></row><row><cell cols="2">propagation (graph) 0.3198</cell><cell>0.6051</cell><cell>0.4980</cell><cell>0.3036</cell></row><row><cell cols="5">model 3: same as model 2, with artificially generated examples (3</cell></row><row><cell>epochs);</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="3">model 4: same as model 3 (4 epochs).</cell><cell></cell><cell></cell></row></table></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">The Anatomy of a Large-Scale Hypertextual Web Search Engine</title>
		<author>
			<persName><forename type="first">Sergey</forename><surname>Brin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lawrence</forename><surname>Page</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Seventh International Conference on World Wide Web 7</title>
				<meeting>the Seventh International Conference on World Wide Web 7<address><addrLine>Brisbane, Australia</addrLine></address></meeting>
		<imprint>
			<publisher>Elsevier Science Publishers B. V</publisher>
			<date type="published" when="1998">1998</date>
			<biblScope unit="page" from="107" to="117" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding</title>
		<author>
			<persName><forename type="first">Jacob</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ming-Wei</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kenton</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kristina</forename><surname>Toutanova</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/N19-1423</idno>
		<ptr target="https://doi.org/10.18653/v1/N19-1423" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</title>
		<title level="s">Long and Short Papers</title>
		<meeting>the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies<address><addrLine>Minneapolis, Minnesota</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="4171" to="4186" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Node2vec: Scalable Feature Learning for Networks</title>
		<author>
			<persName><forename type="first">Aditya</forename><surname>Grover</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jure</forename><surname>Leskovec</surname></persName>
		</author>
		<idno type="DOI">10.1145/2939672.2939754</idno>
		<ptr target="https://doi.org/10.1145/2939672.2939754" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD &apos;16)</title>
				<meeting>the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD &apos;16)<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="855" to="864" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Link spam alliances</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Gyngyi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Garcia-Molina</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 31st international conference on Very large data bases</title>
				<meeting>the 31st international conference on Very large data bases<address><addrLine>Trondheim, Norway</addrLine></address></meeting>
		<imprint>
			<publisher>VLDB</publisher>
			<date type="published" when="2005">2005</date>
			<biblScope unit="page" from="517" to="528" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Representation Learning on Graphs: Methods and Applications</title>
		<author>
			<persName><forename type="first">William</forename><forename type="middle">L</forename><surname>Hamilton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Rex</forename><surname>Ying</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jure</forename><surname>Leskovec</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Computer Society Technical Committee on Data Engineering</title>
		<imprint>
			<date type="published" when="2017">2017. 2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<author>
			<persName><forename type="first">Kamran</forename><surname>Kowsari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kiana</forename><forename type="middle">Jafari</forename><surname>Meimandi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mojtaba</forename><surname>Heidarysafa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sanjana</forename><surname>Mendu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Laura</forename><surname>Barnes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Donald</forename><surname>Brown</surname></persName>
		</author>
		<title level="m">Text Classification Algorithms: A Survey</title>
				<imprint>
			<publisher>Information</publisher>
			<date type="published" when="2019">2019. 2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">RoBERTa: A Robustly Optimized BERT Pretraining Approach</title>
		<author>
			<persName><forename type="first">Yinhan</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Myle</forename><surname>Ott</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Naman</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jingfei</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mandar</forename><surname>Joshi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Danqi</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Omer</forename><surname>Levy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mike</forename><surname>Lewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Luke</forename><surname>Zettlemoyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Veselin</forename><surname>Stoyanov</surname></persName>
		</author>
		<idno>arXiv:cs.CL/1907.11692</idno>
		<imprint>
			<date type="published" when="2019">2019. 2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">MediaEval 2016: A multimodal system for the Verifying Multimedia Use task</title>
		<author>
			<persName><forename type="first">Cédric</forename><surname>Maigrot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Vincent</forename><surname>Claveau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ewa</forename><surname>Kijak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ronan</forename><surname>Sicre</surname></persName>
		</author>
		<idno type="DOI">10.1145/1235</idno>
		<ptr target="https://doi.org/10.1145/1235" />
	</analytic>
	<monogr>
		<title level="m">Verfiying Multimedia Use&quot; task. Hilversum</title>
				<meeting><address><addrLine>Netherlands</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
	<note>MediaEval 2016</note>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Scikit-learn: Machine Learning in Python</title>
		<author>
			<persName><forename type="first">F</forename><surname>Pedregosa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Varoquaux</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gramfort</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Michel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Thirion</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Grisel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Blondel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Prettenhofer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Weiss</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Dubourg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Vanderplas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Passos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Cournapeau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Brucher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Perrot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Duchesnay</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page" from="2825" to="2830" />
			<date type="published" when="2011">2011. 2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">FakeNews: Corona Virus and 5G Conspiracy Task at MediaEval</title>
		<author>
			<persName><forename type="first">Konstantin</forename><surname>Pogorelov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Daniel</forename><forename type="middle">Thilo</forename><surname>Schroeder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Luk</forename><surname>Burchard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Johannes</forename><surname>Moe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Stefan</forename><surname>Brenner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Petra</forename><surname>Filkukova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Johannes</forename><surname>Langguth</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">MediaEval 2020 Workshop</title>
				<imprint>
			<date type="published" when="2020">2020. 2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Language Models are Unsupervised Multitask Learners</title>
		<author>
			<persName><forename type="first">Alec</forename><surname>Radford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jeff</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Rewon</forename><surname>Child</surname></persName>
		</author>
		<author>
			<persName><forename type="first">David</forename><surname>Luan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Dario</forename><surname>Amodei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ilya</forename><surname>Sutskever</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">OpenAI Blog</title>
		<imprint>
			<date type="published" when="2019">2019. 2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">FACT: a Framework for Analysis and Capture of Twitter Graphs</title>
		<author>
			<persName><forename type="first">Thilo</forename><surname>Daniel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Konstantin</forename><surname>Schroeder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Johannes</forename><surname>Pogorelov</surname></persName>
		</author>
		<author>
			<persName><surname>Langguth</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Sixth International Conference on Social Networks Analysis, Management and Security (SNAMS). IEEE</title>
				<imprint>
			<date type="published" when="2019">2019. 2019</date>
			<biblScope unit="page" from="134" to="141" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<title level="m" type="main">Emoji Python library</title>
		<author>
			<persName><forename type="first">Kevin</forename><surname>Wurster</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Taehoon</forename><surname>Kim</surname></persName>
		</author>
		<ptr target="https://pypi.org/project/emoji/" />
		<imprint>
			<date type="published" when="2020">2020. 2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Attention is All you Need</title>
		<author>
			<persName><forename type="first">Ashish</forename><surname>Vaswani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Noam</forename><surname>Shazeer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Niki</forename><surname>Parmar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jakob</forename><surname>Uszkoreit</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Llion</forename><surname>Jones</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Aidan</forename><forename type="middle">N</forename><surname>Gomez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ł</forename><surname>Kaiser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Illia</forename><surname>Polosukhin</surname></persName>
		</author>
		<ptr target="http://papers.nips.cc/paper/7181-attention-is-all-you-need.pdf" />
	</analytic>
	<monogr>
		<title level="m">Advances in Neural Information Processing Systems 30</title>
				<editor>
			<persName><forename type="first">I</forename><surname>Guyon</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">U</forename><forename type="middle">V</forename><surname>Luxburg</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Bengio</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Wallach</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Fergus</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Vishwanathan</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Garnett</surname></persName>
		</editor>
		<imprint>
			<publisher>Curran Associates, Inc</publisher>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="5998" to="6008" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding</title>
		<author>
			<persName><forename type="first">Soroush</forename><surname>Vosoughi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Deb</forename><surname>Roy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sinan</forename><surname>Aral</surname></persName>
		</author>
		<author>
			<persName><forename type="first">;</forename><surname>Alex</surname></persName>
		</author>
		<author>
			<persName><forename type="first">;</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Amanpreet</forename><surname>Singh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Julian</forename><surname>Michael</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Felix</forename><surname>Hill</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Omer</forename><surname>Levy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Samuel</forename><forename type="middle">R</forename><surname>Bowman</surname></persName>
		</author>
		<idno type="DOI">10.1126/science.aap9559</idno>
		<ptr target="https://science.sciencemag.org/content/359/6380/1146." />
	</analytic>
	<monogr>
		<title level="m">7th International Conference on Learning Representations, ICLR 2019</title>
				<meeting><address><addrLine>New Orleans, LA, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018">2018. 2018. 2019. May 6-9, 2019</date>
			<biblScope unit="volume">359</biblScope>
			<biblScope unit="page" from="1146" to="1151" />
		</imprint>
	</monogr>
	<note>The spread of true and false news online</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
