<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Incorporating Context and Knowledge for Better Sentiment Analysis of Narrative Text</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Chenyang</forename><surname>Lyu</surname></persName>
							<email>chenyang.lyu2@mail.dcu.ie</email>
							<affiliation key="aff0">
								<orgName type="department">School of Computing</orgName>
								<orgName type="institution">Dublin City University Dublin</orgName>
								<address>
									<country key="IE">Ireland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Tianbo</forename><surname>Ji</surname></persName>
							<email>tianbo.ji2@mail.dcu.ie</email>
							<affiliation key="aff1">
								<orgName type="department">ADAPT Centre</orgName>
								<orgName type="institution">Dublin City University</orgName>
								<address>
									<settlement>Dublin</settlement>
									<country key="IE">Ireland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Yvette</forename><surname>Graham</surname></persName>
							<email>yvette.graham@dcu.ie</email>
							<affiliation key="aff2">
								<orgName type="department">ADAPT Centre</orgName>
								<orgName type="institution">Dublin City University</orgName>
								<address>
									<settlement>Dublin</settlement>
									<country key="IE">Ireland</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Incorporating Context and Knowledge for Better Sentiment Analysis of Narrative Text</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">E63C527CC1A28E43531D7756A15D3351</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T14:12+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Recent years have witnessed a significant increase in the availability of online narrative texts, such as news articles and online stories. Automatic sentiment classification of such narrative texts will enable better organisation, search and retrieval. Little research has been carried out to date that to automatically predict sentiment of narrative text however despite much work being carried out to date on sentiment analysis of social media and user feedback. In this paper, we present an approach to sentiment analysis of narrative text that employs pre-trained language model, an approach already proven e↵ective for a range of other NLP tasks. However, for the purpose of sentiment analysis of narrative text in particular, we introduce two new features: a contextual feature and extra knowledge feature that prove to aid text understanding for prediction of sentiment. We conduct preliminary experiments on a publicly available dataset of fairy tale texts. Results show an increase of accuracy over vanilla pre-trained language model and baseline models for classification of sentiment for narrative texts.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Sentiment analysis refers to the analysis and prediction of the sentiment and emotion in language. Most research carried out to date in this domain has been based on user-generated content such as social media and user feedback <ref type="bibr" target="#b18">[VC12]</ref>. Recently, with the development of online narrative resources such as web-fiction and news, the interest in sentiment analysis of narrative text is slowly expanding <ref type="bibr" target="#b10">[Moh13]</ref> in order to make better organisation and retrieval. The sentiment and emotion of narrative text can characterised defined as a general feeling reflected in the text and in this current work we use sentiment to stand for both sentiment and emotion.</p><p>Contrary to other types of text, narrative is a perceived sequence of non-randomly connected events <ref type="bibr" target="#b17">[Too12]</ref>. For example, A countryman's son stepped on a snake's tail accidentally. The tail suddenly turned and hit him so that he died. The father was very angry so that he cut o↵ part of the snake's tail. is a piece of narrative text 1 composed of connected events. Due to its complexity, analysing the sentiment of narrative text requires a deep understanding of semantic structure and opinions. Since the events that impact the plot of narrative texts are commonly interrelated, this increases the importance of su ciently taking context into consideration. Furthermore, there are several linguistic phenomena, such as metaphor in sentence The lack of having proper standing army is Achilles heel for this small country, and terminology requiring additional knowledge in order to be correctly interpreted. Incorporating such additional knowledge has the potential to help predict the sentiment of text.</p><p>Nowadays, pre-trained language models, such as those in <ref type="bibr" target="#b4">[DCLT19]</ref>, [PNI + 18] and <ref type="bibr">[RNSS]</ref>, have proven e↵ective for a range of tasks including syntactic and semantic tasks, like textual similarity, sentiment classification, textual entailment, etc. These models are usually trained on large scale corpora and have the ability to capture and model semantic dependencies over long text. Such pre-trained models have the potential to detect and classify the sentiment of narrative text.</p><p>In this paper, we propose a model that employs a state-of-the-art pre-trained language model to classify the sentiment class of narrative text. This model takes context into consideration as well as additional knowledge as features to predict the sentiment class. Firstly, what we call context features are used to generate a contextaware representation of a given sentence, before it is used to determine its sentiment class. We provide a detailed description of the methodology and experiments investigating the accuracy of our approach in Sections 3 and 4, respectively.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Related Work</head><p>Sentiment Analysis is regarded as a text classification or text mining problem, and there are two main approaches to this problem: a rule-based approach and a learning-based approach. At first in sentiment analysis, a rulebased approach was attempted involving pre-defined lexicon with corresponding sentiment polarity to predict the polarity of a given text by capturing word occurrences [HS00, DC07, WWB + 04]. Besides lexicon methods, Zhang et al [ZZL + 09] also proposed a document-level sentiment analysis framework, which determines sentiment polarity of sentences based on word dependency before aggregating sentences for document-level sentiment prediction, in which single sentence polarity is computed independent of context.</p><p>However, sentiment of text can be expressed in a more subtle manner than is possible to detect with such approaches, since they are di cult classify only based on indicative lexical features. Machine learning algorithms have the potential to address the short-comings of rule-based approaches. However, learning-based approaches require a labeled training set to be trained on and then predict new samples. Pang et al <ref type="bibr" target="#b12">[PLV02]</ref> proposed three machine learning models, Naïve Bayes, Support Vector Machine and Maximum Entropy to predict sentiment based on bag-of-word feature. <ref type="bibr" target="#b11">[NIK10]</ref> proposed a dependency tree-based model to capture interactions between words, while <ref type="bibr" target="#b19">[VCB14]</ref> and <ref type="bibr" target="#b9">[KF16]</ref> apply a range of kernel methods.</p><p>Sentiment analysis with neural networks has a long history. For example, Socher et al [SPW + 13] proposed Recursive Neural Tensor Networks(RNTNs) to model sentence-level sentiment recursively from its syntactic tree.</p><p>[dSG14] on the other hand adopted Convolutional Neural Networks (CNNs) to classify sentiment for short text, their model outperforming traditional machine learning models, such as Naïve Bayes and SVM. Some alternate approaches using Recurrent Neural Networks have also been proposed, such as [WHZ + 16] and <ref type="bibr" target="#b24">[ZWX16]</ref>, the latter employing attention-based LSTM models.</p><p>Recently, large-scale pre-trained models have emerged and have been shown very e↵ective on various tasks [PNI + 18, RNSS, DCLT19], including reading comprehension and sentiment classification, etc. Pre-trained models are trained on large unlabeled corpora in unsupervised setting, after pre-training they are able to capture rich semantic patterns and syntactic information from text <ref type="bibr" target="#b8">[JSS19]</ref>. For instance, the noun modifier(The former ) in sentence The 45-year-old former General Electric CO. executive figures it will be easier this time. can be clearly attended to its noun (executive) by self-attention mechanism in BERT <ref type="bibr" target="#b1">[CKLM19]</ref>. Pre-trained model is quite suitable for us to model sentiment of text, we therefore employ BERT(Bidirectional Encoder Representation from Transformer) on our task because of its capability to model bidirectional dependencies among text.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Methodology</head><p>Most current state-of-the-art pre-trained language models are based on a neural architecture known as the transformer[VSP + 17], originally proposed by Vaswani et al.In this study, we use pre-trained model BERT [WDS + 19]. The BERT <ref type="bibr" target="#b4">[DCLT19]</ref> we use has been pre-trained on a large unlabeled corpus containing over 3 billion words. This large volume of data enables it to model complex language dependencies in text and produce very rich representations of text. The input is represented as document = {s 1 , ......, s n }, s i = {w i,1 , ......, w i,m }, where s i is the i-th sentence in document. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">BERT-Bidirectional Encoder Representation from Transformer</head><p>We use BERT as encoder to generate sentiment label for a single sentence, which is based on a transformer block containing one self-attention layer, two layer norms and one feed-forward layer. The transformer we employed consists of 12 stacked transformer blocks, and the output of its last layer is used to get the sentiment label. The overall process of computing sentiment label for a single sentence can be formulated as following:</p><formula xml:id="formula_0">h l = bert encoder(w 1 , w 2 , ......, w k )<label>(1)</label></formula><p>y = sof tmax(linear layer(pooling layer(h l ), context))</p><p>where h l is hidden state vectors from the last layer and it will be fed into pooling layer<ref type="foot" target="#foot_0">2</ref> and linear layer, then passed into softmax layer with context vector to generate the sentiment label y, which will be introduced in section 3.2. The overall architecture of our model is shown in Figure <ref type="figure" target="#fig_0">1</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Contextual Feature</head><p>Usually, the sentiment label of a single sentence in a document is only conditioned on itself:</p><formula xml:id="formula_2">sentiment label = argmax y p(y|s j )<label>(3)</label></formula><p>which does not take advantage of the context. However, we argue that in narrative, the context of a given sentence plays a crucial role in determining its semantic meaning as well as sentiment label.</p><p>We propose to incorporate context features when we compute the sentiment label for a sentence by combining the representation vector of context sentences of a certain sentence, in this way we propose to give sentiment label as following:</p><formula xml:id="formula_3">sentiment label = argmax y p(y|s j , context j ) (4)</formula><p>context j = (s j t , s j t+1 , ......, s j 1 ) (5</p><formula xml:id="formula_4">)</formula><p>where t is the number of sentences we incorporate in context, is the function we use to combine sentence vectors. In our framework, we use the linear combination of context sentences and the target sentence to generate sentiment label. In this work, we use the last sentence of current sentence as it's context, which has the most significant e↵ect on the sentiment of current sentence.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Extra Knowledge</head><p>Besides context feature, we propose to use extra knowledge information to improve understanding of sentiment class of text. In this paper, we incorporate entity information and idiom information into our model. We use idioms dataset SLIDE from <ref type="bibr" target="#b7">[JBBHS18]</ref>, which includes over 5000 idioms with sentiment annotation. In SLIDE dataset, every idiom has a sentiment distribution, which has three entries. Every entry scales from 0 to 1 and the sum of three entries is 1. For our framework, the extra knowledge is represented by a sentiment-term matrix, every entry of a certain term represents their inclination to corresponding sentiment label and all entries sum to one. The sentiment-term matrix is defined as :</p><formula xml:id="formula_5">M i,j = p(y j |term i ), z X j=1 M i,j = 1<label>(6)</label></formula><p>In our model, we incorporate extra knowledge as following:</p><formula xml:id="formula_6">p(y|s j , context j ) = sof tmax(normalize(h j ) + v X i=1 M i ) (<label>7</label></formula><formula xml:id="formula_7">)</formula><p>M i is the sentiment distribution vector of the term that is attended in given sentence, v is the number of all attended terms. We first normalize the hidden state h l and then add normalized h l and all term vectors {M i } v i=1 which a↵ect sentiment of text. In our normalize function, the hidden state h l is normalized and scaled by a factor , which makes the ratio of l 1 norms of h l and M i to . In our experiment, we set to 10.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Experiments and Results</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Data</head><p>For a preliminary validation of our methodology, we evaluate our proposed model with di↵erent features on dataset in <ref type="bibr" target="#b0">[Alm08]</ref>. This dataset includes 176 fairy tales from three sets of children's stories: Beatrix Potter(19 stories), H. C. Andersen(77 stories), and the Brothers Grimm(80 stories). There are totally 15,302 sentences annotated with emotional labels in sentence-level. Emotion label set consists of eight labels that are categorized into emotional and non-emotional. The label distribution is in Table <ref type="table">.</ref>1. The dataset is imbalanced, NEUTRAL was most frequent with 66.26%. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Conclusion and Future Work</head><p>In this paper, we apply pre-trained language models to sentence-level sentiment analysis of narrative texts and introduce two new features into model: context feature and extra knowledge feature. We evaluate the performance of a pre-trained model for sentiment analysis of fairy tales with a range of features. Experiments demonstrate that our method improves classification accuracy, showing the importance of incorporating context for sentiment analysis of narrative texts. In this work, we propose a sentence-level model to classify sentiment of narrative texts however how to build a document-level model based on that remains to be further investigated.</p><p>Our extra knowledge dataset only covers a small portion of text due to limitations with respect to acquisition of the knowledge set with sentiment annotation, hence our incorporation of knowledge did not make a significant improvement. In future work, we aim to build a larger knowledge dataset to improve coverage for narrative text. In addition, we would like to further investigate the alternative ways to inject knowledge into sentiment prediction for narrative text.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Model Architecture for Sentence-Level Sentiment Classification Incorporating Contextual Feature and Knowledge. The blue arrows stand for contextual vectors from previous sentences.</figDesc><graphic coords="3,101.59,84.15,412.43,177.19" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 :</head><label>1</label><figDesc>Label distribution in dataset, where H is HAPPY,Su+ is POSITIVELY SURPRISED, F is FEARFUL, A is ANGRY, Sa is SAD, D is DISGUSTED, Su-is NEGATIVELY SURPRISED, N is NEUTRALWe firstly split the dataset described in Sec.4.1 into training and test sets. Since in our study we take context into consideration, we select 10% stories from each set of stories as test set. Our test set contains 16 stories, 1,189 sentences in total. The label distribution in training set and test set is in Table.2.</figDesc><table><row><cell></cell><cell></cell><cell cols="2">Emotional</cell><cell></cell><cell></cell><cell></cell><cell cols="2">Non-Emotional</cell></row><row><cell>H</cell><cell>Su+</cell><cell>F</cell><cell>A</cell><cell>Sa</cell><cell>D</cell><cell>Su-</cell><cell>N</cell><cell></cell></row><row><cell cols="7">10.52% 2.15% 4.55% 4.77% 5.43% 3.03% 3.29%</cell><cell>66.26%</cell><cell></cell></row><row><cell>4.2 Experiment Setup</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell cols="7">Table 2: Label distribution in training set and test set.</cell><cell></cell></row><row><cell>label</cell><cell>H</cell><cell>Su+</cell><cell>F</cell><cell>A</cell><cell>Sa</cell><cell>D</cell><cell>Su-</cell><cell>N</cell></row><row><cell cols="9">training set 10.58% 2.12% 4.44% 4.32% 5.31% 3.08% 3.19% 66.96%</cell></row><row><cell>test set</cell><cell cols="8">9.84% 2.52% 5.88% 10.17% 6.90% 2.35% 4.46% 57.86%</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_0">We use "pooling layer" in accordance with huggingface and this layer simply selects the representation vector of token [CLS], which is a special token added in front of every sentence.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_1">https://huggingface.co/transformers/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_2">https://scikit-learn.org/stable/index.html</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgements</head><p>This work was supported by Science Foundation Ireland. The authors would like to thank Jennifer Foster and three anonymous reviewers for their helpful comments.</p></div>
			</div>

			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The pre-trained model we use is BERT-Base-Cased 3 . The four baselines we select are naïve majority classifier, always predicting the most likely label NEUTRAL, and Support Vector Machines (SVM) <ref type="bibr" target="#b2">[CV95]</ref> implemented by scikit-learn 4 [PVG + 11], we use TF-IDF and word2vec embeddings as input to SVM, for word2vec we sum all word embeddings in sentences and then fed it into model. The last baseline is Naïve Bayes. We fine-tune pre-trained model on our training set for 2 epochs, and learning rate of 3e-6, which are recommended settings in original BERT paper <ref type="bibr" target="#b4">[DCLT19]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Classification Results</head><p>Table <ref type="table">3</ref> shows our first set of experimental results for classifying sentences to either emotional or non-emotional (i.e., for detecting emotional contents in narratives). In evaluation results, pre-training language model with context feature outperforms all the other models, which proves the e↵ectiveness of incorporating contextual feature. BERT(WF) performs even worse than the naïve classifier, that shows the necessity of fine-tuning. BERT(CF) outperforms vanilla BERT by 2.36% in accuracy and 3.38% in precision, and performs 12.03% better in accuracy and 13.87% in precision over the strongest baselines model Naïve Bayes. We can see from the Table <ref type="table">3</ref>. the improvement of incorporating extra knowledge isn't obvious, we argue that may caused by the limited coverage of our knowledge base and the way we incorporate knowledge. Experiments were also run for Bi-LSTMs and CNNs, however the results are not included. Overall, they performed worse than all models listed. Furthermore, we conduct experiments on multi-class classification of all labels including HAPPY, POSI-TIVELY SURPRISED, FEARFUL, ANGRY, SAD, DISGUSTED, NEGATIVELY SURPRISED, NEUTRAL. Results are shown in Table <ref type="table">4</ref>. In the second experiment, we unfortunately did not employ knowledge features because our knowledge base is incompatible with label set. Results in terms of accuracy, precision, recall and F1score are included. From Table <ref type="table">4</ref>, we can see that BERT with contextual feature outperforms the other models in four metrics. The averaged accuracy of BERT(CF) improved by approximately 6.99% over SVM+Word2Vec, by 7.08% over SVM+TF-IDF and by 3.21% over vanilla BERT. Since our dataset is imbalanced, if more annotated data becomes available the performance of our model seems likely to improve.</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">A↵ect in Text and Speech</title>
		<author>
			<persName><forename type="first">Cecilia</forename><surname>Alm</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
		<respStmt>
			<orgName>University of Illinois at Urbana-Champaign</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">PhD thesis</note>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">What does BERT look at? an analysis of bert&apos;s attention</title>
		<author>
			<persName><forename type="first">Kevin</forename><surname>Clark</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Urvashi</forename><surname>Khandelwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Omer</forename><surname>Levy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Christopher</forename><forename type="middle">D</forename><surname>Manning</surname></persName>
		</author>
		<idno>CoRR, abs/1906.04341</idno>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Support-vector networks</title>
		<author>
			<persName><forename type="first">Corinna</forename><surname>Cortes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Vladimir</forename><surname>Vapnik</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Machine Learning</title>
				<imprint>
			<date type="published" when="1995-09">Sep 1995</date>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="page" from="273" to="297" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Yahoo! for amazon: Sentiment extraction from small talk on the web</title>
		<author>
			<persName><forename type="first">R</forename><surname>Sanjiv</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mike</forename><forename type="middle">Y</forename><surname>Das</surname></persName>
		</author>
		<author>
			<persName><surname>Chen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Management Science</title>
		<imprint>
			<biblScope unit="volume">53</biblScope>
			<biblScope unit="issue">9</biblScope>
			<biblScope unit="page" from="1375" to="1388" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">BERT: Pre-training of deep bidirectional transformers for language understanding</title>
		<author>
			<persName><forename type="first">Jacob</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ming-Wei</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kenton</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kristina</forename><surname>Toutanova</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</title>
		<title level="s">Long and Short Papers</title>
		<meeting>the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies<address><addrLine>Minneapolis, Minnesota</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019-06">June 2019</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="4171" to="4186" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Deep convolutional neural networks for sentiment analysis of short texts</title>
		<author>
			<persName><forename type="first">Santos</forename><surname>Cícero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Maíra</forename><surname>Gatti</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of COLING 2014, the 25th International Conference on Computational Linguistics: Technical Papers</title>
				<meeting>COLING 2014, the 25th International Conference on Computational Linguistics: Technical Papers<address><addrLine>Dublin, Ireland; Dublin City</addrLine></address></meeting>
		<imprint>
			<publisher>University and Association for Computational Linguistics</publisher>
			<date type="published" when="2014-08">August 2014</date>
			<biblScope unit="page" from="69" to="78" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Fuzzy typing for document management</title>
		<author>
			<persName><forename type="first">Alison</forename><surname>Huettner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pero</forename><surname>Subasic</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ACL 2000 Companion Volume: Tutorial Abstracts and Demonstration Notes</title>
				<imprint>
			<date type="published" when="2000">2000</date>
			<biblScope unit="page" from="26" to="27" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">SLIDE -a sentiment lexicon of common idioms</title>
		<author>
			<persName><forename type="first">Charles</forename><surname>Jochim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Francesca</forename><surname>Bonin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Roy</forename><surname>Bar-Haim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Noam</forename><surname>Slonim</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018)</title>
				<meeting>the Eleventh International Conference on Language Resources and Evaluation (LREC 2018)<address><addrLine>Miyazaki, Japan</addrLine></address></meeting>
		<imprint>
			<publisher>ELRA</publisher>
			<date type="published" when="2018-05">May 2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">What does BERT learn about the structure of language?</title>
		<author>
			<persName><forename type="first">Ganesh</forename><surname>Jawahar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Benoît</forename><surname>Sagot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Djamé</forename><surname>Seddah</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics</title>
				<meeting>the 57th Annual Meeting of the Association for Computational Linguistics<address><addrLine>Florence, Italy</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019-07">July 2019</date>
			<biblScope unit="page" from="3651" to="3657" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Detecting opinion polarities using kernel methods</title>
		<author>
			<persName><forename type="first">Rasoul</forename><surname>Kaljahi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jennifer</forename><surname>Foster</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Workshop on Computational Modeling of People&apos;s Opinions, Personality, and Emotions in Social Media (PEOPLES)</title>
				<meeting>the Workshop on Computational Modeling of People&apos;s Opinions, Personality, and Emotions in Social Media (PEOPLES)<address><addrLine>Osaka, Japan</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016-12">December 2016</date>
			<biblScope unit="page" from="60" to="69" />
		</imprint>
	</monogr>
	<note>The COLING 2016. Organizing Committee</note>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<title level="m" type="main">From once upon a time to happily ever after: Tracking emotions in novels and fairy tales</title>
		<author>
			<persName><forename type="first">Saif</forename><surname>Mohammad</surname></persName>
		</author>
		<idno>CoRR, abs/1309.5909</idno>
		<imprint>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Dependency tree-based sentiment classification using crfs with hidden variables</title>
		<author>
			<persName><forename type="first">Tetsuji</forename><surname>Nakagawa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kentaro</forename><surname>Inui</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sadao</forename><surname>Kurohashi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Human Language Technologies: The 2010 Annual Conference of the North American Chapter of the Association for Computational Linguistics</title>
				<imprint>
			<publisher>Association for Computational Linguistics</publisher>
			<date type="published" when="2010">2010</date>
			<biblScope unit="page" from="786" to="794" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Thumbs up?: Sentiment classification using machine learning techniques</title>
		<author>
			<persName><forename type="first">Bo</forename><surname>Pang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lillian</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Shivakumar</forename><surname>Vaithyanathan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the ACL-02 Conference on Empirical Methods in Natural Language Processing -Volume 10</title>
				<meeting>the ACL-02 Conference on Empirical Methods in Natural Language Processing -Volume 10<address><addrLine>Stroudsburg, PA, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2002">2002</date>
			<biblScope unit="page" from="79" to="86" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Deep contextualized word representations</title>
		<author>
			<persName><forename type="first">Mark</forename><surname>Pni + ; Matthew E Peters</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mohit</forename><surname>Neumann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Matt</forename><surname>Iyyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Christopher</forename><surname>Gardner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kenton</forename><surname>Clark</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Luke</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><surname>Zettlemoyer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of NAACL-HLT</title>
				<meeting>NAACL-HLT</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="2227" to="2237" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Scikit-learn: Machine learning in Python</title>
		<author>
			<persName><forename type="first">;</forename><forename type="middle">F</forename><surname>Pvg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Pedregosa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Varoquaux</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Gramfort</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Michel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Thirion</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Grisel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Blondel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Prettenhofer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Weiss</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Dubourg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Vanderplas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Passos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Cournapeau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Brucher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Perrot</surname></persName>
		</author>
		<author>
			<persName><surname>Duchesnay</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page" from="2825" to="2830" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<title level="m" type="main">Improving language understanding by generative pre-training</title>
		<author>
			<persName><forename type="first">Alec</forename><surname>Radford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Karthik</forename><surname>Narasimhan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tim</forename><surname>Salimans</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ilya</forename><surname>Sutskever</surname></persName>
		</author>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Recursive deep models for semantic compositionality over a sentiment treebank</title>
		<author>
			<persName><surname>Spw + ; Richard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alex</forename><surname>Socher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jean</forename><surname>Perelygin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jason</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Christopher</forename><forename type="middle">D</forename><surname>Chuang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Andrew</forename><surname>Manning</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Christopher</forename><surname>Ng</surname></persName>
		</author>
		<author>
			<persName><surname>Potts</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing</title>
				<meeting>the 2013 Conference on Empirical Methods in Natural Language Processing<address><addrLine>Seattle, Washington, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2013-10">October 2013</date>
			<biblScope unit="page" from="1631" to="1642" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<title level="m" type="main">Narrative: A critical linguistic introduction</title>
		<author>
			<persName><forename type="first">Michael</forename><surname>Toolan</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2012">2012</date>
			<publisher>Routledge</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Sentiment analysis and opinion mining: a survey</title>
		<author>
			<persName><forename type="first">G</forename><surname>Vinodhini</surname></persName>
		</author>
		<author>
			<persName><surname>Chandrasekaran</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page" from="282" to="292" />
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">A context-based model for sentiment analysis in twitter</title>
		<author>
			<persName><forename type="first">Andrea</forename><surname>Vanzo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Danilo</forename><surname>Croce</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Roberto</forename><surname>Basili</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of coling 2014, the 25th international conference on computational linguistics: Technical papers</title>
				<meeting>coling 2014, the 25th international conference on computational linguistics: Technical papers</meeting>
		<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="2345" to="2354" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<title level="m" type="main">Attention is all you need</title>
		<author>
			<persName><surname>Vsp + ; Ashish</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Noam</forename><surname>Vaswani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Niki</forename><surname>Shazeer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jakob</forename><surname>Parmar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Llion</forename><surname>Uszkoreit</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Aidan</forename><forename type="middle">N</forename><surname>Jones</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lukasz</forename><surname>Gomez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Illia</forename><surname>Kaiser</surname></persName>
		</author>
		<author>
			<persName><surname>Polosukhin</surname></persName>
		</author>
		<idno>CoRR, abs/1706.03762</idno>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<monogr>
		<title level="m" type="main">Huggingface&apos;s transformers: State-of-the-art natural language processing</title>
		<author>
			<persName><surname>Wds + ; Thomas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lysandre</forename><surname>Wolf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Victor</forename><surname>Debut</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Julien</forename><surname>Sanh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Clement</forename><surname>Chaumond</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Anthony</forename><surname>Delangue</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pierric</forename><surname>Moi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tim</forename><surname>Cistac</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R'emi</forename><surname>Rault</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Morgan</forename><surname>Louf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jamie</forename><surname>Funtowicz</surname></persName>
		</author>
		<author>
			<persName><surname>Brew</surname></persName>
		</author>
		<idno>ArXiv, abs/1910.03771</idno>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Attention-based lstm for aspect-level sentiment classification</title>
		<author>
			<persName><surname>Whz + ; Yequan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Minlie</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Li</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><surname>Zhao</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2016 conference on empirical methods in natural language processing</title>
				<meeting>the 2016 conference on empirical methods in natural language processing</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="606" to="615" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Learning subjective language</title>
		<author>
			<persName><forename type="first">Janyce</forename><surname>Wwb + ;</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Theresa</forename><surname>Wiebe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Rebecca</forename><surname>Wilson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Matthew</forename><surname>Bruce</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Melanie</forename><surname>Bell</surname></persName>
		</author>
		<author>
			<persName><surname>Martin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computational linguistics</title>
		<imprint>
			<biblScope unit="volume">30</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="277" to="308" />
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Attention-based lstm network for cross-lingual sentiment classification</title>
		<author>
			<persName><forename type="first">Xinjie</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Xiaojun</forename><surname>Wan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jianguo</forename><surname>Xiao</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2016 conference on empirical methods in natural language processing</title>
				<meeting>the 2016 conference on empirical methods in natural language processing</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="247" to="256" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Sentiment analysis of chinese documents: From sentence to document level</title>
		<author>
			<persName><surname>Zzl + ; Changli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Daniel</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jiexun</forename><surname>Zeng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Fei-Yue</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Wanli</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><surname>Zuo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">J. Am. Soc. Inf. Sci. Technol</title>
		<imprint>
			<biblScope unit="volume">60</biblScope>
			<biblScope unit="issue">12</biblScope>
			<biblScope unit="page" from="2474" to="2487" />
			<date type="published" when="2009-12">December 2009</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
