<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="es">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">LABDA at the 2016 TASS challenge task: using word embeddings for the sentiment analysis task * LABDA en la competición TASS 2016: utilizando vectores de palabras para la tarea de análisis de sentimiento</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Antonio</forename><surname>Quirós</surname></persName>
							<email>antonio.quiros@sngular.team</email>
							<affiliation key="aff0">
								<orgName type="department">Departamento de Informática</orgName>
								<orgName type="institution">Universidad Calos III de Madrid Avd. de la Universidad</orgName>
								<address>
									<addrLine>30</addrLine>
									<postCode>28911</postCode>
									<settlement>Leganés, Madrid</settlement>
									<country key="ES">España</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="department">Sngular Data&amp;Analytics</orgName>
								<address>
									<addrLine>Av, LLano Castellano 13, Planta 5</addrLine>
									<postCode>28034</postCode>
									<settlement>Madrid</settlement>
									<country key="ES">España</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Isabel</forename><surname>Segura-Bedmar</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Departamento de Informática</orgName>
								<orgName type="institution">Universidad Calos III de Madrid Avd. de la Universidad</orgName>
								<address>
									<addrLine>30</addrLine>
									<postCode>28911</postCode>
									<settlement>Leganés, Madrid</settlement>
									<country key="ES">España</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Paloma</forename><surname>Martínez</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Departamento de Informática</orgName>
								<orgName type="institution">Universidad Calos III de Madrid Avd. de la Universidad</orgName>
								<address>
									<addrLine>30</addrLine>
									<postCode>28911</postCode>
									<settlement>Leganés, Madrid</settlement>
									<country key="ES">España</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff2">
								<orgName type="department">Análisis de Sentimiento</orgName>
								<address>
									<addrLine>Vectores de palabras</addrLine>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">LABDA at the 2016 TASS challenge task: using word embeddings for the sentiment analysis task * LABDA en la competición TASS 2016: utilizando vectores de palabras para la tarea de análisis de sentimiento</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">8DDF01BB01817387FA9CCF4D93A1A0F0</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T10:55+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>Sentiment Analysis, Word embeddings</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Resumen: Este artículo describe la participación del grupo LABDA en la tarea 1 (Sentiment Analysis at global level) de la competición TASS 2016. En nuestro enfoque, los tweets son representados por medio de vectores de palabras y son clasificados utilizando algoritmos como SVM y regresión logística.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="es">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Knowing the opinion of customers or users has become a priority for companies and organizations in order to improve the quality of their services and products. With the ongoing explosion of social media, it affords a significant opportunity to poll the opinion of many Internet users by processing their comments. However, it should be noted that sentiment analysis, which can be defined as the automatic analysis of opinion in texts <ref type="bibr" target="#b6">(Pang and Lee, 2008)</ref>, is a challenging task because it is not strange that different people assign different polarities to a given text. On Twitter, the task is even more difficult, because the texts are small (only 140 characters) and are charectized by their informal style language, many grammatical errors and spelling mistakes, slang and vulgar vocabulary and abbreviations.</p><p>Since their introduction in 2013, the TASS shared task editions have had as main goal to promote the development of methods and resources for sentiment analysis of tweets in Spanish. This paper describes the participation of the LABDA group at the Task 1 (Sentiment Analysis at global level). In this task, the participating systems have to determine the global polarity of each tweet in the test dataset. There are two different evaluations: one based on 6 different polarity labels (P+, P, NEU, N, N+, NONE) and another based on just 4 labels (P, N, NEU, NONE). A detailed description of the task can be found in the overview paper of TASS 2016 <ref type="bibr" target="#b3">(García-Cumbreras et al., 2016)</ref>. Our approach exploits word embedding representations for tweets and machine learning algorithms such as SVM and logistics regression. The word embedding model can yield significant dimensionality reduction compared to the classical Bag-Of-Word (BoW) model. The dimensionality redution can have several positive effects on our algorithms such as faster training, avoiding overfitting and better performance.</p><p>The paper is organized as follows. Section 2 describes our approach. The experimental results are presented and discussed in Section 3. We conclude in Section 4 with a summary of our findings and some directions for future work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">System</head><p>In this paper, we study the use of word embeddings (also known as word vectors) in order to represent tweets and then examine several machine learning algorithms to classify them. Word embeddings have shown promising results in NLP tasks, such as named entity recognition (Segura-Bedmar, <ref type="bibr" target="#b7">Suárez-Paniagua, and Martınez, 2015)</ref>, relation extraction <ref type="bibr" target="#b0">(Alam et al., 2016)</ref>, sentiment analysis <ref type="bibr" target="#b9">(Socher et al., 2013b)</ref> or parsing <ref type="bibr" target="#b8">(Socher et al., 2013a)</ref>. A word embedding is a function to map words to low dimensional vectors, which are learned from a large collection of texts. At present, Neural Network is one of the most used learning techniques for generating word embeddings <ref type="bibr" target="#b4">(Mikolov and Dean, 2013)</ref>. The essential assumption of this model is that semantically close words will have similar vectors (in terms of cosine similarity). Word embeddings can help to capture semantic and syntactic relationships of the corresponding words.</p><p>While the well-known Bag-of-Words (BoW) model involves a very large number of features (as many as the number of nonstopwords words with at least a minimum number of occurrences in the training data), the word embedding representation allows a significant reduction in the feature set size (in our case, from million to just 300). The dimensionality reduction is a desirable goal, because it helps in avoiding overfitting and leads to a reduction of the training and classification times, without any performance loss.</p><p>As a preprocessing step, tweets must be cleaned. First, we remove all links and urls. We then remove usernames which can be easily recognized because their first character is the symbol @. We then transform the hashtags to words by removing its first character (that is, the symbol #). Taking advantage of regular expressions, the emoticons are detected and classified in order to count the number of positive and negative emoticons in each tweet and then we remove them from the text. Table <ref type="table">1</ref> shows the list of positive and negative emoticons, which were taken from the wikipedia page https://en.wikipedia. org/wiki/List\_of\_emoticons. We con-vert the tweets to lowercase and replace misspelled accented letters with the correct one (for instance "à" with "á"). We also treat elongations (that is, the repetition of a character) by removing the repetition of a character after its second occurrence (for example, "hoooolaaaa" would be translated to "hola"). We then decided to take into account laughs (for instance "jajaja") which turned out to be challenging because of the diverse ways they are expressed (i.e. expressions like "jajajaja" or "jejeje" and even misspelled ones like "jajjajaaj") We addressed this using regular expressions to standardize the different forms (i.e. "jajjjaaj" to "jajaja") and then replace them with the word "risas". Finally we remove all non-letters characters and all stopwords present in tweets<ref type="foot" target="#foot_0">1</ref> . Table <ref type="table">1</ref>: List of positive and negative emoticons</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Orientation Emoticons</head><p>Once the tweets are preprocessed, they are tokenized using the NLKT toolkit (a Python package for NLP); we also performed experimentation by lemmatizing each tweet using MeaningCloud<ref type="foot" target="#foot_1">2</ref> Text Analytic software to compare both approaches. Then, for each token, we search its vector in the word embedding model. We use a pretrained model <ref type="bibr" target="#b1">(Cardellino, 2016)</ref>, which was generated by using the word2vec algorithm <ref type="bibr" target="#b4">(Mikolov and Dean, 2013)</ref> from a collection of Spanish texts with approximately 1.5 billion words. The dimension of the word embedding is 300. It should be noted that these texts were taken from different resources such as Spanish Wikipedia, WikiSource and Wikibooks, but none of them contains tweets. Therefore, it is possible that the main characteristics of the social media texts (such as informal style language, noisy, plenty of grammatical errors and spelling mistakes, slang and vulgar vocabulary, abbreviations, etc) are not correctly represented in this model. One of the main problems is that there is a significant number of words (almost a 13 % of the vocabulary, representing the 6 % of words occurrences) that are not found in the model. We perform a review of a small sample of these words, showing that most of them were mainly hashtags.</p><p>In our approach, a tweet of n tokens (T = w 1 , w 2 , ..., w n ) is represented as the centroid of the word vectors w i of its tokens, as shown in the following equation:</p><formula xml:id="formula_0">T = 1 n n i=1 w i = N j=1 w j .T F (w j , t) N j=1 T F (w j , t)<label>(1)</label></formula><p>where N is the vocabulary size, that is, the total number of distinct words, while T F (w j , t) refers to the number of occurrences of the j-th vocabulary word in the tweet T.</p><p>We also explore the effect of including the inverse document frequencies IDF to represent tweets (see Equation <ref type="formula">2</ref>). This helps to increase the weight of words that occur often, but only in a few documents, while it reduces the relevance of words that occur very frequently in a larger number of texts.</p><formula xml:id="formula_1">T = 1 n n i=1 w i = N j=1 w j .T F (w j , t).IDF (w j ) N j=1 T F (w j , t).IDF (w j )</formula><p>(2) having IDF (w j ) = log|D| |tw∈D:w j ∈tw| where |D| refers to the number of tweets.</p><p>In addition to using the centroid, we assess the impact of complementing the tweet model with the following additional features: posWords: number of positive words present in the tweet. negWords: number of negative words present in the tweet. posEmo: number of positive emoticons present in the tweet. negEmo: number of negative emoticons present in the tweet.</p><p>For the posWords and negWords features we used the iSOL lexicon <ref type="bibr" target="#b5">(Molina-González et al., 2013)</ref>, a list composed by 2,509 positive words and 5,626 negative words. As described before, for the emoticons we used the listed in Table <ref type="table">1</ref>, but also added to the positive ones the number of laughs detected; and also, we included the number of recommendations present in the form of a "Follow Friday" hashtag (#FF), due to its ease of detection and its positive bias.</p><p>Classification is performed using scikitlearn, a Python module for machine learning. This package provides many algorithms such as Random Forest, Support Vector Machine (SVM) and so on. One of its main advantages is that it is supported by extensive documentation. Moreover, it is robust, fast and easy to use.</p><p>As stated before, we have two main training models: Averaged centroids and the averaged centroids including the inverted document frequency, for both the lemmatized and not-lemmatized texts. We performed experiments using three different classifiers: Random Forests, Support Vector Machines and Logistic Regression because these classifiers often achieved the best results for text classification and sentiment analysis.</p><p>Also we evaluated the impact of applying a set of emoticon's rules as a pre-classification stage, similar to <ref type="bibr" target="#b2">(Chikersal et al., 2015)</ref>, in which we determine a first stage polarity for each tweet as follows:</p><p>If posEmo is greater than zero and negEmo is equal to zero, the tweet is marked as "P".</p><p>If negEmo is greater than zero and posEmo is equal to zero, the tweet is marked as "N".</p><p>If both posEmo and negEmo are greater than zero, the tweet is marked as "NEU".</p><p>If both posEmo and negEmo are equal to zero, the tweet is marked as "NONE".</p><p>Then, after the classification takes place we made three tests: i) Applying no rule, ii) honoring the polarity defined by the rule, which means, we keep the predefined polarity if the tweet was marked as "P" or "N", otherwise we take the value estimated by the classifier, and iii) a mixed approach where we give each polarity a value (N+: -2; N: -1; NEU,NONE: 0; P: 1; P+: 2) and performed an arithmetic sum of both the predefined and estimated polarity if and only if they are not equal; with that for instance, if the classifier marked a tweet as "N" and the rules marked it as "P" the tweet will be classified as "NEU".</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Results</head><p>In order to choose the best-performing classifiers, we use 10-fold cross-validation because there is no development dataset and this strategy has become the standard method in practical terms. Our experiments showed that, although the results were similar<ref type="foot" target="#foot_2">3</ref> , the best settings for the 5-levels task are: RUN-1: Support Vector Machine, over the averaged centroids without applying any rules for pre-defining polarities.</p><p>RUN-2: Support Vector Machine, over the averaged centroids and applying the mixed rules approach. RUN-3: Logistic Regression, over the centroids with inverted document frequency and applying the mixed rules approach.</p><p>and for the 3-levels task are: RUN-1: Support Vector Machine, over the averaged centroids and applying the mixed rules approach. RUN-2: Logistic Regression, over the centroids with inverted document frequency and applying the mixed rules approach.</p><p>RUN-3: Logistic Regression, over the averaged centroids and applying the mixed rules approach.</p><p>Tables <ref type="table" target="#tab_1">2 and 3</ref> show the results for these settings provided by the TASS submission system. For each run, accuracy is provided as well as the macro-averaged precision, recall and F1-measure. As expected, the results for 3 levels are higher than for 5 levels because the training dataset is larger. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Run</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Conclusions and future work</head><p>This paper explores the use of word embeddings for the task of sentiment analysis. Instead of using, the bag-of-words model to represent tweets, these are represented as word vectors taken from a pre-trained model of word embeddings. An important advantage of word embedding model compared to the technique of bag-of-words representation is that it achieves a significant dimensional reduction of the feature set needed to represent tweets and leads, therefore, to a reduction of training and testing time of the algorithms.</p><p>In order to use word embedding models properly, a preprocessing stage had to be completed before training a classifier. Due to the unstructured nature of the tweets, this preprocessing proved to be a very important step in order to standardize at some degree the input data. The experimentation showed that the three tested classifiers obtained very similar results, with Random Forest having slight worse performance and Logistic Regression being slightly better and much more faster.</p><p>One of the main drawback of our approach is that many words do not have a word vector in the word embedding model used for our experiments. An analysis showed that many of these words come from hashtags, which are usually short phrases. Therefore, we should apply a more sophisticated method in order to extract the words forming hashtag.</p><p>As future work, we also plan to use a word embedding model trained on a collection of text from Spanish social media. We think that this will have a positive effect of the performance of our system to identify the polarity of tweets because this model will be generated from documents characterized by the main features that describe social media texts (for example, informal style language, plenty of grammatical errors and spelling mistakes, slang and vulgar vocabulary).</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head></head><label></label><figDesc>Positive :-), :), :D, :o), :], D:3, :c), :&gt;, =], 8), =), :}, :ˆ), :-D, 8-D, 8D, x-D, xD, X-D, XD, =-D, =D, =-3, =3, BˆD, :'), :'), :*, :-*, :ˆ*, ;-), ;), *-), *), ;-], ;], ;D, ;ˆ), &gt;:P, :-P, :P, X-P, x-p, xp, XP, :-p, :p, =p, :-b, :b Negative &gt;:[, :-(, :(, :-c, :-&lt;, :&lt;, :-[, :[, :{, ;(, :-||, &gt;:(, :'-(, :'(, D:&lt;, D=, v.v</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 2 :</head><label>2</label><figDesc>Results for Sentiment Analysis at global level (5 levels, Full test corpus)</figDesc><table><row><cell></cell><cell>P</cell><cell>R</cell><cell>F1</cell><cell>Acc</cell></row><row><cell cols="5">RUN-1 0.411 0.449 0.429 0.527</cell></row><row><cell cols="5">RUN-2 0.412 0.448 0.429 0.527</cell></row><row><cell cols="5">RUN-3 0.402 0.436 0.418 0.549</cell></row><row><cell>Run</cell><cell>P</cell><cell>R</cell><cell>F1</cell><cell>Acc</cell></row><row><cell cols="5">RUN-1 0.506 0.510 0.508 0.652</cell></row><row><cell cols="5">RUN-2 0.508 0.508 0.508 0.652</cell></row><row><cell cols="5">RUN-3 0.512 0.511 0.511 0.653</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 3 :</head><label>3</label><figDesc>Results for Sentiment Analysis at global level (3 levels, Full test corpus)With the settings mentioned above, the obtained results are extremely similar, but we can state that, in terms of Accuracy, Logistic Regression report the best results; and, even it's not measured in this work, is worth mentioning that Logistic Regression's performance was observably faster.</figDesc><table /></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">http://snowball.tartarus.org/algorithms/spanish/stop.txt</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">https://www.meaningcloud.com/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">Experiments showed that not-lemmatized text performed better in all settings, hence the best settings reported here is using not-lematized model</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This work was supported by eGovernAbility-Access project (TIN2014-52665-C2-2-R).</p></div>
			</div>


			<div type="funding">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>* This work was supported by eGovernAbility-Access project (TIN2014-52665-C2-2-R).</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">A knowledge-poor approach to chemical-disease relation extraction</title>
		<author>
			<persName><forename type="first">F</forename><surname>Alam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Corazza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Lavelli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Zanoli</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Database</title>
		<imprint>
			<biblScope unit="page">71</biblScope>
			<date type="published" when="2016">2016. 2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">Spanish Billion Words Corpus and Embeddings</title>
		<author>
			<persName><forename type="first">C</forename><surname>Cardellino</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2016-03">2016. March</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Modelling public sentiment in twitter: using linguistic patterns to enhance supervised learning</title>
		<author>
			<persName><forename type="first">P</forename><surname>Chikersal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Poria</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Cambria</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gelbukh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">E</forename><surname>Siong</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Intelligent Text Processing and Computational Linguistics</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="49" to="65" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Overview of tass 2016</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>García-Cumbreras</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Villena-Román</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Martínez-Cámara</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">C</forename><surname>Díaz-Galiano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">T</forename><surname>Martín-Valdivia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">A U</forename><surname>López</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of TASS 2016: Workshop on Sentiment Analysis at SEPLN colocated with the 32nd SEPLN Conference (SEPLN 2016)</title>
				<meeting>TASS 2016: Workshop on Sentiment Analysis at SEPLN colocated with the 32nd SEPLN Conference (SEPLN 2016)<address><addrLine>Salamanca, Spain</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016-09">2016. September</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Distributed representations of words and phrases and their compositionality</title>
		<author>
			<persName><forename type="first">T</forename><surname>Mikolov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Dean</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in neural information processing systems</title>
				<imprint>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Semantic orientation for polarity classification in spanish reviews</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">D</forename><surname>Molina-González</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Martínez-Cámara</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>-T. Martín-Valdivia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Perea-Ortega</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Expert Systems with Applications</title>
		<imprint>
			<biblScope unit="volume">40</biblScope>
			<biblScope unit="issue">18</biblScope>
			<biblScope unit="page" from="7250" to="7257" />
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Opinion mining and sentiment analysis</title>
		<author>
			<persName><forename type="first">B</forename><surname>Pang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Lee</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Foundations and trends in information retrieval</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="issue">1-2</biblScope>
			<biblScope unit="page" from="1" to="135" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Exploring word embedding for drug name recognition</title>
		<author>
			<persName><forename type="first">I</forename><surname>Segura-Bedmar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Suárez-Paniagua</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Martınez</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SIXTH INTERNATIONAL WORKS-HOP ON HEALTH TEXT MINING AND INFORMATION ANALYSIS (LOUHI)</title>
				<imprint>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page">64</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Parsing with compositional vector grammars</title>
		<author>
			<persName><forename type="first">R</forename><surname>Socher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Bauer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Manning</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">Y</forename><surname>Ng</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ACL (1)</title>
				<imprint>
			<date type="published" when="2013">2013a</date>
			<biblScope unit="page" from="455" to="465" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Recursive deep models for semantic compositionality over a sentiment treebank</title>
		<author>
			<persName><forename type="first">R</forename><surname>Socher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Perelygin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">Y</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Chuang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Manning</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">Y</forename><surname>Ng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Potts</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the conference on empirical methods in natural language processing (EMNLP)</title>
				<meeting>the conference on empirical methods in natural language processing (EMNLP)</meeting>
		<imprint>
			<publisher>Citeseer</publisher>
			<date type="published" when="2013">2013b</date>
			<biblScope unit="volume">1631</biblScope>
			<biblScope unit="page">1642</biblScope>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
