<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Automatic Misogyny Identification Using Neural Networks</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">I</forename><surname>Goenaga</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Ixa Taldea EHU</orgName>
								<orgName type="institution" key="instit1">IXA Group</orgName>
								<orgName type="institution" key="instit2">University of the Basque Country (EHU/UPV</orgName>
								<orgName type="institution" key="instit3">UPV Informatika Fakultatea</orgName>
								<address>
									<addrLine>M. Lardizabal 1</addrLine>
									<postCode>20008</postCode>
									<settlement>Donostia</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">A</forename><surname>Atutxa</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Ixa Taldea EHU</orgName>
								<orgName type="institution" key="instit1">IXA Group</orgName>
								<orgName type="institution" key="instit2">University of the Basque Country (EHU/UPV</orgName>
								<orgName type="institution" key="instit3">UPV Informatika Fakultatea</orgName>
								<address>
									<addrLine>M. Lardizabal 1</addrLine>
									<postCode>20008</postCode>
									<settlement>Donostia</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">K</forename><surname>Gojenola</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Ixa Taldea EHU</orgName>
								<orgName type="institution" key="instit1">IXA Group</orgName>
								<orgName type="institution" key="instit2">University of the Basque Country (EHU/UPV</orgName>
								<orgName type="institution" key="instit3">UPV Informatika Fakultatea</orgName>
								<address>
									<addrLine>M. Lardizabal 1</addrLine>
									<postCode>20008</postCode>
									<settlement>Donostia</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">A</forename><surname>Casillas</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Ixa Taldea EHU</orgName>
								<orgName type="institution" key="instit1">IXA Group</orgName>
								<orgName type="institution" key="instit2">University of the Basque Country (EHU/UPV</orgName>
								<orgName type="institution" key="instit3">UPV Informatika Fakultatea</orgName>
								<address>
									<addrLine>M. Lardizabal 1</addrLine>
									<postCode>20008</postCode>
									<settlement>Donostia</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">A</forename><surname>Díaz De Ilarraza</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Ixa Taldea EHU</orgName>
								<orgName type="institution" key="instit1">IXA Group</orgName>
								<orgName type="institution" key="instit2">University of the Basque Country (EHU/UPV</orgName>
								<orgName type="institution" key="instit3">UPV Informatika Fakultatea</orgName>
								<address>
									<addrLine>M. Lardizabal 1</addrLine>
									<postCode>20008</postCode>
									<settlement>Donostia</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">N</forename><surname>Ezeiza</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Ixa Taldea EHU</orgName>
								<orgName type="institution" key="instit1">IXA Group</orgName>
								<orgName type="institution" key="instit2">University of the Basque Country (EHU/UPV</orgName>
								<orgName type="institution" key="instit3">UPV Informatika Fakultatea</orgName>
								<address>
									<addrLine>M. Lardizabal 1</addrLine>
									<postCode>20008</postCode>
									<settlement>Donostia</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">M</forename><surname>Oronoz</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Ixa Taldea EHU</orgName>
								<orgName type="institution" key="instit1">IXA Group</orgName>
								<orgName type="institution" key="instit2">University of the Basque Country (EHU/UPV</orgName>
								<orgName type="institution" key="instit3">UPV Informatika Fakultatea</orgName>
								<address>
									<addrLine>M. Lardizabal 1</addrLine>
									<postCode>20008</postCode>
									<settlement>Donostia</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">A</forename><surname>Pérez</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Ixa Taldea EHU</orgName>
								<orgName type="institution" key="instit1">IXA Group</orgName>
								<orgName type="institution" key="instit2">University of the Basque Country (EHU/UPV</orgName>
								<orgName type="institution" key="instit3">UPV Informatika Fakultatea</orgName>
								<address>
									<addrLine>M. Lardizabal 1</addrLine>
									<postCode>20008</postCode>
									<settlement>Donostia</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">O</forename><surname>Perez De Viñaspre</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Ixa Taldea EHU</orgName>
								<orgName type="institution" key="instit1">IXA Group</orgName>
								<orgName type="institution" key="instit2">University of the Basque Country (EHU/UPV</orgName>
								<orgName type="institution" key="instit3">UPV Informatika Fakultatea</orgName>
								<address>
									<addrLine>M. Lardizabal 1</addrLine>
									<postCode>20008</postCode>
									<settlement>Donostia</settlement>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Automatic Misogyny Identification Using Neural Networks</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">D0319EBE83C56DC615959B462AFA0478</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T20:25+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Shared task</term>
					<term>Misogyny</term>
					<term>Neural Networks</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In this paper we present our approach to automatically identify misogyny in Twitter tweets. That task is one of the two sub-tasks organized by AMI-IberEval 2018 organization. In order to carry out the task, we present a neural network approach. Neural network models have been demonstrated to be capable of achieving remarkable performance in sentence and document modeling. Convolutional neural network (CNN) and recurrent neural network (RNN) are two mainstream architectures for such modeling tasks, which adopt totally different ways of understanding natural languages. In this work we focus on RNN approach using a Bidirectional Long Short Term Memory (Bi-LSTM) with Conditional Random Fields (CRF) and we evaluate the proposed architecture on misogyny identification task (text classification). The experimental results show that the system can achieve good performance on this task obtaining 78.9 accuracy on English tweets and 76.8 accuracy on Spanish tweets.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>In the last couple of years we have started to see deep learning making significant inroads into areas where computers have previously seen limited success. Rather than requiring a set of fixed rules that are defined by the programmer, deep learning uses neural networks that learn rich non-linear relationships directly from data. Deep learning has also seen some success in NLP, for example in text classification. Text classification is an essential component in many applications, such as web searching, information filtering, and sentiment analysis <ref type="bibr" target="#b2">[3]</ref>.</p><p>A key problem in text classification is feature representation, which is commonly based on the bag-of-words (BoW) model, where unigrams, bigrams, ngrams or some specific patterns are extracted as features. Moreover, several feature selection methods, such as pLSA <ref type="bibr" target="#b3">[4]</ref> or LDA <ref type="bibr" target="#b4">[5]</ref> are applied to select more discriminative features. Nevertheless, traditional feature representation methods often have problems when they try to capture the semantics of the words because they ignore contextual information. This is a problem in text classification because contextual information is the key in order to correctly classify a text. Although high-order n-grams and more complex features are designed to capture more contextual information and word orders, the data sparsity problem remains, which heavily affects the classification accuracy.</p><p>In a recurrent neural network approach, the models analyze a text word by word and store the semantics of all the previous text in a fixed-sized hidden layer <ref type="bibr" target="#b5">[6]</ref>. They receive as input a sequence of vectors and return another sequence that represents some information about the sequence at every step in the input. Although RNNs can learn long dependencies, they often fail to do so and tend to be biased towards their most recent inputs in the sequence <ref type="bibr" target="#b7">[8]</ref>. Likewise, Long Short-term Memory Networks (LSTMs) incorporate a memory-cell and have been shown effective capturing long-range dependencies.</p><p>Classic LSTMs create the representation of each word of the sentence using only the left context. Is interesting to use also the right context if we want to create a more complete representation of the words, though. This can be done with a second LSTM that reads the same sequence in reverse. This type of LSTMs are named bidirectional LSTMs (BI-LSTMs) <ref type="bibr" target="#b8">[9]</ref> and they create the representation of the words concatenating the left representation and the right representation. These representations effectively include a representation of a word in context, which is useful for numerous tasks.</p><p>On the other hand, Conditional Random Fields (CRF) are a probabilistic framework for labeling and segmenting structured data, such as sequences and trees. The underlying idea is that of defining a conditional probability distribution over label sequences given a particular observation sequence, rather than a joint distribution over both label and observation sequences. The primary advantage of CRFs is the relaxation of the independence assumption. Independence assumption states that the variables do not depend on each other and they do not affect each other in any way and this is not always the case and, consequently, it can lead to serious inaccuracies. Likewise, CRFs have been shown really effective in different tasks such as POS tagging <ref type="bibr" target="#b6">[7]</ref>, text processing <ref type="bibr" target="#b9">[10]</ref> or computer vision <ref type="bibr" target="#b10">[11]</ref>.</p><p>Taking that into account, in this work we have employed a BI-LSTM with Conditional Random Fields (CRF) <ref type="bibr" target="#b6">[7]</ref> in order to prove its effectiveness in misogynous tweet identification. In this area, one of the last works is <ref type="bibr" target="#b0">[1]</ref> where the authors address the problem of automatic detection and categorization of misogynous language in online social media, and they set the bases to organize AMI-IberEval 2018 shared task <ref type="bibr" target="#b1">[2]</ref>.</p><p>In the rest of this paper we will first present the experimental setup we have used to carry out our experiments in section 2, followed by the results obtained in the shared task test set in section 3 and the main conclusions of the work in section 4.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">System Description</head><p>We have divided this section into three subsections. In subsection 2.1 we explain the preprocessors we have used to tokenize and normalize the tweets, in subsection 2.2 the data resources we have employed in addition to the data shared by the organization, while in subsection 2.3 we focus on the used system.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Preprocessors</head><p>We have made use of one of python's packages designed for preprocessing tweets <ref type="bibr" target="#b11">[12]</ref>. The tool performs tokenization, word normalization, word segmentation (for splitting hash-tags) and spell correction, using word statistics from 2 big corpora (English Wikipedia, twitter -330mil English tweets). In addition, for Spanish we have used a set of simple rules proposed in <ref type="bibr" target="#b12">[13]</ref> for spell correction.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">External Data</head><p>Our AMI System needs word-embeddings in order to create a better word representation for each word we find in the corpus. Thus, we have used wordembeddings extracted from the Spanish Billion Word Corpus <ref type="bibr" target="#b13">[14]</ref> and from Wikipedia 2014 and Gigaword 5 <ref type="bibr" target="#b14">[15]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3">AMI System</head><p>In order to classify the tweets we have employed a neural network based architecture, more precisely a specific Bi-LSTM (an RNN subclass) with a CRF on top of it as proposed in <ref type="bibr" target="#b6">[7]</ref>. This kind of neural network is widely used to pursue sequence to sequence tagging. One of the advantages of using Bi-LSTM in contrast to other machine learning techniques such as SVM, Perceptron or CRFs is that the size of the context is automatically learned by the LSTM and there is no need to perform any complicated text preprocessing to obtain features to feed the tool. As we mentioned previously, our system is a tagger and marks the beginning and the next words of the sequences (IOB) we want to label. In this case we want to predict whether a tweet contains misogynous content or not. Thus, we introduce the tweets and the word-embeddings at the beginning of the process as in <ref type="bibr" target="#b6">[7]</ref>. When a word is missing in the word-embeddings, the system replaces the word with unknown (UNK ) label.</p><p>In all cases the system returns every word of each tweet tagged with Yes label when the tweet contains misogyny and with No label otherwise. If the opposite happened, we would consider a tweet as misogynous if at least has one Yes label. The examples below are the output of the system for two tweets written in English and represent the aforementioned: </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Results and Discussion</head><p>In AMI-IberEval 2018 shared task the participants can try their misogynous content identification systems in two languages: English and Spanish. We have participated in both languages and we have included the results for English track in table <ref type="table" target="#tab_0">1</ref> and the results for Spanish track in table 2.</p><p>English R Team Accu R Team Accu R Team Accu 1 14-exlab-r1 91. If we analyze the results for English track, we observe that our position within all participants of the shared task is tenth with 78.9 of accuracy. Although we are far from winning the shared task (-12.4), we are in the first third of the classification and the two previous systems in the classification are not far (+ 0.4) from us which demonstrates a good performance of our system identifying misogynous tweets in English. Spanish R Team Accu R Team Accu R Team Accu 1 14-exlab-r3 81. On the other hand, our system's accuracy identifying Spanish written tweets is 76.8. This time our position within the all participants is seventeenth just above the shared task's baseline. However, almost all the participants have obtained accuracies between 81.4 and 76.6 which indicates that the vast majority of the systems are close to each other. Likewise, identifying misogynous content in Spanish written tweets is more difficult mostly because the lack of top quality resources (corpus, word-embeddings, preprocessors ...) we can find relatively easy for English.</p><p>Once we analyzed our system's results for both languages, taking into account our system was designed for sequential tagging or sequence labeling we consider the experimental setup has performed well in a task it was not thought for. We realize the best option to do text classification would have been a convolutional neural network (CNN) specially because the best systems of the state of the art employ this type of neural networks. Nevertheless, our main purpose has been to test a BI-LSTM with CRF on text classification task and bearing in mind its constraints the system has achieved reasonable results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Conclusions</head><p>This paper presents our approach to automatically identify misogynous content in Twitter tweets. In order to carry out the task, we have chosen a neural network approach due to their ability to achieving remarkable performance in sentence and document modeling. In this work we focus on RNN approach using a Bidirectional Long Short Term Memory (Bi-LSTM) with Conditional Random Fields (CRF) and experimental results show that the system can achieve good performance identifying misogynous tweets obtaining 78.9 accuracy on English tweets and 76.8 accuracy on tweets.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>[</head><label></label><figDesc>B-Yes]&lt; user &gt; [I-Yes]bitch [I-Yes]is [I-Yes]a [I-Yes]psyco [I-Yes], [I-Yes]* [I-Yes]dry [I-Yes]pussy [I-Yes]detected [I-Yes]* [B-No]you [I-No]give [I-No]me [I-No]life [I-No]! [I-No]&lt; repeated &gt; [I-No]&lt; url &gt;</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>Results obtained by participants for English track using only provided training data (constrained).</figDesc><table><row><cell></cell><cell>3 11 resham-r1</cell><cell cols="2">78.5 21 JoseSebastian-r1 74.9</cell></row><row><cell>2 14-exlab-r2</cell><cell cols="3">90.2 12 AMI-Baseline 78.3 22 Amrita CEN-r3 73.8</cell></row><row><cell>3 14-exlab-r4</cell><cell>89.8 13 vic -r2</cell><cell>78.0 23 vic -r1</cell><cell>70.9</cell></row><row><cell>4 14-exlab-r3</cell><cell>87.8 14 vic -r3</cell><cell>78.0 24 ITT-r1</cell><cell>70.6</cell></row><row><cell>5 SB-r4</cell><cell>87.0 15 vic -r4</cell><cell>78.0 25 vic -r5</cell><cell>64.6</cell></row><row><cell>6 SB-r5</cell><cell cols="3">85.1 16 maybelraul-r3 77.9 26 Amrita CEN-r2 56.3</cell></row><row><cell>7 14-exlab-r5</cell><cell cols="3">82.3 17 maybelraul-r1 77.1 27 GrCML2016-r3 52.7</cell></row><row><cell cols="4">8 AnotherOne-r1 79.3 18 maybelraul-r4 76.9 28 GrCML2016-r2 52.4</cell></row><row><cell cols="4">9 maybelraul-r2 79.3 19 maybelraul-r5 76.0 29 Amrita CEN-r1 51.9</cell></row><row><cell>10 ixaTeam-r1</cell><cell>78.9 20 ITT-r2</cell><cell>75.8</cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 .</head><label>2</label><figDesc>Results obtained by participants for Spanish track using only provided training data (constrained).Proceedings of the Third Workshop on Evaluation of Human Language Technologies for Iberian Languages(IberEval 2018)   </figDesc><table><row><cell></cell><cell>4 10 SB-r3</cell><cell>80.5 19 maybelraul-r1</cell><cell>76.7</cell></row><row><cell cols="2">2 JoseSebastian-r1 81.4 11 SB-r1</cell><cell>80.3 20 vic -r2</cell><cell>76.6</cell></row><row><cell>3 SB-r4</cell><cell>81.3 12 AnotherOne-r1</cell><cell cols="2">80.2 21 Amrita CEN-r3 74.4</cell></row><row><cell>4 14-exlab-r1</cell><cell>81.2 13 maybelraul-r5</cell><cell>79.6 22 vic -r3</cell><cell>65.9</cell></row><row><cell>5 14-exlab-r2</cell><cell>81.2 14 maybelraul-r2</cell><cell cols="2">78.8 23 Amrita CEN-r1 54.2</cell></row><row><cell>6 14-exlab-r4</cell><cell>80.9 15 maybelraul-r3</cell><cell>78.7 24 14-exlab-r5</cell><cell>53.6</cell></row><row><cell>7 SB-r2</cell><cell>80.8 16 maybelraul-r4</cell><cell cols="2">78.2 25 Amrita CEN-r2 52.9</cell></row><row><cell>8 SB-r5</cell><cell>80.6 17 ixaTeam-r1</cell><cell>76.8</cell><cell></cell></row><row><cell>9 vic -r1</cell><cell cols="2">80.5 18 AMI-BASELINE 76.7</cell><cell></cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" xml:id="foot_0">Proceedings of the Third Workshop on Evaluation of Human Language Technologies for Iberian Languages(IberEval 2018)   </note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This work has been partially funded by: -The Spanish ministry (projects TADEEP: TIN2015-70214-P, PROSA-MED: TIN2016-77820-C3-1-R). -The Basque Government (projects DETEAMI: 2014111003, ELKAROLA:KK-2015/00098).</p><p>We gratefully acknowledge the support of NVIDIA Corporation with the donation of the Titan X Pascal GPU used for this research.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Automatic Identification and Classification of Misogynistic Language on Twitter</title>
		<author>
			<persName><forename type="first">M</forename><surname>Anzovino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Fersini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Third Workshop on Evaluation of Human Language Technologies for Iberian Languages</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">M</forename><surname>Silberztein</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Atigui</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">E</forename><surname>Kornyshova</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">E</forename><surname>Métais</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Meziane</surname></persName>
		</editor>
		<meeting>the Third Workshop on Evaluation of Human Language Technologies for Iberian Languages<address><addrLine>IberEval</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018">2018. 2018</date>
			<biblScope unit="volume">10859</biblScope>
		</imprint>
	</monogr>
	<note>Natural Language Processing and Information Systems. NLDB 2018</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Overview of the Task on Automatic Misogyny Identification at IberEval</title>
		<author>
			<persName><forename type="first">E</forename><surname>Fersini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Anzovino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Third Workshop on Evaluation of Human Language Technologies for Iberian Languages (IberEval 2018), co-located with 34th Conference of the Spanish Society for Natural Language Processing (SE-PLN 2018</title>
				<meeting>the Third Workshop on Evaluation of Human Language Technologies for Iberian Languages (IberEval 2018), co-located with 34th Conference of the Spanish Society for Natural Language Processing (SE-PLN 2018<address><addrLine>Seville, Spain</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018-09-18">September 18, 2018</date>
		</imprint>
	</monogr>
	<note>CEUR Workshop Proceedings. CEUR-WS.org</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">SemEval-2016 Task 7: Determining sentiment intensity of English and Arabic phrases</title>
		<author>
			<persName><forename type="first">S</forename><surname>Kiritchenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Mohammad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Salameh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 10th International Workshop on Semantic Evaluation</title>
				<meeting>the 10th International Workshop on Semantic Evaluation<address><addrLine>San Diego, California, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="volume">16</biblScope>
			<biblScope unit="page" from="42" to="51" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Text categorization by boosting automatically extracted concepts</title>
		<author>
			<persName><forename type="first">L</forename><surname>Cai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Hofmann</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SIGIR</title>
				<imprint>
			<date type="published" when="2003">2003</date>
			<biblScope unit="page" from="182" to="189" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Document classification by topic labeling</title>
		<author>
			<persName><forename type="first">S</forename><surname>Hingmire</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Chougule</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">K</forename><surname>Palshikar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Chakraborti</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SIGIR</title>
				<imprint>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="877" to="880" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Finding structure in time</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">L</forename><surname>Elman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Cognitive science</title>
		<imprint>
			<biblScope unit="volume">14</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="179" to="211" />
			<date type="published" when="1990">1990</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">Neural architectures for named entity recognition</title>
		<author>
			<persName><forename type="first">G</forename><surname>Lample</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ballesteros</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Subramanian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Kawakami</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Dyer</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1603.01360</idno>
		<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Learning long-term dependencies with gradient descent is difficult</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Bengio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Simard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Frasconi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="157" to="166" />
			<date type="published" when="1994">1994</date>
		</imprint>
	</monogr>
	<note>Neural Networks</note>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Framewise phoneme classification with bidirectional LSTM networks</title>
		<author>
			<persName><forename type="first">A</forename><surname>Graves</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Schmidhuber</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. IJCNN</title>
				<meeting>IJCNN</meeting>
		<imprint>
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Abner: an open source tool for automatically tagging genes, proteins, and other entity names in text</title>
		<author>
			<persName><forename type="first">B</forename><surname>Settles</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Bioinformatics</title>
		<imprint>
			<biblScope unit="volume">21</biblScope>
			<biblScope unit="issue">14</biblScope>
			<biblScope unit="page" from="3191" to="3192" />
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Multiscale conditional random fields for image labelling</title>
		<author>
			<persName><forename type="first">X</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">S</forename><surname>Zemel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Carreira-Perpinian</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE Computer Society Conference on Computer Vision and Pattern Recognition</title>
				<imprint>
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">DataStories at SemEval-2017 Task 4: Deep LSTM with Attention for Message-level and Topic-based Sentiment Analysis</title>
		<author>
			<persName><forename type="first">C</forename><surname>Baziotis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Pelekis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Doulkeridis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 11th International Workshop on Semantic Evaluation</title>
				<meeting>the 11th International Workshop on Semantic Evaluation<address><addrLine>SemEval-</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2017">2017. 2017</date>
			<biblScope unit="page" from="747" to="754" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<author>
			<persName><forename type="first">I</forename><surname>Alegria</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Etxeberria</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Labaka</surname></persName>
		</author>
		<title level="m">Una Cascada de Transductores Simples para Normalizar Tweets</title>
				<imprint>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
	<note>Tweet-Norm@ SEPLN</note>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<author>
			<persName><forename type="first">C</forename><surname>Cardellino</surname></persName>
		</author>
		<ptr target="http://crscardellino.me/SBWCE/" />
		<title level="m">Spanish Billion Words Corpus and Embeddings</title>
				<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<title level="m" type="main">GloVe: Global Vectors for Word Representation</title>
		<author>
			<persName><forename type="first">J</forename><surname>Pennington</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Socher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Manning</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
