<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Identifying Complaints from Product Reviews: A Case Study on Hindi</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Raghvendra</forename><forename type="middle">Pratap</forename><surname>Singh</surname></persName>
							<email>raghvendra.singh6@mail.dcu.ie</email>
							<affiliation key="aff0">
								<orgName type="department">School of Computing</orgName>
								<orgName type="institution">Dublin City University</orgName>
								<address>
									<settlement>Dublin</settlement>
									<country key="IE">Ireland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Rejwanul</forename><surname>Haque</surname></persName>
							<email>rejwanul.haque@adaptcentre.ie</email>
						</author>
						<author>
							<persName><forename type="first">Mohammed</forename><surname>Hasanuzzaman</surname></persName>
							<email>mohammed.hasanuzzaman@adaptcentre.ie</email>
							<affiliation key="aff1">
								<orgName type="institution">Cork Institute of Technology</orgName>
								<address>
									<settlement>Cork</settlement>
									<country key="IE">Ireland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Andy</forename><surname>Way</surname></persName>
							<email>andy.way@adaptcentre.ie</email>
						</author>
						<title level="a" type="main">Identifying Complaints from Product Reviews: A Case Study on Hindi</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">FA6DAEFF4A6F4ED00AA0705EB0218276</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-25T02:38+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Random walk</term>
					<term>LSTM</term>
					<term>fastText</term>
					<term>Dice coefficient</term>
					<term>SMOTE</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>When an expectation does not meet reality in a real-world situation, the difference is usually expressed and communicated via an act which is complaint. Customers often post reviews on the products or services they purchase on the retailer websites and different social media platforms, and the reviews may reflect complaints about the products or services. Automatic recognition of customers' complaints on products or services that they purchase can be crucial for the organisations, multinationals and online retailers since they can exploit this information to fulfil the customers' expectations including managing and resolving the complaints. In this work, we present the supervised and semi-supervised learning strategies to identify users' complaints from the language they use to post their reviews. In other words, we automatically identify complaints from the opinionated texts (reviews) about products posted in Hindi. For this, first we automatically crawled the Hindi reviews on different products from the the websites of the retail giant Amazon and the popular social media platform YouTube, and prepared a gold-standard data set via a systematic manual annotation process. We use state-ofthe-art classification algorithms for the complaints identification task and our classification models achieve reasonable classification accuracy (F1 = 68.38%) on a gold-standard evaluation test set.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Text classification is an active field of natural language processing (NLP) and data mining. Almost all online retailers allow users to freely express their opinions and thoughts on products via their websites and the relevant social media platforms. The customers who intend to purchase a product may take purchasing decisions based on the reviews of the product. Accordingly, the commercial and retail companies considers product reviews as an important source of information, and could exploit this information to build their marketing tool and strategy, and to resolve any issues in relation to the product. This could also benefit the users with the suggestions on the quality of the products or services that they want to purchase. As for the number of reviews of a product posted by the users, they could range from several hundreds to tens of thousands. The e-commerce companies and online retailers want to identify complaints given the reviews of a product for their own benefit. Likewise, the customers who want to buy a product or service may need such information while avoiding having to go through thousands of reviews about the product.</p><p>In this context, Gupta et al. <ref type="bibr" target="#b4">[5]</ref> identified the relationship between users' purchase intent from their social media forums such as Quora <ref type="foot" target="#foot_0">1</ref> and Yahoo! Answers. <ref type="foot" target="#foot_1">2</ref> They primarily carried out text analysis to detect purchase intent from user-generated content (UGC). Wang et al. <ref type="bibr" target="#b12">[13]</ref> investigated the problem of identifying purchase intent. In particular, the authors proposed a graph-based learning approach to identify intent tweets and classify them into six categories. For this, they retrieved tweets with a bootstrap method, with using a list of seed intent-indicators (e.g. 'want to'), and manually created training examples from the collected tweets. Haque et al. <ref type="bibr" target="#b5">[6]</ref> extends the work of Wang et al. <ref type="bibr" target="#b12">[13]</ref> while increasing the coverage of the purchase intent indicators with the distributed vector representation of words using the continuous skip-gram model <ref type="bibr" target="#b9">[10]</ref>.</p><p>Recently, Preotiuc-Pietro et al. <ref type="bibr" target="#b10">[11]</ref> automatically identify complaints from the tweets posted by the social media users and the potential customers. To the best of our knowledge, the most relevant works to ours come from Preotiuc-Pietro et al. <ref type="bibr" target="#b10">[11]</ref>. In fact, to a certain extent, our proposed methods can be viewed as the extension of Preotiuc-Pietro et al. <ref type="bibr" target="#b10">[11]</ref> with mainly the following additions as far as this task is concerned: (i) we considered product reviews instead of tweets as in Preotiuc-Pietro et al. <ref type="bibr" target="#b10">[11]</ref> and (ii) we explore a resourcepoor and less-explored language, Hindi, for our investigation, (iii) we applied a state-of-the-art sampling strategy <ref type="bibr" target="#b1">[2]</ref> in order to encounter class imbalance problem in the training data, and (iv) we explore applying a semi-supervised classification algorithm in the complaint identification task. Moreover, one of the key contributions to this work is creation of the gold-standard dataset in Hindi.</p><p>The remainder of the paper is organised as follows. In Section 2, we detail how we semi-automatically created training data for our experiments. In Section 3, we present our experimental methodology and setups. Section 4 presents our evaluation results, with some discussions. Section 5 concludes and provides avenues for further work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Dataset Creation</head><p>This section details the creation of training data that has been used in this task.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Collecting Hindi Reviews</head><p>To the best of our knowledge, there is no existing (freely available) annotated review data (complaint and non-complaint) for Hindi. For this task we needed an annotated review dataset for Hindi. In order to create an annotated review dataset for Hindi, we first collected Hindi reviews posted online. The reviews were taken from two different sources: (i) the websites of the retail giant Amazon, and (ii) YouTube. The users usually post their Hindi reviews on these two platforms.</p><p>In order to collect the reviews from Amazon, we used amazon-reviews-scraper Python library <ref type="foot" target="#foot_2">3</ref> which takes a product name as input and provides the reviews about the product across the different languages. Similarly, in order to collect the reviews from YouTube, we used youtube-comment-downloader Python library. <ref type="foot" target="#foot_3">4</ref>This script provided us reviews on the products across the different languages.</p><p>In order to remove noise (e.g. HTML tags, special characters) from reviews, we applied a number cleaning scripts including a language identifier. <ref type="foot" target="#foot_4">5</ref> We also removed emojis from the review texts. Each of the collected clean reviews is manually tagged with a particular category, namely complaint or non-complaint. The annotation scheme and results are presented in the next section.  </p><formula xml:id="formula_0">माउस काम नहीं कर रहा स व स बे कार है 1 1 1 1 क्या खू ब लखा है मजा आ गया सत्य भै या ऐसे ही लखते रहे 0 0 0 0 काफ रोमां चक है और कु छ हट के भी 0 0 0 0 कहानी का शानदार आग़ाज़ बे हतरीन अं दाज़ आप आगे पढ़ने के लए मजबू र होते हैं 0 0 0 0 चा ज ग करते समय बहुत गरम हो जाता है , फटने का दर है 1 1 1 1</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Annotation</head><p>The annotation task is performed in-house by our three undergraduate students. Each reviews is presented to the annotators. The annotators are expected to answer two questions for a given review. While the first question is related to the decision as to whether a review is complaints or non-complaints, the motivation behind the second question is to collect a more fine-grained (book, phone, tv etc.) gold standard dataset. <ref type="foot" target="#foot_5">6</ref>To have a concrete idea about the agreement between annotators, we calculated the majority class for each review in our dataset. A review belongs to a majority class k if the most frequent annotation for the review was selected by at least k annotators. As a consequence, a large percentage of review belonging to high majority classes are symptomatic of good inter-annotator agreement. Similar to earlier studies, we consider all annotations with a majority class greater than 2 as reliable. In this case, for the complaints or non-complaints annotation scheme, over 88% of the review were annotated identically by the majority of annotators, while for fine-grained annotation scheme,<ref type="foot" target="#foot_6">7</ref> over 85% of the annotations fell into this case. As such, we can be confident that the annotation process was successful and the dataset is reliable. A sample of our annotated dataset is presented in Table <ref type="table" target="#tab_0">1</ref>, and statistics about the dataset are presented in the next section.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3">Data Statistics</head><p>We report the statistics of the our gold standard data set in Table <ref type="table" target="#tab_1">2</ref>. The table shows the numbers of the complaint and non-complaint reviews which are listed according to different product types.  3 Methodology</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">The LSTM Network</head><p>Nowadays, recurrent neural network (RNN), in particular with long-short term memory (LSTM) <ref type="bibr" target="#b6">[7]</ref> hidden units, has been proved to be an effective model for many classification tasks in NLP, e.g. sentiment analysis <ref type="bibr" target="#b13">[14]</ref>, text classification <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b14">15]</ref>. RNN is an extension of the feed-forward neural network (NN), which has the gradient vanishing or exploding problems. LSTM deals with the exploding and vanishing gradient problems of RNN. An RNN composed of LSTM hidden units is often called an LSTM network. A common LSTM unit is composed of a cell, an input gate, an output gate and a forget gate. More formally, each cell in LSTM can be computed as follows:</p><formula xml:id="formula_1">X = [ h t−1 x t ]<label>(1)</label></formula><formula xml:id="formula_2">f t = σ(W f • X + b f )<label>(2)</label></formula><formula xml:id="formula_3">i t = σ(W i • X + b i )<label>(3)</label></formula><formula xml:id="formula_4">o t = σ(W o • X + b o )<label>(4)</label></formula><formula xml:id="formula_5">c t = f t ⊙ c t−1 + i t ⊙ tanh (W c • X + b c )<label>(5)</label></formula><formula xml:id="formula_6">h t = o t ⊙ tanh (c t )<label>(6)</label></formula><p>where</p><formula xml:id="formula_7">W i , W f , W o ∈ R d×2d are the weighted matrices and b i , b f , b o ∈ R d are</formula><p>biases of LSTM, which need to be learned during training, parameterising the transformations of the input, forget and output gates, respectively. σ is the sigmoid function, and ⊙ stands for element-wise multiplication. x t includes the inputs of LSTM cell unit. The vector of hidden layer is h t . The final hidden vector h N represents the whole input review, which is passed to softmax layer after linearising it into a vector whose length is equal to the number of class labels. In our work, the set of class labels includes complaint and non-complaint categories.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">fastText</head><p>We used fastText <ref type="bibr" target="#b7">[8]</ref>, a fast and efficient classifier, in our task. fastText is often on par with the state-of-the-art deep learning classifiers in terms of accuracy. It is also found to be faster for training and evaluation. fastText has many advance features, e.g. hierarchical softmax, hidden states shared among features and classes, and bag of n-grams features.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Graph-based Classifier</head><p>We also used a semi-supervised classification algorithm, i.e. random walk on graph. A graph, G is a pair of vertices V and edges E. We use a weighted graph for the classification. A weighted graph is a graph where each edge, E i , is linked to a numeric value W ij where i, j ∈ V . For our approach, it is necessary that</p><formula xml:id="formula_8">W ij = W ji<label>(7)</label></formula><p>Random walk is based on the concept of randomly determined processes. It considers a transition probability of the random walker that it will reach from a vertex V i to the other vertex V j with each step being either +1 or -1 with equal probability. As for implementing random walk for this task, we follow the approach and python implementation of the algorithm described in <ref type="bibr" target="#b11">[12]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4">Classical Supervised Classification Models</head><p>Furthermore, we compare the deep learning model (the LSTM network), fastText and semi-supervised classification model (random walk) presented above with the classical supervised classification models. We employ the following classical supervised classification techniques in our task:</p><p>-Logistic Regression (LR) -Decision Tree (DT) -Random Forest (RF) -Naïve Bayes (NB) -Support Vector Machine (SVM) These classical learning models (LR, DT, RF, NB and SVM) can be viewed as the baselines in this task. Thus, we obtain a comparative overview on the performances of different supervised and semi-supervised classification models including the LSTM network.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.5">Training Setup</head><p>In order to build LR, DT, RF and NB classification models, we use the wellknown scikit-learn machine learning library, <ref type="foot" target="#foot_7">8</ref> and performed all the experiments with default parameters set by scikit-learn. As for the representation space, each review was represented as a vector of word unigrams weighted by their frequency in the reviews.</p><p>For the classifiers based on the neural networks, we use a 300-Dimensional word embeddings from fastText. We use sigmoid activation function with Adam optimizer <ref type="bibr" target="#b8">[9]</ref> and binary cross entropy loss function. The size of input layer of the NN is 300. We employ layer normalisation <ref type="bibr" target="#b0">[1]</ref> in the model. Dropout <ref type="bibr" target="#b3">[4]</ref> between layers is set to 0.10. The size of embedding and hidden layers are 300. The learning-rate is set to 0.0003, and the training examples were reshuffled for each epoch.</p><p>As for the random walk algorithm, we use the similarity between the sentences as the transition probability. Again, we create a network with the training examples. We take a set of labelled vertices L and unlabelled vertices U , where (L, U ) ⊆ G (G : set of all the labels). The random walk algorithm considers the vertices of the network as the states of a Markov chain. We calculate the similarity measure between the data points with the Dice coefficient (DC) <ref type="bibr" target="#b2">[3]</ref> and reach to the final decision according to the probability. The DC measure has been widely used for evaluating the degree of association between words.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.6">Handling Class Imbalance</head><p>We recall Tables <ref type="table" target="#tab_2">2 and 3</ref> where we can see the presence of class imbalance in the training data. In order to encounter the class imbalance problem in our training data, we followed Chawla et al. <ref type="bibr" target="#b1">[2]</ref> who presented a combination of oversampling of minority class with under-sampling of majority class. This is also called SMOTE which is a kind of synthetic sampling technique and found to be effective in our problem.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Results and Discussion</head><p>We evaluate the performance our classifiers on the gold-standard test set (cf. Table <ref type="table" target="#tab_2">3</ref>) and report the evaluation results in this section. In order to measure classifier's accuracy on the test set, we use three widely-used evaluation metrics: precision, recall and F 1 measures. The results obtained are reported in Table <ref type="table" target="#tab_3">4</ref>. The first five rows of Table <ref type="table" target="#tab_3">4</ref> represent our baseline classifiers (i.e. the classical supervised classification models). The next row represents the random walk algorithm. We see from the table that these classifiers performs below par and SVM is the best-performing method among them (SVM: a 47.94 F 1 score)) according to the F 1 scores.</p><p>As for NN-based classifiers, the LSTM network trained on fastText embeddings performed reasonable as we see from Table <ref type="table" target="#tab_3">4</ref> that it produces a moderate F 1 score (68.38 F 1 ) on the test set. The fastText classifier also performs reasonably well; however, it could not surpass the LSTM network.</p><p>In Table <ref type="table" target="#tab_4">5</ref>, we show the F 1 scores of the best-performing classifiers of the complaint identification task presented in <ref type="bibr" target="#b10">[11]</ref>. As can be seen from Table <ref type="table" target="#tab_4">5</ref>,  <ref type="table" target="#tab_3">4</ref>). Note that, unlike us, Preotiuc-Pietro et al. <ref type="bibr" target="#b10">[11]</ref> carried out experiments on English and the data sets including the evaluation test set are different too. Moreover, the nature of data is also different to us. Preotiuc-Pietro et al. <ref type="bibr" target="#b10">[11]</ref> focused on identifying complaints on tweets and we focused on identifying the same on the product reviews posted on Amazon and YouTube. Naturally, the scores presented in Preotiuc-Pietro et al. <ref type="bibr" target="#b10">[11]</ref> cannot be directly compared to those presented in this paper. Given the fact that English is a high resource language and there are a plenty of linguistic tools and resources freely available in English, Preotiuc-Pietro et al. <ref type="bibr" target="#b10">[11]</ref> exploited many linguistic resources, models and tools in their task for feature engineering, e.g. sentiment and emotion analysis, temporal model, part-of-speech information. In Hindi, it is difficult to obtain such resources and standard tools, and many such tools are not available to use. Moreover, unlike English, Hindi is a morphological complex and highly inflected language. Therefore, we believe that identifying complaints on Hindi texts is to be more challenging in comparison to that in English or other high resource languages. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Conclusion</head><p>In this paper, we presented supervised and semi-supervised learning models to identify customers' complaints from the review data in Hindi, a low-resource, less-explored, and morphological rich highly inflected Indic language. There is no publicly available gold-standard training data as far as identifying complaints from user generated Hindi texts about products is concerned. Accordingly, we semiautomatically created a gold-standard dataset for complaints identification on Hindi. We conducted our experiments with state-of-the-art LSTM classifier and classical supervised classification models. With our LSTM classifier, we achieved a competent accuracy (a F 1 score of 68.38) on the gold-standard evaluation test set. There are limited linguistic resources and tools freely available for research in Hindi in comparison to many high-resource languages including English. We believe that this work would add an additional value to the social media analytics research in low-resource scenarios.</p><p>In future, we intend to test our method on different low-resource and non-English languages. We also plan to investigate applying more sophisticated and linguistic features in our model, e.g. part-of-speech information. We removed emojis from reviews at the time data prereprocessing. Since emojis may capture polarity and encode information regarding users' experiences and complaints, in future, we aim to carry out experiments while keeping emojis in the reviews. Our classifiers were trained on the data which is a mixture of reviews from different product types. We intend to train classifiers on reviews from specific product type, i.e. book or phone reviews. By this, we can compare the classifiers trained on the data from individual product type to the one trained on the data consisting reviews from all product types.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>1 फ़ोन1 1 1 सब</head><label>11</label><figDesc>ड बहुत बे हतरीन होता इस लए मु झे ये बहुत पसं द है .... का कै मरा बहुत अ है ड ले भी और बै टरी तो तबाही हैं 0 0 0 0 पृ सं ख्या के बाद के पृ उल्टे लगाए गए हैं 1 इधर उधर का चोरी कया हुआ लखा है ...मकसद फल्मों में स्क्रप्ट लखने का काम करना... 1 1 1 1</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>Sample Dataset. A1-3: three annotators.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 .</head><label>2</label><figDesc>The collected Hindi reviews listed according to product types.We divided the annotated set of reviews (i.e. 3,711 reviews) into train, development and test sets. The test and development set reviews were randomly sampled from the all reviews. The statistics about the train, development and test set reviews are shown in Table3. The training, test and development data sets have been released publicly and can be downloaded from https://github. com/MrRaghav/Complaints-mining-from-Hindi-product-reviews.</figDesc><table><row><cell cols="3">Category Non-complaints Complaints</cell></row><row><cell>Book</cell><cell>2,722</cell><cell>117</cell></row><row><cell>Phone</cell><cell>335</cell><cell>339</cell></row><row><cell>Headphone</cell><cell>30</cell><cell>18</cell></row><row><cell>Watch</cell><cell>26</cell><cell>17</cell></row><row><cell>Misc.</cell><cell>32</cell><cell>75</cell></row><row><cell>Total</cell><cell>3,145</cell><cell>566</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3 .</head><label>3</label><figDesc>Statistics of the train, development and test sets reviews.</figDesc><table><row><cell></cell><cell cols="2">Reviews Words Complaints</cell></row><row><cell>Train set</cell><cell>2,967 105,322</cell><cell>452</cell></row><row><cell>Dev. set</cell><cell>372 13,966</cell><cell>57</cell></row><row><cell>Test set</cell><cell>372 14,038</cell><cell>57</cell></row><row><cell>Total</cell><cell>3,711 133,326</cell><cell>566</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4 .</head><label>4</label><figDesc>Performance of the classifiers on the evaluation test set. identification models produce F 1 scores in the range of 78-79 which are more than 10 F 1 points higher than the F 1 ) score of our bestperforming complaint identification model (cf. second last row of Table</figDesc><table><row><cell></cell><cell cols="2">Precision Recall</cell><cell>F 1</cell></row><row><cell>NB</cell><cell>24.82</cell><cell cols="2">63.15 35.64</cell></row><row><cell>LR</cell><cell>40.44</cell><cell cols="2">63.15 49.31</cell></row><row><cell>DT</cell><cell>36.25</cell><cell cols="2">50.87 42.33</cell></row><row><cell>SVM</cell><cell>39.32</cell><cell cols="2">61.40 47.94</cell></row><row><cell>RF</cell><cell>52.63</cell><cell cols="2">35.08 42.10</cell></row><row><cell>Random walk</cell><cell>43.84</cell><cell cols="2">23.06 30.22</cell></row><row><cell>LSTM</cell><cell>66.67</cell><cell cols="2">70.18 68.38</cell></row><row><cell>fastText</cell><cell>72.34</cell><cell cols="2">59.64 65.38</cell></row><row><cell>their best complaint</cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 5 .</head><label>5</label><figDesc>The best-performing complaint identification classifiers (English) presented in Preotiuc-Pietro et al.<ref type="bibr" target="#b10">[11]</ref>.</figDesc><table><row><cell>F 1</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">www.quora.com</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">www.answers.yahoo.com</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">https://github.com/philipperemy/amazon-reviews-scraper. Accessed on August 2020</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3"><ref type="bibr" target="#b3">4</ref> https://github.com/egbertbouman/youtube-comment-downloader. Accessed on August 2020</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_4">.<ref type="bibr" target="#b4">5</ref> https://pypi.org/project/pycld2/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_5">The details of the annotation guidelines are out of the scope of this paper.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_6">Annotation of reviews tagged with 'complaints' into product types such as phone, book, tv etc.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_7">https://scikit-learn.org/stable/</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">Layer normalization</title>
		<author>
			<persName><forename type="first">Jimmy</forename><surname>Lei Ba</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jamie</forename></persName>
		</author>
		<author>
			<persName><forename type="first">Ryan</forename><surname>Kiros</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Geoffrey</forename><forename type="middle">E</forename><surname>Hinton</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1607.0645</idno>
		<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">SMOTE: Synthetic Minority Over-sampling Technique</title>
		<author>
			<persName><forename type="first">Kevin</forename><forename type="middle">W</forename><surname>Bowyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Nitesh</forename><forename type="middle">V</forename><surname>Chawla</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lawrence</forename><forename type="middle">O</forename><surname>Hall</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">Philip</forename><surname>Kegelmeyer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Artificial Intelligence Research</title>
		<imprint>
			<biblScope unit="volume">16</biblScope>
			<biblScope unit="page" from="321" to="357" />
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Measures of the Amount of Ecologic Association Between Species</title>
		<author>
			<persName><forename type="first">Lee</forename><forename type="middle">R</forename><surname>Dice</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Ecology</title>
		<imprint>
			<biblScope unit="volume">26</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="297" to="302" />
			<date type="published" when="1945">1945</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">A Theoretically Grounded Application of Dropout in Recurrent Neural Networks</title>
		<author>
			<persName><forename type="first">Yarin</forename><surname>Gal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zoubin</forename><surname>Ghahramani</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in neural information processing systems</title>
				<meeting><address><addrLine>Barcelona, Spain</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="1019" to="1027" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Identifying Purchase Intent from Social Posts</title>
		<author>
			<persName><forename type="first">Vineet</forename><surname>Gupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Devesh</forename><surname>Varshney</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Harsh</forename><surname>Jhamtani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Deepam</forename><surname>Kedia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Shweta</forename><surname>Karwa</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Eighth International AAAI Conference on Weblogs and Social Media</title>
				<meeting>the Eighth International AAAI Conference on Weblogs and Social Media<address><addrLine>Ann Arbor, MI</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="180" to="186" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Mining Purchase Intent in Twitter</title>
		<author>
			<persName><forename type="first">Rejwanul</forename><surname>Haque</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Arvind</forename><surname>Ramadurai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mohammed</forename><surname>Hasanuzzaman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Andy</forename><surname>Way</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computación y Sistemas</title>
		<imprint>
			<biblScope unit="volume">23</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="871" to="881" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Long short-term memory</title>
		<author>
			<persName><forename type="first">Sepp</forename><surname>Hochreiter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jürgen</forename><surname>Schmidhuber</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Neural Computation</title>
				<imprint>
			<publisher>MIT Press</publisher>
			<date type="published" when="1997">1997</date>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page" from="1735" to="1780" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Bag of Tricks for Efficient Text Classification</title>
		<author>
			<persName><forename type="first">Armand</forename><surname>Joulin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Edouard</forename><surname>Grave</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Piotr</forename><surname>Bojanowski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tomas</forename><surname>Mikolov</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics</title>
		<title level="s">Short Papers</title>
		<meeting>the 15th Conference of the European Chapter of the Association for Computational Linguistics<address><addrLine>Valencia, Spain</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="427" to="431" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m" type="main">Adam: A Method for Stochastic Optimization</title>
		<author>
			<persName><forename type="first">P</forename><surname>Diederik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jimmy</forename><surname>Kingma</surname></persName>
		</author>
		<author>
			<persName><surname>Ba</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1412.6980</idno>
		<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Distributed representations of words and phrases and their compositionality</title>
		<author>
			<persName><forename type="first">Tomas</forename><surname>Mikolov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ilya</forename><surname>Sutskever</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kai</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Greg</forename><forename type="middle">S</forename><surname>Corrado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jeff</forename><surname>Dean</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in neural information processing systems</title>
				<meeting><address><addrLine>Lake Tahoe, NV</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="3111" to="3119" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Automatically Identifying Complaints in Social Media</title>
		<author>
			<persName><forename type="first">Daniel</forename><surname>Preotiuc-Pietro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mihaela</forename><surname>Gaman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Nikolaos</forename><surname>Aletras</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 57th Annual Meeting of the Association for Computational Linguistic</title>
				<meeting>the 57th Annual Meeting of the Association for Computational Linguistic<address><addrLine>Los Angeles, CA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="219" to="224" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Partially labeled classification with Markov random walks: Advances in neural information processing systems</title>
		<author>
			<persName><forename type="first">Martin</forename><surname>Szummer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tommi</forename><surname>Jaakkola</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in neural information processing systems</title>
				<meeting><address><addrLine>Vancouver, BC</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2002">2002</date>
			<biblScope unit="page" from="945" to="952" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Mining user intents in twitter: a semi-supervised approach to inferring intent categories for tweets</title>
		<author>
			<persName><forename type="first">Jinpeng</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gao</forename><surname>Cong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Wayne</forename></persName>
		</author>
		<author>
			<persName><forename type="first">Xin</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Xiaoming</forename><surname>Li</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Twenty-Ninth AAAI Conference on Artificial Intelligence</title>
				<meeting><address><addrLine>Austin, TX</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="318" to="324" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Attention-based LSTM for aspect-level sentiment classification</title>
		<author>
			<persName><forename type="first">Yequan</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Minlie</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Xiaoyan</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Li</forename><surname>Zhao</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2016 conference on empirical methods in natural language processing</title>
				<meeting>the 2016 conference on empirical methods in natural language processing<address><addrLine>Austin, TX</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="606" to="615" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<title level="m" type="main">Text classification improved by integrating bidirectional LSTM with twodimensional max pooling</title>
		<author>
			<persName><forename type="first">Peng</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zhenyu</forename><surname>Qi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Suncong</forename><surname>Zheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jiaming</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hongyun</forename><surname>Bao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Bo</forename><surname>Xu</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1611.06639</idno>
		<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
