<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Automatic Aspect-Based Sentiment Analysis (AABSA) from Customer Reviews</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ella Jiaming Xu</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Bo Tang</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Xiao Liu</string-name>
          <email>xliug@stern.nyu.edu</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Feiyu Xiong</string-name>
          <email>feiyu.xfyg@alibaba-inc.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Alibaba Group</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Stern School of Business, New York University</institution>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Online review platforms provide enormous information for users to evaluate products and services. However, the sheer volume of reviews can create information overload that could increase user search costs and cognitive burden. To reduce information overload, in this paper, we propose an Automatic Aspect-Based Sentiment Analysis (AABSA) model to automatically identify key aspects from Chinese online reviews and conduct aspect-based sentiment analysis. We create a hierarchical structure of hypernyms and hyponyms, apply deep-learning-based representation learning and clustering to identify aspects that are the core content in the reviews, and then calculate the sentiment score of each aspect. To evaluate the performance of the identi ed aspects, we use an econometric model to estimate the impact of each aspect on product sales. We collaborate with one of Asia's largest online shopping platforms and employ the model in its product review tagging system to help consumers search for product aspects. Compared with benchmark models, our model is both more e ective, because it creates a more comprehensive list of aspects that are indicative of customer needs, and more e cient because it is fully automated without any human labor cost.</p>
      </abstract>
      <kwd-group>
        <kwd>Aspect-Based Sentiment Analysis</kwd>
        <kwd>Representation Learn- ing</kwd>
        <kwd>Deep Learning</kwd>
        <kwd>Sentiment Analysis</kwd>
        <kwd>Econometric Model</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Online reviews are critical for multiple stakeholders. Consumers can obtain rich
information from reviews to evaluate products and services. Firms can leverage
reviews to gain insights on customer needs and opportunities to improve their
products. Despite the enormous informational value provided by reviews, the
sheer volume of reviews has created a problem of information overload.
Consumers cannot easily process information of thousands of reviews to understand
the strengths and weaknesses of each product fully. Firms cannot easily glean
insights from immense unstructured review content of their own products and
competitors'. Although review rating, usually on a ve-point Likert scale, is a
useful summary statistic, it is a single-dimensional value that fails to capture
the multi-dimensional facets of products.</p>
      <p>
        To overcome the information overload problem, a few online platforms, such
as Yelp and TripAdvisor, have started to provide aspect-based sentiment scores
on review pages, which help customers select reviews containing the selected
aspects. Current literature also pay growing attention to identifying customer
needs from online reviews [
        <xref ref-type="bibr" rid="ref13 ref16 ref17 ref2 ref20 ref6 ref7">2, 28, 33, 30, 37, 6, 34, 7, 25</xref>
        ]. However, both the
platforms and the literature face several fundamental problems:
      </p>
      <p>
        First, many previous solutions applied the supervised learning approach to
de ne aspects manually and then match them to corresponding reviews [
        <xref ref-type="bibr" rid="ref16 ref17 ref2 ref6">6, 2, 34,
33</xref>
        ]. This approach has many drawbacks. First of all, e-commerce platforms, such
as Amazon and Alibaba, often cover a wide range of product categories. The
aspects that are relevant to one product category might be irrelevant for another
category. For example, sound quality is important for the TV category but not for
the sofa category. Therefore, it is time-consuming to identify aspects manually
for each product category. Moreover, e-commerce evolves rapidly. New product
categories constantly arrive, and customer tastes are continuously changing. It is
hard to keep up with the trend and manually identify aspects for each newly
developed category and new customer needs. For example, in 2019, a new product
category, anti-smoking smart lighter, was introduced. And new aspects such as
social connectedness and windproof need to be de ned and added. Furthermore,
even if de ning a set of aspects is feasible, annotating large datasets is highly
demanding on human labor costs and time. Last but not least, with signi cant
human intervention, uncontrollable bias may arise. Second, although some
previous papers also proposed automatic aspect detection, they selected the most
frequently mentioned aspects as the core product aspects [
        <xref ref-type="bibr" rid="ref2 ref23 ref6">40, 6, 2</xref>
        ]. The
drawback of this approach is that it ignores word similarities. For example, although
hue and color each might not be the most frequently mentioned keyword, they
can be constructed as an important aspect jointly. Some follow-up research
applied word embedding techniques, such as word2vec and wordnet clustering, to
capture word similarities [
        <xref ref-type="bibr" rid="ref19 ref21 ref23">40, 36, 38</xref>
        ]. But these works fail to capture complex
semantic relationships of aspect keywords.
      </p>
      <p>
        Third, previous literature assumes that all the aspects are at the same level [
        <xref ref-type="bibr" rid="ref17 ref2 ref6">6,
2, 34</xref>
        ]. However, there are limitations to attening the aspect structure. Consider
a laptop retailer aiming at improving the quality of laptops. Quality, undeniably,
is an important aspect of a laptop, but it is an abstract aspect that consists
of many sub-aspects like durability and speed. A review without the keyword
"quality," but with more speci c words such as "durability" and "speed" could
also re ect a consumer's overall sentiment towards the quality of a laptop.
      </p>
      <p>In summary, the following questions are left unsolved to conduct aspect-based
sentiment analysis:</p>
      <p>1. How can we identify aspects automatically with reduced cost and improved
exibility?</p>
      <p>2. How can we better capture the semantic relationship among keywords to
construct aspects?
3. Does the hierarchical structure among aspects exist, and can the
hierarchical structure improve the comprehensiveness of identi ed aspects?</p>
      <p>In this paper, we develop the Automatic Aspect-Based Sentiment Analysis
(AABSA) model to extract hierarchical-structured product aspects from online
consumer reviews. Speci cally, we provide solutions to the three questions
mentioned above:</p>
      <p>1. We propose a fully automated aspect-based sentiment analysis model
(AABSA). The model can create aspect-based sentiment scores from online
reviews without any human intervention or domain knowledge. In the AABSA
model, we applied k-means clustering to put sentence embeddings into groups
and select the center words from clusters as aspects. A prominent advantage of
the k-means clustering is that it is an unsupervised learning model so that we
do not need to pre-determine the aspects manually. AABSA could automatically
identify the aspects, aspect structure, and the number of aspects. No labels are
needed for the learning process, leaving it on its own to nd structure in its
input. The model saves the time of de ning labels and allows us to identify aspects
automatically.</p>
      <p>
        2. We introduce the Bidirectional Encoder Representations from
Transformers (BERT) model to transfer short reviews into sentence embeddings and cluster
them [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Unlike recent language representation methods, BERT jointly
considers both left and right side context of words in all layers and helps us better
capture the semantic relationship among aspects.
      </p>
      <p>3. We develop a hierarchical aspect structure consisting of hypernym aspects,
which are de ned as the core content that can summarize the semantics, and
hyponym aspects, which are de ned as the sub-aspects of hypernym aspects.
We rst cluster sentence embeddings and identify center words of clusters as
hypernym candidates. We then applied PageRank to build a weighted word
map with synonyms of hypernym candidates and applied PageRank to identify
hypernyms [26]. An essential advantage of the model is that the hypernyms and
hyponyms are not necessarily the words that appear most frequently, but those
that can capture the theme of the entire sentence. The hierarchical structure
signi cantly increases the comprehensiveness and accuracy of identi ed aspects.</p>
      <p>In summary, this paper makes several substantive and methodological
contributions. We propose an innovative method to identify product aspects by
introducing a hierarchical system. We demonstrate three comparative advantages
of the proposed model against benchmark methods: 1) improved
comprehensiveness, 2) better prediction accuracy on sales, and 3) full automation without
timeconsuming hand-coding. The method has been employed in Alibaba's Chinese
product review tagging system to help consumers search for product aspects.
2
2.1</p>
    </sec>
    <sec id="sec-2">
      <title>Literature Review</title>
      <sec id="sec-2-1">
        <title>Aspect Identi cation</title>
        <p>Online review platforms allow customers to express their attitudes towards
products and services freely, and customers rely on online reviews to make decisions.</p>
        <p>
          The sheer volume of online reviews makes it di cult for a human to process
and extract all meaningful information. Hence, research on identifying product
aspects from user-generated content and ranking their relative importance has
been proli c in the past few decades [
          <xref ref-type="bibr" rid="ref14 ref23 ref25">31, 40, 42, 12</xref>
          ]. The most common methods
rely on focus groups, experiential interviews, or ethnography as input. Trained
professional analysts, then review the input, manually identify customer needs,
remove redundancy, and structure the customer needs [
          <xref ref-type="bibr" rid="ref1 ref17">34, 16, 1</xref>
          ]. [
          <xref ref-type="bibr" rid="ref23">40</xref>
          ]
identied important aspects according the frequency and the in uence of consumers'
opinions given to each aspect on their overall opinions by a shallow dependency
parser. [
          <xref ref-type="bibr" rid="ref25">42</xref>
          ] then extended [
          <xref ref-type="bibr" rid="ref23">40</xref>
          ]'s paper by performing extensive evaluations on
more products in more diverse domains and more real-world applications. [12]
applied an automatic clustering approach to aspect identi cation. One common
limitation of these approaches is that they assume the frequency that an
aspect appears is positively correlated with its importance. However, high-level
and abstract concepts, such as "quality," may not appear very frequently in the
reviews. Still, the associated low-level, concrete concepts, such as durability and
conformance, may appear very frequently. The approaches, as mentioned above,
could fail to detect important high-level and abstract aspects. We instead
propose a method that can rely on the hierarchical structure between hypernyms
and hyponyms to detect important aspects. And our method is fully automatic,
not relying on any human labor cost.
        </p>
        <p>
          In the marketing eld, researchers often rely on existing psychological and
economic theory to pre-de ne a list of aspects and then extract the pre-de ned
aspects from user-generated reviews [
          <xref ref-type="bibr" rid="ref10 ref18 ref24 ref8">41, 24, 19, 35, 8, 10, 20</xref>
          ]. However, this
approach is theory-driven instead of data-driven. Therefore, it is hard to generalize
across contexts. For example, one paper that extracted the "health" aspect from
weight-loss products might not be relevant for another product category, such
as automobiles. In contrast, we propose a data-driven method that can extract
the most relevant aspects tailored to the speci c context. And our method is
domain knowledge agnostic, not relying on human expertise.
2.2
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Aspect Sentiment Analysis</title>
        <p>
          Sentiment analysis is a type of subjectivity analysis that aims to identify
opinions, emotions, and evaluations expressed in natural language [27]. The main
goal is to predict the sentiment orientation by analyzing opinion words and
expressions and detect trends. Sentiment analysis plays an important role in
identifying customer's attitudes towards brands, and recent studies are paying more
attention to developing more ne-grained aspect-based sentiment analysis on
user-generated content. Previously, researchers studied extraction of evaluating
expressions from customer opinions [
          <xref ref-type="bibr" rid="ref15 ref26 ref4">4, 17, 27, 32, 43, 14</xref>
          ]. [14] extracted features
and summarized opinions from consumer reviews by part-of-speech tagging and
built an opinion word list. [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] summarized the sentiment of reviews for a local
service and focused on aspect-based summarization models.
        </p>
        <p>
          With the development of machine learning techniques, researchers applied
these advanced techniques to sentiment analysis. [23] introduced support
vector machines (SVMs) and unigram models to sentiment analysis. [22] applied
Naive Bayes to analyze aspect-based sentiment. [18] applied the Maximum
Entropy (MaxEnt) classi cation to classify consumer messages into either positive
or negative. [27] researched the performance of various machine learning
techniques, including MaxEnt classi cation, and showed that MaxEnt classi cation
was powerful with classifying reviews. Researchers then applied deep learning
methods such as XLNet and LSTM to conduct sentiment analysis [
          <xref ref-type="bibr" rid="ref11 ref12 ref22">39, 11, 15,
29</xref>
          ]. In this paper, we tested both MaxEnt classi cation and Fasttext and found
that MaxEnt outperforms Fasttext because the e-commerce platform has
created a rich sentiment vocabulary pertaining to product reviews. The compared
results of MaxEnt and Fasttext is listed in Table 2 of Appendices.
3
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>AABSA</title>
    </sec>
    <sec id="sec-4">
      <title>Model Framework</title>
      <p>In this section, we describe the details of the structure of our aspect-sentiment
analysis model, AABSA. We start with an overview of its framework, which
consists of two main components: aspect identi cation and sentiment analysis.
We then describe the baseline model to be compared with.
3.1</p>
      <sec id="sec-4-1">
        <title>Aspect-based Sentiment Analysis Problem</title>
        <p>
          The aspect-based sentiment analysis problem is to identify product aspects from
a review document, and the aspects represent the most important customer needs
in the document. Having identi ed the aspects, we then need to associate
sentiment scores with every aspect. We make two assumptions. First, we assume that
each sentence is possible to be associated with more than one aspect. Second,
we assume that the hierarchical structure exists among aspects. We classify
aspects into hypernyms and hyponyms. Hypernyms are the core content that can
summarize the theme of the content and are often abstract and involve various
sub-aspects. For example, "battery" is a core theme in the camera category
identi ed by a previous research [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. However, from the retailer's perspective,
"battery" cannot provide them with detailed and comprehensive information on the
direction to improve the battery aspect. As a result, we identify the sub-aspects
as their hyponyms. For example, if the "battery" is identi ed as a hypernym,
then "battery life" and "battery production place" are its possible hyponyms.
Hyponyms could provide more exact direction for product improvement.
3.2
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>AABSA Model Framework</title>
        <p>Our model consists of eight steps:</p>
        <p>1. Pre-process reviews. We collected reviews from Alibaba, one of the biggest
e-commerce platforms in Asia. In our research, we analyze reviews for two
product categories: camera and toothbrush. There are two reasons why we choose
these two categories. First, camera and toothbrush are common products and
they are widely analyzed in marketing literature and we are able to compare our
results with previous works. Second, the camera represent the high-end
product categories and the toothbrush represent the lower-end and more daily-used
products. We can compare the impact of reviews on the sales of them and
generate business insights. We divide the entire review document into short sentences
and identify informative sentences, which were de ned by the company's existing
internal rules. For example, the sentence "Very good" is classi ed as
uninformative, whereas the sentence "The battery can last more than 10 hours" is classi ed
as informative.</p>
        <p>2. Train word embeddings. The hierarchical aspect structure is based on the
relationships between hypernyms and hyponyms, which are represented by word
similarities. Concerning quantitative similarity representations, we convert words
into vectors using the Word2vec algorithm and eliminate the lower-frequency
words in synonym pairs [21].</p>
        <p>
          3. Train sentence embeddings. In order to measure the similarity between
reviews quantitatively, we convert the most frequent 50% short sentences into
sentence embeddings with BERT [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ].
        </p>
        <p>4. Select hypernym candidates. We assume that a few core words, which
are de ned as hypernym candidates, could summary each sentence. Hence, we
apply k-means clustering to generate semantic categories and select the most
important words, whose accumulated cosine distances to their cluster centers
are the shortest within the clusters, as hypernym candidates. We then lter
invalid words among hypernyms candidates.</p>
        <p>5. Further, we introduce the concept of hyponyms to assist in the subsequent
sentiment analysis step. We select the words closest to the hypernym candidates
in each cluster as hyponyms candidates. We then select the words closest to
the hyponym candidates as their subordinates. Then we construct a weighted
word network comprising hyponym candidates, hypernym candidates, and their
subordinates. We then apply PageRank to select hyponym candidates according
to their relative importance.</p>
        <p>6. Merge hypernyms and hyponym candidates. We nd that there are overlaps
among hypernyms and hyponym candidates. To avoid redundancy, we rank all
the hypernyms according to their importance and merge the hypernym and
hyponym candidates to nalize hypernyms. If a hypernym belongs to several
hypernym sets, then we merge the hypernym with its hyponym candidates to
the hyponym set of the highest-ranked hypernym.</p>
        <p>7. Match hypernyms to reviews. We select the words closest to hypernyms
and hyponyms from the content and then apply a regular expression matching
to match them to reviews.</p>
        <p>8. We use the Maximum Entropy (MaxEnt) classi cation to classify reviews
sentences associated with each aspect into positive, neutral, or negative. The
sentiment score of reviews of each product is aggregated at the week level.</p>
        <p>The framework and algorithm of AABSA model is shown in Algorithm 1 and
Figure 1.
Algorithm 1 Aspect Identi cation</p>
        <p>I
Input: Reviews: fRigi=1
Output: Hypernym set H1 and hyponym set H2
1: Learn word vectors from reviews: fW VjgjJ=1 = Word2Vec(fRigiI=1)
2: Divide all reviews fRigi=1 into short reviews fSRigiS=I1</p>
        <p>
          I
3: Calculate vector representation of short reviews fSRVigiS=I1 with BERT
4: Cluster fSRVigiS=I1 into k clusters with k-means clustering
5: Calculate the center of each cluster m: Cm
6: for m in clusters do
7: for word wj in cluster m do
8: Calculate the frequency of wj in cluster m: Nmj
9: for instance nwj of word wj do
10: Calculate the cosine distance between wj and Cm: Dn(wj; m)
11: Calculate importance of wj in m: Fmj = PnNwmjj=1 Dn(wj; m)
Pre-process reviews In general, online reviews are complex sentences
consisting of complicate sentiments and are composed of both informative and
uninformative contents [
          <xref ref-type="bibr" rid="ref17">34</xref>
          ]. For example, in a review such as \I just got this camera
today, and it looks fantastic but it's too heavy for me!", the rst clause is
uninformative since it is irrelevant to the camera's aspects, while the second clause
describes the customer's positive attitude towards its appearance and negative
attitude towards its weight. To better identify sentiments and informative
contents, we separate original comments into single sentences and then
automatically eliminate uninformative single sentences with regular expression matching
with prede ned regulations. Then we automatically eliminate the stop-words,
numbers, brand names, and punctuation.
        </p>
        <p>Train word embeddings To measure the similarities and also gure out
synonyms quantitatively, we need to transfer words into vectors with word
embed</p>
        <sec id="sec-4-2-1">
          <title>Input:</title>
        </sec>
        <sec id="sec-4-2-2">
          <title>Raw Consumer</title>
        </sec>
        <sec id="sec-4-2-3">
          <title>Reviews</title>
        </sec>
        <sec id="sec-4-2-4">
          <title>Merge Hypernym and</title>
        </sec>
        <sec id="sec-4-2-5">
          <title>Hyponym Candidates</title>
        </sec>
        <sec id="sec-4-2-6">
          <title>Match Hypernyms</title>
          <p>and Hyponyms to</p>
        </sec>
        <sec id="sec-4-2-7">
          <title>Reviews</title>
          <p>ding. Word embedding is a representation of document vocabulary utilizing the
context of a word, including semantic and syntactic similarity and word
relationships. With word embedding, words used in similar contexts have similar
representations, and the cosine similarity between word vectors could
quantitatively represent similarities between words. We apply a skip-gram word2vec
model to train word embeddings [21]. Skip-gram takes as its input a large
corpus of text and produces a vector space, typically of several hundred dimensions,
with each unique word in the corpus being assigned a corresponding vector in
the space [21].</p>
          <p>Train Sentence Embedding (with BERT) Sentence embeddings are
useful for keyword expansion and are used to identify the relationship between
words and sentences. In order to quantify the relationship between the sentences
and discover the latent customer needs, we formulate sentence embeddings and
extract keywords from the sentence clusters afterward. Consider the following
examples:
\The toothbrush hair is super soft, and it really protects my son's teeth!"
\I am really disappointed that its toothbrush hair too soft, and it cannot
clean my teeth."</p>
          <p>
            These two sentences are di erent expressions of opposite attitudes towards
the same product aspect, but they are similar in the semantic structure. In
earlier works, researchers often created sentence embeddings by directly taking
the average of word embeddings, which ignores the semantic and concatenate
relationships between sentences [
            <xref ref-type="bibr" rid="ref9">9</xref>
            ]. For example, word2vec would produce the
same word embedding for the word \soft" in both sentences. Language models
training word embeddings only use directionless or unidirectional context and
match each word to a xed representation regardless of the context within which
the word appears. Discussing the same aspect in a similar semantic structure
might have di erent meanings. In this paper, we apply BERT (Bidirectional
Encoder Representations from Transformers) to obtain sentence embeddings.
          </p>
          <p>BERT is a deep learning-based model in natural language processing, and the
architecture is a multi-layer bidirectional Transformer encoder. It is designed to
learn deep bidirectional representations from the unlabeled text by cooperatively
considering both sides of context. BERT trains contextual representations on
text corpus and produces word representations that are dynamically informed by
the words around them. In contrast to previous e orts that read text sequentially
either from left to right or right to left, BERT introduces more comprehensive
and global word relationships to the word representation. BERT is bidirectional,
generalizable, has high-performance, and universal. Since the pre-training
procedure is comparatively hardware-demanding and time-consuming, we use BERT's
own pre-built pre-training model, Chinese L-12 H-768 A-12, which was trained
by Google with Chinese Wikipedia data, as our pre-training model.
Select Hypernym Candidates After training sentence embeddings, we
cluster them with the k-means clustering algorithm. We assume that sentence vectors
within a cluster describe the same customer needs, and a limited number of core
words could summarize the opinions of each cluster. To exploit variety and
comprehensiveness, we select the non-repeated central words of each of the top 10
largest clusters as hypernym candidates. Both silhouette coe cients and BIC
determine the optimal number of clusters.</p>
          <p>The process of selecting hypernym candidates is as follows. Denote
embedding of sentence i in cluster m as smi, word j as wj , and if wj appears in smi then
indicator amij is 1, otherwise is 0. The number of sentence embeddings in cluster
m is Nm and the number of wj appearance in smi is nmij . We rst calculate
the cosine distance between smi to its cluster center, dmi, and the distance is
proportional to its representativeness. We sum up the cosine similarities between
wj and the cluster center as its importance in cluster m:</p>
          <p>The cosine distance also represents the similarities between words in a
sentence and its cluster center. Since words repeatedly appear in di erent sentences,
we sum up the cosine distances between words and their cluster center as their
nal distances. The words with the largest similarity within each cluster are the
most core words, and we select the top two words from each cluster as
hypernym candidates. The whole process of selecting hypernym candidates is shown
in Figure 2. Hypernyms candidates are then nalized after eliminating repeated
candidates chosen from all clusters.</p>
          <p>Sentence 1
Sentence 2
Sentence N</p>
          <p>Hypernym Candidate 1
Hypernym Candidate 2
Hypernym Candidate M</p>
          <p>CID 1
CID M</p>
          <p>Sentence 1
Sentence 2</p>
          <p>CID 1</p>
          <p>CID m
Sentence N</p>
          <p>CID M
Sentence_11
Sentence_12
Sentence_1x1
Sentence_m1
Sentence_m2</p>
          <p>
            Sentence_MxM
Recall Hyponym Candidates (with PageRank) As we mentioned earlier,
hyponyms provide retailers with more detailed and granular information about
a product. Another purpose of introducing hyponyms is that also they help
matching hypernyms to more related reviews. We select the closest words to each
hypernym candidate as second-order related words and again select the closest
words to each second-order related words as third-order ones. Then we construct
weighted directed wordnet where weights are determined by distances between
pre-trained word embeddings. The process of building the wordnet is shown in
Figure 3. Then we apply the PageRank algorithm to generate the nal hyponyms
according to their relative closeness and importance. PageRank is an iterative
algorithm that determines the importance of a web page based on the importance
of its parent page [
            <xref ref-type="bibr" rid="ref5">26, 5, 13</xref>
            ]. The core idea of PageRank is that the rank of an
element is divided among its forward links evenly to contribute to the ranks of
the pages they point to. After PageRank of each element is obtained, we select
words with the highest PageRank between hypernyms as hyponym candidates.
Compared with selecting the closest words to hypernyms and hyponyms, the
main advantage of using PageRank is that it uses the entire graph rather than
a small subset to estimate relative relationships between words. As a result, it
enlarges the diversity, reliability, and richness of identi ed aspects.
          </p>
          <p>The procedure of calculating PageRank is described as follows. Let Fi be the
set of words that word i points to and Bi be the set of words that points to
i. Let Ni = jFij be the number of links from i and let c be a factor used for
normalization. PageRank of i is then</p>
          <p>Hypernym
Candidate
2nd-order
related word
2nd-order
related word
2nd-order
related word</p>
          <p>R(i) = c X R(v)
v2Bi</p>
          <p>Nv
3rd-order
related word
3rd-order
related word
3rd-order
related word
Merge Hypernyms and Hyponyms After nalizing the recall process, we
noticed that there are overlaps between hypernyms and hyponyms. For example,
hypernym candidate A is also a hyponym of hypernym candidate B. Overlaps
would cause redundancy and confusion when mapping sentiment to aspects in
the following steps. In order to further improve the precision of the constructed
aspect lexicon and investigate the internal similarity between hypernym
candidates, we merge hypernyms and hyponym candidates.</p>
          <p>Match hypernyms to Reviews The next step is matching hypernyms to
the reviews discussing corresponding aspects. With the pre-trained word
embeddings, we select the closest words to hyponyms and match them with hypernyms
and hyponyms to reviews with regular expression matching.</p>
          <p>
            Sentiment Analysis In the next step of the AABSA model, we need to
identify the sentiment evaluation of identi ed aspects. We applied the MaxEnt
classi cation algorithm to the sentiment classi cation problem. MaxEnt models are
feature-based models and could solve feature selection and model selection.
MaxEnt classi cation is proved to be e ective in a number of natural language
processing applications [
            <xref ref-type="bibr" rid="ref3">27, 3</xref>
            ]. The goal is to assign a class c to a given document d
to maximize P (cjd), which is calculated as below:
          </p>
          <p>PME (cjd) =</p>
          <p>1 exp(X
Z(d)
i
i;cfi;c(d; c))
(3)
where Z(d) is a normalization function. Fi;c is a aspect function for aspect fi
and class c. Fi;c(d; c0) = 1 if ni(d) &gt; 0 and c0 = c. The i;c is a aspect-weighted
parameter and a large i;c means that fi is considered a strong indicator for
class c.</p>
          <p>Now, each review can be represented as a vector consists of aspects and
sentiment evaluations. We measure the overall sentiment evaluation of aspect i
in week t as:
sentimentit =</p>
          <p>Pnit
j=1</p>
          <p>Pkm=ij1t sentimentkjt</p>
          <p>
            mijt
nit
where nit is the number of reviews that mention aspect i in week t and mijt is
the time of i's appearance in review j in week t.
The baseline model is developed by [
            <xref ref-type="bibr" rid="ref2">2</xref>
            ] on deriving product aspects by mining
consumer reviews. It mainly consists of four steps: pre-process content, eliminate
synonyms, obtain core word candidates, and select hypernyms. After splitting
reviews into sentences and remove stopwords, they calculate the TF-IDF value of
each word and convert words into one-hot vectors with TF-IDF values of context
words. Then they cluster word vectors with k-means clustering and choose the
center words of clusters as hypernyms. The major di erences are the training
process of word vectors and the application of hierarchical structure of aspects.
The architecture of the baseline model is indicated in Figure 4.
(4)
In this section, we evaluate the AABSA model with review data drawn from
product categories \Toothbrush" and \Camera" provided by Alibaba. In section
4.1, we rst describe our data set. Then in section 4.2, we describe the identi ed
aspects with AABSA.
Alibaba Group is one of the largest e-commerce companies in Asia, which was
rst launched in 1999 in China. It is commonly referred to as the \Chinese
Amazon." As of June 2019, Alibaba has 755 million active users in more than
200 countries. It has three biggest digital shopping platforms, Alibaba, Taobao,
and Tmall, which focus on B2B, C2B, and B2C business separately. Since 2010,
Alibaba has launched sales on singles day in November and Spring Festival in
January or February. In our work, we used panel data from 20 weeks between
March and July to avoid uctuation caused by the sales e ect. For each item, we
observe reviews, ratings, weekly sales, and essential attributes (e.g., price, weight,
popularity), which are de ned by retailers when the products were launched. The
full data set consists of 295,628 reviews of 13,944 camera products and 18,550,956
reviews of 147,337 toothbrush products. In the pre-processing step, we use the
reviews to build a vocabulary of nouns from which we select hypernyms and
hyponyms. In the sentiment analysis step, we hired human taggers to classify
aspect sentiments into three categories: positive, neutral, and negative.
4.2
          </p>
        </sec>
      </sec>
      <sec id="sec-4-3">
        <title>Selecting Aspects</title>
        <p>Tables 1 and Table 2 describe the top 10 hypernyms and 50 hyponyms from
camera and toothbrush reviews. We nd that aspects obtained from our model
provide more detailed and comprehensive information on product aspects and
customer needs. Each aspect captured by the AABSA model represents a
detailed aspect, and it could provide clear instructions for rms to perform product
improvement. For example, a positive "pixel" aspect indicates that the photo
taken by the camera is clear. However, some words obtained by the baseline
model, such as "cell phone" and "camera" are broad-de ned aspects, and it is
hard for rms to make speci c improvements given this information.</p>
        <p>To make an apple-to-apple comparison, among the aspects identi ed by the
baseline model, we select the 10 most frequently-mentioned aspects. They are
shown in Table 3.
5</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Experimental Results</title>
      <p>We describe two main sets of results: i) performance comparison of our AABSA
and the benchmark model and ii) the marketing insights. First, we select the
most popular and frequent 10 aspects from all hypernyms and hyponyms for
AABSA and the baseline model. The aspects are shown in Table 4.</p>
      <p>Accuracy To compare the performance of AABSA and the benchmark
model, we create an econometric model to estimate the impact of each aspect
on product sales. The intuition is that if the identi ed aspects are more useful
for consumers and rms, they should be better predictors of product sales. We
calculate the percentage of positive reviews of an aspect in the past 180 days of
the week t as the support rate of the aspect and then use the support rate to
predict product sales in week t in linear regression. In our model, there are 9
hypernyms and 1 hyponym of \price," \o ine." We then report the performance
of our model and the baseline model in terms of sales prediction accuracy and
analyze the prediction power of each aspect. The regression result of cameras is
shown in table 5. The rst column reports the estimates from AABSA, and the
second column reports estimates from the baseline model.</p>
      <p>We can make several inferences from the regression coe cients. First, in our
model, coe cients for every aspect are signi cant, and the adjust r-squared is
5% higher than that of the baseline model. Second, we nd that while positive
reviews on most aspects have positive e ects on sales, positives reviews on o ine
stores have negative e ects on sales. One plausible explanation for this e ect is
that the o ine and online stores are of competitive relationships, and customers
would tend to switch to o ine stores if they read related positive reviews on
online retailing platforms.</p>
      <p>The regression result of toothbrushes is shown in table 6. In the toothbrush
category, the coe cients are all signi cant, and our model also out-performances
the baseline model by 2%. However, the improvement is not as much as in the
camera category. A plausible explanation is that toothbrushes are daily
necessities, and they are much cheaper than cameras. As a result, when consumers
purchase toothbrushes, they would spare less time to read textual reviews, and
the sales prediction power of reviews is weakened.</p>
      <p>
        Comprehensiveness We then compare the comprehensiveness of the
aspects identi ed by our model to develop some intuition of what drives the
performance discrepancy. [
        <xref ref-type="bibr" rid="ref17">34</xref>
        ] identi ed 6 primary customer needs and 22 secondary
customer needs of oral care products with a machine-learning hybrid method and
then classi ed the needs into the primary group and the secondary group. We
compare the toothbrush aspects extracted from the AABSA model with aspects
extracted from [
        <xref ref-type="bibr" rid="ref17">34</xref>
        ]'s model. The comparison table is listed in Table 1 in the
Appendices.
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref17">34</xref>
        ]'s results, \feel clean and fresh" captures the customer's own oral
feeling while and after using oral care products, and in AABSA's results, \easy to
clean" hypernym aspect also describes the customers' feelings' of the oral smell
after brushing teeth, and \toothbrush hair" and \toothbrush head" captures
the comfort while using the toothbrush; \strong teeth and gums" describes the
aspect of preventing gingivitis and protecting the gum, and we identi ed \gum
pain"; \Product e cacy" describes the e cacy of oral care products, which focus
on a more subjective aspect. It matches \e cacy" in our model, which re ects
the e ect of using the toothbrush; \Convenience" describes the convenience of
using the oral product to reach the cleaning perspective, and AABSA identi ed
\easy to clean" which also describes the toothbrush's ability to clean teeth;
and \Shopping/product choice" describes the competitiveness between brands.
From the above results, we can conclude that our model can create a more
comprehensive list of aspects than [
        <xref ref-type="bibr" rid="ref17">34</xref>
        ].
6
      </p>
    </sec>
    <sec id="sec-6">
      <title>Discussions</title>
      <p>In this paper, we propose an innovative method to identify product aspects by
introducing a hierarchical system. Compared with the previous aspect
identication and sentiment analysis model, the AABSA model improves the
comprehensiveness and the prediction accuracy on sales, and it is fully automatic
in aspect-identi cation without time-consuming hand-coding. The method has
been adopted by Alibaba's product review tagging system to help consumers
search for product aspects.
12. Hadano, M., Shimada, K., Endo, T.: Aspect identi cation of sentiment
sentences using a clustering algorithm. Procedia - Social and Behavioral Sciences
27, 22 { 31 (2011). https://doi.org/https://doi.org/10.1016/j.sbspro.2011.10.579,
http://www.sciencedirect.com/science/article/pii/S1877042811024062,
computational Linguistics and Related Fields
13. Haveliwala, T.: E cient computation of pagerank. Tech. rep., Stanford (1999)
14. Hu, M., Liu, B.: Mining opinion features in customer reviews. In: AAAI. vol. 4,
pp. 755{760 (2004)
15. Johnson, R., Zhang, T.: Supervised and semi-supervised text categorization using
lstm for region embeddings. arXiv preprint arXiv:1602.02373 (2016)
16. Kaulio, M.A.: Customer, consumer and user involvement in product development:
A framework and a review of selected methods. Total quality management 9(1),
141{149 (1998)
17. Kobayashi, N., Inui, K., Matsumoto, Y.: Extracting aspect-evaluation and
aspectof relations in opinion mining. In: Proceedings of the 2007 Joint Conference on
Empirical Methods in Natural Language Processing and Computational Natural
Language Learning (EMNLP-CoNLL). pp. 1065{1074 (2007)
18. Lee, H.Y., Renganathan, H.: Chinese sentiment analysis using maximum entropy.</p>
      <p>In: Proceedings of the Workshop on Sentiment Analysis where AI meets
Psychology (SAAIP 2011). pp. 89{93. Asian Federation of Natural Language Processing,
Chiang Mai, Thailand (Nov 2011), https://www.aclweb.org/anthology/W11-3713
19. Lee, T.Y., Bradlow, E.T.: Automated marketing research using online customer
reviews. Journal of Marketing Research 48(5), 881{894 (2011)
20. Liu, X., Lee, D., Srinivasan, K.: Large scale cross category analysis of consumer
review content on sales conversion leveraging deep learning. Available at SSRN
2848528 (2017)
21. Mikolov, T., Sutskever, I., Chen, K., Corrado, G.S., Dean, J.: Distributed
representations of words and phrases and their compositionality. In: Advances in neural
information processing systems. pp. 3111{3119 (2013)
22. Mubarok, M.S., Adiwijaya, Aldhi, M.D.: Aspect-based sentiment analysis to review
products using nave bayes. In: AIP Conference Proceedings. vol. 1867, p. 020060.</p>
      <p>AIP Publishing (2017)
23. Mullen, T., Collier, N.: Sentiment analysis using support vector machines with
diverse information sources. In: Proceedings of the 2004 conference on empirical
methods in natural language processing. pp. 412{418 (2004)
24. Netzer, O., Feldman, R., Goldenberg, J., Fresko, M.: Mine your own business:
Market-structure surveillance through text mining. Marketing Science 31(3), 521{
543 (2012)
25. Nguyen, T.H., Shirai, K.: Phrasernn: Phrase recursive neural network for
aspectbased sentiment analysis. In: Proceedings of the 2015 Conference on Empirical
Methods in Natural Language Processing. pp. 2509{2514 (2015)
26. Page, L., Brin, S., Motwani, R., Winograd, T.: The pagerank citation ranking:</p>
      <p>Bringing order to the web. Tech. rep., Stanford InfoLab (1999)
27. Pang, B., Lee, L., Vaithyanathan, S.: Thumbs up?: sentiment classi cation using
machine learning techniques. In: Proceedings of the ACL-02 conference on
Empirical methods in natural language processing-Volume 10. pp. 79{86. Association for
Computational Linguistics (2002)
28. Pontiki, M., Galanis, D., Papageorgiou, H., Androutsopoulos, I., Manandhar, S.,
Mohammad, A.S., Al-Ayyoub, M., Zhao, Y., Qin, B., De Clercq, O., et al.:
Semeval2016 task 5: Aspect based sentiment analysis. In: Proceedings of the 10th
international workshop on semantic evaluation (SemEval-2016). pp. 19{30 (2016)</p>
    </sec>
    <sec id="sec-7">
      <title>Appendix</title>
      <p>easy to clean
smell
toothbrush hair,
toothbrush head
gums, e cacy disease</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Alam</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Perry</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>A customer-oriented new service development process</article-title>
          .
          <source>Journal of services Marketing</source>
          <volume>16</volume>
          (
          <issue>6</issue>
          ),
          <volume>515</volume>
          {
          <fpage>534</fpage>
          (
          <year>2002</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Archak</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ghose</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ipeirotis</surname>
            ,
            <given-names>P.G.</given-names>
          </string-name>
          :
          <article-title>Show me the money!: deriving the pricing power of product features by mining consumer reviews</article-title>
          .
          <source>In: Proceedings of the 13th ACM SIGKDD international conference on Knowledge discovery and data mining</source>
          . pp.
          <volume>56</volume>
          {
          <fpage>65</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Berger</surname>
            ,
            <given-names>A.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pietra</surname>
            ,
            <given-names>V.J.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pietra</surname>
            ,
            <given-names>S.A.D.:</given-names>
          </string-name>
          <article-title>A maximum entropy approach to natural language processing</article-title>
          .
          <source>Computational linguistics 22(1)</source>
          ,
          <volume>39</volume>
          {
          <fpage>71</fpage>
          (
          <year>1996</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Blair-Goldensohn</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hannan</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McDonald</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Neylon</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Reis</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Reynar</surname>
          </string-name>
          , J.:
          <article-title>Building a sentiment summarizer for local service reviews (</article-title>
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Brin</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Page</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>The anatomy of a large-scale hypertextual web search engine</article-title>
          .
          <source>Computer networks and ISDN systems 30(1-7)</source>
          ,
          <volume>107</volume>
          {
          <fpage>117</fpage>
          (
          <year>1998</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Chakraborty</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sudhir</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>Attribute sentiment scoring with online text reviews: Accounting for language structure and attribute self-selection (</article-title>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Che</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhao</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Guo</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Su</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          , Liu,
          <string-name>
            <surname>T.</surname>
          </string-name>
          :
          <article-title>Sentence compression for aspectbased sentiment analysis</article-title>
          .
          <source>IEEE/ACM Transactions on audio, speech, and language processing</source>
          <volume>23</volume>
          (
          <issue>12</issue>
          ),
          <volume>2111</volume>
          {
          <fpage>2124</fpage>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Chevalier</surname>
            ,
            <given-names>J.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mayzlin</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>The e ect of word of mouth on sales: Online book reviews</article-title>
          .
          <source>Journal of marketing research 43(3)</source>
          ,
          <volume>345</volume>
          {
          <fpage>354</fpage>
          (
          <year>2006</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Devlin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>M.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Toutanova</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>Bert: Pre-training of deep bidirectional transformers for language understanding</article-title>
          . arXiv preprint arXiv:
          <year>1810</year>
          .
          <volume>04805</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Dhar</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>E.A.</given-names>
          </string-name>
          :
          <article-title>Does chatter matter? the impact of user-generated content on music sales</article-title>
          .
          <source>Journal of Interactive Marketing</source>
          <volume>23</volume>
          (
          <issue>4</issue>
          ),
          <volume>300</volume>
          {
          <fpage>307</fpage>
          (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Gray</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Radford</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kingma</surname>
            ,
            <given-names>D.P.</given-names>
          </string-name>
          :
          <article-title>Gpu kernels for block-sparse weights</article-title>
          .
          <source>arXiv preprint arXiv:1711.09224</source>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          29.
          <string-name>
            <surname>Radford</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jozefowicz</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sutskever</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Learning to generate reviews and discovering sentiment</article-title>
          .
          <source>arXiv preprint arXiv:1704.01444</source>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          30.
          <string-name>
            <surname>Ruder</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , Gha ari, P.,
          <string-name>
            <surname>Breslin</surname>
            ,
            <given-names>J.G.</given-names>
          </string-name>
          :
          <article-title>A hierarchical model of reviews for aspectbased sentiment analysis</article-title>
          .
          <source>arXiv preprint arXiv:1609.02745</source>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          31.
          <string-name>
            <surname>Snyder</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Barzilay</surname>
          </string-name>
          , R.:
          <article-title>Multiple aspect ranking using the good grief algorithm</article-title>
          .
          <source>In: Human Language Technologies</source>
          <year>2007</year>
          :
          <article-title>The Conference of the North American Chapter of the Association for Computational Linguistics;</article-title>
          <source>Proceedings of the Main Conference</source>
          . pp.
          <volume>300</volume>
          {
          <issue>307</issue>
          (
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          32.
          <string-name>
            <surname>Tadano</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shimada</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Endo</surname>
          </string-name>
          , T.:
          <article-title>E ective construction and expansion of a sentiment corpus using an existing corpus and evaluative criteria estimation</article-title>
          .
          <source>In: Proceedings of the 11th Conference of the Paci c Association for Computational Linguistics (PACLING2009)</source>
          . pp.
          <volume>211</volume>
          {
          <fpage>216</fpage>
          .
          <string-name>
            <surname>Citeseer</surname>
          </string-name>
          (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          33.
          <string-name>
            <surname>Thet</surname>
          </string-name>
          , T.T.,
          <string-name>
            <surname>Na</surname>
            ,
            <given-names>J.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Khoo</surname>
            ,
            <given-names>C.S.</given-names>
          </string-name>
          :
          <article-title>Aspect-based sentiment analysis of movie reviews on discussion boards</article-title>
          .
          <source>Journal of information science 36</source>
          (
          <issue>6</issue>
          ),
          <volume>823</volume>
          {
          <fpage>848</fpage>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          34.
          <string-name>
            <surname>Timoshenko</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hauser</surname>
            ,
            <given-names>J.R.</given-names>
          </string-name>
          :
          <article-title>Identifying customer needs from user-generated content</article-title>
          .
          <source>Marketing Science</source>
          <volume>38</volume>
          (
          <issue>1</issue>
          ),
          <volume>1</volume>
          {
          <fpage>20</fpage>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          35.
          <string-name>
            <surname>Tirunillai</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tellis</surname>
            ,
            <given-names>G.J.:</given-names>
          </string-name>
          <article-title>Mining marketing meaning from online chatter: Strategic brand analysis of big data using latent dirichlet allocation</article-title>
          .
          <source>Journal of Marketing Research</source>
          <volume>51</volume>
          (
          <issue>4</issue>
          ),
          <volume>463</volume>
          {
          <fpage>479</fpage>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          36.
          <string-name>
            <surname>Tsai</surname>
            ,
            <given-names>Y.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>Y.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chung</surname>
            ,
            <given-names>C.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Su</surname>
            ,
            <given-names>S.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tsai</surname>
          </string-name>
          , R.T.H.:
          <article-title>Aspect-categorybased sentiment classi cation with aspect-opinion relation</article-title>
          .
          <source>In: 2016 Conference on Technologies and Applications of Arti cial Intelligence (TAAI)</source>
          . pp.
          <volume>162</volume>
          {
          <fpage>169</fpage>
          .
          <string-name>
            <surname>IEEE</surname>
          </string-name>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          37.
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pan</surname>
            ,
            <given-names>S.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dahlmeier</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xiao</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          :
          <article-title>Recursive neural conditional random elds for aspect-based sentiment analysis</article-title>
          .
          <source>arXiv preprint arXiv:1603.06679</source>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          38.
          <string-name>
            <surname>Xue</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fu</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shaobin</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          :
          <article-title>A study on sentiment computing and classi cation of sina weibo with word2vec</article-title>
          .
          <source>In: 2014 IEEE International Congress on Big Data</source>
          . pp.
          <volume>358</volume>
          {
          <fpage>363</fpage>
          .
          <string-name>
            <surname>IEEE</surname>
          </string-name>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          39.
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dai</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          , Carbonell, J.,
          <string-name>
            <surname>Salakhutdinov</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Le</surname>
            ,
            <given-names>Q.V.</given-names>
          </string-name>
          :
          <article-title>Xlnet: Generalized autoregressive pretraining for language understanding</article-title>
          . arXiv preprint arXiv:
          <year>1906</year>
          .
          <volume>08237</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          40.
          <string-name>
            <surname>Yu</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zha</surname>
            ,
            <given-names>Z.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chua</surname>
            ,
            <given-names>T.S.:</given-names>
          </string-name>
          <article-title>Aspect ranking: Identifying important product aspects from online consumer reviews</article-title>
          .
          <source>In: Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies - Volume 1</source>
          . pp.
          <volume>1496</volume>
          {
          <fpage>1505</fpage>
          . HLT '
          <volume>11</volume>
          ,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computational Linguistics, Stroudsburg, PA, USA (
          <year>2011</year>
          ), http://dl.acm.org/citation.cfm?id=
          <volume>2002472</volume>
          .
          <fpage>2002654</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          41.
          <string-name>
            <surname>Zha</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yu</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chua</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Product aspect ranking and its applications</article-title>
          .
          <source>IEEE Transactions on Knowledge and Data Engineering</source>
          <volume>26</volume>
          (
          <issue>5</issue>
          ),
          <volume>1211</volume>
          { 1224 (May
          <year>2014</year>
          ). https://doi.org/10.1109/TKDE.
          <year>2013</year>
          .136
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          42.
          <string-name>
            <surname>Zha</surname>
            ,
            <given-names>Z.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yu</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chua</surname>
            ,
            <given-names>T.S.:</given-names>
          </string-name>
          <article-title>Product aspect ranking and its applications</article-title>
          .
          <source>IEEE transactions on knowledge and data engineering 26(5)</source>
          ,
          <volume>1211</volume>
          {
          <fpage>1224</fpage>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          43.
          <string-name>
            <surname>Zhu</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tsou</surname>
            ,
            <given-names>B.K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhu</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>Multi-aspect opinion polling from textual reviews</article-title>
          .
          <source>In: Proceedings of the 18th ACM conference on Information and knowledge management</source>
          . pp.
          <volume>1799</volume>
          {
          <year>1802</year>
          . ACM (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>