<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Feature Enhanced Dual-GRU for Aspect-based Sentiment Analysis</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Meng Zhao</string-name>
          <email>zhao_meng@hrbeu.edu.cn</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jing Yang</string-name>
          <email>yangjing@hrbeu.edu.cn</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Shuo Wang</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jiaqi Liu</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Harbin Engineering University</institution>
          ,
          <addr-line>Harbin, Heilongjiang 150001</addr-line>
          ,
          <country country="CN">China</country>
        </aff>
      </contrib-group>
      <fpage>92</fpage>
      <lpage>101</lpage>
      <abstract>
        <p>Aspect-based sentiment analysis (ABSA) aims to predict the sentiment polarity with the different aspect terms or categories, which play an important role to guide the representation of context vector. Previous studies have used concatenation operation as a common means of information aggregation, which increase irrelevant noise and lose the dependence between the original features. In this paper, we propose a lightweight feature enhanced dual-GRU to selectively learn the feature relevance between aspect terms and context. The dual-GRU contains an extended aspect-related GRU and a position-related GRU to generate relevant information adaptively. Meanwhile, we construct a context-related GRU to enhance the dependency between aspect terms and context. Extensive experimental results demonstrate that the proposed model is reliable and effective in improving the performance of the two tasks of ABSA.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Aspect-based sentiment analysis</kwd>
        <kwd>GRU</kwd>
        <kwd>Attention networks</kwd>
        <kwd>Position Information</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>1. Introduction 1
embeddings, which is simple and effective but prone to context-irrelated noise from position
information.</p>
      <p>
        To address this problem, inspired by Xing et al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], we propose a feature enhanced dual-GRU
(FEGRU) approach relying on the internal gating mechanism to control the information of the hidden state
learned from input features. It is inappropriate to incorporate the extra information into the hidden state
of GRU[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] (or LSTM[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]). Consequently, we extended different GRU-based variants to guide the
extraction of necessary information, and applied attention mechanisms to highlight the important
sequence information on the dependencies of the sentences.
      </p>
      <p>The main contributions of our paper are as follows:
1. We extend different GRU variants, Aspect-related GRU and Position-related GRU, to
selectively embed the aspect and the position information.
2. We construct the Context-related GRU to enhance the dependency between aspect terms and
context.
3. We present a lightweight feature enhanced dual-GRU (FE-GRU) to guide such related
information to predict sentiment polarity.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Works</title>
      <p>
        Aspect based sentiment analysis [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] has received widespread concern in recent years. Many
approaches have emerged to deal with the classification of sentiments, among which deep-learning
network is current the mainstream approach for sentiment analysis. There are too many neural network
structures based on RNN, especially the LSTM and its variant GRU. Neural networks have the ability
to learn by themselves through their own features and be designed different structures for various tasks.
      </p>
      <p>
        Long Short-Term Memory is a modified version of recurrent neural networks to solve the vanishing
gradient problem of RNN. At the same time, attention mechanism [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ][
        <xref ref-type="bibr" rid="ref26">26</xref>
        ] can avoid the long-term
dependence experienced by the LSTM model. They have been explored in various NLP tasks and have
shown a good performance for NLP tasks. Wang et al. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] respectively concatenated the aspect
embeddings to the hidden state generated by LSTM and the input word embeddings, and utilized
attention mechanism to capture the key part of sentence. Tang et al. [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] used two LSTMs with target as
the end point to model the preceding and following contexts respectively and concatenated the last
hidden vectors of the two LSTMs. The attention mechanism has the ability to distinguish the importance
of each sequence of information and pays more attention to this particular sequence with the given
aspect. Chen et al. [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] leveraged multiple-attention mechanism to capture sentiment features separated
by a long distance through a Bi-LSTM. Huang et al. [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] introduced an attention-over-attention (AOA)
[
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] to generate mutual attentions not only from aspect-to-text but also text-to-aspect.
      </p>
      <p>
        Convolutional neural network (CNN) is also a means of extracting sentence features. It is better at
extracting local and position-invariant features than RNN. Xue et al. [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] proposed a gated convolutional
network with aspect embedding (GCAE) based on a CNN and gating mechanisms. Li et al. [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] used
bidirectional long short-term memory (Bi-LSTM) to produce the context information and designed a
target-specific transformation and context-preserving mechanism to learn integrated word
representation and target representation rather than directly concatenating them. Target-specific
Transformation Networks (TNet) finally adopt position-aware convolutional layer instead of vanilla
convolutional layer. Liu et al. [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ] proposed a novel neural network framework, namely the Gated
Alternate Neural Network (GANN), which was aimed to enhance the capability of model in capturing
long-distance dependency and modeling sequence information. Compared with traditional RNN, LSTM
and Memory Network [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] has the ability of long-term memory. For aspect-level sentiment analysis.
many improved models [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ][
        <xref ref-type="bibr" rid="ref17">17</xref>
        ][
        <xref ref-type="bibr" rid="ref18">18</xref>
        ][
        <xref ref-type="bibr" rid="ref28">28</xref>
        ] based on memory networks have emerged to fit the memory
of the features themselves.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Our Framework</title>
      <p>This section presents a novel feature enhanced dual-GRU model for ABSA. The architecture of the
proposed model is illustrated in Fig. 1.
3.1.</p>
    </sec>
    <sec id="sec-4">
      <title>Variants of GRU</title>
      <p>Given a sentence  the aspect (aspect terms or categories) contains m (m ＜ n) words  =
 ,  , … ,  , the embedding vectors of the given aspect  =  ,  , … ,  , and the embedding
vectors of a sentence  =  ,  , … ,  , the purpose of ABSA is to predict sentiment polarity  =
0,1,2,3 or  = 0,1,2 for sentence S, where 0, 1, 2 and 3 denote the “negative”, “neutral”, “positive”
and “conflict” sentiment polarities, respectively.</p>
    </sec>
    <sec id="sec-5">
      <title>3.1.1. Aspect-related GRU</title>
      <p>
        To reduce the noise from aspect-irrelevant, we design the Aspect- related GRU (A-GRU), which is
also a variant of the AA-LSTM [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Similarly, we add the aspect-reset gate and aspect-update gate to
control how much aspect information flows into the hidden state. Fig. 2 (a) illustrates the architecture
of the Aspect-related GRU. The core structure of A-GRU are the reset gate and the update gate of aspect
information and input information. Through the core gate structure, the aspect and input information
combine with the previous hidden state to output a new hidden state at each time step. Therefore, the
hidden state can carry aspect-related information throughout the processing of the time sequence.
      </p>
      <p>At time step t, the aspect-reset gate ra , reset gate rt and the candidate ha . are computed as follows:
where hta−1 denotes the previous hidden state, a is the aspect embedding vectors, σ is the activation
function. The candidate ha calculated by rt represents the new aspect-related input information at the
current moment. The aspect-update gate za , update gate zt and the new hidden state hta−1 are computed
as follows:
(1)
(2)
(3)
(4)
ra =σ (Wra[a; hta−1] + bra）
rt =σ (Wrt [xt ; hta−1] + ra  a + brt )
ha = tanh(Wha[rt * hta−1; xt ])
za =σ (Wza[a; hta−1] + bza )
where the aspect-update gate za combine input information with the aspect information to produce
the aspect-related context information, which may contain context-irrelated noise. The aspect-update
gate is used to control how much aspect-related information is brought into the update gate zt while the
update gate zt decides how much aspect-related input information to add and how much
aspectirrelated input information to throw away.</p>
      <p>where p is the position embedding vectors, which is invariable through the sequence chain. At time
step t, the position-reset gate rp is used to calculate how much sensitive position information flows into
the hidden state. At the next step, the position information is directly controlled by the update gate rt
(a): Aspect-related GRU
(b): Position-related GRU</p>
      <p>The extended aspect gate allows the GRU to forget certain irrelated parts of the aspect information
and keep previous semantic information simultaneously. We regard certain irrelated parts of the aspect
information and context information as noise, which is a factor influencing the result of predicting the
given aspect’s sentiment polarity. Our extended A-GRU can solve this problem by aspect gates to learn
to reduce the noise in the process of information transfer.</p>
    </sec>
    <sec id="sec-6">
      <title>3.1.2. Position-related GRU</title>
      <p>We construct Position-related GRU (P-GRU) to dynamically learn the position information of
aspects in each sentence. Fig. 2 (b) illustrates the architecture of the Position-related GRU. We add a
position gate to control the inflow of position information like aspect gates of A-GRU and leave out the
position update gate. P-GRU can be formalized as follows:</p>
      <p>rp = σ (Wrp[ p; htp−1] + brp )
rt = σ (Wrt [xt ; htp−1] + rp  p + bpt )</p>
      <p>zt =σ (Wzt [xt ; hta−1] + bzt )
hp = tanh(W a [rt * htp−1; xt ])</p>
      <p>h
htp = (1 − zt ) * htp−1 + zt * h p
(5)
(6)
(7)
(8)
(9)
(10)
(11)
of input information rather than the position update gate. rt is to ensure that the position information is
reserved for the aspect words during the learning process of the input information.</p>
    </sec>
    <sec id="sec-7">
      <title>3.1.3. Context-related GRU</title>
      <p>Compared to LSTM, GRU has fewer parameters and is easier to calculate when adding aspect and
position gates. We design Context-related GRU to take into account the context-related information of
the sentence, while also associating the aspect with the information of the sentence. Context-related
GRU (C-GRU), not alike A-GRU, add the aspect information directly to hidden state through gate
mechanism.</p>
      <p>Comparatively, C-GRU is a reverse A-GRU, which swaps the contextual sequence input and aspect
input, and adds control gates to ensure the original features of the aspect terms. Max pool layer selects
the maximum value in local features, by this means we can extract the important aspect-related
information. Thus, we introduce Maxout Layer to compress the aspect-aware information and add the
pristine aspect features, which can be obtained as the following:
c = Maxout(htc )</p>
      <p>c′ = β * a + c
where htc is the output of C-GRU and β is a trade-off parameter.
3.2.</p>
    </sec>
    <sec id="sec-8">
      <title>Information fusion</title>
      <p>The final loss function consists of two parts: the Euclidean Loss and the Cross-entropy Loss. For the
C-GRU, we iteratively minimize the squared Euclidean between the original aspect terms and the c′ ,
which is defined as:
n
d (c′ , a) =  (c′ − a)2</p>
      <p>i=1
For the ABSA task, we take the Cross-entropy Loss as the final loss function:</p>
      <p>
        For ABSA tasks, we explicitly set position index of each aspect words in the sentence to zero, and
define the relative distance to indicate the importance of each word relative to the aspect terms[
        <xref ref-type="bibr" rid="ref12">12</xref>
        ][
        <xref ref-type="bibr" rid="ref21">21</xref>
        ].
The final context-rich information is expressed as follows:
      </p>
      <p>H = GRU (X ) (14)
H a = A-GRU (H1, Aspect)
H p = P-GRU (H1, Position)</p>
      <p>H final = [H p ; H a ]</p>
      <p>We introduce the attention mechanisms to highlight the dependency between the rich contextual
features and context-related aspect information. The extent of concern is expressed by the weight of
words, which can be expressed as follows:</p>
      <p>The final layer we use a softmax layer to output the same number of nodes as the number of
sentiment class:</p>
      <p>ci = attention(h pa , c′ )
yi = softmax(W * ci + b)
where the softmax operator is used to obtain the probabilities yi of each class label, W and b are
the model parameters.
3.3.</p>
      <p>Objective function
(12)
(13)
(15)
(16)
(17)
(18)
(19)
(20)</p>
      <p>N
 = − ( yilog( yi )) + d
i=0
(21)
where yi is the predicted sentiment distribution of each class. yi is the true sentiment polarity, N
is the number of all training sentences.</p>
    </sec>
    <sec id="sec-9">
      <title>4. Experiments</title>
      <p>
        In this section, we will introduce four datasets to verify the effectiveness of our model and provide
details of parameter settings. In addition, we extend the ATAE [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] and AOA [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] model and design
different variants of FE-GRU to evaluate an ablation study.
4.1.
      </p>
    </sec>
    <sec id="sec-10">
      <title>Datasets</title>
      <p>
        We experiment on widely used datasets of SemEval: Restaurant 2014–2016 and Laptop 2014[
        <xref ref-type="bibr" rid="ref22">22</xref>
        ]
for ABSA task. Datasets of SemEval consist of laptop and restaurant reviews. We retain the reviews
with sentiment polarity of “conflict” and divide the reviews into four sentiment polarities: positive,
neutral, negative, and conflict because it is unreasonable to set conflict reviews as positive or negative
sentiment polarities or even remove the conflict reviews[
        <xref ref-type="bibr" rid="ref23">23</xref>
        ]. The details of the datasets for ABSA task
are shown in Table 1.
      </p>
      <p>
        In our experiments, the pre-trained Glove [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] embedding is adopted as word embeddings in the
datasets and the dimensions of word embeddings, aspect embeddings and position embeddings are set
to 300, the maximum sequence length of a sentence is set to 83. For the out of vocabulary words and
weight matrices, we initialize using a uniform distribution U(-0.25,0.25). Note that we remove the
PGRU to assess the influence of position information in the ACSA task, because the aspect category
does not appear in the sentence. The batch size is set to {16,32} and the learning rate is set to
{0.003,0.007}.
4.2.
      </p>
    </sec>
    <sec id="sec-11">
      <title>Variants of FE-GRU</title>
      <p>We utilize some baseline approaches to evaluate the effectiveness of the proposed FE-GRU model,
and compare different variants of GRU to verify the effectiveness of the aspect gates and the position
gates.</p>
      <p>•</p>
      <p>A-GRU is a particularly simple model to control the flow of aspect information via the aspect
gates.</p>
      <p>PA-GRU introduces the position gates and combine the hidden states of P-GRU and A-GRU to
output the final sentiment class.</p>
      <p>ATAE_A replaces the hidden state of the original LSMT with the state of aspect information
based on ATAE-LSTM.</p>
      <p>ATAE_PA adds the position gates to control the extent of position information flowing into the
hidden state based on ATAE_A.</p>
      <p>AOA_A uses the A-GRU instead of the original LSTM and outputs the hidden state with
aspectrelated information.</p>
      <p>AOA_PA concatenates the hidden state of A-GRU through the position-related information
generated by P-GRU.
w/o_GRU removes the A-GRU and P-GRU cells and use the word embeddings directly as input
to the model.
w/o_P-GRU removes the A-GRU of our FE-GRU.
w/o_A-GRU removes the P-GRU of our FE-GRU.</p>
      <p>w/o_C-GRU removes the C-GRU of our FE-GRU.
4.3.</p>
    </sec>
    <sec id="sec-12">
      <title>Results and analyses</title>
      <p>The results of experiment comparing the baseline model are shown in Table 2 and Table 3. It's
obviously that the proposed FE-GRU has a better result than the baseline models, because our model
has the ability to selectively learn the dependency information of the aspect terms. ATAE_LSTM and
GCAE in the baseline models have a good classification accuracy partly because the aspect information
is directly used to concatenate the context information. It's a very common choice for various existing
models but lacks the ability to distinguish the irrelevant information. AA_LSTM is designed to influent
the information flow instead of integrating the aspect information into the hidden state vectors.
ATAE_AA_LSTM and IAN_AA_LSTM based on AA_LSTM effectively retain the aspect-related
information compared to the original model. However, AA_LSTM introduces three gates in the LSTM
increasing the training parameters and making the training process more difficult.</p>
      <p>It is proved that the optimized gates can control the extent of information flowing into the hidden
state. It is also not appropriate to combine position information directly into context in the existing
models. In addition, the designed C-GRU associates aspect information into contexts as the clue to
guide aspect features, which aims to enhance contextual dependencies.</p>
      <p>F1
45.42</p>
      <p>We chose the simple A-GRU and PA-GRU models to compare to prove the significant contribution
of position information in sentiment classification, the experimental results are shown in Table 4. The
results of the comparative experiments show that P-GRU is sensitive to the specific position of aspect
terms in the sentence, which is learned and determined by P-GRU.</p>
      <p>In addition, the experimental effect of the ATAE_LSTM model with position information is much
better than the original model, which concatenates the aspect embedding and the hidden outputs of
LSTM to generate attention weights. We change the hidden outputs of LSTM with the aspect-related
context information and concatenate the position-related context information to calculate the weight of
each sequence. The only difference between ATAE_A and ATAE_PA is that the latter has the ability
to learn position information, and the experiment of two groups of ATAE demonstrate the reliability of
learning position information through gate mechanisms. We use the hidden state of PA-GRU to replace
the original LSTM output, which implies that the position-related information and the aspect-related
information of the AOA model can improve the effect of sentiment classification.</p>
    </sec>
    <sec id="sec-13">
      <title>5. Conclusions</title>
      <p>In this paper, we have proposed a feature enhanced dual-GRU (FE-GRU) for ABSA task. The
purpose of the proposed FE-GRU is to improve the validity and correlation of the context and the aspect
information as much as possible. We designed different feature embedding strategies in the lightweight
GRU to associate and enhance the specific information dependencies, which is a novel approach to
aggregating information instead of the single concatenation operation. We conducted massive
experiments with extensive models on the ACSA and ATSA tasks and achieved significant
improvements on most datasets. The experimental results have proved the rationality and effectiveness
of the proposed approach.</p>
    </sec>
    <sec id="sec-14">
      <title>6. Acknowledgements</title>
      <p>This paper is supported by the National Natural Science Foundation of China under Grant
nos.61672179, 61370083.</p>
    </sec>
    <sec id="sec-15">
      <title>7. References</title>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.</given-names>
            <surname>Pennington</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Socher</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C.</given-names>
            <surname>Manning</surname>
          </string-name>
          , “Glove:
          <article-title>Global vectors for word representation</article-title>
          ,”
          <source>in Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)</source>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>S.</given-names>
            <surname>Hochreiter</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Schmidhuber</surname>
          </string-name>
          ,
          <article-title>“Long short-term memory,” Neural Comput</article-title>
          ., vol.
          <volume>9</volume>
          , no.
          <issue>8</issue>
          , pp.
          <fpage>1735</fpage>
          -
          <lpage>1780</lpage>
          ,
          <year>1997</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J.</given-names>
            <surname>Chung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Gulcehre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Cho</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bengio</surname>
          </string-name>
          , “
          <article-title>Empirical evaluation of gated recurrent neural networks on sequence modeling,” arXiv [cs</article-title>
          .NE],
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>B.</given-names>
            <surname>Xing</surname>
          </string-name>
          et al., “
          <article-title>Earlier attention? Aspect-aware LSTM for aspect-based sentiment analysis</article-title>
          ,
          <source>” in Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>B.</given-names>
            <surname>Keith Norambuena</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. F.</given-names>
            <surname>Lettura</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C. M.</given-names>
            <surname>Villegas</surname>
          </string-name>
          , “
          <article-title>Sentiment analysis and opinion mining applied to scientific paper reviews,” Intell. Data Anal.</article-title>
          , vol.
          <volume>23</volume>
          , no.
          <issue>1</issue>
          , pp.
          <fpage>191</fpage>
          -
          <lpage>214</lpage>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>M.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhao</surname>
          </string-name>
          , and
          <string-name>
            <given-names>W.</given-names>
            <surname>Zhou</surname>
          </string-name>
          , “
          <article-title>Hierarchical human-like strategy for aspect-level sentiment classification with sentiment linguistic knowledge and reinforcement learning,” Neural Netw</article-title>
          ., vol.
          <volume>117</volume>
          , pp.
          <fpage>240</fpage>
          -
          <lpage>248</lpage>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhu</surname>
          </string-name>
          , and
          <string-name>
            <given-names>L</given-names>
            .
            <surname>Zhao</surname>
          </string-name>
          , “
          <article-title>Attention-based LSTM for Aspect-level Sentiment Classification,”</article-title>
          <source>in Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing</source>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>D.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Qin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Feng</surname>
          </string-name>
          , and T. Liu, “
          <article-title>Effective LSTMs for Target-Dependent Sentiment Classification,” arXiv [cs</article-title>
          .CL],
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>W.</given-names>
            <surname>Xue</surname>
          </string-name>
          and
          <string-name>
            <given-names>T.</given-names>
            <surname>Li</surname>
          </string-name>
          , “
          <article-title>Aspect based sentiment analysis with gated convolutional networks,” in Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics</article-title>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>P.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Bing</surname>
          </string-name>
          , and
          <string-name>
            <given-names>W.</given-names>
            <surname>Yang</surname>
          </string-name>
          , “
          <article-title>Recurrent attention network on memory for aspect sentiment analysis</article-title>
          ,
          <source>” in Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>B.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Ou</surname>
          </string-name>
          , and
          <string-name>
            <surname>K. M. Carley</surname>
          </string-name>
          , “
          <article-title>Aspect level sentiment classification with attention-overattention neural networks</article-title>
          ,” in Social, Cultural, and Behavioral Modeling, Cham: Springer International Publishing,
          <year>2018</year>
          , pp.
          <fpage>197</fpage>
          -
          <lpage>206</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>N.</given-names>
            <surname>Liu</surname>
          </string-name>
          and
          <string-name>
            <given-names>B.</given-names>
            <surname>Shen</surname>
          </string-name>
          , “
          <article-title>Aspect-based sentiment analysis with gated alternate neural network</article-title>
          ,
          <source>” Knowl. Based Syst.</source>
          , vol.
          <volume>188</volume>
          , no.
          <issue>105010</issue>
          , p.
          <fpage>105010</fpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Cui</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          , T. Liu, and G. Hu, “
          <article-title>Attention-over-attention neural networks for reading comprehension</article-title>
          ,”
          <source>in Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>X.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Bing</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Lam</surname>
          </string-name>
          , and
          <string-name>
            <given-names>B.</given-names>
            <surname>Shi</surname>
          </string-name>
          , “
          <article-title>Transformation networks for target-oriented sentiment classification,” in Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics</article-title>
          (Volume
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>D.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Qin</surname>
          </string-name>
          , and T. Liu, “
          <article-title>Aspect level sentiment classification with deep memory network,”</article-title>
          <source>in Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing</source>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zou</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C.</given-names>
            <surname>Gan</surname>
          </string-name>
          , “
          <article-title>The optimally designed dynamic memory networks for targeted sentiment classification</article-title>
          ,
          <source>” Neurocomputing</source>
          , vol.
          <volume>309</volume>
          , pp.
          <fpage>36</fpage>
          -
          <lpage>45</lpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mazumder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zhou</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chang</surname>
          </string-name>
          , “
          <article-title>Target-sensitive memory networks for aspect sentiment classification,” in Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics</article-title>
          (Volume
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>N.</given-names>
            <surname>Liu</surname>
          </string-name>
          and
          <string-name>
            <given-names>B.</given-names>
            <surname>Shen</surname>
          </string-name>
          , “
          <article-title>ReMemNN: A novel memory neural network for powerful interaction in aspect-based sentiment analysis</article-title>
          ,
          <source>” Neurocomputing</source>
          , vol.
          <volume>395</volume>
          , pp.
          <fpage>66</fpage>
          -
          <lpage>77</lpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>D.</given-names>
            <surname>Ma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , and
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          , “
          <article-title>Interactive Attention Networks for Aspect-Level Sentiment Classification,”</article-title>
          <source>in Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>L.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Zhou</surname>
          </string-name>
          , “
          <article-title>Hierarchical attention based position-aware network for aspect-level sentiment analysis</article-title>
          ,
          <source>” in Proceedings of the 22nd Conference on Computational Natural Language Learning</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. X.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q. V.</given-names>
            <surname>Hu</surname>
          </string-name>
          , and
          <string-name>
            <given-names>L.</given-names>
            <surname>He</surname>
          </string-name>
          , “
          <article-title>Position-aware hierarchical transfer model for aspect-level sentiment classification,” Inf</article-title>
          . Sci. (Ny), vol.
          <volume>513</volume>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>16</lpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>M.</given-names>
            <surname>Pontiki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Galanis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Pavlopoulos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Papageorgiou</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Androutsopoulos</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Manandhar</surname>
          </string-name>
          , “
          <fpage>SemEval</fpage>
          -2014
          <source>Task 4: Aspect Based Sentiment Analysis,” in Proceedings of the 8th International Workshop on Semantic Evaluation (SemEval</source>
          <year>2014</year>
          ),
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>X.</given-names>
            <surname>Tan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Cai</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C.</given-names>
            <surname>Zhu</surname>
          </string-name>
          , “
          <article-title>Recognizing conflict opinions in aspect-level sentiment classification with dual attention networks</article-title>
          ,
          <source>” in Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>L.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Lu</surname>
          </string-name>
          , and L. Bing, “
          <article-title>Position-aware tagging for aspect sentiment triplet extraction</article-title>
          ,”
          <source>in Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>G.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , T. Zhang,
          <string-name>
            <given-names>S.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Meng</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Chen</surname>
          </string-name>
          , “
          <article-title>Aspect-level sentiment classification based on attention-BiLSTM model and transfer learning</article-title>
          ,
          <source>” Knowl. Based Syst.</source>
          , vol.
          <volume>245</volume>
          , no.
          <issue>108586</issue>
          , p.
          <fpage>108586</fpage>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhou</surname>
          </string-name>
          and
          <string-name>
            <given-names>F.</given-names>
            <surname>Liu</surname>
          </string-name>
          , “
          <article-title>Filter gate network based on multi-head attention for aspect-level sentiment classification</article-title>
          ,
          <source>” Neurocomputing</source>
          , vol.
          <volume>441</volume>
          , pp.
          <fpage>214</fpage>
          -
          <lpage>225</lpage>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Rao</surname>
          </string-name>
          , and
          <string-name>
            <given-names>X.</given-names>
            <surname>Quan</surname>
          </string-name>
          , “GSMNet:
          <article-title>Global semantic memory network for aspect-level sentiment classification,” IEEE Intell</article-title>
          . Syst., vol.
          <volume>36</volume>
          , no.
          <issue>5</issue>
          , pp.
          <fpage>122</fpage>
          -
          <lpage>130</lpage>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>P.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Yang</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Lai</surname>
          </string-name>
          , “
          <article-title>Deep selective memory network with selective attention and interaspect modeling for aspect level sentiment classification,” IEEE ACM Trans</article-title>
          .
          <article-title>Audio Speech Lang</article-title>
          . Process., vol.
          <volume>29</volume>
          , pp.
          <fpage>1093</fpage>
          -
          <lpage>1106</lpage>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>B.</given-names>
            <surname>Huang</surname>
          </string-name>
          et al., “
          <article-title>Aspect-level sentiment analysis with aspect-specific context position information</article-title>
          ,
          <source>” Knowl. Based Syst.</source>
          , vol.
          <volume>243</volume>
          , no.
          <issue>108473</issue>
          , p.
          <fpage>108473</fpage>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>D.</given-names>
            <surname>Shao</surname>
          </string-name>
          et al., “
          <article-title>Aspect-level sentiment analysis for based on joint aspect and position hierarchy attention mechanism network,”</article-title>
          <string-name>
            <given-names>J.</given-names>
            <surname>Intell</surname>
          </string-name>
          . Fuzzy Syst., vol.
          <volume>42</volume>
          , no.
          <issue>3</issue>
          , pp.
          <fpage>2207</fpage>
          -
          <lpage>2218</lpage>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>