<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Italian Conference on Big Data and Data Science, September</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Detection using Multi-View Deep Learning Combining Content and Behavioral Features</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Giuseppina Andresini</string-name>
          <email>giuseppina.andresini@uniba.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andrea Iovine</string-name>
          <email>andrea.iovine@uniba.it</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Roberto Gasbarro</string-name>
          <email>r.gasbarro1@studenti.uniba.it</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Marco Lomolino</string-name>
          <email>m.lomolino@studenti.uniba.it</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Marco de Gemmis</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Annalisa Appice</string-name>
          <email>annalisa.appice@uniba.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Consorzio Interuniversitario Nazionale per l'Informatica - CINI</institution>
          ,
          <addr-line>Bari</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Computer Science, University of Bari Aldo Moro</institution>
          ,
          <addr-line>Bary</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2022</year>
      </pub-date>
      <volume>2</volume>
      <fpage>0</fpage>
      <lpage>21</lpage>
      <abstract>
        <p>Nowadays, online reviews are the main source to customer opinions. They are especially important in the realm of e-commerce, where reviews regarding products and services influence the purchase decisions of customers, as well as the reputation of the commerce websites. Unfortunately, not all the online reviews are truthful and trustworthy. Therefore, it is crucial to develop machine learning techniques to detect review spam. This study describes EUPHORIA - a novel classification approach to distinguish spam from truthful reviews. This approach couples multi-view learning to deep learning, in order to gain accuracy by accounting for the variety of information possibly associated with both the reviews' content and the reviewers' behavior. Experiments carried out on two real review datasets from Yelp.com - Hotel and Restaurant - show that the use of multi-view learning can improve the performance of a deep learning classifier trained for review spam detection.</p>
      </abstract>
      <kwd-group>
        <kwd>review spam detection</kwd>
        <kwd>deep learning</kwd>
        <kwd>multi-view learning</kwd>
        <kwd>word embedding</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Content and Behavioral Features</p>
    </sec>
    <sec id="sec-2">
      <title>1. Introduction</title>
      <p>
        In the last two decades, the widespread difusion of customer reviews has raised the risk of review
spam attacks towards several commerce websites (e.g., Amazon.com, Yelp.com). Customer
reviews are user-contributed consumer opinions posted to commerce websites and originated
from the users’ experiences regarding specific products or services. They represent the most
valuable source of information that can be used to determine the public opinion on the reviewed
products or services. In fact, customer reviews are one of the primary factors in a customer’s
decision to purchase a product or service [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Furthermore, there are increasing eforts to
incorporate the rich information embedded in reviews into the process of user modeling and
recommendation generation or justification [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ]. On the other hand, as anyone can easily
produce opinions and post fake reviews, i.e., spam reviews, to social media with no constraints,
certain product vendors or service providers may abuse this situation to promote their products
and services, or to criticize their competitors unfairly. Due to the real risk of review spam
attacks, developing efective review spam detection approaches is a crucial task to secure the
reliability of online opinions. Today, distinguishing spam reviews from truthful (non-spam)
reviews is a challenging task that attracts growing attention in the machine learning community,
since it is dificult, if not impossible, to recognize fake reviews by manually reading their content
[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>
        Although content features are widely investigated in the review spam detection literature,
several studies assess that a purely content-based approach is not suficient to train a review
spam classifier with adequate accuracy performance [
        <xref ref-type="bibr" rid="ref4 ref5">5, 4</xref>
        ]. The emerging research trend is to
improve the accuracy of review spam classifiers by taking additional features into account. These
features include the review’s post date/time, the behavior associated with the review’s writer,
and the deviation from other reviews concerning the same product or service [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Coupling
review content features with reviewer behavioral features has proven to be more efective than
using each type of feature alone [
        <xref ref-type="bibr" rid="ref4 ref5">5, 4</xref>
        ].
      </p>
      <p>
        In this paper, we investigate the efects of jointly performing supervised learning on both
content and behavioral features for the task of review spam detection. We define a novel review
classification approach, named EUPHORIA (nEural mUlti-view aPproacH fOr RevIew spAm)
that couples deep learning with multi-view learning, in order to elaborate knowledge pertaining
to both the review content and the reviewer behavior. With regard to the feature extraction
technique, we employ word embedding models used to derive a feature vector representation of
the review content. This step is performed to capture complex global semantic information that
is hidden in reviews and dificult to express using traditional bag-of-words features. In particular,
we explore the performance of two word embedding models of the review content, namely
Word2Vec [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] and BERT [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], and a temporal-aware representation of the reviewer behavior.
In addition, we identify a set of behavioral reviewer features, based on the related literature,
that can aid in enhancing the performance of the review content-based classification. Another
distinctive characteristic of the proposed solution is that, with regard to the classification
technique, we adopt a combination of deep learning and multi-view learning. In fact, we handle
both content and behavioral features as multiple views of the same review corpus and propose
a deep neural network architecture to learn an accurate classification model to distinguish spam
reviews from non-spam reviews. Specifically, we train a multi-input neural network that is
able to share knowledge among review content-based and reviewer behavior-based views. This
architecture allows us to gain in classification performance. The original contributions of this
work is reported in [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
      </p>
      <p>The paper is organized as follows. Section 2 illustrates the related work. Section 3 presents
the proposed EUPHORIA approach. Section 4 describes the data collections processed in
the experiments, the experimental setting and the relevant results. Finally, Section 5 draws
conclusions and outlines the future directions of this work.</p>
    </sec>
    <sec id="sec-3">
      <title>2. Related work</title>
      <p>
        Several supervised approaches are formulated in the machine learning literature to distinguish
spam reviews from non-spam reviews [
        <xref ref-type="bibr" rid="ref4">4, 10</xref>
        ]. The seminal studies in [11, 12] started the
investigation of the review spam detection task in the context of product reviews. Starting
from these studies, many researchers have investigated the performance of several feature
representation models adopted to describe structural properties of review content or delineate
behavioral characteristics of reviewers. In particular, state-of-the-art approaches for review
spam detection can be categorized into two main groups: methods based on features created
from the review content and methods based on features created on the reviewer’s behavior
[13]. Content-based approaches extract features based on the review text. The majority of
content-based approaches adopt the traditional bag-of-words model, which represents text
as sets of words, ignoring their order. However, recent studies have started to explore word
embedding models that allow us to overcome the limitations of the bag-of-words model by
capturing complex global information contained in the text. For example, TopicSpam [14]
uses the Latent Dirichlet Allocation (LDA) algorithm to identify slight diferences between
the distribution of the keywords in the spam and non-spam reviews. More recently, large
pre-trained models have been employed for this task. An example of such model is BERT [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], a
powerful Transformer-based encoder model trained to generate a bidirectional representation of
text. BERT has already been proven to be successful in a variety of NLP tasks. The application
of BERT on the review spam detection task has been studied in [15], which aims to capture the
semantic relevance in the review’s sentences. The experimental results reported in this study
show that BERT can generate a text representation with richer content information compared
to traditional text representation approaches based on bag-of-words/n-gram features.
      </p>
      <p>
        Behavior-based approaches focus on studying the behavior of the reviewers, rather than
analyzing their reviews as individual units [13]. Starting from the seminal study of [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], various
behavioral features (e.g., maximum number of reviews per day, percentage of positive reviews,
total number of reviews for a reviewer, content similarity) have been studied to improve the
classification performance achieved with content features. A recently emerging research trend
has combined the two approaches, improving the performance of classifiers trained to detect
spam reviews [11, 13, 16, 17]. For example, in [16], both content features (e.g., uni-gram,
bigram, similarity scores with other reviews) and behavioral features (e.g., authority score of a
reviewer, rating deviation score) are analyzed for the review classification. Classification with
both behavioral features and content features has been recently investigated also in [13] and
[17].
      </p>
      <p>Finally, few recent studies have shown the superiority of deep learning approaches compared
to traditional classification methods in several problems of review spam detection [ 18, 19, 20].
In [19], a Recurrent Neural Network with Attention mechanism (GRNN) is trained to capture
the non-local information over sentence vectors. In [18], a deep learning-based approach is used
to learn a document-level representation to identify spam reviews. In particular, the approach
combines Word2Vec and Convolutional Neural Networks in order to learn the representation of
the reviews, and extract higher-level n-gram features of the review content. A Bi-directional
LSTM is finally used for the review classification. The work in [ 20] uses word embeddings trained
on a large Amazon review dataset using the Continuous Bag-of-Words (CBOW) algorithm,
and trains a model that combines Convolutional Neural Networks and Gated Recurrent Neural
Networks.</p>
      <p>Even though recent literature in review spam detection strongly suggests that the analysis
of the reviewer behavior covers a crucial role in improving review classification performance,
prior studies exploring deep learning methods for review spam detection are mostly limited to
One of the best hotels I've
staid at. Cool rooms and
awesome lakefront view.</p>
      <p>Threw a New Years party in
one of their suits.</p>
      <sec id="sec-3-1">
        <title>Review</title>
      </sec>
      <sec id="sec-3-2">
        <title>Reviewer</title>
        <sec id="sec-3-2-1">
          <title>Word2Vec</title>
        </sec>
        <sec id="sec-3-2-2">
          <title>Behavior</title>
        </sec>
        <sec id="sec-3-2-3">
          <title>Profiling</title>
        </sec>
      </sec>
      <sec id="sec-3-3">
        <title>Feature</title>
        <p>extraction
tr
o
c
e
V
g
n
i
d
d
e
b
m
E
tr
o
c
e
V
g
n
i
d
d
e
b
m
E
s
e
tr
u
a
e
F
l
a
r
o
i
v
a
h
e
B
rye tu
La po
FC roD
rye tu
La op
FC roD
rye tu
La po
FC roD
e
ta re
tcnnae yaLC
o F
C</p>
      </sec>
      <sec id="sec-3-4">
        <title>Multi Input NN</title>
        <p>processing the review content.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>3. Proposed method</title>
      <p>EUPHORIA uses the multi-view learning approach in order to detect fake reviews more
effectively, by exploiting multiple (overlapping) views of the data. In our approach, each data
view is represented as a distinct feature vector, which is then processed by the multi-input
neural network. Currently, the model supports three diferent data views, two related to textual
features (Word2Vec and BERT), and one related to behavioral features. Both Word2Vec and
BERT are used to extract text-centric features from text. The two techniques represent to
diferent views of the textual data, which focus on diferent aspects of the review text</p>
      <p>Word2Vec is one of the first attempts to learn dense and continuous textual representations in
the form of a numerical vector, which describes the meaning of a particular word. The meaning
of a specific word is derived from the context that the word appears in, i.e., the words that
commonly co-occur with it. To represent a sentence or a paragraph, word embeddings are
combined through an aggregation function, where the representation of a review is obtained
by averaging the word embeddings of all tokens contained in that review. For the purposes of
this study, we decided to train a new set of word embeddings using Continuous Bag-of-Words
(CBOW) algorithm. In the CBOW algorithm, a feed-forward single-layer neural network is
trained to predict a specific word, given a set of history words (which precede the word to be
predicted) and a set of future words (which follow the word to be predicted). The size of the
context window, i.e., the number of history and future words given as input is considered as a
hyperparameter With these embeddings, we are able to obtain a more domain-focused view of
the text data, which we can use to identify patterns that are specific to the review domain.</p>
      <p>
        It is important to note that while the meaning of each word depends on its context
(cooccurring words), the embedding of a specific word computed with Word2Vec does not change
depending on the sentence the word appears in. This can cause problems with polysemous
words, which have multiple meanings. Contextualized word embeddings may overcome this issue
by adding the ability to produce diferent vector representations for the same word, based on
the context of the entire sentence. BERT [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], that stands for Bidirectional Encoder Representation
from Transformers, is a state-of-the-art pre-trained model for generating contextual
representations of text. In particular, BERT is based on the bidirectional Transformer architecture [21],
which features an encoder and a decoder with several multi-head attention layers. Due to its
bidirectional and contextual nature, BERT can encode the meaning of an entire
sentence/paragraph in a single vector. The vector representation of a sentence was constructed by using the
ifnal hidden state of the special classification (CLS) token, which is the common approach for
representing sequences of text for classification purposes with BERT [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Finally, the vector
representation of the entire review was obtained by averaging the vectors of each sentence.
      </p>
      <p>Finally, we calculated six behavioral features, which represent the profile of the reviewer over
the course of time. These features are reviewer-centric, as they consider all the reviews written
by the reviewer in the specified time range.</p>
      <p>The reviewer profile includes the following behavioral dimensions:
• Maximum Number of Reviews per Day (MRD), that measures the maximum number of
reviews that the reviewer  has written in a single day up to the current time  . A reviewer
who writes a large amount of reviews in a single day can thus be a potential threat.
• Positive Ratio (PR), that is the percentage of positive reviews written by a reviewer up to
the time  . Specifically, we consider a review as positive if its rating is higher than 3 out of
5.
• Average Review Length (ARL), that is the average length of the reviews (number of words)
written by  up to the time  . It has been noted that fake reviews tend to be generally
shorter than legitimate ones.
• Reviewer Deviation (RD) that is defined as the absolute diference between the average
rating obtained by the item, and the rating that  has assigned to the same product. In
fact, it has been observed that the ratings of fake reviews tend to deviate much more
frequently from the average rating.
• Average Review Similarity (ARS), that measures the similarity of all reviews written by 
up to the time  . The idea is to detect reviewers who use templates to write large amounts
of reviews, which will be similar in both syntax and meaning.
• Maximum Review Similarity (MRS), that measures the maximum similarity between
the current review  and the reviews previously written by the same reviewer  . The
similarity is again calculated as the cosine between the Word2Vec feature vectors of the
two reviews. Similarly to ARS, MRS can be used to detect reviewers that use standard
templates to generate fake reviews.</p>
    </sec>
    <sec id="sec-5">
      <title>4. Empirical evaluation and discussion</title>
      <p>
        We evaluated the accuracy performance of the proposed approach by performing several
experiments on two benchmark review datasets. These experiments aimed to investigate the
achievements of the multi-view learning approach, as well as the accuracy performance of
the deep learning architecture. The datasets are presented in Section 4.1. The implementation
details of the trained neural network architecture are reported in Section 4.2. The experimental
setting is described in Section 4.3, while the results are discussed in Section 4.4.
4.1. Data
Obtaining gold standard datasets for detecting spam review is a challenging problem. Due to the
large amount of reviews online, manual labeling for ground truth reviews is complex and costly.
In our study, we consider two datasets from Yelp.com, which are first used in [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Yelp.com is a
well-known large-scale online review site that provides labeled datasets for the evaluation of
review spam detection approaches. The two datasets, namely Hotel and Restaurant, contain
reviews across 85 hotels and 130 restaurants, respectively, in the Chicago area. T Both datasets
collect reviews for the target services of this experimental study, i.e., hotels and restaurants, as
well as reviews obtained from the reviewers’ profile pages for any product or service that they
wrote a review for. In this study, we classify reviews of hotels and restaurants separately, as
any domain has specific characteristics to take into account for review spam analysis.
      </p>
      <p>However, we also investigate the performance that we can achieve by integrating the user
profile with reviews coming from other domains. Dataset statistics are reported in Table 1. The
class distribution is imbalanced in both datasets.
4.2. Implementation details
EUPHORIA was implemented in Python 3. In particular, the multi-input neural network
architecture was realized using the high-level neural network API Keras 2.4 with TensorFlow as
back-end. For each dataset, we performed automatic hyper-parameter optimization using the
tree-structured Parzen estimator algorithm, as implemented in the Hyperopt library.1
Hyperparameter optimization was performed by using 20% of the entire training as a validation set.
We selected the configuration of the hyper-parameters that achieved the lowest validation set
loss. The hyper-parameter search space is reported in Table 2. The standard Rectified Linear
Unit (ReLU) was selected as the activation function for each hidden layer, while for the last
layer the softmax activation function has been used. The neural network was trained with
mini-batches by back-propagation, while the gradient-based optimization was performed using
the Adam update rule</p>
      <p>To preprocess of the text of each review, we adopted the Natural Language Toolkit (NLTK),
a suite of libraries for natural language processing for English language 2. The Word2Vec
algorithm adopted in the proposed method is included in the Gensim library 3. Finally, we
used the existing implementation of BERT ofered by the Transformers library 4, which was
pre-trained on a large corpus derived from the Toronto Book Corpus and Wikipedia.
4.3. Experimental setting
For each dataset, reviews were sorted according to their post date/time. The first ordered 80%
of reviews were used as training set, while the remaining 20% of the reviews were used as
testing set. The classification models were learned on each training set, and their accuracy was
evaluated on the corresponding testing set.</p>
      <p>Notice that the experimental setting adopted in this study is diferent from the one commonly
adopted in the review spam detection literature, where training-testing splits are generated
randomly, therefore neglecting the post date/time of reviews. We believe that ignoring the
temporal ordering of the reviews might reduce the fairness of the evaluation phase, because
at test time the model may make use of information that was produced after the review had
been written, which would not normally be possible in a realistic scenario. Therefore, in the
validation and test sets, we update the values of the behavioral features for each review, based
on the order in which they were written. Really, at testing time the model can only access
information that was collected up to the post date/time of the current review. This setup also
correctly captures the behavior of reviewers, which naturally changes over the time as they
1https://github.com/hyperopt/hyperopt
2https://www.nltk.org/
3https://radimrehurek.com/gensim/models/word2vec.html
4https://huggingface.co/docs/transformers/model_doc/bert
post new reviews. We measured F-score, G-mean and AUC-ROC to evaluate the accuracy
performance of the compared approaches.These metrics were measured by considering the class
“spam” as the positive class of the classification.
4.4. Results and discussion
The empirical validation was done to answer the following questions:
• To what extent each individual view influences the accuracy of the classification model?
(Section 4.4.1)
• Is a multi-input network more powerful than a single-input network? (Section 4.4.2)
• Is the analysis of reviews written on any domain helpful in improving the accuracy of
the spam classifier learned in the domain of a specific product or service? (Section 4.4.3)
• How does the multi-input neural network compare to recent review spam detection
methods that use traditional classification algorithms? (Section 4.4.4)
4.4.1. Data view analysis (Q1)
We analyzed how the knowledge enclosed in both the content and behavioral views can influence
the performance of EUPHORIA . To this aim, we performed an ablation study, where we
measured the accuracy of the following baselines of EUPHORIA :
• Word2Vec that elaborated content features extracted through Word2Vec.
• BERT that elaborated content features extracted through BERT.
• Behav that elaborated Behavioral features.
• Word2Vec + Behav that elaborated both Word2Vec features and Behavioral features.
• Word2Vec + BERT that elaborated both Word2Vec and BERT features.
• EUPHORIA (Word2Vec + BERT + Behav) that elaborated both Word2Vec and BERT
features, as well as Behavioral features.</p>
      <p>The results collected in Table 3 show that the behavioral features convey the most relevant
information to detect fake reviews: Behav outperforms Word2Vec, BERT in both datasets. On
the other hand, content features extracted with Word2Vec and BERT disclose non-redundant
knowledge on the review text. In fact, the trained neural networks gained accuracy when the
content features extracted through both Word2Vec and BERT were processed jointly instead than
separately (Word2Vec + BERT outperforms both Word2Vec and BERT). Finally, the main outcome
of this experiment is that the richer the processed information, the higher the learning ability
of the trained the accuracy of the learned multi-input deep neural network is actually improved
when both content and behavioral knowledge are taken into account. In fact, EUPHORIA
achieved the highest accuracy in both datasets.
4.4.2. Single-input versus multi-input analysis (Q2)
We continue the study of the efectiveness of the multi-view learning schema of EUPHORIA
by investigating the improvement achieved by learning a “multi-input” neural network. To
this purpose, we considered a single-input baseline (denoted as Single NN) of EUPHORIA . For
Single NN, we first computed the feature vectors on the three distinct views (i.e., Word2Vec,
BERT and Behav). Then we concatenated the three feature vectors in a single input vector.
Finally, we processed the concatenated data as input of a single-input neural network trained to
learn the classification model.</p>
      <p>Table 4 reports the F-score, G-mean and AUC-ROC of both Single NN and EUPHORIA. These
results show that the multi-input neural network of EUPHORIA can actually take advantage of
the richness of multi-view data. In fact, they confirmed that processing data of the individual
views though a multi-input architecture is more powerful than pre-concatenating these multiple
inputs and training a single-input neural network, whose performances are worsened by the
efects of the curse of dimensionality.
4.4.3. Multi-domain review analysis (Q3)
In the proposed approach, the classifier is trained to detect fake reviews written for a specific
domain of products or services (e.g., hotels or restaurants). However, the behavioral features can
be computed over time by accounting for reviews written by the same reviewer on a multitude
of products. To evaluate the efectiveness of this choice, we formulated the counterpart (denoted
as EUPHORIA(B)) of EUPHORIA, which computed behavioral features by only using reviews
written in the target domain (i.e., only restaurant reviews in the Restaurant dataset, and only
hotel reviews in the Hotel dataset).</p>
      <p>The results in Table 5 confirmed that the more the knowledge take into account to represent
the behavior of a reviewer, the higher the accuracy gained leveraging this knowledge to detect
spamming phenomena.
4.4.4. Deep learning analysis
Finally, we evaluate the efectiveness of the adopted neural network architecture for the
considered classification task, by comparing the performance of EUPHORIA to that of competitors
with SVM as classification algorithm. In particular, we measured the accuracy of two SVM-based
competitors:
• SVM that concatenated the three feature vectors produced by Word2Vec, BERT and
Behav in a single input vector and processed the concatenated data as the input of a SVM
classifier;
• Ens-SVM that trained an ensemble of three separate SVMs trained from the three feature
vectors produced by Word2Vec, BERT and Behav, respectively, and used the ensemble
majority rule for the final classification.</p>
      <p>The results in Table 6 show that the highest accuracy was achieved with EUPHORIA by
training a multi-input neural network. Notably, the main finding of this experiment is that
learning separate classifiers for each view performs worse than learning a single classifier by
concatenating all views together.</p>
    </sec>
    <sec id="sec-6">
      <title>5. Conclusions</title>
      <p>In this paper, we have illustrated a novel predictive approach for review spam detection, which
is able to take advantage of both content and behavioral characteristics possibly hidden in online
product reviews. In particular, we have proposed to process these multi-view data in their raw
format, leaving the task of sharing information (and consequently relationships) across multiple
views to the deep learning architecture. We have coupled a multi-view learning approach with
a deep learning architecture, in order to gain predictive accuracy from the diversity of data in
each view without sufering from the curse of dimensionality. The experiments performed on
two benchmark datasets confirm the efectiveness of the proposed approach. One limitation of
our methodology is the absence of an online learning phase able to periodically increment the
trained classification model as new reviews are recorded over time. In fact, the field of online
learning with deep neural networks is still mostly unexplored in the review spam detection
literature. An interesting avenue for future work is to explore further fine-tuning strategies,
which were recently explored in other tasks such as network intrusion detection [22].</p>
      <p>Finally, we plan to investigate the use of graph embedding techniques, in order to dynamically
represent relationships between reviewers and products.
review spam detection, Journal of Computational Mathematics and Data Science 3 (2022)
100036.
[10] N. Hussain, H. Turab Mirza, G. Rasool, I. Hussain, M. Kaleem, Spam review detection
techniques: A systematic literature review, Applied Sciences 9 (2019).
[11] N. Jindal, B. Liu, Opinion spam and analysis, in: Proceedings of the 2008 International
Conference on Web Search and Data Mining, WSDM 2008, Association for Computing
Machinery, New York, NY, USA, 2008, p. 219–230.
[12] N. Jindal, B. Liu, Analyzing and detecting review spam, in: Seventh IEEE International</p>
      <p>Conference on Data Mining (ICDM 2007), 2007, pp. 547–552.
[13] G. Satia Budhi, R. Chiong, Z. Wang, S. Dhakal, Using a hybrid content-based and
behaviourbased featuring approach in a parallel environment to detect fake reviews, Electronic
Commerce Research and Applications 47 (2021) 101048.
[14] J. Li, C. Cardie, S. Li, Topicspam: a topic-model based approach for spam detection, in:</p>
      <p>ACL, 2013.
[15] Y. Shang, M. Liu, T. Zhao, J. Zhou, T-bert: A spam review detection model combining
group intelligence and personalized sentiment information, in: I. Farkaš, P. Masulli,
S. Otte, S. Wermter (Eds.), Artificial Neural Networks and Machine Learning – ICANN
2021, Springer International Publishing, Cham, 2021, pp. 409–421.
[16] F. Li, M. Huang, Y. Yang, X. Zhu, Learning to identify review spam., 2011, pp. 2488–2493.
[17] N. Hussain, H. Mirza, I. Hussain, F. Iqbal, I. Memon, Spam review detection using the
linguistic and spammer behavioral methods, IEEE Access PP (2020) 1–1.
[18] P. Bhuvaneshwari, A. Rao, H. Robinson, Spam review detection using self attention based
cnn and bi-directional lstm, Multimedia Tools and Applications 80 (2021) 1–18.
[19] Y. Ren, Y. Zhang, Deceptive opinion spam detection using neural network, in: Proceedings
of COLING 2016, the 26th International Conference on Computational Linguistics:
Technical Papers, The COLING 2016 Organizing Committee, Osaka, Japan, 2016, pp. 140–150.
[20] Y. Ren, D. Ji, Neural networks for deceptive opinion spam detection: An empirical study,</p>
      <p>Information Sciences 385-386 (2017) 213–224.
[21] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, \. Kaiser, I.
Polosukhin, Attention is all you need, in: Advances in neural information processing systems,
2017, pp. 5998–6008.
[22] G. Andresini, F. Pendlebury, F. Pierazzi, C. Loglisci, A. Appice, L. Cavallaro, INSOMNIA:
towards concept-drift robustness in network intrusion detection, in: N. Carlini, A.
Demontis, Y. Chen (Eds.), AISec@CCS 2021: Proceedings of the 14th ACM Workshop on Artificial
Intelligence and Security, Virtual Event, ACM, 2021, pp. 111–122.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Heydari</surname>
          </string-name>
          , M. ali
          <string-name>
            <surname>Tavakoli</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Salim</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Heydari</surname>
          </string-name>
          ,
          <article-title>Detection of review spam: A survey</article-title>
          ,
          <source>Expert Systems with Applications</source>
          <volume>42</volume>
          (
          <year>2015</year>
          )
          <fpage>3634</fpage>
          -
          <lpage>3642</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>C.</given-names>
            <surname>Musto</surname>
          </string-name>
          , M. de Gemmis, P. Lops, G. Semeraro,
          <article-title>Generating post hoc review-based natural language justifications for recommender systems, User Model</article-title>
          .
          <source>User Adapt. Interact</source>
          .
          <volume>31</volume>
          (
          <year>2021</year>
          )
          <fpage>629</fpage>
          -
          <lpage>673</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>L.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Recommender systems based on user reviews: the state of the art, User Modeling and User-Adapted Interaction 25 (</article-title>
          <year>2015</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>M.</given-names>
            <surname>Crawford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. M.</given-names>
            <surname>Khoshgoftaar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. D.</given-names>
            <surname>Prusa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. N.</given-names>
            <surname>Richter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. A.</given-names>
            <surname>Najada</surname>
          </string-name>
          ,
          <article-title>Survey of review spam detection using machine learning techniques</article-title>
          ,
          <source>J. Big Data</source>
          <volume>2</volume>
          (
          <year>2015</year>
          )
          <fpage>23</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>E. Ferrara,</surname>
          </string-name>
          <article-title>The history of digital spam</article-title>
          ,
          <source>Commun. ACM</source>
          <volume>62</volume>
          (
          <year>2019</year>
          )
          <fpage>82</fpage>
          -
          <lpage>91</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Mukherjee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Venkataraman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. S.</given-names>
            <surname>Glance</surname>
          </string-name>
          ,
          <article-title>What yelp fake review filter might be doing?</article-title>
          , in: E. Kiciman,
          <string-name>
            <given-names>N. B.</given-names>
            <surname>Ellison</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Hogan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Resnick</surname>
          </string-name>
          , I. Soborof (Eds.),
          <source>Proceedings of the Seventh International Conference on Weblogs and Social Media</source>
          ,
          <string-name>
            <surname>ICWSM</surname>
          </string-name>
          <year>2013</year>
          , Cambridge, Massachusetts, USA, July
          <volume>8</volume>
          -
          <issue>11</issue>
          ,
          <year>2013</year>
          , The AAAI Press,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>T.</given-names>
            <surname>Mikolov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q. V.</given-names>
            <surname>Le</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Sutskever</surname>
          </string-name>
          ,
          <article-title>Exploiting similarities among languages for machine translation</article-title>
          ,
          <source>ArXiv abs/1309</source>
          .4168 (
          <year>2013</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , Bert:
          <article-title>Pre-training of deep bidirectional transformers for language understanding</article-title>
          , ArXiv abs/
          <year>1810</year>
          .04805 (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>G.</given-names>
            <surname>Andresini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Andrea</surname>
          </string-name>
          <string-name>
            <surname>Iovine</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Gasbarro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lomolino</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. de Gemmis</surname>
          </string-name>
          , A. Appice,
          <article-title>Euphoria: A neural multi-view approach to combine content and behavioral features in</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>