<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Contrasting Fake Reviews in TripAdvisor (discussion paper)</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Francesco Buccafurri</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Michela Fazzolari</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Gianluca Lax</string-name>
          <email>laxg@unirc.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Marinella Petrocchi</string-name>
          <email>m.petrocchig@iit.cnr.it</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>DIIES, University Mediterranea of Reggio Calabria Via Graziella, Localita Feo di Vito 89122 Reggio Calabria</institution>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Istituto di Informatica e Telematica - CNR Via G. Moruzzi</institution>
          ,
          <addr-line>1 56124 Pisa</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2018</year>
      </pub-date>
      <fpage>24</fpage>
      <lpage>27</lpage>
      <abstract>
        <p>Fake reviews are a concrete problem still a ecting the reliability of systems like TripAdvisor, especially in the case of few reviews, thus in the rst, most vulnerable, activity period of operators. In this work-in-progress paper, we present a model aimed to contrast this problem, based on a sort of normalization of scores given by users, to take into account the level of assurance of the reviews. This is done by considering two di erent dimensions, combined each other, which are the level of assurance of the identity of the review's author and the level of assurance of the occurrence of the evaluated experience. The paper presents a rst validation of the approach conducted on real-life data, giving us very encouraging results.</p>
      </abstract>
      <kwd-group>
        <kwd>reputation model</kwd>
        <kwd>trust management</kwd>
        <kwd>TripAdvisor</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Over the last years, online reviews became very important since they re ect the
customers' experience about a product or a service and nowadays constitute
the basis on which the reputation of an organization is built. Online reviews
have a great in uence on the purchase decisions of other customers, who are
increasingly relying on them [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>
        Unfortunately, the con dence in such reviews is often misplaced, due to the
fact that scammers are tempted to write fake information in exchange for some
reward or to mislead consumers for obtaining business advantages [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. These
reviews are called opinion spam or fake reviews [
        <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
        ].
      </p>
      <p>The identi cation of fake reviews is not an easy task, since they can be
identical to genuine ones. Nevertheless, several automatic techniques have been
proposed in recent years. Fake reviews detection involves the identi cation of
a set of features, linked with the content (review centric features) or with the
review author (reviewer centric features).</p>
      <p>
        Most of the existing machine learning approaches are not su ciently e ective
in spotting fake reviews, nevertheless they are more reliable than manual
detection. There exist several studies in the literature that rely on machine learning
approaches and consider di erent set of review features [
        <xref ref-type="bibr" rid="ref10 ref7">7, 10</xref>
        ]. Further
studies consider also reviewer centric features [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], which cannot be extracted from
the text of a single review. In addition, graph-theory based approaches have
been investigated to nd relationships between reviews and their corresponding
authors [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. The spam detection techniques that combine reviews features and
reviewers behaviors normally lead to better results [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
      </p>
      <p>
        In this work-in-progress paper we propose an approach di erent from the
previous fake review detection approaches. Moreover, we introduce a
reputation model and its preliminary experimental validation, designed to contrast the
phenomenon of fake reviews in TripAdvisor. TripAdvisor is a very famous travel
Website collecting reviews of travel-related contents. On the basis of these
reviews, an aggregate score of each content is shown. Due to the economic value
related to the e ects of this system, and despite the e orts declared by
TripAdvisor, the system is not immune from the problem of dishonest reviews, aimed
either to ctitiously promote a given operator (i.e., self promoting attack) or to
denigrate a competitor (i.e., slandering attack). Therefore, the noise occurring
in the reviews derives not only from physiological subjectivity of the users [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
Starting from a preliminary proposal [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], we de ne a model and a consequent
methodology aimed to (partially) purify the system from the noise coming from
fake reviews, in order to obtain more reliable normalized scores. This is done by
considering two di erent dimensions, combined each other, which are the level
of assurance of the identity of the review's author and the level of assurance of
the occurrence of the evaluated experience.
      </p>
      <p>It is worth noting that the approach here proposed is heuristic and any
feature used to compute the trust is based on reasonable argumentations and
ad hoc observations of the phenomenon, with no a speci c validation of any
single feature. Anyway, the paper is just aimed to give a general experimental
validation of the whole approach and, thus, of the global combination of any
feature.</p>
      <p>
        The novel contributions, w.r.t. the initial proposal presented in [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], is that
the model does not introduces new features required to TripAdvisor (in favor
of the practical relevance of the proposal) and that this paper includes also an
experimental validation of real-life TripAdvisor data.
      </p>
      <p>The structure of the paper is the following. In the next section, we de ne the
proposed reputation model. In Section 3, we describe the experiments carried
out to validate our proposal. Finally, our conclusions are summarized in Section
4.</p>
    </sec>
    <sec id="sec-2">
      <title>The Reputation Model</title>
      <p>In this section, we describe our reputation model, whose components are listed
in the following.
1. A set U of users, corresponding to the set of travelers, potentially covering
all Web users.
2. A set S of service providers, corresponding to the set of restaurants, bars,
hotels, and other operators registered to TripAdvisor.
3. For each service provider s 2 S, a list of feedbacks R(s) (which are the
reviews), each corresponding to a transaction. A feedback rs 2 R(S) for
the service provider s is a tuple hu; d; v; k; t; I i, where u is the author of the
feedback, d is the time of the feedback, v is the time of the transaction, k
is the score given by u on s (it is the aggregate score { also detailed into
di erent dimensions), t is a text motivating k (it is the text included by
the user to describe the experience) and I is the set of additional resources
(it is the images posted by the user). Concerning u, we rely on the set of
attributes which can be drawn from the system (and also by using external
sources, like social-network sources).</p>
      <p>The transaction corresponding to a feedback rs is denoted by tr. To
implement the notion of certi ed reputation in this model, for a given feedback rs, we
need two basic measures, which are both numbers ranging in the interval [0; 1]:
(1) the trustworthiness of the identity of u, denoted by trust(u), and (2) the
trustworthiness of the transaction tr, denoted by trust(tr).</p>
      <p>The rst measure trust(u) is related to identity management issues and is a
measure of the level of assurance of identity proo ng given by the registration
phase into the reputation system. Observe that we do not consider the level of
assurance of the authentication phase, thus assuming that no attacks on accounts
of users occur. This measure may take into account also external information
associated with the digital identity of the user in the system (e.g., information
coming from di erent sources as online social networks). We remark that the
trustworthiness of the identity of u is directly related to misbehaving users, as
any malicious activity is facilitated when the level of assurance of user identity
is low. The trustworthiness of the transaction, denoted by trust(tr) is equally
important. Indeed, fake reviews usually correspond to transactions that never
occurred.</p>
      <p>On the basis of the two measures above, we measure the trustworthiness
trust(rs) of a feedback rs = hu; d; v; k; t; I i associated with a transaction tr, by
the function trust(rs) = f (trust(u); trust(tr)) such that the higher trust(u) and
trust(tr), the higher trust(rs). In this paper, we experiment as rst attempt a
simple function f which is the linear combination of the two contributions. Once
trust(rs) has been computed, the score k of the feedback rs is corrected on the
basis of the overall trustworthiness trust(rs) by means of a function g(trust(rs)).
This function is build is such a way that the score k is much closer to the
average score obtained by the service provider s as the value trust(rs) is low.
Speci cally, g(k) = Prs0(2R+(tSr)u(st+(rtsr)u)skt(rs0)) , where is s suitable (small) o set to
avoid null terms. This way, the mean computed over all the scores obtained by the
operator s is just the mean weighted by trust values (shifted by ) of each review.
Therefore, we obtain the normalized feedback rs as rs = hu; d; v; g(k); t; Ii.</p>
      <p>Let us explain now how trust(u), representing the level of assurance of
identity the user u authoring the review, is computed. Of course, a signi cant part
of the current weakness of the reputation system of TripAdvisor is based on the
weakness of its digital identity management system.</p>
      <p>Concerning trust(u), we observe that the registration phase of TripAdvisor
does not force the user to provide any non-self declared credential. Anyway, the
possibility of registering via an existing Facebook pro le is also allowed.</p>
      <p>To compute the level of assurance of the identity, we consider in our model
the following features (which are, in turn, numbers from 0 to 1).
{ re (reviewer experience): it measures the seniority of the user in the system.
{ tc (text coherence): it measures the coherence of the known information about
the user (for example, the gender) and the text.
{ rc (review count): it measures the number of reviews of the user. This feature
is related to the fact that often fake reviews are done through accounts aimed
to a speci c goal (for example, for self-promotion or slandering attacks), so
they are not reused massively, also to avoid the linkage with de-anonymizing
information.
{ f i (Facebook identity): it measures the level of assurance of the Facebook
identity of the user (it is trivially 0 if the TripAdvisor account is not
associated with a Facebook pro le). Concerning this feature, we argue that a
fake reviewer does not have any interest in allowing linkage of her/his
TripAdvisor account with other (even fake) accounts, because probably tends
to operate as much as possible in an anonymous way.
trust(u) is then obtained as a linear combination of the above components.</p>
      <p>To compute the level of assurance of the transaction trust(tr), we consider
in our model the following features (which are, again, numbers from 0 to 1).
{ dc (data coherence): it measures the coherence between the date of the review
and the date of the transaction (this measure is based on the fact, according
to our estimation, 50% of the reviews is done within 23 days, 75% within 34
days and 80% within 40).
{ ip (image proof): it measures the degree of proof given by posted images
that the transaction really occurred (trivially, no posted images corresponds
to 0, the presence of images recognized as coherent with the other posted
by the majority of users corresponds to the maximum value). Indeed, the
standard behavior of a faker is to hide as much as possible any information
that could be a potential risk for de-anonymization, also the publication of
images, which is a typical action done by honest reviewers to give a proof of
their claims.
{ rl (review locality): it measures the presence of other reviews by the same user
in the same location (city or region) if di erent from the place of residence
corresponding to transactions experienced in the same period.
{ or (operator reaction): it measures the presence of a reaction posted by the
operator which is an outlier w.r.t. the standard behavior of the operator.
trust(tr) is also obtained as a linear combination of the above components.
3</p>
    </sec>
    <sec id="sec-3">
      <title>Experiments</title>
      <p>In this section, we describe the experiments carried out to validate the proposal:
they are based on real-life reviews of restaurants extracted from the TripAdvisor
site. First, we describe the data collection procedure and the error metrics used
in the experiments. Then, we discuss how a score quantifying the (actual) quality
of a restaurant has been computed (it is used as ground truth) and the method
adopted to tune the weight of the reputation model parameters. Finally, we
show the improvements obtained by our proposal in measuring the score of a
restaurant.
3.1</p>
      <sec id="sec-3-1">
        <title>Test Bed</title>
        <p>For the study presented in this contribution, we consider data taken from
TripAdvisor. Data were collected in January 2017, by developing an ad-hoc scraping
software and by using the available API to crawl information. The web scraping
process was performed by a Python script that navigated through the
restaurants available on the Province of Lucca web page. The metadata related to a
review, such as the language of the review, the rating, etc, were obtained by
the available APIs. We also stored the reviewers' pro les, which include, when
available, the age and country of origin.</p>
        <p>At the end of the extraction phase, we obtained a dataset composed of 1.499
restaurants, 60.613 reviewers, and 107.556 reviews. Each review judges the
quality of a restaurant by an integer score from 1 to 5.</p>
        <p>In our experiments, we de ne the bias of a review w as Bw = jsw 5GT rj ,
where sw is the review score of the restaurant r, the operator jxj denotes the
absolute value of the number x, and Qr (expected quality) is a real number in
the interval [1; 5] representing the quality of the restaurant r, which is computed
as the average of the scores of that restaurant. Bw represents how much the
review score is far from the score of the restaurant and it is normalized w.r.t.
the maximum score (i.e., 5). For example, Bw = 0 means that the review score
is coherent with the quality of the restaurant.</p>
        <p>We de ne review bias of a restaurant r as</p>
        <p>RBr =</p>
        <p>Pin=0 jBWi j
n
(1)
https://www.tripadvisor.com/Tourism-g187898-Lucca Province of Lucca Tusc
any-Vacations.html
where Wi is the i-th of the n reviews of the restaurant r. In words, it is used to
quantify the amount of variation of the reviews of a restaurant w.r.t. the expected
quality of that restaurant. For example, RBr = 0 means that all review scores
coincide and their value re ects the quality of r.</p>
        <p>Given a reputation model tu, we de ne its error %</p>
        <p>Eu =</p>
        <p>Pm
i=0 RBir
m
100
(2)
where m is the number of restaurants used in the reputation model and ri is the
i-th one. Clearly, it is the average of the review bias computed for all restaurants.</p>
        <p>Finally, to compare the accuracy of two reputation models t1 and t2, we
de ne the improvement % of t1 (w.r.t. t2) as I1 = E1E1E2 100. Observe that the
improvement can be negative in case the reputation model is less accurate than
the compared one.
3.2</p>
      </sec>
      <sec id="sec-3-2">
        <title>Parameter Setting</title>
        <p>The reputation model presented in this paper uses only ve of the parameters
de ned in Section 2, because they are the only ones evaluable from the collected
dataset: two based on the identity of the reviewer and three based on the
transaction. They assume a rate from 0 to 1 and how their rate is computed is now
discussed.
f1 This parameter is related to the number of reviews done by the reviewer.</p>
        <p>We computed an average of 36 reviews for each reviewer and we assigned
1 to this parameter to reviewers with at least twice the average value. This
values is linearly reduced to 0 for reviewers with only 1 review.
f2 This parameter is 1 if the reviewer signs in by Facebook, 0 otherwise.
f3 This parameter is 1 if the review contains at least an image, 0 otherwise.
f4 This parameter is related to length of the reviewer membership. As We have
reviewers with activity up to 180 months, we assigned 1 to the rate of ID2
= 1 to reviewers with at least 90 months of activity. This values is linearly
reduced to 0 for reviewers registered very recently (clearly, recently w.r.t.
the period in which data have been collected)
f5 This parameter is related to the coherence between review date and visit
date. We set f5 = 1 if time delta &lt; 15 days, else f5 = :75 if time delta &lt; 30
days, else f5 = :25 if time delta &lt; 45 days, 0 otherwise.</p>
        <p>In the next experiment, we analyze the performance of our model in which
a single parameter is considered. The result of this experiment is reported in
Fig. 1.(a) and shows that some parameters are useful to improve the reputation
model (namely, f1, f3, and f4), others reduce its performance.</p>
        <p>The next task is to combine all parameters and, for this purpose, we need
to give a weight to each of them. Such weights are computed by applying a</p>
        <p>Observe that this number includes many reviews that are not included in our dataset.
(a) Improvement of the reputation model (b) Performance of the proposed
reputaenabling only one parameter at a time tion model
multivariable regression model with the ve parameters f1 : : : f5 which returned
the following weights w1 = 0:6068, w2 = 0:2520, w3 = 0:0605, w4 = 0:2864,
w5 = 0:5778, where wx is the weight of the parameter fx with 1 x 5. These
weights will be used in the next experiment.
In this experiment, we measure the performance of the reputation model of
TripAdvisor and the performance obtained by using the reputation model proposed
in this paper, setting the parameter weight to the values obtained in Section 3.2.
We measured the review bias for the rst n reviews, with n ranging from 10 to
100: we limited the upper bound to 100 because, for higher values, the review
bias is very low. In Fig. 1.(b), we report the improvement % of the proposed
reputation model w.r.t. that of TripAdvisor. It is possible to see that our proposal
always gives the best results: the improvement is higher when a restaurant has
few reviews, that is when it is more vulnerable to fake reviews.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Conclusion</title>
      <p>TripAdvisor, as well as many other reputation systems, su ers from self-promoting
and slandering attacks typically performed by using fake accounts just created
for this purpose, which post fake reviews not corresponding to real experiences.
In this paper, we de ned a new reputation model for the reviews of TripAdvisor.
Our proposal aims at evaluate the dependability of a review on the basis of a
level of assurance for both identity proo ng and truth of the transaction. We
validate our proposal by measuring the improvement obtained by enabling our
reputation model on a real-life dataset reviews referring to restaurants of the
Province of Lucca, Italy. Anyway, the paper is just aimed to give a general
experimental validation of the whole approach and, thus, of the global combination
of any feature. As a future work, a selective validation of the di erent features
can be also performed.</p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgment</title>
      <p>This work has been partially supported by the project \Secure Citizen Remote
Identi cation" (CUP J88C17000150006) and by the project \Stabilimento
Virtuale" (CUP J38C17000100006), both funded by Regione Calabria - POR
Calabria FESR-FSE 2014-2020 - Asse I and by the European Union under grants
675320 (NECS-Network of Excellence in Cybersecurity) and 700294
(C3ISPCollaborative and Con dential Information Sharing and Analysis for Cyber
Protection) and by Fondazione Cassa di Risparmio di Lucca that partially funds the
ReviewLand project.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <given-names>J. K.</given-names>
            <surname>Ayeh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Au</surname>
          </string-name>
          , and
          <string-name>
            <given-names>R.</given-names>
            <surname>Law</surname>
          </string-name>
          .
          <article-title>Do We Believe in TripAdvisor? Examining Credibility Perceptions and Online Travelers' Attitude toward Using User-Generated Content</article-title>
          .
          <source>Journal of Travel Research</source>
          ,
          <volume>52</volume>
          (
          <issue>4</issue>
          ):
          <volume>437</volume>
          {
          <fpage>452</fpage>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>F.</given-names>
            <surname>Buccafurri</surname>
          </string-name>
          , G. Lax,
          <string-name>
            <given-names>S.</given-names>
            <surname>Nicolazzo</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Nocera</surname>
          </string-name>
          .
          <article-title>A model implementing certi ed reputation and its application to tripadvisor</article-title>
          .
          <source>In Availability, Reliability and Security (ARES)</source>
          ,
          <year>2015</year>
          10th International Conference on, pages
          <volume>218</volume>
          {
          <fpage>223</fpage>
          . IEEE,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <given-names>E.</given-names>
            <surname>Choo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Yu</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Chi</surname>
          </string-name>
          .
          <article-title>Detecting opinion spammer groups through community discovery and sentiment analysis</article-title>
          . In P. Samarati, editor,
          <source>Data and Applications Security and Privacy XXIX</source>
          , pages
          <volume>170</volume>
          {
          <fpage>187</fpage>
          ,
          <string-name>
            <surname>Cham</surname>
          </string-name>
          ,
          <year>2015</year>
          . Springer International Publishing.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <given-names>M.</given-names>
            <surname>Crawford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. M.</given-names>
            <surname>Khoshgoftaar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. D.</given-names>
            <surname>Prusa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. N.</given-names>
            <surname>Richter</surname>
          </string-name>
          , and
          <string-name>
            <given-names>H. Al</given-names>
            <surname>Najada</surname>
          </string-name>
          .
          <article-title>Survey of review spam detection using machine learning techniques</article-title>
          .
          <source>Journal of Big Data</source>
          ,
          <volume>2</volume>
          (
          <issue>1</issue>
          ):
          <fpage>23</fpage>
          ,
          <string-name>
            <surname>Oct</surname>
          </string-name>
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>A.</given-names>
            <surname>Heydari</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>a</article-title>
          . Tavakoli,
          <string-name>
            <given-names>N.</given-names>
            <surname>Salim</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Z.</given-names>
            <surname>Heydari</surname>
          </string-name>
          .
          <article-title>Detection of review spam</article-title>
          .
          <source>Expert Syst. Appl.</source>
          ,
          <volume>42</volume>
          (
          <issue>7</issue>
          ):
          <volume>3634</volume>
          {
          <fpage>3642</fpage>
          , May
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>N.</given-names>
            <surname>Jindal</surname>
          </string-name>
          and
          <string-name>
            <given-names>B.</given-names>
            <surname>Liu</surname>
          </string-name>
          .
          <article-title>Opinion spam and analysis</article-title>
          .
          <source>In Proceedings of the 2008 International Conference on Web Search and Data Mining, WSDM '08</source>
          , pages
          <fpage>219</fpage>
          {
          <fpage>230</fpage>
          , New York, NY, USA,
          <year>2008</year>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <given-names>F.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yang</surname>
          </string-name>
          , and
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhu</surname>
          </string-name>
          .
          <article-title>Learning to identify review spam</article-title>
          .
          <source>In Proceedings of the Twenty-Second International Joint Conference on Arti cial Intelligence -</source>
          Volume Volume Three,
          <source>IJCAI'11</source>
          , pages
          <fpage>2488</fpage>
          {
          <fpage>2493</fpage>
          . AAAI Press,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <given-names>B. Liu. Sentiment</given-names>
            <surname>Analysis</surname>
          </string-name>
          and
          <string-name>
            <given-names>Opinion</given-names>
            <surname>Mining</surname>
          </string-name>
          . Morgan &amp; Claypool Publishers,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <given-names>A.</given-names>
            <surname>Mukherjee</surname>
          </string-name>
          .
          <article-title>Detecting deceptive opinion spam using linguistics, behavioral and statistical modeling</article-title>
          .
          <source>In ACL</source>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <given-names>A.</given-names>
            <surname>Mukherjee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Liu</surname>
          </string-name>
          , and
          <string-name>
            <given-names>N.</given-names>
            <surname>Glance</surname>
          </string-name>
          .
          <article-title>Spotting fake reviewer groups in consumer reviews</article-title>
          .
          <source>In Proceedings of the 21st International Conference on World Wide Web, WWW '12</source>
          , pages
          <fpage>191</fpage>
          {
          <fpage>200</fpage>
          , New York, NY, USA,
          <year>2012</year>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <given-names>S.</given-names>
            <surname>Rayana</surname>
          </string-name>
          and
          <string-name>
            <given-names>L.</given-names>
            <surname>Akoglu</surname>
          </string-name>
          .
          <article-title>Collective opinion spam detection using active inference</article-title>
          .
          <source>In Proceedings of the 2016 SIAM International Conference on Data Mining</source>
          , pages
          <volume>630</volume>
          {
          <fpage>638</fpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>