Comparing Fake Review Tools on Amazon.com Masterarbeit - unipub

Page created by Kyle Perry
 
CONTINUE READING
Comparing Fake Review Tools on Amazon.com Masterarbeit - unipub
Marija Basic

      Comparing Fake Review Tools on Amazon.com

                               Masterarbeit
        zur Erlangung des akademischen Grades eines Master of Science
  der Studienrichtung Betriebswirtschaft an der Karl-Franzens-Universität Graz

Ao.Univ.-Prof. Mag. Dipl.-Ing. Dr. rer.soc.oec. Christian Schlögl
Institute for Information Science and Information Systems
at Karl Franzens University Graz

                                                                    Graz, April 2020
Comparing Fake Review Tools on Amazon.com Masterarbeit - unipub
Table of Contents

1.    Introduction                                            1
1.1 Social influence on opinions and their impact             4
1.2 Ethics, incentives and organizations                      5
1.3 Problem statement                                         7
1.4 Organization of the thesis                                8
2.    Spam detection                                          9
2.1 Different types of reviews                                9
2.2 Different types of reviewers                              11
2.3 About Amazon.com                                          12
2.3.1 The review system and Amazon dataset                    13
2.3.2 Review rating, product review and feedback              14
2.4 Data sources for fake review detection                    16
3.    Empirical part                                          18
3.1 State of the art                                          18
3.2 Objective and research questions                          21
3.3 Methodology                                               22
3.4. Results                                                  25
3.4.1 Selection of the tools                                  25
3.4.2 Description of the selected tools                       26
3.4.2.1 ReviewMeta                                            26
3.4.2.1.1 Usability of ReviewMeta                             26
3.4.2.1.2 Scope, depth and form of outcomes of the reports    30
3.4.2.1.3 Authority of ReviewMeta                             45
3.4.2.1.4 Curency of ReviewMeta                               46
3.4.2.1.5 Export and share function of ReviewMeta             46
3.4.2.2 FakeSpot                                              47
3.4.2.2.1 Usability of FakeSpot                               47
3.4.2.2.2 Scope, depth and form of outcomes of the reports    51
3.4.2.2.3 Authority of FakeSpot                               55
3.4.2.2.4 Currency of FakeSpot                                56
3.4.2.2.5 Export and share function of FakeSpot               56
3.4.2.3 TheReviewIndex                                        56

                                                                   i
Comparing Fake Review Tools on Amazon.com Masterarbeit - unipub
3.4.2.3.1 Usability of TheReviewIndex                                            57
3.4.2.3.2 Scope, depth and forms of the outcomes of the TheReviewIndex reports   59
3.4.2.3.3 Authority of TheReviewIndex                                            64
3.4.2.3.4 Currency of TheReviewIndex                                             65
3.4.2.3.5 Export and share function                                              65
3.4.3 Comparison of the tools based on criteria                                  66
3.4.4 Comparison of the outcomes of the tools                                    71
4.    Critical reflection                                                        74
References                                                                       76

                                                                                      ii
Comparing Fake Review Tools on Amazon.com Masterarbeit - unipub
Figures
Figure 1 Key statistical indicators ......................................................................................................1
Figure 2 The structure of posted reviews by buyers and non-buyers ..................................................2
Figure 3 Increase in suspicious reviews…………………………………………………………..….3
Figure 4 Types of Spams ..................................................................................................................10
Figure 5 An incentivized review .......................................................................................................13
Figure 6 Review rating system on Amazon ......................................................................................15
Figure 7 A review feedback ..............................................................................................................15
Figure 8 Popularity of the topic in academic publications ................................................................19
Figure 9 Research model of the work ................................................................................................22
Figure 10 ReviewMeta extension .....................................................................................................27
Figure 11 The URL search bar on RevieMeta website ......................................................................28
Figure 12 The FAQ browser window on ReviewMeta website ........................................................28
Figure 13 The contact us browser windows on ReviewMeta ............................................................29
Figure 14 Country-based Amazon websites supported by the tool ...................................................30
Figure 15 The final scorecard of the ReviewMeta tool ....................................................................31
Figure 16 Review adjustment function of the report .........................................................................32
Figure 17 An overview of all outcomes in the report .......................................................................33
Figure 18 The most vs trusted reviews in the report ..........................................................................34
Figure 19 The presentation of the deleted reviews in the report ........................................................35
Figure 20 The presentation of suspicious reviewers .........................................................................36
Figure 21 The reviewer ease of the product and the product in its category .....................................37
Figure 22 The presentation of word count comparison .....................................................................38
Figure 23 The rating trend of reviewers ...........................................................................................39
Figure 24 The presentation of reviewer participation ........................................................................40
Figure 25 The presentation of the unverified purchases ....................................................................40
Figure 26 Examples of unverified reviews in unverified purchase report .........................................41
Figure 27 Phrase repetition test .........................................................................................................42
Figure 28 Examples of the posted reviews with repetitive phrases ...................................................42
Figure 29 The percentage of the overlapping reviews .......................................................................43
Figure 30 Percentage of reviews about other products from the same brand ...................................44
Figure 31 The presentation of incentivized reviews .........................................................................45
Figure 32 The currency of the report .................................................................................................46
Figure 33 Browser extension of the FakeSpot tool ............................................................................47
Figure 34 FakeSpot grade on Amazon product page .........................................................................48
Figure 35 FakeSpot grade in a Google search result .........................................................................48
Figure 36 Search bar of the FakeSpot tool.........................................................................................49
Figure 37 The FAQ window .............................................................................................................50
Figure 38 Contact us browser window ..............................................................................................50
Figure 39 Final scorecard of the FakeSpot report ..............................................................................52
Figure 40 The company review report as a part of the final scorecard ..............................................52
Figure 41 The analysis overview section in the report ......................................................................53
Figure 42 The presentation of the review analytic section of the report............................................54
Figure 43 The presentation of the review insights section of the report ............................................55
Figure 44 The information about the currency of the report..............................................................56
Figure 45 The browser extension of the TheReviewIndex tool .........................................................57
Figure 46 The URL search bar of the tool .........................................................................................58
Figure 47 The scorecard overview of the tool ...................................................................................60
Figure 48 The presentation of one of the product performances summarized by the tool ................60
                                                                                                                                               iii
Comparing Fake Review Tools on Amazon.com Masterarbeit - unipub
Figure 49 The presentation of warn outcome of the spam test in the report......................................61
Figure 50 The presentation of the passed outcomes of the spam test in the report ...........................62
Figure 51 The specifications section in the report .............................................................................62
Figure 52 Recommendations offered by the tool in the report ..........................................................63
Figure 53 A new browser window of the recommendation section ..................................................63
Figure 54 The web store product link as a part of the report .............................................................64
Figure 55 Last update of a report...………………………………………………………………... 65

                                                                                                                                     iv
Comparing Fake Review Tools on Amazon.com Masterarbeit - unipub
Tables
Table 1 Comparison of the tools based on proposed criteria ........................................................66
Table 2 The final outcomes of the compared tools for the Samsung computer ............................71

                                                                                                                        v
Comparing Fake Review Tools on Amazon.com Masterarbeit - unipub
1 Introduction

1. Introduction

Our beliefs and perceptions are widely influenced by opinions of others. These opinions,
consisting of different types of plenty of valuable information, are one of the primary
factors in making our choices and decisions. We rely on others not to make wrong decisions
under uncertainty.1 Generally, how others see and value the things and the world around
themselves largely shapes human behaviour and activities.
Over the years, technology has revolutionized the way how we communicate and created
tools and resources, putting each person’s most useful information at their fingertips.
Modern technologies such as customer review sites, online shops, blogs, social networking
sites, and online communities have allowed individuals to share and read opinions with
other individuals.2 New technologies have influenced the use, scope and effectiveness of
social media in our lives. In favour of this, the new 2018 Global Digital Report reveals that
more than 4 billion people use the internet and more than 3 billion people around the world
use social media (see Figure 1).3 People are prone to be active in online networks in order
to receive opinions and information and are likely to trust others. This is distinctive not
only for consumers as individuals when they are making a purchase decision but also for
firms and organizations when they are making business decisions.

                    Figure 1 Key statistical indicators (Source: Screenshot from wearesocial.com)

The number of online shops, products and product reviews rises with the growth of users.

1
  See Sherif (1935), pp. 47.
2
  See Constantinides et. al (2008), pp. 233.
3
  See Kemp (2018), n. p.
                                                                                                             1
Comparing Fake Review Tools on Amazon.com Masterarbeit - unipub
1 Introduction

Thus, a 2017 review report by statista.com reveals that the increased number of product
reviews in online shops has become a major factor in making decisions to buy.4 For
instance, if one wants to buy a product, one will look for relevant information from a large
number of review websites. If most opinions or reviews are positive, one is likely to buy it.
Also, if the opinions of online participants are mostly negative, one will almost certainly
not buy it. Accordingly, if we make our decisions based on fake positive or negative
reviews, then we will make a wrong decision.
There are different reasons for such dishonest actions of users. Recent findings from 2018
made by reviemonitoring.com reveal that almost 18% of reviewers commented on a
product they did not buy in order to receive a gift by friends/family, or they are fans of the
brand. Also, there are individuals who were paid to do it.5 The structure of buyers and non-
buyers is presented below in Figure 2.

             Figure 2 The structure of posted reviews by buyers and non-buyers (Source: Screenshot from
                                                       reviewmonitoring.com)

Similar to findings from reviewmonitoring.com, 68% of the customers left a review after
a local business asked them to do that. Manufacturers and sellers are today more likely to
trick the system misusing fair business practices by creating incentives for individuals to
write positive deceptive reviews when they are faced with increased competition or when
they have a less established reputation. Positive fake reviews are related to economic

4   See Clement (2018), n. p.
5   See Fake Reviews- Is It Really An Issue (2008), n p.
                                                                                                            2
Comparing Fake Review Tools on Amazon.com Masterarbeit - unipub
1 Introduction

benefits and may result in significant financial gains, whereas negative fake reviews could
be very harmful to potential buyers and manufacturers which can cause further strong
incentives for writing fake reviews.
The confidence in online reviews is misplaced. According to the Local Consumer Review
Survey 2017, 74% of the consumers have read a fake online review.6 In addition to this
problem, an online tool for detection fake reviews, ReviewMeta.com, reported that the
number of suspicious reviews posted especially in June, July and August 2017 (see Figure
3) have been largely increased, as the average review weight (blue line) drooped off
considerably. The average review weight indicates the measure of trustworthy reviews,
whereas the red line indicates the average rating of suspicious reviews.

                      Figure 3 Increase in suspicious reviews (Source: Screenshot from forbes.com)

The problem is widespread. According to statistics and many practical examples, fake
reviews will get worse and also more complicated. Both sides, producers and sellers, are

6
    See Murphy (2018), n. p.

                                                                                                             3
Comparing Fake Review Tools on Amazon.com Masterarbeit - unipub
1 Introduction

highly concerned with user reviews. Marketers analyse them intending to find product
problems and provide effective solutions. A manufacturer`s reputation in a business world
is widely affected by user reviews.
Detecting spam reviews will become more and more demanding. That is why it becomes
an absolute necessity to identify and prevent such spamming activities. Therefore, it is
necessary to think about how they can be detected and what are the reliable methods,
approaches, applications or tools available for their identification.

1.1.      Social influence on opinions and their impact

Social factors have high relevance in everyday life. In literature, there are numerous studies
and theories that are concerned with the social influence analysing the impact of different
social mechanisms on a change in an individual’s behaviour and how other people’s
opinions affect one’s decisions.
According to psychologists, there are two types of social influence: normative and
informational. Our decisions, opinions and therefore reviews work in both ways as norms
and information (social proof).
Normative social influence is defined in social psychology as “...the influence of other
people that leads us to conform in order to be liked and accepted by them”7. For instance,
if a product gathers more reviews, individuals will be more likely to actually buy it. Thus,
reviews also serve as valuable information. Informational social influence is where a person
conforms to gain knowledge because they believe that someone else is right.8
Psychologist Herbert C. Kelman recognizes three different modes of social influence that
impact customers’ purchase decision, and therefore, there are three different modes of
behaviour of individuals: 9
     1. People may be influenced by someone following someone else’s decision or
           suggestion despite their own private opinion because of the need to be liked and
           not to be different what is also seen in “Asch’s line experiment”10. According to
           Kelman, this mode of social influence is called compliance.

7
  See Aronson et. al (2005), pp. 253.
8
  See Deutsch et. al (1955), pp. 629.
9
  See Kelman (1958), pp. 53.
10
   See Asch (1955), pp. 5.
                                                                                            4
1 Introduction

        2. The second mode of social influence or internalization occurs when people agree
             with a decision of people or groups and adopt it because of the need not to make
             wrong decisions. This is particularly seen in “Sherif’s autokinetic experiment”11.
        3. The last mode of social influence is identification where a person adopts the
             behaviour of a widely respected and admired person, in order to achieve a friendly
             relationship with another individual or group.
According to some evidence, people may imitate behaviour and accept opinions of others.
Imitating others’ behaviour can increase people’s efficiency and what was good for one
person will still probably be good for another person.12
All these studies have shown that individuals’ opinions are influenced by opinions of others
not only in real-world but also in online environments. As a consequence, opinions help
people to decide which products or services they could buy and have slowly become an
integral part of the business on the internet.
The digitalization and existence of social media platforms have not largely changed social
behaviours since Kelman’s research. Moreover, the dramatic development of technology
over the past decades has changed only the sources of social influence. They will continue
to have an impact.
Simply, company products cannot compete without other opinions and reviews. The social
media environment is used for advertising, creation of an efficient marketing strategy,
which in turn could help them gain more customers. On the other hand, both positive and
negative deceptive reviews may have a further impact on the revenue, brand awareness,
company reputation and profitability of organizations selling products. These deceptive
reviews by consumers may result in wrong decisions by managers. If managers heavily
rely their tactics and business decisions on them rather than on expert opinions, then they
could make wrong decisions with adverse economic consequences.

1.2.         Ethics, incentives and organizations

According to previous findings, reviews have become increasingly popular over the past
decade, and now this phenomenon exists for nearly every product and service. There is a

11
     See Sherif (1935b), pp. 49.
12
     See McLeod (2011), n. p.
                                                                                                  5
1 Introduction

large number of consumer reviews sites such as Amazon, Angie’s List, Choice,
TripAdvisor and Yelp which bring consumers and businesses benefits.
Their profitability increases with the growth of their popularity or visits too. Higher
profitability caused by the growth of the number of visitors raises the possibility of the
appearance of non-ethical behaviour from organizations and visitors. Sometimes the line
between a good business idea and an unethical or even illegal business practice is not
perfectly clear. In a real market, firms are likely to use illegal means for marketing
purposes, especially when they are faced with an increased competition or established a
bad reputation.
Here are some examples of different forms of unethical behaviour conducted by individuals
and companies:
          •    Firms are likely to offer customers a product without any payment or to sell them
               at a discount in exchange for a review (incentivized reviews).13
          •    A group of reviewers who collaborate to post fake reviews have been found as
               groups with damaging intentions and may be regarded as illegitimate groups.14
          •    A large number of reviews posted in a very short period may be defined as fake.15
          •    According to Morning Advertiser, a criminal court in Italy ruled that writing
               reviews using a false identity is criminal conduct under Italian law.16
          •    Online technologies such as automated services (referred to as “bots”) that are
               made to create fake identities using a fake person`s name or email were reported
               by the European Commission in 2018.17
          •    Buyers may receive e-mails appearing to come from Amazon, that are actually
               false e-mails, sometimes called ‘spoof e-mails’ or ‘phishing e-mails’. The
               purpose is to log into an account and to change some parts of their account.18
               They can be used for posting fake reviews.
          •    BBC discovered that Samsung paid students to write negative reviews about
               phones made by Taiwan’s HTC.19

13
   See McCabe (2018), n. p.
14
   See Saradha et.al. (2018), pp. 113.
15
   See Mukherjee et.al. (2012a), pp. 1.
16
   See Robinson (2018), n. p.
17
   See European Commission (2018), pp. 5.
18
   See Evans (2018), n. p.
19
     See Samsung probed in Taiwan over ‘fake web reviews’ (2003), n. p.
                                                                                                6
1 Introduction

           •    Federal Trade Commission in 2010 took action against a public relations business
                because of asking their own employees to post reviews as consumers for its
                clients.20
           •    A Dallas business owner accused Yelp of hiding good reviews of his coffee shop
                after he refused to pay them for advertising.21
As can be seen from these examples and evidence, fake reviews concern all participants in
a business chain, consumers and sellers. It can cause an erosion of seller credibility and
have an incremental negative impact on the business. To deal with this issue, fake online
reviews should be paid much more attention, as the number of online shoppers has
dramatically increased, and their action may have unethical intentions.

1.3.           Problem statement

The challenge of this work is to find the most adequate application or tool for the detection
of fake reviews with the best performance because sometimes it is not possible to recognize
truthful opinions and reviews by merely reading them. There are some of the following
reasons for that:
           •    Reviews may be written by experts, retailers or any other organization that cannot
                be easily distinguished from genuine, honest reviews;
           •    Organizations may incentivize users to post genuine looking reviews.
The objective of this thesis is to explore online tools available for the detection of fake
reviews. Each tool should be appropriate to identify fake reviews and current habits and
intentions of reviewers. Specifically, I investigated how much these tools may differ from
one another. Also, the work intends to make a comparison using a previously defined list
of criteria and finally to discuss the outcome in more detail. An important aspect of this
research is to analyse their credibility in their detection. This allowed me to infer which
tool is the most reliable, how various tools may increase their performance and what are
the strengths and weaknesses of each of them.

20
     See Valant (2015), pp. 6.
21
     See Paul (2018), n. p.
                                                                                                 7
1 Introduction

1.4 Organization of the thesis

Chapter 1 gives fundamental explanations of some of the elements and social factors that
influence opinions in the decision-making process. The effect of social influence on
reviews is explained using existing theories. Furthermore, this Chapter highlights different
forms of unethical behaviour of individuals and companies or even possible illegal
activities.
Various types of reviews, spammers, spamming, review ratings, review feedback and
features are considered in Chapter 2. It also presents the online web store amazon.com
focusing more on the main characteristics of Amazon’s dataset that will be used for spam
detection.
Chapter 3 highlights state of the art and presents the research questions and goals of this
master thesis. All steps of the proposed framework for the research are introduced in this
chapter. The steps of the proposed framework are followed by the detailed description of
the tools, their comparison according to well- defined criteria. Finally, the results of the
analysis will be shown.
Finally, Chapter 4 is concerned with conclusions about each analysed tool, their strengths
and weaknesses. This should help not only individual consumers but also businesses and
organizations to choose the most appropriate tool for their purposes. Furthermore, this
chapter suggests directions for possible future work on fake review detection.

                                                                                           8
2 Theoretical part

2. Spam detection

This chapter should help us better understand reviews and show the differences between
truthful and untruthful reviews. The focus is on definitions and characteristics of reviews,
fake reviews, review ratings and review feedbacks. To inspect the reviews a dataset from
Amazon.com will be used, and their characteristics will be discussed.

2.1.           Different types of reviews

In literature, the opinion spam problem was first determined by Jindal and Liu. Basically,
according to Jindal and Lui, there are three types of review spam22:
Type 1 (Untruthful or fake reviews):
This type of reviews are untruthful positive reviews of products that have been posted by
customers or users, with a hidden intention to promote the product or untruthful negative
reviews written with the intention of harming their reputation. These reviews are meant to
mislead readers in their opinions. Usually, they are written with an intention to have an
economic impact.
Type 2 (Reviews on brands only):
These reviews refer to brands, businesses or sellers of a product and do not comment on a
product’s features. For this reason, they are considered as biased. For instance, a reviewer
may write for a specific Samsung product: “I do not like Samsung, and I will never
purchase any of Samsung’s products”.
Type 3 (Non-reviews):
This type encompasses all reviews written for marketing purposes and may not refer to the
product. They can be roughly categorized into two subtypes:
           •     Advertisement,
           •     irrelevant text or reviews having random texts, answers, etc.

Most studies in the literature have shown that types 2 and 3 of spam reviews can be very
easily detected by merely reading them. Even a reader with little or no knowledge may
recognize them easily while reading. However, type 1 is tough to detect because those

22See   Jindal et. al. (2008a), n. p.

                                                                                                 9
2 Theoretical part

reviews are very similar to truthful reviews. For this reason, the identification of type 1 is
a challenging task as it is hard to distinguish between fake and real reviews by reading
them. This type of reviews could be posted by experts who are paid to write believable
reviews. Neither positive nor negative fake reviews are based on reality. At best, they are
based on a shared misunderstanding; at worst they are outright lies. Both
positive and negative opinions have been shown to influence others in the decision-making
process directly.
Different classifications of fake reviews were the subject of research of many authors. For
instance, there is a classification of opinion spam into deceptive and disruptive opinion
spam.23 Deceptive opinion spam refers to type 1, whereas disruptive opinion spam refers
to types 2 and 3.

                                          Figure 4 Types of Spam (Source: Screenshot)24

Figure 4 presents six different forms of fake reviews. Positive fake reviews for a good
quality product (Type 1) or negative fake reviews for a bad quality product (Type 6) are
not very harmful, because they may be partly true. But there are reviews that demonstrate
the conflict of the review level and product quality and may have hidden motives such as
damaging the reputation of the product or brand with a final result of creating losses or
profits (Type 2, 4) or to promote the product (Type 3 and 5). The primary focus is the
detection of these types of damaging fake reviews using free based web tools.
Also, in literature, there are many researchers that define fake reviews as a form of
“deception”. Buller and Burgoon define deception as “a message knowingly transmitted
by a sender to foster a false belief or conclusion by the receiver”.25 The traditional meaning
of deception, which refers to a false pretence of facts of feelings, has become an integrated
part of the online environment. For instance, a deceptive reviewer writes in a certain way,

23 See Ott, M. et.al. (2011a), pp. 309.
24 See Lim, E. et.al. (2010a), pp. 1.
25 See Buller, D. B., (1996), pp. 239.

                                                                                                         10
2 Theoretical part

using special signals in the text or often using words in order to conceal the truth.26 Many
researchers developed a set of cues for their identification.
Instead of fake reviews, in the literature, a “fraud review” can be used as a term for a fake
one with the same meaning. Fraud reviews occur when online vendors, individuals or
reviewers write “consumer” reviews while presenting themselves as real customers.27 The
Concise Oxford Dictionary defines fraud as “criminal deception; the use of false
representations to gain an unjust advantage”.

2.2.        Different types of reviewers

A fake reviewer or spammer may be anyone who writes fake reviews, any type of business
or organization or even a competitor with an intention to provide a kind of reward to those
who post fake reviews for the competitor`s products. Basically, they may act individually
or in a group, without any indication of their secret behaviour at first sight. This is a
collective act, where the members of a group cooperate blindingly.
In the research work of Mukherjee, fake reviewers can be divided into two groups: an
individual spammer and a group of spammers.28 They differ according to the following
characteristics:
        •       An individual spammer is considered to be an individual, who works exclusively
                alone, and therefore has its own user account.
        •       A group of spammers is a group of individuals who act together, and they can be
        •       categorized further into two sub-groups:
            -    A group of individuals who act in a team with a purpose to build up or to demote
                 the reputation of a target product.
            -    A single person may be behind different user IDs that help him avoid his
                 detection and gain rewards by writing many fake reviews.
The distinct behavioural features of fake reviewers were also the subject of the research of
many authors:29
        •       Spammers are likely to write more reviews for a single product.
        •       Spammers are likely to write more reviews also for a group of products that

26
   See Zhou, L. et. al. (2008), pp. 1.
27
   See Hu, N. et. al. (2011), pp. 614.
28
   See Mukherjee et.al. (2012b), pp. 1.
29
   See Lim, E. et.al. (2010b), pp. 3
                                                                                              11
2 Theoretical part

              belong to a brand.
        •     The rating score of a group of fake reviewers differs from the rating score of
              honest reviewers.
        •     Spammers tend to rate the product in the early phase after a product’s release.
Taking into account features such as time, frequency and content, the following possible
attributes of spammers were discovered:30
        •     They act during the same time of the day.
        •     They post the same text over and over.
        •     They use different email IDs.
        •     They are prone to sharing their contacts when they post.

2.3. About Amazon.com

The analysis of this thesis is based on reviews from amazon.com. Amazon has a consumer
review platform where users may review all types of products such as books, music, videos,
electronic cards, auctioned items and many other products offered by sellers.
The company officially launched its websites in Seattle by Jeff Bezos in 1994 and today is
one of the most successful e-commerce companies in the world. The main goal of Amazon
is to sell different products through the same means, from the same storages or through
subsidiaries through a common web site for browsing, searching, and buying. The
simplicity, easiness, and low cost of the shopping experience is the driving motive of this
company.
In October 2018, Amazon offered more than 3 billion products across 11 marketplaces
worldwide.31 According to the statista.com report from October 2018, Amazon received
approximately 2.4 billion32 unique visits and contains more than 7 million reviews.
Amazon counts more than 76 million user accounts and 1.3 million active sellers.
One of the concerns of Amazon is “incentivized reviews”.33 Incentivized reviews are
product reviews posted by buyers that received compensation, such as a discount or free
item, for writing a review (see Figure 5).

30
   See Husna, H., et. Al. (2008), pp.1
31
   See Scrape Hero (2018), n. p.
32
   See Statista. (2019), n. p.
33
   See Amazon.com. (n. Y.-b), n. p.
                                                                                                12
2 Theoretical part

                   Figure 5 An incentivized review (Source: Screenshot from landingcube.com)

Amazon offers a few ways for sellers to get an initial set of reviews for newly launched
products        such       as     “Early        Reviewer         Program”       and   “Vine    Program”.
Through “Early Reviewer Program”, Amazon provides direct incentives (e.g. small
discount) for verified buyers if they share their real and honest experience about products.34
Actually, vendors ask reviewers through Amazon for their real feedback, but without any
direct contact between them.
The “Amazon Vine Program” is exclusively an invitation program for honest reviews.35
Amazon invites the members of Vine program to post reviews about a new product. The
members of Vine program are the most trusted reviewers or top-ranked reviewers. Their
opinions may help those who are not prone to comment, to get information about products.

2.3.1. The review system and Amazon dataset

One of the key reasons for the company’s success is that Amazon has review systems,
which may help both customers and organizations. Basically, a review system is a part of
the reputation of every online company because it may increase the trust in internet
interactions.36
The first of Amazon’s Leadership Principles is “Customer Obsession”37 and to be
customer-focused. Indeed, customer reviews have been a crucial part of Amazon for over
20 years. The 5-star rating system allows users to rank products (1 is the worst, 5 is the
best), write or read comments. In this way, ranked products with stars demonstrate the level
of popularity or quality, which is often useful to other buyers when they cannot touch or
test products before purchasing.

34
   SeeAmazon.com Help: What is the Early Reviewer Program? (n. Y.), n. p.
35
   See Amazon.de- Amazon Vine- Club der Produkttester Programm (n. Y.), n. p.
36
   See Resnick, P. (2000), pp. 45.
37
   See Amazon.com. (n. Y.-a). n. p.
                                                                                                       13
2 Theoretical part

Amazon presents a compelling setting in which fake reviews can be investigated. Amazon
counts more than 7 million reviews. In particular, the reasons for using the Amazon dataset
are the following:
          •    Amazon is the top starting point for the product search by consumers, according
               to the recent finding of ReviewMonitoring.com38
          •    the dataset is large and transparent,
          •    it has a rating system,
          •    it is easy to access the data, and
          •    it covers a very wide range of products.
Each of Amazon’s review dataset consists of the following features:
          •    reviewer’s name (a real name or a nickname),
          •    product name,
          •    rating score (from one to five stars),
          •    review title,
          •    review body,
          •    number of feedbacks,
          •    number of helpful feedbacks, and
          •    the date of the review.

2.3.2. Review rating, product review and feedback

Each review contains three parts: product review and feedback. The main distinctions
between online reviews and feedback are the kind of data being collected and how it is
processed.
Product reviews refer to consumer evaluations and judgments about a particular product in
the form of 5-star scale (with 1 being the worst, 5 being the best). A low rating indicates a
very poor opinion about a product, whereas a high rating indicates a positive opinion. Those
evaluations contain quantitative and qualitative information on the product’s attributes.
Overall, the final rating of the product is calculated as the average of all the ratings (see
Figure 6).

38
     See ReviewMonitoring.com. (n. Y.), n. p.
                                                                                           14
2 Theoretical part

                        Figure 6 Review rating system on Amazon (Screenshot from amazon.com)

Review feedback presents truthful or untruthful information about prior online experience
of an individual posted at web sites. Other users may use that information to adjust and
improve future decisions. Moreover, these feedback results provide businesses with deep
insights into consumers’ experience, which marketers can analyse and use to adjust or
improve their services and products.
A review product differs from review feedback. According to Amazon, review feedback is
considered to be a review for the seller or the purchasing experience that emanates after
purchase, whereas a product review refers exclusively to an item purchased (see Figure
7).39 One typical example for feedback appointed to a seller may be the following: “The
phone case was delivered in time that I expected, and the product was exactly as I saw it
pictured. Thank you!”

                            Figure 7 A review feedback (Source: Screenshot from amazon.com)

In order to be legitimate and not manipulated, Amazon has defined guidelines for feedbacks

39   See Amazon.com (n. Y-d), n. p.
                                                                                                             15
2 Theoretical part

strictly. If feedback contains “promotional content, obscene or abusive language, personal
information”40, the product reviews may be removed at Amazon’s discretion.

2.4. Data sources for fake review detection

There are three important data sources which can be used for fake review detection:41
        •       Review content refers basically to the content of a review. On this level, one
                learns essentials about the content (e.g. that texts are very similar for different
                products, or that a critique of a book was simply copied), about the
                professionalism (e.g. text length) or the words used for lies.
        •       Meta-data informs users about the given stars (for example, a reviewer gave 5-
                stars for the brand and 1 star for the competitor’s brand), the user ID (a reviewer
                has more user IDs), the time of posting (the average rating of a product has been
                changed suddenly or a considerable number of reviews appear suddenly), IP and
                MAC addresses, location of the computer, and the number of clicks.
        •       Product information reports information about the product, e.g., sales volume,
                the rank of a product or a product’s characteristics. For instance, a product is
                selling well but has many negative reviews.
Taking into account these three data sources, Liu, in his book, suggests the following
indicators for detection:42
        •       Review content:
            -    Headline of review
            -    Text
        •       Meta-data
            -    Review URL
            -    Rating
            -    Number of reviews
            -    Number of ratings with five stars
            -    Number of ratings with a star
            -    Reviewer

40
   See Amazon.com (n. Y-e), n. p.
41
   See Liu, B. (2012-b), pp. 126–127.
42
   See Liu, B. (2012-c), pp. 126–127.

                                                                                                16
2 Theoretical part

          -    User ID
          -    Verified purchase
          -    Review was helpful
          -    Date
      •       Product information
          - Product description
          - Sales rank
          - Sales volume.
Considering all these elements of reviews, the intentions and behaviour of spammers can
be possibly discovered. For their detection, there are free web tools that report fake reviews,
reviewers and groups of spammers on the basis of the features described above. They may
explain untruthful behavioural patterns based on this data. In this work, several fake review
tools were analysed and compared to see how reliable they are.

                                                                                            17
3 Empirical part

3. Empirical part

This Chapter is divided into three parts. The first section is devoted to the state-of-the-art
with an aim to present an overview of the studies in this field of analysis. For that purpose,
I searched publications in Thomson Reuter’s online database Web of Science. This is
followed by the description of the research object and the research questions. The second
part’s purpose is to explain the different steps that were carried out in order to conduct
the tasks and analysis of this thesis. More precisely, conducting the tasks requires steps
such as well-defined criteria for the selection of web tools that will be used for fake review
detection, description of the web tools, data gathering and finally, testing the tools.

3.1.   State of the art

This section gives an overview of the state of the art and a literature review of the most
important research papers on fake review detection so far. For that purpose, I conducted
the following steps: First, I conducted a search query on Thomson Reuters Web of
Science. Second, as there has been much research on this topic in recent years, after
examining titles, abstracts, and knowledge domains and used synonyms, I narrowed the
search results. In the third step, the significant research methodologies on fake reviews
are presented. In order to conduct a meaningful search in Web of Science, it was necessary
to analyse frequently used synonyms. In the course of the literature search, besides the
“fake review”, the following synonyms were used: fake reviews, deceptive reviews,
bogus reviews, fraudulent reviews, and spam. In order to find the methods, which were
used for detection or identification of fake reviews, I used the advanced search function
of the Web of Science. The original search term “TS = (Fake OR Spam OR Deceptive
OR Bogus OR Fraudulent) AND (Review*)” was extended by “AND TS = (Detect* OR
Identify*) AND TS = (Method* OR Tool* OR Application* OR Criteria OR
Requirement*)”. Thus, the search result includes tools and software, which can be used
for the identification of fake reviews. Figure 8 reveals the trend in articles on this topic
from 2000 to 2019 showing especially significant growth of publications in 2017 with
significant slowdown in 2018 and 2019.

                                                                                           18
3 Empirical part

Figure 8 Popularity of the topic in academic publications (n=270) (Source: Screenshot from webofscience.com)

After examining titles, abstracts, knowledge domains and topics that thematically related
to the subject of the thesis considering the comparison of methods, tools and software,
almost 270 articles were related to the fake review meaning. Additionally, the references
of these articles have been looked at and some of the sources have been used for the
literature review as well.
The subject of fake review, review spam or opinion spam has been studied by many
authors. Various methods for the identification of spam reviews have been proposed by
different authors. However, a very small number of authors have based fake review
detections on tools. For instance, Ott, Choi, Cardie & Hancock have developed an open
tool reviewsceptic.com43. In their study, they claim that the tool may detect a fake review
with 90% accuracy by using psychological and linguistic criteria.44
Methods dedicated to fake review detection can be grouped into three categories:
     •     Review Centric:
           Most works in this field propose a set of features that will give information about

43
     See Review Sceptic (2013), n. p.
44
     See Ott, M. et.al. (2011-b), pp. 316
                                                                                                         19
3 Empirical part

         the reviews. These features refer to characteristics, sentiment and content extracted
         from the review. Thus, various combinations of clues engineered from review
         content have been studied by Jindal et al.45, Lie et al.46, Ott et al.47, Fei et al.48,
         Shojaee et al.49, Li et al.50 , Ong et al.51, Kim et al.52
     •   Reviewer Centric:
         In terms of reviewer centric, many authors focus their studies on reviewer features.
         These features give information about the behaviour of reviewers and reviewers’
         profile features. For instance, differences between rating and honest reviewer
         meanings have been shown by a group of the authors Savage, Zhang, Yu, Chou,
         and Wang. 53
     •   Time Centric:
         Time centric works use time-related clues to develop time-based patterns such as a
         spam attack in bursts for one product. Many authors suggest well-defined criteria
         as indicators for detecting review spammers in review bursts (e.g. frequency of a
         review). For instance, Fei, Mukherjee, Liu, Castellanos and Ghosh have
         implemented an algorithm which detects a massive number of reviews within a
         short time.54
After examining many articles and studies, it can be concluded, that research in this field
is pretty much dedicated to the detection of fake reviews by proposing different methods
for their analysing. However, a very small number of authors have based their study on
tools.
In terms of a comparison of web tools for fake reviews detection, there is only a negligible
number of authors. For instance, Ana Pieper, in her paper, has suggested an approach,
ReviewAlarm, and compared it with an already developed automated tool for identifying
non-hotel fake reviews, Reviewsceptic. This new approach additionally considers
behaviour based and time-based clues. Finally, based on the previously defined dataset,

45
   See Jindal et. al. (2008-b), n. p.
46
   See Li, F. et. al. (2011), pp.2488–2493.
47
   See Ott, M. et. al. (2011-b), pp. 310
48
   See Fei, G. et.al. (2013a), pp. 176.
49
   See Shojaee, S., et. al., (2013), pp. 55.
50
   See L, and J., et. al. (2014), pp. 1566.
51
   See Ong, T., et. al. (2014), pp. 69–78.
52
   See Kim, S., et. al. (2015), pp. 2–10.
53
   See Savage, D., et. al. (2015), pp. 8650–8657.
54
   See Fei, G. et.al. (2013b), pp. 176.
                                                                                             20
3 Empirical part

the performances of Reviewsceptic and ReviewAlarm were tested.55 Additionally,
Crawford, Khoshgoftaar, Prusa, Richter and Najada have compared the results of the
techniques and methods used for fake review detection based on determined criteria but
without considering any web tools.56
Despite the vast interest in the subject and the wide popularity of some methods, there are
no studies which compared existing fake review detection tools. However, there are
plenty of authors that compare different social media tools according to advanced defined
criteria. For instance, a group of authors presented a systematic approach for the
evaluation criteria of testing tools. Their model relies on the ISO/IEC 9126 quality model
with the addition of other criteria such as functionality, reliability, usability, efficiency
and many others.57 Similarly, several Business Intelligence tools were compared,
whereby a set of features were defined in order to choose the best one.58
Since there is no study on this topic, there is a strong need for comparing fake review
detection tools. This comparison may contribute to a better understanding of the potential
limitations, advantages, and disadvantages of popular free web tools.

3.2.         Objective and research questions

According to many findings, consumers are increasingly relying on web tools in their
decision-making process. Free web tools can empower consumers and bring transparency
to the reviews. Therefore, the main objective of this thesis is to explore, test and analyse
the current Web platforms for fake review detection.
The objectives of this thesis are:
         •    To investigate which tools can be used for spam detection;
         •    To build an extensive list of criteria that can be used to compare free web tools;
         •    To test the reliability of web tools in detecting fake reviews considering a list
              of criteria for that;
         •    To propose recommendations on how to increase the performance of these
              tools.

55
   See Pieper, A. T. (2016), pp. 2.
56
   See Crawford, M., et. al. (2015), n. p.
57
   See Illes, A., et. al. (2005), n. p.
58
   See Brando, A., et. al. (2016), pp. 6–9.
                                                                                             21
3 Empirical part

Therefore, this study aims to answer the following research question:
       1. Which open-access tools for spam detection are currently available?
       2. How strongly do the tools differ from each other?
       3. What are the strengths and weaknesses of free web tools?

3.3.       Methodology

This Chapter provides a detailed description of the procedures and methodology set in
this thesis. Following a broad definition of the problem and its associated objects, the
proposed methodology aims to provide an effective framework for comparing online fake
review tools. More precisely, the analysis of each employed web tool is conducted in
order to distinguish respective contributions and performances of each tool.
For this purpose, the research model consists of the following four steps: selection,
description, comparison and analysis (see Figure 9).

                                                                            Comparison of
   Selection of tools               Description of tools
                                                                               tools

                     Figure 9 Research model of the work (Source: self-presentation)

1. Selection of the tools
The first step of the proposed framework involves the selection of free online tools for
fake review detection according to previously well- defined criteria. All tools that fully
utilize information from the online web platform Amazon.com are included in analysis.
Generally, detection should be presented by providing a report of product reviews. This
means that each tool can be regarded as a filtering system where a product review is an
input, and the result is a report indicates whether a review is truthful or spam. The
selection of the tools is based on the following criteria:
       •    The tool can be used by everyone (sellers, buyers, users) and not only for sellers;
                                                                                                    22
3 Empirical part

          •    It is a free source and grants users permission without any costs to use and study
               detection results;
          •    It should allow further investigation in the form of a produced report and does
               not deliver the result whether a review is truthful or untruthful.

2.        Description of selected web Tools:
Each selected tool has its own individual features and characteristics. Therefore, the
presentation of each tool is followed by pointing out useful characteristics about usability,
scope and depth as well as giving some hints on the functionality of each tool. More
precisely, conducting the description requires the analysis of each tool according to a
previously defined set of criteria. For this purpose, different criteria from different articles
have been looked at and on this basis of different criteria, tools are compared.
The selected tools will be described and compared according to the following set of
criteria:
Criteria 1          Usability
•    Accessibility59
      -       How quickly can users use the tool?
      -       Is registration required?
      -       How detailed is the registration?
•     Supported browsers and devices
      -       Works across different browsers and operating systems?
      -       How many browsers?
      -       Works on devices (mobile phones or tablets) that users prefer to use?
      -       In order to use the tool, does anything have to be downloaded and installed?
•    Ease of navigation/search60
      -       Can users find the fake review detection search bar quickly?
•     User support61
      -       Is there any help function for users?
                o Availability of support and knowledgebase (specific help section, FAQ) for
                     users?

59
   See Hassan, S., et. al. (2001a), pp 4.
60
   See Hassan, S., et. al. (2001b), pp 4.
61
   See Hassan, S., et. al. (2001-c), pp 4.
                                                                                                 23
3 Empirical part

         -    Does the tool offer any options to report possible errors?
                 o Is there any contact address for technical concerns available?
•       Language support62
         -    Language of the reports and user interface
                 o Are the results of the tool supported in different languages?
                 o Which languages are supported?
                 o Is the used language mentioned on the website?
         -    Language of the reviews
                 o Can the tool analyse reviews from different country-based Amazon sites?
                 o Which Amazon sites are supported?

Criteria 2: Scope, depth and forms of the outcomes of the reports
•       Scope
         -    How many different tests does the tool run?
•       Presentation of the results
         -    Is the report well-structured?
         -    How does each tool summarize the results (through letter grade A, B, C, D, E, F
              or through PASS, WARN or FAIL)?
         -    Is it easy to understand the results?
                 o Does the tool use simple sentences for the presentation of the results?
                 o Does the report include multimedia elements (image, voice and video data)
                       for the purpose of results presentation?
•       Depth
         -    Is the report supported by evidence such as statistics? Is each test supported by
              evidence?
         -    Does the report of the tool provide explanations for users to better understand the
              discovered outcomes?

Criteria 3: Developer of the tool
•       Availability of author information
         -    Is it clear who the developer is?

62
     See Nantel, J., et.al. (2008), pp. 112–122.
                                                                                              24
3 Empirical part

Criteria 4: Currency
•     When was the tool developed?
•     When was the tool updated last time? Is there information provided on what was
      changed?
•     Is it possible to conduct an analysis with current data?

Criteria 5: Export and share function
•     Is it possible to export the report?
      -   Which formats can reports be exported?

3. Comparison of selected web Tools:
The next step of the proposed framework is dedicated to the comparison of the selected
tools. More precisely, the described tools will be compared according to the criteria.
Indeed, the performance of each tool reflects the degree of the tool quality. For instance,
the higher the degree of usability, the higher is the quality of the tool. The goal of the
proposed method for comparing different tools should be useful to pinpoint the strengths
and weaknesses of the tools in terms of their effectiveness.

3.4       Results

In this Chapter, each step of the proposed framework will be performed, described and
presented. For that purpose, the product reviews will be used from Amazon.com. All
product reviews will be further divided among five product categories, i.e., from those
with minor reviewing activity to those with substantial reviewing activity. Based on these,
evaluation conclusions will be drawn about tool performance and how it varies for
products with different volumes of reviewing activity.

3.4.1 Selection of the tools

In the selection process, all free online tools that meet all the condition stated above in
Section 3.3 are taken into account. They are:
          •   ReviewMeta.com

                                                                                        25
3 Empirical part

               •     FakeSpot.com
               •     TheReviewIndex.com

     3.4.2 Description of the selected tools

     3.4.2.1 ReviewMeta

     Tommy Noonan has developed a fake review website with the full name
     “ReviewMeta.com - Amazon Review Checker” in May 2016. Inspired by the success of
     his previous website SupplementReview.com for the review of supplement products, he
     developed the ReviewMeta platform to analyse and filter the product reviews transparent
     for consumers before they make a purchasing decision on Amazon.com.63The Web
     address of the platform is reviewmeta.com. The visitors may analyse the product reviews
     and ratings from Amazon.com, Amazon.co.uk, and Amazon.ca and Bodybuilding.64

     3.4.2.1.1 Usability of ReviewMeta

     Accessibility
     The tool is designed to work for all people whatever their hardware, software, location,
     or ability. There are no fees charged for its use. To access the tool, registration is not
     required. It is suggested only for a full report in order to be sent to a visitor’s email address
     for free. Also, for the purpose of using the forum section, the registration of the user is
     required.
     Analysing the extent to which the tool is compatible with different browsers, it was found
     that the content and design of the website work on the following web browsers:
               -     Internet Explorer 11,
               -     Google Chrome 72,
               -     FireFox 65,
               -     Opera 58,
               -     Safari 5.
     For a full analysis, this service works by copying any Amazon product URL into the

63   See ReviewMeta.com Press & Media - ReviewMeta Blog. (2018a), n. p.
     64 See ReviewMeta.com Press & Media - ReviewMeta Blog. (2018b), n. p.

                                                                                                   26
3 Empirical part

search bar on ReviewMeta.com or by using the ReviewMeta browser extension (Chrome,
Firefox, Safari or Explorer) in the toolbar. The tool is designed to accommodate also
portable devices (mobile sites) for Android and IOS. To use an extension, users must
only download it, and the extension is added to an offered browser. This allows users to
read an adjusted rating of the product directly during the visit of the Amazon website (see
Figure 10). The adjusted rating is colour based on the authenticity of reviews. Red colour
indicates unnatural behaviour of reviews (fake reviews), whereas green coloured adjusted
ratings refer to the reviews which passed.
Yellow coloured adjusted rating suggests neutral behaviour of reviews. To see a detailed
report, users should click the ReviewMeta extension in the toolbar.

              Figure 10 ReviewMeta extension (Source: Screenshot from reviewmeta.com)

Ease of navigation/search
Navigating around ReviewMeta is not difficult as there are two search bars available.
They are very distinctive, and it is easy to find them. One is located at the top of the page
and the other at the centre of the home page. This enables users an intuitive, simple and
easy to use system (see Figure 11). If users analyse product reviews with the browser
extension, they can quickly scan an adjusted rating in a horizontal movement across the
upper part of the ReviewMeta page.

                                                                                               27
3 Empirical part

      Figure 11 The URL search bar on RevieMeta website (Source: Screenshot on reviewmeta.com)

Interactivity and user support
Although there is no particular “help section” provided on this website, the
ReviewMetatool provides several help links or options available for user support. For
instance, how visitors should use the tool is described in the “How it works” section,
located at the top of the website. Additionally, a detailed knowledge base is also available
in its “FAQ” section which helps users to find their own answers, explanations and
concerns related to the platform and results of the report (see Figure 12). Also, on the
home page of the tool, it is clearly described how to use the service.

  Figure 12 The FAQ browser window on ReviewMeta website (Source: Screenshot from reviewmeta.com)
                                                                                                 28
You can also read