SlideShare une entreprise Scribd logo
1  sur  24
Télécharger pour lire hors ligne
FAKE REVIEW DETECTION
What are they saying about you ? Are they real …
Guided By : Dr. Animesh Mukherjee
ONLINE REVIEW
• Captures testimonials of “real” people (unlike advertisements).
• Shapes decision making of customers.
• Positive reviews – financial gains and fame for business.
• Deceptive opinion spamming – to promote or discredit some target product
and services.[1]
• Opinion Spammers admitted for being paid to write fake reviews.(Kost, 2012)
• Yelp.com – “Sting operation” : publically shame business who buy fake
reviews.[2]
[1] Jindal and Liu 2008 : http://www.cs.uic.edu/~liub/FBS/opinion-spam-WSDM-08.pdf
[2] Yelp official blog : http://officialblog.yelp.com/2013/05/how-yelp-protects-consumers-from-fake-reviews.html
Consumer Alerts !!
• Amazon Mechanical Turk – crowd sourced online workers(turkers) to write fake
reviews ($1 per review) portraying 20 hotels of Chicago in positive light.[1]
• 400 fake positive reviews collected and 400 non-fake reviews on same 20 hotels
using Tripadvisor.com
• Yelp’s filtered and unfiltered reviews collected to understand working of Yelp
review classification algorithm.
• Approach : Linguistic n-gram features and some supervised learning method.[2]
[1] Amazon Mechanical Turk : https://www.mturk.com/
[2] Ott et al. 2011 : https://www.cs.cornell.edu/courses/CS4740/2012sp/lectures/op_spamACL2011.pdf
Dataset Collection and Approach for Analysis
Linguistic Approach : Results
• Using only Bigram features :
accuracy of 89.6% on AMT data.[1]
• Using same n-gram features :
accuracy of 67.8% on Yelp Data.[2]
• Table 1: Class distribution of Yelp
data is skewed – imbalanced data
produces poor model.(Chawla et al,
2004)
• Good model for imbalanced data –
under sampling. (Drummond and
Holte, 2003)
[1] Ott et al. 2011 : https://www.cs.cornell.edu/courses/CS4740/2012sp/lectures/op_spamACL2011.pdf
[2] A Mukherjee - 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf
Linguistic Approach Results : Explained
• For AMT data, word distribution of fake and non-fake reviews are very different,
which explains the high accuracy using n-gram.
• Reason for the different word distribution : Domain knowledge absence and little
gain in writing fake reviews($1 per review).
• Poor performance of n-grams for Yelp data because spammers according to
Yelp filter used very similar language in fake review as non-fake review :
linguistically very similar.[1]
• Inefficiency in linguistic in detecting fake reviews filtered by Yelp encourages to
do behavioral study of reviews.
[1] Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf
))(/)((2log)()||( iNiFiFNFKL i
Information Theoretic Analysis
• To explain huge accuracy difference : analysis of word distribution of AMT and
Yelp data.
• Good-Turing smoothed unigram language models.
• Computation of word distribution difference across fake & non-fake reviews :
Kullback-Leibler(KL) Divergence :
where F(i) and N(i) are respective probabilities of word i in fake and non-fake reviews.[1]
• Here KL(F||N) gives quantitative estimate of how much fake reviews linguistically
differ from non-fake reviews
[1] Kullback Leibler Divergence : http://www.cs.buap.mx/~dpinto/research/CICLing07_1/Pinto06c/node2.html
Information Theoretic Analysis (Cont..)
• KL Divergence is Asymmetric :
• We have,
• For AMT data : and
• For Yelp data : and
• [1]
   FNKLNFKL |||| 
   FNKLNFKL ||||  0KL
   FNKLNFKL ||||  0KL
[1] Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf
   FNKLNFKLKL |||| 
Information Theoretic Analysis (Cont..)
• KL(F||N) Definition implies that words having higher probability in F and very low
probability in N contribute most to KL-Divergence.
• To study word-wise contribution to ∆KL, word ∆KL calculated as :
where,
• Contribution of top k words to ∆KL for k= 200 and k=300
)||()||( iiwordiiword
i
Word FNKLNFKLKL 
  






)(
)(
log)||( 2
iN
iF
iFNFKL iiword
Turkers didn’t do a good job at “Faking”!
Word-wise Difference of KL-
Divergence across top 200 words.
Equally Dense : |E| = |G|
• Symmetric Distribution of for top k-words for AMT
data implies the existence of two set of words:
1. set of words E, appearing more in fake reviews than
in non-fake, Ɐi € E, F(i) > N(i) resulting in
2. set of words G, appearing more in non-fake reviews
than in fake, Ɐi € G, N(i) > F(i) resulting in
i
WordKL
0 Ei
WordKL
0 Ei
WordKL
• Additionally top k=200 words only contribute 20% to ∆KL for the AMT data.
There are many words in AMT data having higher probabilities in fake than non-
fake and vice-versa.
• This implies fake and non-fake reviews in AMT data consist of words with very
different frequencies. Turkers didn’t do a good job in faking.[1]
[1] Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf
Yelp Spammers are Smart but Overdid “Faking”!
• Yelp fake review data shows KL(F||N) is much larger than KL(N||F) and ∆KL>1.
• Given graphs from b-e shows among top 200 words which contribute to major
(=70%) most words have and only a few have .
Word-wise Difference of KL-Divergence across top 200 words.[1]
0 i
WordKL 0 i
WordKL
[1] Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf
Yelp Spammers are Smart but Overdid “Faking”! (Cont..)
• Consider A be set of top words contributing most to ∆KL. We partition A as:
where and
• The curve above y=0 is dense and below it is sparse which implies
• clearly indicates that there exists specific words which contribute
most to ∆KL by appearing in fake reviews with much higher frequencies than
in non-fake reviews.
• Spammers made smart effort to ensure that their fake reviews have most
words that also appear in non-fake reviews to sound convincing.
 ))()(,.,(0| iFiNAiieKLiA Ni
Word
N

 ))()(,.,(0| iNiFAiieKLiA Fi
Word
F

NF
AAA   NF
AA
NF
AA 
NF
AA 
Yelp Spammers are Smart but Overdid “Faking”! (Cont..)
• While making their reviews sound convincing, psychologically they
happened to OVERUSE come words resulting higher frequency of certain
words in fake review than non-fake reviews.
• A quick lookup yields {us, price, stay, feel, deal, comfort} in hotel domain and
{options, went, seat, helpful, overall, serve, amount etc.} in restaurants.
• Prior personality work shown that deception/lying usually involves more use
of personal pronouns(eg. us) and associated action(eg. Went, feel) towards
specific targets(eg. option, price, stay) with objective of incorrect
projection(lying or faking).[1]
• Spammers caught by Yelp left behind linguistic footprints which can be
caught by precise behavioral study.
[1] Newman et al , 2003 : http://www.communicationcache.com/uploads/1/0/8/8/10887248/lying_words-_predicting_deception_from_linguistic_styles.pdf
Spamming Behavior Analysis
1. Maximum Number of Reviews(MNR) : Writing too many reviews in a day is
abnormal.
2. Percentage of Positive Reviews(PR) : Deception words in fake reviews
indicates projection in positive light. CDF of positive(4-5 stars) reviews
among all reviews is plotted to illustrate analysis.
3. Review Length(RL) : While writing fake experiences, there is probably not
much to write and also spammer not want to spend too much time in it.
4. Maximum Content Similarity(MCS) : To examine if some posted reviews are
similar to previous reviews, we computed cosine similarity between two
reviews of a reviewer. Non-spammers mostly write new contents.
Maximum Number of Reviews Percentage of Positive Reviews
Review Length Maximum Content Similarity
Spamming Behavior Analysis CDFs
[1] Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf
Challenges with Supervised Evaluation
• Very difficult to find gold-standard data of fake and non-fake reviews for
model building – too difficult to manually recognize/label fake/non-fake
reviews by mere reading.
• Duplicate and near duplicate assumed to be fake, which is unreliable.
• Usage of manually labeled dataset – reliability issues because it have been
shown that accuracy of human labeling of fake reviews is very poor.[1]
• AMT crowdsourcing fake reviews by paying are fake yet they don’t reflect the
dynamics of fake reviews in commercial website.[2]
• This lack of labeled data, motivates to look after the unsupervised methods of
classification.[3]
[1] Ott M, Choi, Y, Cardie, C. and Hancock, J.T. 2011. Finding Deceptive Opinion Spam by Any Stretch of the Imagination.
[2] Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf
[3] Mukherjee et al : http://delivery.acm.org/10.1145/2490000/2487580/p632-mukherjee.pdf
Unsupervised Evaluation Model
• Since human labeling for supervised learning is difficult, problem was
proposed by modeling spamicity (degree of spamming) as latent with other
observed behavioral features.
• Unsupervised model – Author Spamicity Model (ASM) proposed.[1]
• Taken a fully Bayesian approach and formulated opinion spam detection as
clustering problem.
• Opinion spammers have different behavioral distributions than non-spammers.
• Causes distributional divergence between latent population distributions of two
clusters: spammers and non-spammers.[1]
• Model inference results in learning the population distributions of two clusters.
[1] Mukherjee et al : http://delivery.acm.org/10.1145/2490000/2487580/p632-mukherjee.pdf
• Formulates spam detection as an unsupervised clustering problem in
Bayesian setting.
• Belongs to class of Generative models for clustering based on set of
observed features.
• Models spamicity 𝑠 𝑎
(in range[0, 1]) of an author a; and spam label 𝜋 𝑟
of a
review which is the class variable reflecting the cluster membership (two
cluster K=2, spam and non-spam).[1]
• Each author/reviewer and respectively each review has a set of observed
features(behavioral clues).
• Certain characteristics of abnormal behavior defined which likely to link
with spamming and thus can be exploited in model for learning spam and
non-spam clusters. [1]
Author Spamicity Model
[1] Mukherjee et al : http://delivery.acm.org/10.1145/2490000/2487580/p632-mukherjee.pdf
Author Features
• It have value in range [0, 1] and value close to 1 indicates spamming.
1. Content Similarity : Crafting new review every time is time consuming, spammers
likely to copy reviews across similar product. Choose maximum similarity to
capture worst spamming behavior.[1]
2. Maximum Number of Reviews : Posting many reviews a day is also abnormal.
3. Reviewing Burstiness : Spammers are usually not longtime members of site.
Defined over an activity window(first and last review posting date). If posted over
reasonably long timeframe, it probably a normal activity but all review posted
within short burst likely to be spam.[2]
4. Ratio of First Review : People mostly rely on early reviews and spamming early
impacts hugely on sales. Spammers try to be among first reviewers. [1]
[1] Mukherjee et al : http://delivery.acm.org/10.1145/2490000/2487580/p632-mukherjee.pdf
[2] Mukherjee, A., Liu, B. and Glance, N. 2012. Spotting Fake Reviewer Groups in Consumer Reviews. WWW (2012).
Review Features
• It have five binary review features. Value 1 indicates spamming else 0 is non-spamming.
1. Duplicate /Near Duplicate Reviews: Spammers often post multiple reviews which are
duplicate /near duplicate on same product to boost ratings.
2. Extreme Rating: Spammers mostly like to give extreme ratings(1 or 5) in order to boost
ratings to demote or promote products.
3. Rating Deviation: Spammers usually involve in wrong projection in either positive or
negative and it deviates from average ratings given by other reviewers.
4. Early Time Frame: Early review can greatly impact people’s sentiments on a product.
5. Rating Abuse: Multiple ratings on same product are unusual. Similar to DUP but focuses
on rating dimensions rather than content.
CONCLUSIONS
• We presented an in-depth investigation of nature of fake reviews in Commercial
settings of Yelp.com.
• Our study shows linguistics methods of (Ott et al., 2011) and its high accuracy in
AMT data.
• We presented a behavioral study of spammers for real-life fake reviews.
• We presented a brief introduction to n-gram language model.
• We presented challenges with the supervised evaluation and gave idea about the
unsupervised approach of evaluation.
• We presented a brief introduction to unsupervised Author Spamicity Model (ASM).
REFERENCES
1. Jindal and Liu 2008 : http://www.cs.uic.edu/~liub/FBS/opinion-spam-WSDM-08.pdf
2. Yelp official blog : http://officialblog.yelp.com/2013/05/how-yelp-protects-consumers-from-fake-
reviews.html
3. MIT N-gram Language Model Tutorial:
http://web.mit.edu/6.863/www/fall2012/readings/ngrampages.pdf
4. Amazon Mechanical Turk : https://www.mturk.com/
5. Ott et al. 2011 :
https://www.cs.cornell.edu/courses/CS4740/2012sp/lectures/op_spamACL2011.pdf
6. Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-
submit.pdf
7. Kullback Leibler Divergence :
http://www.cs.buap.mx/~dpinto/research/CICLing07_1/Pinto06c/node2.html
8. Newman et al , 2003 :
http://www.communicationcache.com/uploads/1/0/8/8/10887248/lying_words-
_predicting_deception_from_linguistic_styles.pdf
9. Mukherjee, A., Liu, B. and Glance, N. 2012. Spotting Fake Reviewer Groups in Consumer
Reviews. WWW (2012).
10. Mukherjee et al : http://delivery.acm.org/10.1145/2490000/2487580/p632-mukherjee.pdf
11. Ott M, Choi, Y, Cardie, C. and Hancock, J.T. 2011. Finding Deceptive Opinion Spam by Any
Stretch of the Imagination.
Any Question ?
Thank You !!

Contenu connexe

Tendances

Amazon Product Sentiment review
Amazon Product Sentiment reviewAmazon Product Sentiment review
Amazon Product Sentiment reviewLalit Jain
 
Machine Learning Project
Machine Learning ProjectMachine Learning Project
Machine Learning ProjectAbhishek Singh
 
Fake news detection project
Fake news detection projectFake news detection project
Fake news detection projectHarshdaGhai
 
Practical sentiment analysis
Practical sentiment analysisPractical sentiment analysis
Practical sentiment analysisDiana Maynard
 
Movie recommendation project
Movie recommendation projectMovie recommendation project
Movie recommendation projectAbhishek Jaisingh
 
Facial Emotion Recognition: A Deep Learning approach
Facial Emotion Recognition: A Deep Learning approachFacial Emotion Recognition: A Deep Learning approach
Facial Emotion Recognition: A Deep Learning approachAshwinRachha
 
Detecting Fake News Through NLP
Detecting Fake News Through NLPDetecting Fake News Through NLP
Detecting Fake News Through NLPSakha Global
 
Voice based email for blinds
Voice based email for blindsVoice based email for blinds
Voice based email for blindsArjun AJ
 
Diabetes prediction using machine learning
Diabetes prediction using machine learningDiabetes prediction using machine learning
Diabetes prediction using machine learningdataalcott
 
Presentation on Sentiment Analysis
Presentation on Sentiment AnalysisPresentation on Sentiment Analysis
Presentation on Sentiment AnalysisRebecca Williams
 
Handwritten Digit Recognition(Convolutional Neural Network) PPT
Handwritten Digit Recognition(Convolutional Neural Network) PPTHandwritten Digit Recognition(Convolutional Neural Network) PPT
Handwritten Digit Recognition(Convolutional Neural Network) PPTRishabhTyagi48
 
Machine Learning Project - Email Spam Filtering using Enron Dataset
Machine Learning Project - Email Spam Filtering using Enron DatasetMachine Learning Project - Email Spam Filtering using Enron Dataset
Machine Learning Project - Email Spam Filtering using Enron DatasetAman Singhla
 
Human age and gender Detection
Human age and gender DetectionHuman age and gender Detection
Human age and gender DetectionAbhiAchalla
 
Introduction to emotion detection
Introduction to emotion detectionIntroduction to emotion detection
Introduction to emotion detectionTyler Schnoebelen
 
Machine Learning presentation.
Machine Learning presentation.Machine Learning presentation.
Machine Learning presentation.butest
 
Stock Market Prediction using Machine Learning
Stock Market Prediction using Machine LearningStock Market Prediction using Machine Learning
Stock Market Prediction using Machine LearningAravind Balaji
 

Tendances (20)

EMOTION DETECTION USING AI
EMOTION DETECTION USING AIEMOTION DETECTION USING AI
EMOTION DETECTION USING AI
 
Amazon Product Sentiment review
Amazon Product Sentiment reviewAmazon Product Sentiment review
Amazon Product Sentiment review
 
Machine Learning Project
Machine Learning ProjectMachine Learning Project
Machine Learning Project
 
Housing price prediction
Housing price predictionHousing price prediction
Housing price prediction
 
Fake news detection project
Fake news detection projectFake news detection project
Fake news detection project
 
Practical sentiment analysis
Practical sentiment analysisPractical sentiment analysis
Practical sentiment analysis
 
Movie recommendation project
Movie recommendation projectMovie recommendation project
Movie recommendation project
 
Facial Emotion Recognition: A Deep Learning approach
Facial Emotion Recognition: A Deep Learning approachFacial Emotion Recognition: A Deep Learning approach
Facial Emotion Recognition: A Deep Learning approach
 
Detecting Fake News Through NLP
Detecting Fake News Through NLPDetecting Fake News Through NLP
Detecting Fake News Through NLP
 
Voice based email for blinds
Voice based email for blindsVoice based email for blinds
Voice based email for blinds
 
Diabetes prediction using machine learning
Diabetes prediction using machine learningDiabetes prediction using machine learning
Diabetes prediction using machine learning
 
Presentation on Sentiment Analysis
Presentation on Sentiment AnalysisPresentation on Sentiment Analysis
Presentation on Sentiment Analysis
 
Handwritten Digit Recognition(Convolutional Neural Network) PPT
Handwritten Digit Recognition(Convolutional Neural Network) PPTHandwritten Digit Recognition(Convolutional Neural Network) PPT
Handwritten Digit Recognition(Convolutional Neural Network) PPT
 
Machine Learning Project - Email Spam Filtering using Enron Dataset
Machine Learning Project - Email Spam Filtering using Enron DatasetMachine Learning Project - Email Spam Filtering using Enron Dataset
Machine Learning Project - Email Spam Filtering using Enron Dataset
 
FAKE NEWS DETECTION (1).pptx
FAKE NEWS DETECTION (1).pptxFAKE NEWS DETECTION (1).pptx
FAKE NEWS DETECTION (1).pptx
 
Human age and gender Detection
Human age and gender DetectionHuman age and gender Detection
Human age and gender Detection
 
Introduction to emotion detection
Introduction to emotion detectionIntroduction to emotion detection
Introduction to emotion detection
 
Sms spam-detection
Sms spam-detectionSms spam-detection
Sms spam-detection
 
Machine Learning presentation.
Machine Learning presentation.Machine Learning presentation.
Machine Learning presentation.
 
Stock Market Prediction using Machine Learning
Stock Market Prediction using Machine LearningStock Market Prediction using Machine Learning
Stock Market Prediction using Machine Learning
 

Similaire à FAKE Review Detection

Fraud Detection in Online Reviews using Machine Learning Techniques
Fraud Detection in Online Reviews using Machine Learning TechniquesFraud Detection in Online Reviews using Machine Learning Techniques
Fraud Detection in Online Reviews using Machine Learning Techniquesijceronline
 
The study of factors affecting customer’s satisfaction with the three star ho...
The study of factors affecting customer’s satisfaction with the three star ho...The study of factors affecting customer’s satisfaction with the three star ho...
The study of factors affecting customer’s satisfaction with the three star ho...INFOGAIN PUBLICATION
 
Digital technology analysis
Digital technology analysisDigital technology analysis
Digital technology analysis66xiu
 
Yelp an Assessment of Customer Satisfaction
Yelp an Assessment of Customer SatisfactionYelp an Assessment of Customer Satisfaction
Yelp an Assessment of Customer Satisfactionsunnycmui
 
Spot deceptive TripAdvisor Reviews
Spot deceptive TripAdvisor ReviewsSpot deceptive TripAdvisor Reviews
Spot deceptive TripAdvisor ReviewsYousef Fadila
 
Using your User Experience (UX) Super Powers for Good or Evil - Theo Mandel, ...
Using your User Experience (UX) Super Powers for Good or Evil - Theo Mandel, ...Using your User Experience (UX) Super Powers for Good or Evil - Theo Mandel, ...
Using your User Experience (UX) Super Powers for Good or Evil - Theo Mandel, ...Theo Mandel, PhD
 
Yelp's Review Filtering Algorithm Paper
Yelp's Review Filtering Algorithm PaperYelp's Review Filtering Algorithm Paper
Yelp's Review Filtering Algorithm PaperYao Yao
 
IRJET- Implementation of Review Selection using Deep Learning
IRJET-  	  Implementation of Review Selection using Deep LearningIRJET-  	  Implementation of Review Selection using Deep Learning
IRJET- Implementation of Review Selection using Deep LearningIRJET Journal
 
IRJET- Fake Review Detection using Opinion Mining
IRJET- Fake Review Detection using Opinion MiningIRJET- Fake Review Detection using Opinion Mining
IRJET- Fake Review Detection using Opinion MiningIRJET Journal
 
IRJET- Improving Performance of Fake Reviews Detection in Online Review’s usi...
IRJET- Improving Performance of Fake Reviews Detection in Online Review’s usi...IRJET- Improving Performance of Fake Reviews Detection in Online Review’s usi...
IRJET- Improving Performance of Fake Reviews Detection in Online Review’s usi...IRJET Journal
 
Digital presentation yelp
Digital presentation yelpDigital presentation yelp
Digital presentation yelpwdgg307
 
Pitfalls and potholes of content moderation for chatbots, Elayne Ruane
Pitfalls and potholes of content moderation for chatbots, Elayne RuanePitfalls and potholes of content moderation for chatbots, Elayne Ruane
Pitfalls and potholes of content moderation for chatbots, Elayne RuaneAlan Quayle
 
Capturing and managing non-traditional information for Competitive Advantage....
Capturing and managing non-traditional information for Competitive Advantage....Capturing and managing non-traditional information for Competitive Advantage....
Capturing and managing non-traditional information for Competitive Advantage....Traction Software
 

Similaire à FAKE Review Detection (20)

Fraud Detection in Online Reviews using Machine Learning Techniques
Fraud Detection in Online Reviews using Machine Learning TechniquesFraud Detection in Online Reviews using Machine Learning Techniques
Fraud Detection in Online Reviews using Machine Learning Techniques
 
The study of factors affecting customer’s satisfaction with the three star ho...
The study of factors affecting customer’s satisfaction with the three star ho...The study of factors affecting customer’s satisfaction with the three star ho...
The study of factors affecting customer’s satisfaction with the three star ho...
 
Digital technology analysis
Digital technology analysisDigital technology analysis
Digital technology analysis
 
CUROP Poster (002)
CUROP Poster (002)CUROP Poster (002)
CUROP Poster (002)
 
Yelp an Assessment of Customer Satisfaction
Yelp an Assessment of Customer SatisfactionYelp an Assessment of Customer Satisfaction
Yelp an Assessment of Customer Satisfaction
 
Spot deceptive TripAdvisor Reviews
Spot deceptive TripAdvisor ReviewsSpot deceptive TripAdvisor Reviews
Spot deceptive TripAdvisor Reviews
 
Marriott Interviews
Marriott InterviewsMarriott Interviews
Marriott Interviews
 
Mahendra nath
Mahendra nathMahendra nath
Mahendra nath
 
Sm chap 1
Sm chap 1Sm chap 1
Sm chap 1
 
Using your User Experience (UX) Super Powers for Good or Evil - Theo Mandel, ...
Using your User Experience (UX) Super Powers for Good or Evil - Theo Mandel, ...Using your User Experience (UX) Super Powers for Good or Evil - Theo Mandel, ...
Using your User Experience (UX) Super Powers for Good or Evil - Theo Mandel, ...
 
Yelp's Review Filtering Algorithm Paper
Yelp's Review Filtering Algorithm PaperYelp's Review Filtering Algorithm Paper
Yelp's Review Filtering Algorithm Paper
 
Can You Identify Fake or Authentic Reviews? An fsQCA Approach
Can You Identify Fake or Authentic Reviews? An fsQCA ApproachCan You Identify Fake or Authentic Reviews? An fsQCA Approach
Can You Identify Fake or Authentic Reviews? An fsQCA Approach
 
IRJET- Implementation of Review Selection using Deep Learning
IRJET-  	  Implementation of Review Selection using Deep LearningIRJET-  	  Implementation of Review Selection using Deep Learning
IRJET- Implementation of Review Selection using Deep Learning
 
IRJET- Fake Review Detection using Opinion Mining
IRJET- Fake Review Detection using Opinion MiningIRJET- Fake Review Detection using Opinion Mining
IRJET- Fake Review Detection using Opinion Mining
 
IRJET- Improving Performance of Fake Reviews Detection in Online Review’s usi...
IRJET- Improving Performance of Fake Reviews Detection in Online Review’s usi...IRJET- Improving Performance of Fake Reviews Detection in Online Review’s usi...
IRJET- Improving Performance of Fake Reviews Detection in Online Review’s usi...
 
Greg Sterling - Advanced Search Summit Napa 2021
Greg Sterling - Advanced Search Summit Napa 2021Greg Sterling - Advanced Search Summit Napa 2021
Greg Sterling - Advanced Search Summit Napa 2021
 
Prevalence of Deception in Online Customer Reviews
Prevalence of Deception in Online Customer ReviewsPrevalence of Deception in Online Customer Reviews
Prevalence of Deception in Online Customer Reviews
 
Digital presentation yelp
Digital presentation yelpDigital presentation yelp
Digital presentation yelp
 
Pitfalls and potholes of content moderation for chatbots, Elayne Ruane
Pitfalls and potholes of content moderation for chatbots, Elayne RuanePitfalls and potholes of content moderation for chatbots, Elayne Ruane
Pitfalls and potholes of content moderation for chatbots, Elayne Ruane
 
Capturing and managing non-traditional information for Competitive Advantage....
Capturing and managing non-traditional information for Competitive Advantage....Capturing and managing non-traditional information for Competitive Advantage....
Capturing and managing non-traditional information for Competitive Advantage....
 

FAKE Review Detection

  • 1. FAKE REVIEW DETECTION What are they saying about you ? Are they real … Guided By : Dr. Animesh Mukherjee
  • 2. ONLINE REVIEW • Captures testimonials of “real” people (unlike advertisements). • Shapes decision making of customers. • Positive reviews – financial gains and fame for business. • Deceptive opinion spamming – to promote or discredit some target product and services.[1] • Opinion Spammers admitted for being paid to write fake reviews.(Kost, 2012) • Yelp.com – “Sting operation” : publically shame business who buy fake reviews.[2] [1] Jindal and Liu 2008 : http://www.cs.uic.edu/~liub/FBS/opinion-spam-WSDM-08.pdf [2] Yelp official blog : http://officialblog.yelp.com/2013/05/how-yelp-protects-consumers-from-fake-reviews.html
  • 4. • Amazon Mechanical Turk – crowd sourced online workers(turkers) to write fake reviews ($1 per review) portraying 20 hotels of Chicago in positive light.[1] • 400 fake positive reviews collected and 400 non-fake reviews on same 20 hotels using Tripadvisor.com • Yelp’s filtered and unfiltered reviews collected to understand working of Yelp review classification algorithm. • Approach : Linguistic n-gram features and some supervised learning method.[2] [1] Amazon Mechanical Turk : https://www.mturk.com/ [2] Ott et al. 2011 : https://www.cs.cornell.edu/courses/CS4740/2012sp/lectures/op_spamACL2011.pdf Dataset Collection and Approach for Analysis
  • 5. Linguistic Approach : Results • Using only Bigram features : accuracy of 89.6% on AMT data.[1] • Using same n-gram features : accuracy of 67.8% on Yelp Data.[2] • Table 1: Class distribution of Yelp data is skewed – imbalanced data produces poor model.(Chawla et al, 2004) • Good model for imbalanced data – under sampling. (Drummond and Holte, 2003) [1] Ott et al. 2011 : https://www.cs.cornell.edu/courses/CS4740/2012sp/lectures/op_spamACL2011.pdf [2] A Mukherjee - 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf
  • 6. Linguistic Approach Results : Explained • For AMT data, word distribution of fake and non-fake reviews are very different, which explains the high accuracy using n-gram. • Reason for the different word distribution : Domain knowledge absence and little gain in writing fake reviews($1 per review). • Poor performance of n-grams for Yelp data because spammers according to Yelp filter used very similar language in fake review as non-fake review : linguistically very similar.[1] • Inefficiency in linguistic in detecting fake reviews filtered by Yelp encourages to do behavioral study of reviews. [1] Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf
  • 7. ))(/)((2log)()||( iNiFiFNFKL i Information Theoretic Analysis • To explain huge accuracy difference : analysis of word distribution of AMT and Yelp data. • Good-Turing smoothed unigram language models. • Computation of word distribution difference across fake & non-fake reviews : Kullback-Leibler(KL) Divergence : where F(i) and N(i) are respective probabilities of word i in fake and non-fake reviews.[1] • Here KL(F||N) gives quantitative estimate of how much fake reviews linguistically differ from non-fake reviews [1] Kullback Leibler Divergence : http://www.cs.buap.mx/~dpinto/research/CICLing07_1/Pinto06c/node2.html
  • 8. Information Theoretic Analysis (Cont..) • KL Divergence is Asymmetric : • We have, • For AMT data : and • For Yelp data : and • [1]    FNKLNFKL ||||     FNKLNFKL ||||  0KL    FNKLNFKL ||||  0KL [1] Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf    FNKLNFKLKL |||| 
  • 9. Information Theoretic Analysis (Cont..) • KL(F||N) Definition implies that words having higher probability in F and very low probability in N contribute most to KL-Divergence. • To study word-wise contribution to ∆KL, word ∆KL calculated as : where, • Contribution of top k words to ∆KL for k= 200 and k=300 )||()||( iiwordiiword i Word FNKLNFKLKL           )( )( log)||( 2 iN iF iFNFKL iiword
  • 10. Turkers didn’t do a good job at “Faking”! Word-wise Difference of KL- Divergence across top 200 words. Equally Dense : |E| = |G| • Symmetric Distribution of for top k-words for AMT data implies the existence of two set of words: 1. set of words E, appearing more in fake reviews than in non-fake, Ɐi € E, F(i) > N(i) resulting in 2. set of words G, appearing more in non-fake reviews than in fake, Ɐi € G, N(i) > F(i) resulting in i WordKL 0 Ei WordKL 0 Ei WordKL • Additionally top k=200 words only contribute 20% to ∆KL for the AMT data. There are many words in AMT data having higher probabilities in fake than non- fake and vice-versa. • This implies fake and non-fake reviews in AMT data consist of words with very different frequencies. Turkers didn’t do a good job in faking.[1] [1] Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf
  • 11. Yelp Spammers are Smart but Overdid “Faking”! • Yelp fake review data shows KL(F||N) is much larger than KL(N||F) and ∆KL>1. • Given graphs from b-e shows among top 200 words which contribute to major (=70%) most words have and only a few have . Word-wise Difference of KL-Divergence across top 200 words.[1] 0 i WordKL 0 i WordKL [1] Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf
  • 12. Yelp Spammers are Smart but Overdid “Faking”! (Cont..) • Consider A be set of top words contributing most to ∆KL. We partition A as: where and • The curve above y=0 is dense and below it is sparse which implies • clearly indicates that there exists specific words which contribute most to ∆KL by appearing in fake reviews with much higher frequencies than in non-fake reviews. • Spammers made smart effort to ensure that their fake reviews have most words that also appear in non-fake reviews to sound convincing.  ))()(,.,(0| iFiNAiieKLiA Ni Word N   ))()(,.,(0| iNiFAiieKLiA Fi Word F  NF AAA   NF AA NF AA  NF AA 
  • 13. Yelp Spammers are Smart but Overdid “Faking”! (Cont..) • While making their reviews sound convincing, psychologically they happened to OVERUSE come words resulting higher frequency of certain words in fake review than non-fake reviews. • A quick lookup yields {us, price, stay, feel, deal, comfort} in hotel domain and {options, went, seat, helpful, overall, serve, amount etc.} in restaurants. • Prior personality work shown that deception/lying usually involves more use of personal pronouns(eg. us) and associated action(eg. Went, feel) towards specific targets(eg. option, price, stay) with objective of incorrect projection(lying or faking).[1] • Spammers caught by Yelp left behind linguistic footprints which can be caught by precise behavioral study. [1] Newman et al , 2003 : http://www.communicationcache.com/uploads/1/0/8/8/10887248/lying_words-_predicting_deception_from_linguistic_styles.pdf
  • 14. Spamming Behavior Analysis 1. Maximum Number of Reviews(MNR) : Writing too many reviews in a day is abnormal. 2. Percentage of Positive Reviews(PR) : Deception words in fake reviews indicates projection in positive light. CDF of positive(4-5 stars) reviews among all reviews is plotted to illustrate analysis. 3. Review Length(RL) : While writing fake experiences, there is probably not much to write and also spammer not want to spend too much time in it. 4. Maximum Content Similarity(MCS) : To examine if some posted reviews are similar to previous reviews, we computed cosine similarity between two reviews of a reviewer. Non-spammers mostly write new contents.
  • 15. Maximum Number of Reviews Percentage of Positive Reviews Review Length Maximum Content Similarity Spamming Behavior Analysis CDFs [1] Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf
  • 16. Challenges with Supervised Evaluation • Very difficult to find gold-standard data of fake and non-fake reviews for model building – too difficult to manually recognize/label fake/non-fake reviews by mere reading. • Duplicate and near duplicate assumed to be fake, which is unreliable. • Usage of manually labeled dataset – reliability issues because it have been shown that accuracy of human labeling of fake reviews is very poor.[1] • AMT crowdsourcing fake reviews by paying are fake yet they don’t reflect the dynamics of fake reviews in commercial website.[2] • This lack of labeled data, motivates to look after the unsupervised methods of classification.[3] [1] Ott M, Choi, Y, Cardie, C. and Hancock, J.T. 2011. Finding Deceptive Opinion Spam by Any Stretch of the Imagination. [2] Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf [3] Mukherjee et al : http://delivery.acm.org/10.1145/2490000/2487580/p632-mukherjee.pdf
  • 17. Unsupervised Evaluation Model • Since human labeling for supervised learning is difficult, problem was proposed by modeling spamicity (degree of spamming) as latent with other observed behavioral features. • Unsupervised model – Author Spamicity Model (ASM) proposed.[1] • Taken a fully Bayesian approach and formulated opinion spam detection as clustering problem. • Opinion spammers have different behavioral distributions than non-spammers. • Causes distributional divergence between latent population distributions of two clusters: spammers and non-spammers.[1] • Model inference results in learning the population distributions of two clusters. [1] Mukherjee et al : http://delivery.acm.org/10.1145/2490000/2487580/p632-mukherjee.pdf
  • 18. • Formulates spam detection as an unsupervised clustering problem in Bayesian setting. • Belongs to class of Generative models for clustering based on set of observed features. • Models spamicity 𝑠 𝑎 (in range[0, 1]) of an author a; and spam label 𝜋 𝑟 of a review which is the class variable reflecting the cluster membership (two cluster K=2, spam and non-spam).[1] • Each author/reviewer and respectively each review has a set of observed features(behavioral clues). • Certain characteristics of abnormal behavior defined which likely to link with spamming and thus can be exploited in model for learning spam and non-spam clusters. [1] Author Spamicity Model [1] Mukherjee et al : http://delivery.acm.org/10.1145/2490000/2487580/p632-mukherjee.pdf
  • 19. Author Features • It have value in range [0, 1] and value close to 1 indicates spamming. 1. Content Similarity : Crafting new review every time is time consuming, spammers likely to copy reviews across similar product. Choose maximum similarity to capture worst spamming behavior.[1] 2. Maximum Number of Reviews : Posting many reviews a day is also abnormal. 3. Reviewing Burstiness : Spammers are usually not longtime members of site. Defined over an activity window(first and last review posting date). If posted over reasonably long timeframe, it probably a normal activity but all review posted within short burst likely to be spam.[2] 4. Ratio of First Review : People mostly rely on early reviews and spamming early impacts hugely on sales. Spammers try to be among first reviewers. [1] [1] Mukherjee et al : http://delivery.acm.org/10.1145/2490000/2487580/p632-mukherjee.pdf [2] Mukherjee, A., Liu, B. and Glance, N. 2012. Spotting Fake Reviewer Groups in Consumer Reviews. WWW (2012).
  • 20. Review Features • It have five binary review features. Value 1 indicates spamming else 0 is non-spamming. 1. Duplicate /Near Duplicate Reviews: Spammers often post multiple reviews which are duplicate /near duplicate on same product to boost ratings. 2. Extreme Rating: Spammers mostly like to give extreme ratings(1 or 5) in order to boost ratings to demote or promote products. 3. Rating Deviation: Spammers usually involve in wrong projection in either positive or negative and it deviates from average ratings given by other reviewers. 4. Early Time Frame: Early review can greatly impact people’s sentiments on a product. 5. Rating Abuse: Multiple ratings on same product are unusual. Similar to DUP but focuses on rating dimensions rather than content.
  • 21. CONCLUSIONS • We presented an in-depth investigation of nature of fake reviews in Commercial settings of Yelp.com. • Our study shows linguistics methods of (Ott et al., 2011) and its high accuracy in AMT data. • We presented a behavioral study of spammers for real-life fake reviews. • We presented a brief introduction to n-gram language model. • We presented challenges with the supervised evaluation and gave idea about the unsupervised approach of evaluation. • We presented a brief introduction to unsupervised Author Spamicity Model (ASM).
  • 22. REFERENCES 1. Jindal and Liu 2008 : http://www.cs.uic.edu/~liub/FBS/opinion-spam-WSDM-08.pdf 2. Yelp official blog : http://officialblog.yelp.com/2013/05/how-yelp-protects-consumers-from-fake- reviews.html 3. MIT N-gram Language Model Tutorial: http://web.mit.edu/6.863/www/fall2012/readings/ngrampages.pdf 4. Amazon Mechanical Turk : https://www.mturk.com/ 5. Ott et al. 2011 : https://www.cs.cornell.edu/courses/CS4740/2012sp/lectures/op_spamACL2011.pdf 6. Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera- submit.pdf 7. Kullback Leibler Divergence : http://www.cs.buap.mx/~dpinto/research/CICLing07_1/Pinto06c/node2.html 8. Newman et al , 2003 : http://www.communicationcache.com/uploads/1/0/8/8/10887248/lying_words- _predicting_deception_from_linguistic_styles.pdf 9. Mukherjee, A., Liu, B. and Glance, N. 2012. Spotting Fake Reviewer Groups in Consumer Reviews. WWW (2012). 10. Mukherjee et al : http://delivery.acm.org/10.1145/2490000/2487580/p632-mukherjee.pdf 11. Ott M, Choi, Y, Cardie, C. and Hancock, J.T. 2011. Finding Deceptive Opinion Spam by Any Stretch of the Imagination.