TY - JOUR
T1 - Can end-user feedback in social media be trusted for software evolution
T2 - Exploring and analyzing fake reviews
AU - Khan, Javed Ali
AU - Ullah, Tahir
AU - Khan, Arif Ali
AU - Yasin, Affan
AU - Akbar, Muhammad Azeem
AU - Aurangzeb, Khursheed
N1 - Publisher Copyright:
© 2023 John Wiley & Sons Ltd.
PY - 2024/5/1
Y1 - 2024/5/1
N2 - End-user feedback in social media platforms, particularly in the app stores, is increasing exponentially with each passing day. Software researchers and vendors started to mine end-user feedback by proposing text analytics methods and tools to extract useful information for software evolution and maintenance. In addition, research shows that positive feedback and high-star app ratings attract more users and increase downloads. However, it emerged in the fake review market, where software vendors started incorporating fake reviews against their corresponding applications to improve overall software ratings. For this purpose, we conducted an exploratory study to understand how end-users register and write fake reviews in the Google Play Store. We curated a research data set containing 68,000 end-user comments from the Google Play Store and a fake review generator, that is, the Testimonial generator (TG). Its purpose is to understand fake reviews on these platforms and identify the common patterns potential end-users and professionals use to report fake reviews by critically analyzing the end-user feedback. We conducted a detailed survey at the University of Science and Technology Bannu, Pakistan, to identify the intelligence and accuracy of crowd-users in manually identifying fake reviews. In addition, we developed a ground truth to be compared with the results obtained from the automated machine and deep learning (M&DL) classifier experiment. In the survey, 512 end-users participated and recorded their responses in identifying fake reviews. Finally, various M&DL classifiers are employed to classify and identify end-user reviews into real and fake to automate the process. Unlike humans, the M&DL classifiers performed well in automatically classifying reviews into real and fake by obtaining much higher accuracy, precision, recall, and f-measures. The accuracy of manually identifying fake reviews by the crowd-users is 44.4%. In contrast, the M&DL classifiers obtained an average accuracy of 96%. The experimental results obtained with various M&DL classifiers are encouraging. It is the first step towards identifying fake reviews in the app store by studying its implications in software and requirements engineering.
AB - End-user feedback in social media platforms, particularly in the app stores, is increasing exponentially with each passing day. Software researchers and vendors started to mine end-user feedback by proposing text analytics methods and tools to extract useful information for software evolution and maintenance. In addition, research shows that positive feedback and high-star app ratings attract more users and increase downloads. However, it emerged in the fake review market, where software vendors started incorporating fake reviews against their corresponding applications to improve overall software ratings. For this purpose, we conducted an exploratory study to understand how end-users register and write fake reviews in the Google Play Store. We curated a research data set containing 68,000 end-user comments from the Google Play Store and a fake review generator, that is, the Testimonial generator (TG). Its purpose is to understand fake reviews on these platforms and identify the common patterns potential end-users and professionals use to report fake reviews by critically analyzing the end-user feedback. We conducted a detailed survey at the University of Science and Technology Bannu, Pakistan, to identify the intelligence and accuracy of crowd-users in manually identifying fake reviews. In addition, we developed a ground truth to be compared with the results obtained from the automated machine and deep learning (M&DL) classifier experiment. In the survey, 512 end-users participated and recorded their responses in identifying fake reviews. Finally, various M&DL classifiers are employed to classify and identify end-user reviews into real and fake to automate the process. Unlike humans, the M&DL classifiers performed well in automatically classifying reviews into real and fake by obtaining much higher accuracy, precision, recall, and f-measures. The accuracy of manually identifying fake reviews by the crowd-users is 44.4%. In contrast, the M&DL classifiers obtained an average accuracy of 96%. The experimental results obtained with various M&DL classifiers are encouraging. It is the first step towards identifying fake reviews in the app store by studying its implications in software and requirements engineering.
KW - CrowdRE
KW - fake reviews
KW - Google Play Store
KW - machine and deep learning
KW - user feedback
UR - http://www.scopus.com/inward/record.url?scp=85180230046&partnerID=8YFLogxK
U2 - 10.1002/cpe.7990
DO - 10.1002/cpe.7990
M3 - Article
AN - SCOPUS:85180230046
SN - 1532-0626
VL - 36
JO - Concurrency and Computation: Practice and Experience
JF - Concurrency and Computation: Practice and Experience
IS - 10
M1 - e7990
ER -