A Method for the Detection of Fake Reviews Based on Temporal Features of Reviews and Comments

被引:16
作者
Liu W. [1 ]
He J. [1 ]
Han S. [1 ]
Cai F. [1 ]
Yang Z. [1 ]
Zhu N. [1 ]
机构
[1] Faculty of Information Technology, Beijing University of Technology, Beijing
来源
IEEE Engineering Management Review | 2019年 / 47卷 / 04期
基金
中国国家自然科学基金;
关键词
Fake reviews; isolation forest algorithm; products speculation; review records;
D O I
10.1109/EMR.2019.2928964
中图分类号
学科分类号
摘要
Online reviews and comments after product sales have become very important for making buying and selling decisions. Fake reviews will affect such decisions due to deceptive information, leading to financial losses for the consumers. Identification of fake reviews has thus received a great deal of attention in recent years. However, most websites have only focused on dealing with problematic reviews and comments. Amazon and Yelp would only remove possible fake reviews without questioning the sellers who could continue posting deceptive reviews for business purposes. In this paper, we propose a method for the detection of fake reviews based on review records associated with products. We first analyze the characteristics of review data using a crawled Amazon China dataset, which shows that the patterns of review records for products are similar in normal situations. In the proposed method, we first extract the review records of products to a temporal feature vector and then develop an isolation forest algorithm to detect outlier reviews by focusing on the differences between the patterns of product reviews to identify outlier reviews. We will verify the effectiveness of our method and compare it to some existing temporal outlier detection methods using the crawled Amazon China dataset. We will also study the impact caused by the parameter selection of the review records. Our work provides a new perspective of outlier review detection and our experiment demonstrates the effectiveness of our proposed method. © 1973-2011 IEEE.
引用
收藏
页码:67 / 79
页数:12
相关论文
共 35 条
  • [1] Streitfeld D., Fake Reviews, Real Problem, the New York Times Company, (2012)
  • [2] Rayana S., Akoglu L., Collective opinion spam detection: Bridging review networks and metadata, Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 985-994, (2015)
  • [3] Catal C., Guldan S., Product review management software based on multiple classifiers, IET Softw., 11, 3, pp. 89-92, (2017)
  • [4] Zuo L., Carass A., Han S., Prince J.L., Automatic outlier detection using hidden markov model for cerebellar lobule segmentation, Proceedings of International Conference on Medical Applications in Molecular, Structural, and Functional Imaging, pp. 105780D1-105780D8, (2018)
  • [5] Mukherjee A., Venkataraman V., Liu B., Glance N.S., What yelp fake review filter might be doing?, Proceedings of the 7th International AAAI Conference on Weblogs and Social Media, pp. 409-418, (2013)
  • [6] Spirin N., Han J., Survey on web spam detection: Principles and algorithms, ACM SIGKDD Explorations Newslett, 13, pp. 50-64, (2012)
  • [7] Chirita P.A., Diederich J., MailRank N.W., Using ranking for spam detection, Proceedings of the 14th ACM International Conference on Information and Knowledge Management, pp. 373-380, (2005)
  • [8] Yang W., Kwok L., Improving blog spam filters via machine learning, Int. J. Data Anal. Techn. Strategies, 9, pp. 99-121, (2017)
  • [9] Tan E., Guo L., Chen S., Zhang X., Zhao Y., Unik: Unsupervised social network spam detection, Proceedings of the 22nd ACM International Conference on Information & Knowledge Management, pp. 479-488, (2013)
  • [10] Liu B., Sentiment analysis and opinion mining, Synthesis Lectures on Human Lang. Technol, 5, pp. 1-167, (2012)