Data Augmentation for Fake Reviews Detection in Multiple Languages and Multiple Domains
By: Ming Liu, Massimo Poesio
Potential Business Impact:
Finds fake online reviews better.
With the growth of the Internet, buying habits have changed, and customers have become more dependent on the online opinions of other customers to guide their purchases. Identifying fake reviews thus became an important area for Natural Language Processing (NLP) research. However, developing high-performance NLP models depends on the availability of large amounts of training data, which are often not available for low-resource languages or domains. In this research, we used large language models to generate datasets to train fake review detectors. Our approach was used to generate fake reviews in different domains (book reviews, restaurant reviews, and hotel reviews) and different languages (English and Chinese). Our results demonstrate that our data augmentation techniques result in improved performance at fake review detection for all domains and languages. The accuracy of our fake review detection model can be improved by 0.3 percentage points on DeRev TEST, 10.9 percentage points on Amazon TEST, 8.3 percentage points on Yelp TEST and 7.2 percentage points on DianPing TEST using the augmented datasets.
Similar Papers
Unmask It! AI-Generated Product Review Detection in Dravidian Languages
Computation and Language
Finds fake online reviews written by computers.
Large Language Models as 'Hidden Persuaders': Fake Product Reviews are Indistinguishable to Humans and Machines
Computation and Language
Computers can't tell fake reviews from real ones.
Code Review Without Borders: Evaluating Synthetic vs. Real Data for Review Recommendation
Software Engineering
Teaches computers to check new code automatically.