Identifying Bias in Machine-generated Text Detection
By: Kevin Stowe , Svetlana Afanaseva , Rodolfo Raimundo and more
Potential Business Impact:
Detectors unfairly flag some students' writing as fake.
The meteoric rise in text generation capability has been accompanied by parallel growth in interest in machine-generated text detection: the capability to identify whether a given text was generated using a model or written by a person. While detection models show strong performance, they have the capacity to cause significant negative impacts. We explore potential biases in English machine-generated text detection systems. We curate a dataset of student essays and assess 16 different detection systems for bias across four attributes: gender, race/ethnicity, English-language learner (ELL) status, and economic status. We evaluate these attributes using regression-based models to determine the significance and power of the effects, as well as performing subgroup analysis. We find that while biases are generally inconsistent across systems, there are several key issues: several models tend to classify disadvantaged groups as machine-generated, ELL essays are more likely to be classified as machine-generated, economically disadvantaged students' essays are less likely to be classified as machine-generated, and non-White ELL essays are disproportionately classified as machine-generated relative to their White counterparts. Finally, we perform human annotation and find that while humans perform generally poorly at the detection task, they show no significant biases on the studied attributes.
Similar Papers
AI-generated Text Detection: A Multifaceted Approach to Binary and Multiclass Classification
Computation and Language
Finds if writing is from a person or AI.
Evaluating LLMs for Demographic-Targeted Social Bias Detection: A Comprehensive Benchmark Study
Computation and Language
Finds unfairness in AI's words.
On the Detectability of LLM-Generated Text: What Exactly Is LLM-Generated Text?
Computation and Language
Finds if computers wrote the words.