Improving ML Training Data with Gold-Standard Quality Metrics
By: Leslie Barrett, Michael W. Sherman
Hand-tagged training data is essential to many machine learning tasks. However, training data quality control has received little attention in the literature, despite data quality varying considerably with the tagging exercise. We propose methods to evaluate and enhance the quality of hand-tagged training data using statistical approaches to measure tagging consistency and agreement. We show that agreement metrics give more reliable results if recorded over multiple iterations of tagging, where declining variance in such recordings is an indicator of increasing data quality. We also show one way a tagging project can collect high-quality training data without requiring multiple tags for every work item, and that a tagger burn-in period may not be sufficient for minimizing tagger errors.
Similar Papers
How Data Quality Affects Machine Learning Models for Credit Risk Assessment
Machine Learning (CS)
Makes loan decisions more accurate even with bad data.
QualiTagger: Automating software quality detection in issue trackers
Software Engineering
Finds bad code quality in computer programs.
Datasheets Aren't Enough: DataRubrics for Automated Quality Metrics and Accountability
Machine Learning (CS)
Makes computer learning data better and easier to check.