Compliance Rating Scheme: A Data Provenance Framework for Generative AI Datasets
By: Matyas Bohacek, Ignacio Vilanova Echavarri
Potential Business Impact:
Checks if AI training data is safe and fair.
Generative Artificial Intelligence (GAI) has experienced exponential growth in recent years, partly facilitated by the abundance of large-scale open-source datasets. These datasets are often built using unrestricted and opaque data collection practices. While most literature focuses on the development and applications of GAI models, the ethical and legal considerations surrounding the creation of these datasets are often neglected. In addition, as datasets are shared, edited, and further reproduced online, information about their origin, legitimacy, and safety often gets lost. To address this gap, we introduce the Compliance Rating Scheme (CRS), a framework designed to evaluate dataset compliance with critical transparency, accountability, and security principles. We also release an open-source Python library built around data provenance technology to implement this framework, allowing for seamless integration into existing dataset-processing and AI training pipelines. The library is simultaneously reactive and proactive, as in addition to evaluating the CRS of existing datasets, it equally informs responsible scraping and construction of new datasets.
Similar Papers
Do Not Trust Licenses You See: Dataset Compliance Requires Massive-Scale AI-Powered Lifecycle Tracing
Computers and Society
AI checks if data can be legally used.
Data-Centric Safety and Ethical Measures for Data and AI Governance
Computers and Society
Makes AI safer by improving how data is used.
AI Data Development: A Scorecard for the System Card Framework
Computers and Society
Checks AI data for fairness and mistakes.