Towards Meaningful Transparency in Civic AI Systems
By: Dave Murray-Rust, Kars Alfrink, Cristina Zaga
Potential Business Impact:
Lets people understand and change AI decisions.
Artificial intelligence has become a part of the provision of governmental services, from making decisions about benefits to issuing fines for parking violations. However, AI systems rarely live up to the promise of neutral optimisation, creating biased or incorrect outputs and reducing the agency of both citizens and civic workers to shape the way decisions are made. Transparency is a principle that can both help subjects understand decisions made about them and shape the processes behind those decisions. However, transparency as practiced around AI systems tends to focus on the production of technical objects that represent algorithmic aspects of decision making. These are often difficult for publics to understand, do not connect to potential for action, and do not give insight into the wider socio-material context of decision making. In this paper, we build on existing approaches that take a human-centric view on AI transparency, combined with a socio-technical systems view, to develop the concept of meaningful transparency for civic AI systems: transparencies that allow publics to engage with AI systems that affect their lives, connecting understanding with potential for action.
Similar Papers
Towards Transparent Ethical AI: A Roadmap for Trustworthy Robotic Systems
Computers and Society
Makes robots do the right thing.
An Analysis of the New EU AI Act and A Proposed Standardization Framework for Machine Learning Fairness
Computers and Society
Makes AI fair and clear for everyone.
Human Resource Management and AI: A Contextual Transparency Database
Human-Computer Interaction
Helps hiring tools show how they work.