LLM Harms: A Taxonomy and Discussion
By: Kevin Chen , Saleh Afroogh , Abhejay Murali and more
This study addresses categories of harm surrounding Large Language Models (LLMs) in the field of artificial intelligence. It addresses five categories of harms addressed before, during, and after development of AI applications: pre-development, direct output, Misuse and Malicious Application, and downstream application. By underscoring the need to define risks of the current landscape to ensure accountability, transparency and navigating bias when adapting LLMs for practical applications. It proposes mitigation strategies and future directions for specific domains and a dynamic auditing system guiding responsible development and integration of LLMs in a standardized proposal.
Similar Papers
Guardians and Offenders: A Survey on Harmful Content Generation and Safety Mitigation of LLM
Computation and Language
Makes AI safer and less likely to say bad things.
Guardians and Offenders: A Survey on Harmful Content Generation and Safety Mitigation
Computation and Language
Makes AI safer and less likely to say bad things.
From Superficial Outputs to Superficial Learning: Risks of Large Language Models in Education
Computers and Society
Finds dangers of AI in schools.