The Thin Line Between Comprehension and Persuasion in LLMs
By: Adrian de Wynter, Tangming Yuan
Potential Business Impact:
Computers can argue well but don't understand.
Large language models (LLMs) are excellent at maintaining high-level, convincing dialogues. They are being fast deployed as chatbots and evaluators in sensitive areas, such as peer review and mental health applications. This, along with the disparate accounts on their reasoning capabilities, calls for a closer examination of LLMs and their comprehension of dialogue. In this work we begin by evaluating LLMs' ability to maintain a debate--one of the purest yet most complex forms of human communication. Then we measure how this capability relates to their understanding of what is being talked about, namely, their comprehension of dialogical structures and the pragmatic context. We find that LLMs are capable of maintaining coherent, persuasive debates, often swaying the beliefs of participants and audiences alike. We also note that awareness or suspicion of AI involvement encourage people to be more critical of the arguments made. When polling LLMs on their comprehension of deeper structures of dialogue, however, they cannot demonstrate said understanding. Our findings tie the shortcomings of LLMs-as-evaluators to their (in)ability to understand the context. More broadly, for the field of argumentation theory we posit that, if an agent can convincingly maintain a dialogue, it is not necessary for it to know what it is talking about. Hence, the modelling of pragmatic context and coherence are secondary to effectiveness.
Similar Papers
Can (A)I Change Your Mind?
Computation and Language
Computers can change your mind like people.
Advances in LLMs with Focus on Reasoning, Adaptability, Efficiency and Ethics
Computation and Language
Makes computers smarter, faster, and more helpful.
Can LLM Agents Maintain a Persona in Discourse?
Computation and Language
Makes AI chat buddies act like real people.