Evidence of conceptual mastery in the application of rules by Large Language Models
By: José Luiz Nunes, Guilherme FCF Almeida, Brian Flanagan
Potential Business Impact:
Makes AI understand rules like people do.
In this paper we leverage psychological methods to investigate LLMs' conceptual mastery in applying rules. We introduce a novel procedure to match the diversity of thought generated by LLMs to that observed in a human sample. We then conducted two experiments comparing rule-based decision-making in humans and LLMs. Study 1 found that all investigated LLMs replicated human patterns regardless of whether they are prompted with scenarios created before or after their training cut-off. Moreover, we found unanticipated differences between the two sets of scenarios among humans. Surprisingly, even these differences were replicated in LLM responses. Study 2 turned to a contextual feature of human rule application: under forced time delay, human samples rely more heavily on a rule's text than on other considerations such as a rule's purpose.. Our results revealed that some models (Gemini Pro and Claude 3) responded in a human-like manner to a prompt describing either forced delay or time pressure, while others (GPT-4o and Llama 3.2 90b) did not. We argue that the evidence gathered suggests that LLMs have mastery over the concept of rule, with implications for both legal decision making and philosophical inquiry.
Similar Papers
Human-like conceptual representations emerge from language prediction
Computation and Language
Computers learn ideas like people from words.
Less Is More for Multi-Step Logical Reasoning of LLM Generalisation Under Rule Removal, Paraphrasing, and Compression
Artificial Intelligence
Computers struggle with missing or wrong logic.
LLMs model how humans induce logically structured rules
Artificial Intelligence
Computers learn logic like humans do.