A Comparative Approach to Assessing Linguistic Creativity of Large Language Models and Humans
By: Anca Dinu, Andra-Maria Florescu, Alina Resceanu
Potential Business Impact:
Computers invent new words better than people.
The following paper introduces a general linguistic creativity test for humans and Large Language Models (LLMs). The test consists of various tasks aimed at assessing their ability to generate new original words and phrases based on word formation processes (derivation and compounding) and on metaphorical language use. We administered the test to 24 humans and to an equal number of LLMs, and we automatically evaluated their answers using OCSAI tool for three criteria: Originality, Elaboration, and Flexibility. The results show that LLMs not only outperformed humans in all the assessed criteria, but did better in six out of the eight test tasks. We then computed the uniqueness of the individual answers, which showed some minor differences between humans and LLMs. Finally, we performed a short manual analysis of the dataset, which revealed that humans are more inclined towards E(extending)-creativity, while LLMs favor F(ixed)-creativity.
Similar Papers
Evaluating the Creativity of LLMs in Persian Literary Text Generation
Computation and Language
Computers write creative Persian stories.
Deep Associations, High Creativity: A Simple yet Effective Metric for Evaluating Large Language Models
Computation and Language
Tests AI's imagination like a human.
Style Over Story: A Process-Oriented Study of Authorial Creativity in Large Language Models
Computation and Language
AI writing tools prefer style over story.