Coreference as an indicator of context scope in multimodal narrative
By: Nikolai Ilinykh , Shalom Lappin , Asad Sayeed and more
Potential Business Impact:
Helps computers tell stories like people.
We demonstrate that large multimodal language models differ substantially from humans in the distribution of coreferential expressions in a visual storytelling task. We introduce a number of metrics to quantify the characteristics of coreferential patterns in both human- and machine-written texts. Humans distribute coreferential expressions in a way that maintains consistency across texts and images, interleaving references to different entities in a highly varied way. Machines are less able to track mixed references, despite achieving perceived improvements in generation quality. Materials, metrics, and code for our study are available at https://github.com/GU-CLASP/coreference-context-scope.
Similar Papers
Coreference Resolution for Vietnamese Narrative Texts
Computation and Language
Helps computers understand Vietnamese stories better.
Disambiguating Reference in Visually Grounded Dialogues through Joint Modeling of Textual and Multimodal Semantic Structures
Computation and Language
Helps computers understand what you mean in chats.
Enhancing Coreference Resolution with Pretrained Language Models: Bridging the Gap Between Syntax and Semantics
Computation and Language
Helps computers understand who or what "they" refers to.