Towards Generating Automatic Anaphora Annotations
By: Dima Taji, Daniel Zeman
Potential Business Impact:
Teaches computers to understand tricky word meanings.
Training models that can perform well on various NLP tasks require large amounts of data, and this becomes more apparent with nuanced tasks such as anaphora and conference resolution. To combat the prohibitive costs of creating manual gold annotated data, this paper explores two methods to automatically create datasets with coreferential annotations; direct conversion from existing datasets, and parsing using multilingual models capable of handling new and unseen languages. The paper details the current progress on those two fronts, as well as the challenges the efforts currently face, and our approach to overcoming these challenges.
Similar Papers
Enhancing Coreference Resolution with Pretrained Language Models: Bridging the Gap Between Syntax and Semantics
Computation and Language
Helps computers understand who or what "they" refers to.
Coreference Resolution for Vietnamese Narrative Texts
Computation and Language
Helps computers understand Vietnamese stories better.
Large Language Models as Span Annotators
Computation and Language
Computers can now find and label text parts.