Synthetic Data Augmentation for Cross-domain Implicit Discourse Relation Recognition
By: Frances Yung , Varsha Suresh , Zaynab Reza and more
Potential Business Impact:
Computers better understand how sentences connect.
Implicit discourse relation recognition (IDRR) -- the task of identifying the implicit coherence relation between two text spans -- requires deep semantic understanding. Recent studies have shown that zero- or few-shot approaches significantly lag behind supervised models, but LLMs may be useful for synthetic data augmentation, where LLMs generate a second argument following a specified coherence relation. We applied this approach in a cross-domain setting, generating discourse continuations using unlabelled target-domain data to adapt a base model which was trained on source-domain labelled data. Evaluations conducted on a large-scale test set revealed that different variations of the approach did not result in any significant improvements. We conclude that LLMs often fail to generate useful samples for IDRR, and emphasize the importance of considering both statistical significance and comparability when evaluating IDRR models.
Similar Papers
From Reviews to Dialogues: Active Synthesis for Zero-Shot LLM-based Conversational Recommender System
Information Retrieval
Creates smart helpers from less information.
Similarity-Based Domain Adaptation with LLMs
Computation and Language
Teaches computers new tasks without needing old examples.
LLM-RecG: A Semantic Bias-Aware Framework for Zero-Shot Sequential Recommendation
Information Retrieval
Helps online stores suggest items in new areas.