Still Not There: Can LLMs Outperform Smaller Task-Specific Seq2Seq Models on the Poetry-to-Prose Conversion Task?
By: Kunal Kingkar Das , Manoj Balaji Jagadeeshan , Nallani Chakravartula Sahith and more
Potential Business Impact:
Makes old Sanskrit poems understandable as normal sentences.
Large Language Models (LLMs) are increasingly treated as universal, general-purpose solutions across NLP tasks, particularly in English. But does this assumption hold for low-resource, morphologically rich languages such as Sanskrit? We address this question by comparing instruction-tuned and in-context-prompted LLMs with smaller task-specific encoder-decoder models on the Sanskrit poetry-to-prose conversion task. This task is intrinsically challenging: Sanskrit verse exhibits free word order combined with rigid metrical constraints, and its conversion to canonical prose (anvaya) requires multi-step reasoning involving compound segmentation, dependency resolution, and syntactic linearisation. This makes it an ideal testbed to evaluate whether LLMs can surpass specialised models. For LLMs, we apply instruction fine-tuning on general-purpose models and design in-context learning templates grounded in Paninian grammar and classical commentary heuristics. For task-specific modelling, we fully fine-tune a ByT5-Sanskrit Seq2Seq model. Our experiments show that domain-specific fine-tuning of ByT5-Sanskrit significantly outperforms all instruction-driven LLM approaches. Human evaluation strongly corroborates this result, with scores exhibiting high correlation with Kendall's Tau scores. Additionally, our prompting strategies provide an alternative to fine-tuning when domain-specific verse corpora are unavailable, and the task-specific Seq2Seq model demonstrates robust generalisation on out-of-domain evaluations.
Similar Papers
From Plain Text to Poetic Form: Generating Metrically-Constrained Sanskrit Verses
Computation and Language
Teaches computers to write old Sanskrit poems.
State of the Art in Text Classification for South Slavic Languages: Fine-Tuning or Prompting?
Computation and Language
New AI understands many languages for tasks.
Regional Tiny Stories: Using Small Models to Compare Language Learning and Tokenizer Performance
Computation and Language
Helps small computers understand Indian languages.