Simplify-This: A Comparative Analysis of Prompt-Based and Fine-Tuned LLMs
By: Eilam Cohen , Itamar Bul , Danielle Inbar and more
Potential Business Impact:
Makes complex writing easier to understand.
Large language models (LLMs) enable strong text generation, and in general there is a practical tradeoff between fine-tuning and prompt engineering. We introduce Simplify-This, a comparative study evaluating both paradigms for text simplification with encoder-decoder LLMs across multiple benchmarks, using a range of evaluation metrics. Fine-tuned models consistently deliver stronger structural simplification, whereas prompting often attains higher semantic similarity scores yet tends to copy inputs. A human evaluation favors fine-tuned outputs overall. We release code, a cleaned derivative dataset used in our study, checkpoints of fine-tuned models, and prompt templates to facilitate reproducibility and future work.
Similar Papers
Comparative Evaluation of Prompting and Fine-Tuning for Applying Large Language Models to Grid-Structured Geospatial Data
Computation and Language
Helps computers understand maps and time better.
Resource-Efficient Adaptation of Large Language Models for Text Embeddings via Prompt Engineering and Contrastive Fine-tuning
Computation and Language
Makes computers understand text meaning better.
Fine-Tune an SLM or Prompt an LLM? The Case of Generating Low-Code Workflows
Machine Learning (CS)
Small AI models make specific tasks better than big ones.