Meta-aware Learning in text-to-SQL Large Language Model
By: Wenda Zhang
Potential Business Impact:
Helps computers understand business data better.
The advancements of Large language models (LLMs) have provided great opportunities to text-to-SQL tasks to overcome the main challenges to understand complex domain information and complex database structures in business applications. In this paper, we propose a meta-aware learning framework to integrate domain knowledge, database schema, chain-of-thought reasoning processes, and metadata relationships to improve the SQL generation quality. The proposed framework includes four learning strategies: schema-based learning, Chain-of-Thought (CoT) learning, knowledge-enhanced learning, and key information tokenization. This approach provides a comprehensive understanding of database structure and metadata information towards LLM through fine-tuning to improve its performance on SQL generation within business domains. Through two experimental studies, we have demonstrated the superiority of the proposed methods in execution accuracy, multi-task SQL generation capability, and reduction of catastrophic forgetting.
Similar Papers
MageSQL: Enhancing In-context Learning for Text-to-SQL Applications with Large Language Models
Databases
Helps computers understand questions to find data.
An LLM-Based Approach for Insight Generation in Data Analysis
Artificial Intelligence
Finds hidden patterns in data automatically.
Knowledge Distillation with Structured Chain-of-Thought for Text-to-SQL
Computation and Language
Teaches small computers to write accurate database answers.