Rethinking Table Pruning in TableQA: From Sequential Revisions to Gold Trajectory-Supervised Parallel Search
By: Yu Guo , Shenghao Ye , Shuangwu Chen and more
Potential Business Impact:
Finds important data in tables faster.
Table Question Answering (TableQA) benefits significantly from table pruning, which extracts compact sub-tables by eliminating redundant cells to streamline downstream reasoning. However, existing pruning methods typically rely on sequential revisions driven by unreliable critique signals, often failing to detect the loss of answer-critical data. To address this limitation, we propose TabTrim, a novel table pruning framework which transforms table pruning from sequential revisions to gold trajectory-supervised parallel search. TabTrim derives a gold pruning trajectory using the intermediate sub-tables in the execution process of gold SQL queries, and trains a pruner and a verifier to make the step-wise pruning result align with the gold pruning trajectory. During inference, TabTrim performs parallel search to explore multiple candidate pruning trajectories and identify the optimal sub-table. Extensive experiments demonstrate that TabTrim achieves state-of-the-art performance across diverse tabular reasoning tasks: TabTrim-8B reaches 73.5% average accuracy, outperforming the strongest baseline by 3.2%, including 79.4% on WikiTQ and 61.2% on TableBench.
Similar Papers
Towards Question Answering over Large Semi-structured Tables
Computation and Language
Finds answers in huge computer tables faster.
Evidence-Guided Schema Normalization for Temporal Tabular Reasoning
Computation and Language
Makes computers understand old, changing information better.
When TableQA Meets Noise: A Dual Denoising Framework for Complex Questions and Large-scale Tables
Computation and Language
Cleans messy tables for smarter answers.