Chunked TabPFN: Exact Training-Free In-Context Learning for Long-Context Tabular Data
By: Renat Sergazinov, Shao-An Yin
Potential Business Impact:
Lets computers learn from much bigger lists of info.
TabPFN v2 achieves better results than tree-based models on several tabular benchmarks, which is notable since tree-based models are usually the strongest choice for tabular data. However, it cannot handle more than 10K context tokens because transformers have quadratic computation and memory costs. Unlike existing approaches that rely on context compression, such as selecting representative samples via K-nearest neighbors (KNN), we introduce a tiled-block strategy to compute attention within the TabPFN framework. This design is compatible with standard GPU setups and, to the best of our knowledge, is the first to enable TabPFN to process long contexts without any pre-processing. We demonstrate the effectiveness of our approach on the standard TabArena benchmark, with code available at https://github.com/mrsergazinov/chunk_tabpfn.
Similar Papers
TabPFN-2.5: Advancing the State of the Art in Tabular Foundation Models
Machine Learning (CS)
Makes computers learn from bigger, more complex data.
nanoTabPFN: A Lightweight and Educational Reimplementation of TabPFN
Machine Learning (CS)
Makes smart computer models easy to learn.
Gradient Free Deep Reinforcement Learning With TabPFN
Machine Learning (CS)
Teaches robots to learn without slow trial and error.