Processing large tables provided in-context to LLMs is challenging due to token limits and information overload. While RetrievalAugmented Generation can select relevant subsets externally, this work explores Key-Value (KV) cache compression as an alternative, applied directly to the linearized table during inPaolo Papotti EURECOM,France Figure 1: High-level overview of attention-guided KV ference. We show that the LLM’s internal attention scores over the table context guides the retention of essential KV pairs, effectively compressing the processing context while preserving crucial relational information needed for complex queries. Experiments on Spider, WikitableQA, and QTSumm datasets validate the compression approach for in-context table processing, offering a promising path for improved table representation learning in LLMs.
TableKV: KV cache compression for in-context table processing
TRL 2025, 4th Table Representation Learning Workshop at ACL 2025, 27 July-1 August 2025, Vienna, Austria
Type:
Conference
City:
Vienna
Date:
2025-07-27
Department:
Data Science
Eurecom Ref:
8415
Copyright:
Copyright ACL. Personal use of this material is permitted. The definitive version of this paper was published in TRL 2025, 4th Table Representation Learning Workshop at ACL 2025, 27 July-1 August 2025, Vienna, Austria and is available at : https://doi.org/10.18653/v1/2025.trl-1.13
See also:
PERMALINK : https://www.eurecom.fr/publication/8415