Reinforcement learning to enable reasoning LLMs for Text2SQL

Papotti, Paolo
TADA 2025, Keynote talk in 3rd International Workshop on Tabular Data Analysis (TaDA), collocated with the 51th International Conference on Very Large Data Bases (VLDB 2025), 5 September 2025, London, UK

The ability to interact with complex databases using natural language (NL) is a key step in democratizing data access, a long-standing goal in the enterprise world. While Large Language Models (LLMs) have shown remarkable promise in translating NL questions into SQL queries (Text2QL), their performance stall when faced with the complexities of real-world enterprise databases. This talk will report a promising solution to enhance the reasoning capabilities of LLMs for this task. Our "Think2SQL" methodology investigates various strategies for improving LLM performance, including Zero-Shot Learning (ZSL), Supervised Fine-Tuning (SFT), and Reinforcement Learning (RL). RL, using rewards crafted around SQL execution accuracy, significantly boosts the performance of small LLMs, achieving results comparable to those of much larger models on complex datasets. Finally, we will highlight the path forward for Text2SQL systems capable of navigating the nuances of human language, such as ambiguity, in a real-world enterprise context.


Type:
Talk
City:
London
Date:
2025-09-05
Department:
Data Science
Eurecom Ref:
8411
Copyright:
© EURECOM. Personal use of this material is permitted. The definitive version of this paper was published in TADA 2025, Keynote talk in 3rd International Workshop on Tabular Data Analysis (TaDA), collocated with the 51th International Conference on Very Large Data Bases (VLDB 2025), 5 September 2025, London, UK and is available at :
See also:

PERMALINK : https://www.eurecom.fr/publication/8411