A survey on text-to-sql parsing: Concepts, methods, and future directions

B Qin, B Hui, L Wang, M Yang, J Li, B Li… - arxiv preprint arxiv …, 2022 - arxiv.org
Text-to-SQL parsing is an essential and challenging task. The goal of text-to-SQL parsing is
to convert a natural language (NL) question to its corresponding structured query language …

Din-sql: Decomposed in-context learning of text-to-sql with self-correction

M Pourreza, D Rafiei - Advances in Neural Information …, 2023 - proceedings.neurips.cc
There is currently a significant gap between the performance of fine-tuned models and
prompting approaches using Large Language Models (LLMs) on the challenging task of text …

Recent advances in text-to-SQL: a survey of what we have and what we expect

N Deng, Y Chen, Y Zhang - arxiv preprint arxiv:2208.10099, 2022 - arxiv.org
Text-to-SQL has attracted attention from both the natural language processing and database
communities because of its ability to convert the semantics in natural language into SQL …

Text-to-sql empowered by large language models: A benchmark evaluation

D Gao, H Wang, Y Li, X Sun, Y Qian, B Ding… - arxiv preprint arxiv …, 2023 - arxiv.org
Large language models (LLMs) have emerged as a new paradigm for Text-to-SQL task.
However, the absence of a systematical benchmark inhibits the development of designing …

DS-1000: A natural and reliable benchmark for data science code generation

Y Lai, C Li, Y Wang, T Zhang, R Zhong… - International …, 2023 - proceedings.mlr.press
We introduce DS-1000, a code generation benchmark with a thousand data science
problems spanning seven Python libraries, such as Numpy and Pandas. Compared to prior …

Resdsql: Decoupling schema linking and skeleton parsing for text-to-sql

H Li, J Zhang, C Li, H Chen - Proceedings of the AAAI Conference on …, 2023 - ojs.aaai.org
One of the recent best attempts at Text-to-SQL is the pre-trained language model. Due to the
structural property of the SQL queries, the seq2seq model takes the responsibility of parsing …

Selective annotation makes language models better few-shot learners

H Su, J Kasai, CH Wu, W Shi, T Wang, J **n… - arxiv preprint arxiv …, 2022 - arxiv.org
Many recent approaches to natural language tasks are built on the remarkable abilities of
large language models. Large language models can perform in-context learning, where they …

PICARD: Parsing incrementally for constrained auto-regressive decoding from language models

T Scholak, N Schucher, D Bahdanau - arxiv preprint arxiv:2109.05093, 2021 - arxiv.org
Large pre-trained language models for textual data have an unconstrained output space; at
each decoding step, they can produce any of 10,000 s of sub-word tokens. When fine-tuned …

Codes: Towards building open-source language models for text-to-sql

H Li, J Zhang, H Liu, J Fan, X Zhang, J Zhu… - Proceedings of the …, 2024 - dl.acm.org
Language models have shown promising performance on the task of translating natural
language questions into SQL queries (Text-to-SQL). However, most of the state-of-the-art …

A comprehensive evaluation of ChatGPT's zero-shot Text-to-SQL capability

A Liu, X Hu, L Wen, PS Yu - arxiv preprint arxiv:2303.13547, 2023 - arxiv.org
This paper presents the first comprehensive analysis of ChatGPT's Text-to-SQL ability. Given
the recent emergence of large-scale conversational language model ChatGPT and its …