Retrieval-augmented generation for large language models: A survey
Y Gao, Y **ong, X Gao, K Jia, J Pan, Y Bi, Y Dai… - arxiv preprint arxiv …, 2023 - arxiv.org
Large language models (LLMs) demonstrate powerful capabilities, but they still face
challenges in practical applications, such as hallucinations, slow knowledge updates, and …
challenges in practical applications, such as hallucinations, slow knowledge updates, and …
Survey of hallucination in natural language generation
Natural Language Generation (NLG) has improved exponentially in recent years thanks to
the development of sequence-to-sequence deep learning technologies such as Transformer …
the development of sequence-to-sequence deep learning technologies such as Transformer …
Siren's song in the AI ocean: a survey on hallucination in large language models
While large language models (LLMs) have demonstrated remarkable capabilities across a
range of downstream tasks, a significant concern revolves around their propensity to exhibit …
range of downstream tasks, a significant concern revolves around their propensity to exhibit …
Unlimiformer: Long-range transformers with unlimited length input
Since the proposal of transformers, these models have been limited to bounded input
lengths, because of their need to attend to every token in the input. In this work, we propose …
lengths, because of their need to attend to every token in the input. In this work, we propose …
Retentive network: A successor to transformer for large language models
In this work, we propose Retentive Network (RetNet) as a foundation architecture for large
language models, simultaneously achieving training parallelism, low-cost inference, and …
language models, simultaneously achieving training parallelism, low-cost inference, and …
Exploring the limits of chatgpt for query or aspect-based text summarization
Text summarization has been a crucial problem in natural language processing (NLP) for
several decades. It aims to condense lengthy documents into shorter versions while …
several decades. It aims to condense lengthy documents into shorter versions while …
Megalodon: Efficient llm pretraining and inference with unlimited context length
The quadratic complexity and weak length extrapolation of Transformers limits their ability to
scale to long sequences, and while sub-quadratic solutions like linear attention and state …
scale to long sequences, and while sub-quadratic solutions like linear attention and state …
Self-critiquing models for assisting human evaluators
We fine-tune large language models to write natural language critiques (natural language
critical comments) using behavioral cloning. On a topic-based summarization task, critiques …
critical comments) using behavioral cloning. On a topic-based summarization task, critiques …
Effective long-context scaling of foundation models
We present a series of long-context LLMs that support effective context windows of up to
32,768 tokens. Our model series are built through continual pretraining from Llama 2 with …
32,768 tokens. Our model series are built through continual pretraining from Llama 2 with …
Retrieval meets long context large language models
Extending the context window of large language models (LLMs) is getting popular recently,
while the solution of augmenting LLMs with retrieval has existed for years. The natural …
while the solution of augmenting LLMs with retrieval has existed for years. The natural …