Reinforcement learning for generative ai: State of the art, opportunities and open research challenges
Abstract Generative Artificial Intelligence (AI) is one of the most exciting developments in
Computer Science of the last decade. At the same time, Reinforcement Learning (RL) has …
Computer Science of the last decade. At the same time, Reinforcement Learning (RL) has …
Self-refine: Iterative refinement with self-feedback
Like humans, large language models (LLMs) do not always generate the best output on their
first try. Motivated by how humans refine their written text, we introduce Self-Refine, an …
first try. Motivated by how humans refine their written text, we introduce Self-Refine, an …
Augmented language models: a survey
This survey reviews works in which language models (LMs) are augmented with reasoning
skills and the ability to use tools. The former is defined as decomposing a potentially …
skills and the ability to use tools. The former is defined as decomposing a potentially …
Language models meet world models: Embodied experiences enhance language models
While large language models (LMs) have shown remarkable capabilities across numerous
tasks, they often struggle with simple reasoning and planning in physical environments …
tasks, they often struggle with simple reasoning and planning in physical environments …
Reasoning with language model prompting: A survey
Reasoning, as an essential ability for complex problem-solving, can provide back-end
support for various real-world applications, such as medical diagnosis, negotiation, etc. This …
support for various real-world applications, such as medical diagnosis, negotiation, etc. This …
Rl4f: Generating natural language feedback with reinforcement learning for repairing model outputs
Despite their unprecedented success, even the largest language models make mistakes.
Similar to how humans learn and improve using feedback, previous work proposed …
Similar to how humans learn and improve using feedback, previous work proposed …
Retrieve-rewrite-answer: A kg-to-text enhanced llms framework for knowledge graph question answering
Despite their competitive performance on knowledge-intensive tasks, large language
models (LLMs) still have limitations in memorizing all world knowledge especially long tail …
models (LLMs) still have limitations in memorizing all world knowledge especially long tail …
Intentqa: Context-aware video intent reasoning
In this paper, we propose a novel task IntentQA, a special VideoQA task focusing on video
intent reasoning, which has become increasingly important for AI with its advantages in …
intent reasoning, which has become increasingly important for AI with its advantages in …
Large language models are versatile decomposers: Decomposing evidence and questions for table-based reasoning
Table-based reasoning has shown remarkable progress in a wide range of table-based
tasks. It is a challenging task, which requires reasoning over both free-form natural language …
tasks. It is a challenging task, which requires reasoning over both free-form natural language …
Complex QA and language models hybrid architectures, Survey
This paper reviews the state-of-the-art of language models architectures and strategies for"
complex" question-answering (QA, CQA, CPS) with a focus on hybridization. Large …
complex" question-answering (QA, CQA, CPS) with a focus on hybridization. Large …