Challenges and applications of large language models

J Kaddour, J Harris, M Mozes, H Bradley… - arxiv preprint arxiv …, 2023 - arxiv.org
Large Language Models (LLMs) went from non-existent to ubiquitous in the machine
learning discourse within a few years. Due to the fast pace of the field, it is difficult to identify …

A survey of large language models

WX Zhao, K Zhou, J Li, T Tang, X Wang, Y Hou… - arxiv preprint arxiv …, 2023 - arxiv.org
Language is essentially a complex, intricate system of human expressions governed by
grammatical rules. It poses a significant challenge to develop capable AI algorithms for …

Toolqa: A dataset for llm question answering with external tools

Y Zhuang, Y Yu, K Wang, H Sun… - Advances in Neural …, 2023 - proceedings.neurips.cc
Abstract Large Language Models (LLMs) have demonstrated impressive performance in
various NLP tasks, but they still suffer from challenges such as hallucination and weak …

Pal: Program-aided language models

L Gao, A Madaan, S Zhou, U Alon… - International …, 2023 - proceedings.mlr.press
Large language models (LLMs) have demonstrated an impressive ability to perform
arithmetic and symbolic reasoning tasks, when provided with a few examples at test time (" …

Evoprompting: Language models for code-level neural architecture search

A Chen, D Dohan, D So - Advances in neural information …, 2023 - proceedings.neurips.cc
Given the recent impressive accomplishments of language models (LMs) for code
generation, we explore the use of LMs as general adaptive mutation and crossover …

Faithful chain-of-thought reasoning

Q Lyu, S Havaldar, A Stein, L Zhang, D Rao… - arxiv preprint arxiv …, 2023 - arxiv.org
While Chain-of-Thought (CoT) prompting boosts Language Models'(LM) performance on a
gamut of complex reasoning tasks, the generated reasoning chain does not necessarily …

Teaching arithmetic to small transformers

N Lee, K Sreenivasan, JD Lee, K Lee… - arxiv preprint arxiv …, 2023 - arxiv.org
Large language models like GPT-4 exhibit emergent capabilities across general-purpose
tasks, such as basic arithmetic, when trained on extensive text data, even though these tasks …

Large language models are few-shot health learners

X Liu, D McDuff, G Kovacs, I Galatzer-Levy… - arxiv preprint arxiv …, 2023 - arxiv.org
Large language models (LLMs) can capture rich representations of concepts that are useful
for real-world tasks. However, language alone is limited. While existing LLMs excel at text …

Evaluating and improving tool-augmented computation-intensive math reasoning

B Zhang, K Zhou, X Wei, X Zhao… - Advances in …, 2023 - proceedings.neurips.cc
Chain-of-thought prompting (CoT) and tool augmentation have been validated in recent
work as effective practices for improving large language models (LLMs) to perform step-by …

Towards faithful model explanation in nlp: A survey

Q Lyu, M Apidianaki, C Callison-Burch - Computational Linguistics, 2024 - direct.mit.edu
End-to-end neural Natural Language Processing (NLP) models are notoriously difficult to
understand. This has given rise to numerous efforts towards model explainability in recent …