Leveraging explanations in interactive machine learning: An overview

S Teso, Ö Alkan, W Stammer, E Daly - Frontiers in Artificial …, 2023 - frontiersin.org
Explanations have gained an increasing level of interest in the AI and Machine Learning
(ML) communities in order to improve model transparency and allow users to form a mental …

Distilling step-by-step! outperforming larger language models with less training data and smaller model sizes

CY Hsieh, CL Li, CK Yeh, H Nakhost, Y Fujii… - arxiv preprint arxiv …, 2023 - arxiv.org
Deploying large language models (LLMs) is challenging because they are memory
inefficient and compute-intensive for practical applications. In reaction, researchers train …

Challenging big-bench tasks and whether chain-of-thought can solve them

M Suzgun, N Scales, N Schärli, S Gehrmann… - arxiv preprint arxiv …, 2022 - arxiv.org
BIG-Bench (Srivastava et al., 2022) is a diverse evaluation suite that focuses on tasks
believed to be beyond the capabilities of current language models. Language models have …

Chain-of-thought prompting elicits reasoning in large language models

J Wei, X Wang, D Schuurmans… - Advances in neural …, 2022 - proceedings.neurips.cc
We explore how generating a chain of thought---a series of intermediate reasoning steps---
significantly improves the ability of large language models to perform complex reasoning. In …

Can language models learn from explanations in context?

AK Lampinen, I Dasgupta, SCY Chan… - arxiv preprint arxiv …, 2022 - arxiv.org
Language Models (LMs) can perform new tasks by adapting to a few in-context examples.
For humans, explanations that connect examples to task principles can improve learning …

Cross-task generalization via natural language crowdsourcing instructions

S Mishra, D Khashabi, C Baral, H Hajishirzi - arxiv preprint arxiv …, 2021 - arxiv.org
Humans (eg, crowdworkers) have a remarkable ability in solving different tasks, by simply
reading textual instructions that define them and looking at a few examples. Despite the …

Cumulative reasoning with large language models

Y Zhang, J Yang, Y Yuan, ACC Yao - arxiv preprint arxiv:2308.04371, 2023 - arxiv.org
While language models are powerful and versatile, they often fail to address highly complex
problems. This is because solving complex problems requires deliberate thinking, which has …

Symbolic chain-of-thought distillation: Small models can also" think" step-by-step

LH Li, J Hessel, Y Yu, X Ren, KW Chang… - arxiv preprint arxiv …, 2023 - arxiv.org
Chain-of-thought prompting (eg," Let's think step-by-step") primes large language models to
verbalize rationalization for their predictions. While chain-of-thought can lead to dramatic …

Flex: Unifying evaluation for few-shot nlp

J Bragg, A Cohan, K Lo… - Advances in Neural …, 2021 - proceedings.neurips.cc
Few-shot NLP research is highly active, yet conducted in disjoint research threads with
evaluation suites that lack challenging-yet-realistic testing setups and fail to employ careful …

Local interpretations for explainable natural language processing: A survey

S Luo, H Ivison, SC Han, J Poon - ACM Computing Surveys, 2024 - dl.acm.org
As the use of deep learning techniques has grown across various fields over the past
decade, complaints about the opaqueness of the black-box models have increased …