Think twice before trusting: Self-detection for large language models through comprehensive answer reflection

M Li, W Wang, F Feng, F Zhu, Q Wang… - Findings of the …, 2024 - aclanthology.org
Abstract Self-detection for Large Language Models (LLMs) seeks to evaluate the
trustworthiness of the LLM's output by leveraging its own capabilities, thereby alleviating the …

Dawn-icl: Strategic planning of problem-solving trajectories for zero-shot in-context learning

X Tang, X Wang, WX Zhao, JR Wen - arxiv preprint arxiv:2410.20215, 2024 - arxiv.org
Zero-shot in-context learning (ZS-ICL) aims to conduct in-context learning (ICL) without
using human-annotated demonstrations. Most ZS-ICL methods use large language models …

Towards Scalable Automated Alignment of LLMs: A Survey

B Cao, K Lu, X Lu, J Chen, M Ren, H **ang… - arxiv preprint arxiv …, 2024 - arxiv.org
Alignment is the most critical step in building large language models (LLMs) that meet
human needs. With the rapid development of LLMs gradually surpassing human …

TuringQ: Benchmarking AI Comprehension in Theory of Computation

PS Zahraei, E Asgari - arxiv preprint arxiv:2410.06547, 2024 - arxiv.org
We present TuringQ, the first benchmark designed to evaluate the reasoning capabilities of
large language models (LLMs) in the theory of computation. TuringQ consists of 4,006 …