[HTML][HTML] A survey of GPT-3 family large language models including ChatGPT and GPT-4

KS Kalyan - Natural Language Processing Journal, 2024 - Elsevier
Large language models (LLMs) are a special class of pretrained language models (PLMs)
obtained by scaling model size, pretraining corpus and computation. LLMs, because of their …

Evaluating large language models: A comprehensive survey

Z Guo, R **, C Liu, Y Huang, D Shi, L Yu, Y Liu… - arxiv preprint arxiv …, 2023 - arxiv.org
Large language models (LLMs) have demonstrated remarkable capabilities across a broad
spectrum of tasks. They have attracted significant attention and been deployed in numerous …

Efficient streaming language models with attention sinks

G **ao, Y Tian, B Chen, S Han, M Lewis - arxiv preprint arxiv:2309.17453, 2023 - arxiv.org
Deploying Large Language Models (LLMs) in streaming applications such as multi-round
dialogue, where long interactions are expected, is urgently needed but poses two major …

Efficient and effective text encoding for chinese llama and alpaca

Y Cui, Z Yang, X Yao - arxiv preprint arxiv:2304.08177, 2023 - arxiv.org
Large Language Models (LLMs), such as ChatGPT and GPT-4, have dramatically
transformed natural language processing research and shown promising strides towards …

RULER: What's the Real Context Size of Your Long-Context Language Models?

CP Hsieh, S Sun, S Kriman, S Acharya… - arxiv preprint arxiv …, 2024 - arxiv.org
The needle-in-a-haystack (NIAH) test, which examines the ability to retrieve a piece of
information (the" needle") from long distractor texts (the" haystack"), has been widely …

Benchmarking foundation models with language-model-as-an-examiner

Y Bai, J Ying, Y Cao, X Lv, Y He… - Advances in …, 2024 - proceedings.neurips.cc
Numerous benchmarks have been established to assess the performance of foundation
models on open-ended question answering, which serves as a comprehensive test of a …

∞ Bench: Extending long context evaluation beyond 100k tokens

X Zhang, Y Chen, S Hu, Z Xu, J Chen… - Proceedings of the …, 2024 - aclanthology.org
Processing and reasoning over long contexts is crucial for many practical applications of
Large Language Models (LLMs), such as document comprehension and agent construction …

Lm-infinite: Simple on-the-fly length generalization for large language models

C Han, Q Wang, W **ong, Y Chen, H Ji… - arxiv preprint arxiv …, 2023 - arxiv.org
In recent years, there have been remarkable advancements in the performance of
Transformer-based Large Language Models (LLMs) across various domains. As these LLMs …

Leave no document behind: Benchmarking long-context llms with extended multi-doc qa

M Wang, L Chen, F Cheng, S Liao… - Proceedings of the …, 2024 - aclanthology.org
Long-context modeling capabilities of Large Language Models (LLMs) have garnered
widespread attention, leading to the emergence of LLMs with ultra-context windows …

[PDF][PDF] Soaring from 4k to 400k: Extending llm's context with activation beacon

P Zhang, Z Liu, S **ao, N Shao, Q Ye… - arxiv preprint arxiv …, 2024 - openreview.net
The utilization of long contexts poses a big challenge for large language models due to their
limited context window length. Although the context window can be extended through fine …