RULER: What's the Real Context Size of Your Long-Context Language Models?
The needle-in-a-haystack (NIAH) test, which examines the ability to retrieve a piece of
information (the" needle") from long distractor texts (the" haystack"), has been widely …
information (the" needle") from long distractor texts (the" haystack"), has been widely …
Video-xl: Extra-long vision language model for hour-scale video understanding
Although current Multi-modal Large Language Models (MLLMs) demonstrate promising
results in video understanding, processing extremely long videos remains an ongoing …
results in video understanding, processing extremely long videos remains an ongoing …
Longwriter: Unleashing 10,000+ word generation from long context llms
Current long context large language models (LLMs) can process inputs up to 100,000
tokens, yet struggle to generate outputs exceeding even a modest length of 2,000 words …
tokens, yet struggle to generate outputs exceeding even a modest length of 2,000 words …
Longalign: A recipe for long context alignment of large language models
Extending large language models to effectively handle long contexts requires instruction fine-
tuning on input sequences of similar length. To address this, we present LongAlign--a recipe …
tuning on input sequences of similar length. To address this, we present LongAlign--a recipe …
Found in the middle: How language models use long contexts better via plug-and-play positional encoding
This paper aims to overcome the" lost-in-the-middle" challenge of large language models
(LLMs). While recent advancements have successfully enabled LLMs to perform stable …
(LLMs). While recent advancements have successfully enabled LLMs to perform stable …
Eigen attention: Attention in low-rank space for kv cache compression
Large language models (LLMs) represent a groundbreaking advancement in the domain of
natural language processing due to their impressive reasoning abilities. Recently, there has …
natural language processing due to their impressive reasoning abilities. Recently, there has …
Novelqa: A benchmark for long-range novel question answering
The rapid advancement of Large Language Models (LLMs) has introduced a new frontier in
natural language processing, particularly in understanding and processing long-context …
natural language processing, particularly in understanding and processing long-context …
Training-free long-context scaling of large language models
The ability of Large Language Models (LLMs) to process and generate coherent text is
markedly weakened when the number of input tokens exceeds their pretraining length …
markedly weakened when the number of input tokens exceeds their pretraining length …
Triforce: Lossless acceleration of long sequence generation with hierarchical speculative decoding
With large language models (LLMs) widely deployed in long content generation recently,
there has emerged an increasing demand for efficient long-sequence inference support …
there has emerged an increasing demand for efficient long-sequence inference support …
CodeS: Natural Language to Code Repository via Multi-Layer Sketch
The impressive performance of large language models (LLMs) on code-related tasks has
shown the potential of fully automated software development. In light of this, we introduce a …
shown the potential of fully automated software development. In light of this, we introduce a …