Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
[PDF][PDF] A survey of large language models
Ever since the Turing Test was proposed in the 1950s, humans have explored the mastering
of language intelligence by machine. Language is essentially a complex, intricate system of …
of language intelligence by machine. Language is essentially a complex, intricate system of …
Yi: Open foundation models by 01. ai
We introduce the Yi model family, a series of language and multimodal models that
demonstrate strong multi-dimensional capabilities. The Yi model family is based on 6B and …
demonstrate strong multi-dimensional capabilities. The Yi model family is based on 6B and …
You only cache once: Decoder-decoder architectures for language models
We introduce a decoder-decoder architecture, YOCO, for large language models, which only
caches key-value pairs once. It consists of two components, ie, a cross-decoder stacked …
caches key-value pairs once. It consists of two components, ie, a cross-decoder stacked …
RULER: What's the Real Context Size of Your Long-Context Language Models?
The needle-in-a-haystack (NIAH) test, which examines the ability to retrieve a piece of
information (the" needle") from long distractor texts (the" haystack"), has been widely …
information (the" needle") from long distractor texts (the" haystack"), has been widely …
Long-context llms struggle with long in-context learning
Large Language Models (LLMs) have made significant strides in handling long sequences.
Some models like Gemini could even to be capable of dealing with millions of tokens …
Some models like Gemini could even to be capable of dealing with millions of tokens …
Mlvu: A comprehensive benchmark for multi-task long video understanding
The evaluation of Long Video Understanding (LVU) performance poses an important but
challenging research problem. Despite previous efforts, the existing video understanding …
challenging research problem. Despite previous efforts, the existing video understanding …
[PDF][PDF] Leave no context behind: Efficient infinite context transformers with infini-attention
This work introduces an efficient method to scale Transformer-based Large Language
Models (LLMs) to infinitely long inputs with bounded memory and computation. A key …
Models (LLMs) to infinitely long inputs with bounded memory and computation. A key …
Longvila: Scaling long-context visual language models for long videos
Long-context capability is critical for multi-modal foundation models, especially for long
video understanding. We introduce LongVILA, a full-stack solution for long-context visual …
video understanding. We introduce LongVILA, a full-stack solution for long-context visual …
Minference 1.0: Accelerating pre-filling for long-context llms via dynamic sparse attention
The computational challenges of Large Language Model (LLM) inference remain a
significant barrier to their widespread deployment, especially as prompt lengths continue to …
significant barrier to their widespread deployment, especially as prompt lengths continue to …
Longalign: A recipe for long context alignment of large language models
Extending large language models to effectively handle long contexts requires instruction fine-
tuning on input sequences of similar length. To address this, we present LongAlign--a recipe …
tuning on input sequences of similar length. To address this, we present LongAlign--a recipe …