Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
The llama 3 herd of models
Modern artificial intelligence (AI) systems are powered by foundation models. This paper
presents a new set of foundation models, called Llama 3. It is a herd of language models …
presents a new set of foundation models, called Llama 3. It is a herd of language models …
Llamafactory: Unified efficient fine-tuning of 100+ language models
Efficient fine-tuning is vital for adapting large language models (LLMs) to downstream tasks.
However, it requires non-trivial efforts to implement these methods on different models. We …
However, it requires non-trivial efforts to implement these methods on different models. We …
From crowdsourced data to high-quality benchmarks: Arena-hard and benchbuilder pipeline
The rapid evolution of Large Language Models (LLMs) has outpaced the development of
model evaluation, highlighting the need for continuous curation of new, challenging …
model evaluation, highlighting the need for continuous curation of new, challenging …
Magpie: Alignment data synthesis from scratch by prompting aligned llms with nothing
High-quality instruction data is critical for aligning large language models (LLMs). Although
some models, such as Llama-3-Instruct, have open weights, their alignment data remain …
some models, such as Llama-3-Instruct, have open weights, their alignment data remain …
Simple and scalable strategies to continually pre-train large language models
Large language models (LLMs) are routinely pre-trained on billions of tokens, only to start
the process over again once new data becomes available. A much more efficient solution is …
the process over again once new data becomes available. A much more efficient solution is …
Livebench: A challenging, contamination-free llm benchmark
Test set contamination, wherein test data from a benchmark ends up in a newer model's
training set, is a well-documented obstacle for fair LLM evaluation and can quickly render …
training set, is a well-documented obstacle for fair LLM evaluation and can quickly render …
Qwen2. 5-math technical report: Toward mathematical expert model via self-improvement
In this report, we present a series of math-specific large language models: Qwen2. 5-Math
and Qwen2. 5-Math-Instruct-1.5 B/7B/72B. The core innovation of the Qwen2. 5 series lies in …
and Qwen2. 5-Math-Instruct-1.5 B/7B/72B. The core innovation of the Qwen2. 5 series lies in …
Unifying the perspectives of nlp and software engineering: A survey on language models for code
Z Zhang, C Chen, B Liu, C Liao, Z Gong, H Yu… - arxiv preprint arxiv …, 2023 - arxiv.org
In this work we systematically review the recent advancements in software engineering with
language models, covering 70+ models, 40+ evaluation tasks, 180+ datasets, and 900 …
language models, covering 70+ models, 40+ evaluation tasks, 180+ datasets, and 900 …
A survey on large language models for software engineering
Software Engineering (SE) is the systematic design, development, maintenance, and
management of software applications underpinning the digital infrastructure of our modern …
management of software applications underpinning the digital infrastructure of our modern …
Scaling synthetic data creation with 1,000,000,000 personas
We propose a novel persona-driven data synthesis methodology that leverages various
perspectives within a large language model (LLM) to create diverse synthetic data. To fully …
perspectives within a large language model (LLM) to create diverse synthetic data. To fully …