The llama 3 herd of models

A Dubey, A Jauhri, A Pandey, A Kadian… - arxiv preprint arxiv …, 2024 - arxiv.org
Modern artificial intelligence (AI) systems are powered by foundation models. This paper
presents a new set of foundation models, called Llama 3. It is a herd of language models …

Wizardcoder: Empowering code large language models with evol-instruct

Z Luo, C Xu, P Zhao, Q Sun, X Geng, W Hu… - arxiv preprint arxiv …, 2023 - arxiv.org
Code Large Language Models (Code LLMs), such as StarCoder, have demonstrated
exceptional performance in code-related tasks. However, most existing models are solely …

Unifying the perspectives of nlp and software engineering: A survey on language models for code

Z Zhang, C Chen, B Liu, C Liao, Z Gong, H Yu… - arxiv preprint arxiv …, 2023 - arxiv.org
In this work we systematically review the recent advancements in software engineering with
language models, covering 70+ models, 40+ evaluation tasks, 180+ datasets, and 900 …

Gollie: Annotation guidelines improve zero-shot information-extraction

O Sainz, I García-Ferrero, R Agerri… - arxiv preprint arxiv …, 2023 - arxiv.org
Large Language Models (LLMs) combined with instruction tuning have made significant
progress when generalizing to unseen tasks. However, they have been less successful in …

Scaling laws of rope-based extrapolation

X Liu, H Yan, S Zhang, C An, X Qiu, D Lin - arxiv preprint arxiv …, 2023 - arxiv.org
The extrapolation capability of Large Language Models (LLMs) based on Rotary Position
Embedding is currently a topic of considerable interest. The mainstream approach to …

Training large language models for reasoning through reverse curriculum reinforcement learning

Z **, W Chen, B Hong, S **, R Zheng, W He… - arxiv preprint arxiv …, 2024 - arxiv.org
In this paper, we propose R $^ 3$: Learning Reasoning through Reverse Curriculum
Reinforcement Learning (RL), a novel method that employs only outcome supervision to …

Controlled text generation via language model arithmetic

J Dekoninck, M Fischer, L Beurer-Kellner… - arxiv preprint arxiv …, 2023 - arxiv.org
As Large Language Models (LLMs) are deployed more widely, customization with respect to
vocabulary, style, and character becomes more important. In this work, we introduce model …

Mhpp: Exploring the capabilities and limitations of language models beyond basic code generation

J Dai, J Lu, Y Feng, D Huang, G Zeng, R Ruan… - arxiv preprint arxiv …, 2024 - arxiv.org
Recent advancements in large language models (LLMs) have greatly improved code
generation, specifically at the function level. For instance, GPT-4o has achieved a 91.0 …

Dissecting learning and forgetting in language model finetuning

X Zhang, J Wu - The Twelfth International Conference on Learning …, 2024 - openreview.net
Finetuning language models on domain-specific corpus is a common approach to enhance
their domain knowledge and capability. While improving performance on domain tasks, it …

Reflectioncoder: Learning from reflection sequence for enhanced one-off code generation

H Ren, M Zhan, Z Wu, A Zhou, J Pan, H Li - arxiv preprint arxiv …, 2024 - arxiv.org
Code generation plays a crucial role in various tasks, such as code auto-completion and
mathematical reasoning. Previous work has proposed numerous methods to enhance code …