Palm 2 technical report

R Anil, AM Dai, O Firat, M Johnson, D Lepikhin… - arxiv preprint arxiv …, 2023 - arxiv.org
We introduce PaLM 2, a new state-of-the-art language model that has better multilingual and
reasoning capabilities and is more compute-efficient than its predecessor PaLM. PaLM 2 is …

Scaling data-constrained language models

N Muennighoff, A Rush, B Barak… - Advances in …, 2023 - proceedings.neurips.cc
The current trend of scaling language models involves increasing both parameter count and
training dataset size. Extrapolating this trend suggests that training dataset size may soon be …

Octopack: Instruction tuning code large language models

N Muennighoff, Q Liu, A Zebaze, Q Zheng… - arxiv preprint arxiv …, 2023 - arxiv.org
Finetuning large language models (LLMs) on instructions leads to vast performance
improvements on natural language tasks. We apply instruction tuning using code …

Madlad-400: A multilingual and document-level large audited dataset

S Kudugunta, I Caswell, B Zhang… - Advances in …, 2024 - proceedings.neurips.cc
We introduce MADLAD-400, a manually audited, general domain 3T token monolingual
dataset based on CommonCrawl, spanning 419 languages. We discuss the limitations …

Knowledge transfer from high-resource to low-resource programming languages for code llms

F Cassano, J Gouwar, F Lucchetti… - Proceedings of the …, 2024 - dl.acm.org
Over the past few years, Large Language Models of Code (Code LLMs) have started to have
a significant impact on programming practice. Code LLMs are also emerging as building …

BioCoder: a benchmark for bioinformatics code generation with large language models

X Tang, B Qian, R Gao, J Chen, X Chen… - …, 2024 - academic.oup.com
Pretrained large language models (LLMs) have significantly improved code generation. As
these models scale up, there is an increasing need for the output to handle more intricate …

BioCoder: a benchmark for bioinformatics code generation with contextual pragmatic knowledge

X Tang, B Qian, R Gao, J Chen, X Chen… - arxiv preprint arxiv …, 2023 - arxiv.org
Pre-trained language models like ChatGPT have significantly improved code generation. As
these models scale up, there is an increasing need for the output to handle more intricate …

Ircoder: Intermediate representations make language models robust multilingual code generators

I Paul, G Glavaš, I Gurevych - arxiv preprint arxiv:2403.03894, 2024 - arxiv.org
Code understanding and generation have fast become some of the most popular
applications of language models (LMs). Nonetheless, research on multilingual aspects of …

Solmover: Smart contract code translation based on concepts

R Karanjai, L Xu, W Shi - Proceedings of the 1st ACM International …, 2024 - dl.acm.org
Large language models (LLMs) have showcased remarkable skills, rivaling or even
exceeding human intelligence in certain areas. Their proficiency in translation is notable, as …

Batched low-rank adaptation of foundation models

Y Wen, S Chaudhuri - arxiv preprint arxiv:2312.05677, 2023 - arxiv.org
Low-Rank Adaptation (LoRA) has recently gained attention for fine-tuning foundation
models by incorporating trainable low-rank matrices, thereby reducing the number of …