Palm 2 technical report
We introduce PaLM 2, a new state-of-the-art language model that has better multilingual and
reasoning capabilities and is more compute-efficient than its predecessor PaLM. PaLM 2 is …
reasoning capabilities and is more compute-efficient than its predecessor PaLM. PaLM 2 is …
Scaling data-constrained language models
The current trend of scaling language models involves increasing both parameter count and
training dataset size. Extrapolating this trend suggests that training dataset size may soon be …
training dataset size. Extrapolating this trend suggests that training dataset size may soon be …
Octopack: Instruction tuning code large language models
Finetuning large language models (LLMs) on instructions leads to vast performance
improvements on natural language tasks. We apply instruction tuning using code …
improvements on natural language tasks. We apply instruction tuning using code …
Madlad-400: A multilingual and document-level large audited dataset
We introduce MADLAD-400, a manually audited, general domain 3T token monolingual
dataset based on CommonCrawl, spanning 419 languages. We discuss the limitations …
dataset based on CommonCrawl, spanning 419 languages. We discuss the limitations …
Knowledge transfer from high-resource to low-resource programming languages for code llms
Over the past few years, Large Language Models of Code (Code LLMs) have started to have
a significant impact on programming practice. Code LLMs are also emerging as building …
a significant impact on programming practice. Code LLMs are also emerging as building …
BioCoder: a benchmark for bioinformatics code generation with large language models
X Tang, B Qian, R Gao, J Chen, X Chen… - …, 2024 - academic.oup.com
Pretrained large language models (LLMs) have significantly improved code generation. As
these models scale up, there is an increasing need for the output to handle more intricate …
these models scale up, there is an increasing need for the output to handle more intricate …
BioCoder: a benchmark for bioinformatics code generation with contextual pragmatic knowledge
Pre-trained language models like ChatGPT have significantly improved code generation. As
these models scale up, there is an increasing need for the output to handle more intricate …
these models scale up, there is an increasing need for the output to handle more intricate …
Ircoder: Intermediate representations make language models robust multilingual code generators
Code understanding and generation have fast become some of the most popular
applications of language models (LMs). Nonetheless, research on multilingual aspects of …
applications of language models (LMs). Nonetheless, research on multilingual aspects of …
Solmover: Smart contract code translation based on concepts
Large language models (LLMs) have showcased remarkable skills, rivaling or even
exceeding human intelligence in certain areas. Their proficiency in translation is notable, as …
exceeding human intelligence in certain areas. Their proficiency in translation is notable, as …
Batched low-rank adaptation of foundation models
Low-Rank Adaptation (LoRA) has recently gained attention for fine-tuning foundation
models by incorporating trainable low-rank matrices, thereby reducing the number of …
models by incorporating trainable low-rank matrices, thereby reducing the number of …