A comprehensive overview of large language models
Large Language Models (LLMs) have recently demonstrated remarkable capabilities in
natural language processing tasks and beyond. This success of LLMs has led to a large …
natural language processing tasks and beyond. This success of LLMs has led to a large …
Ammus: A survey of transformer-based pretrained models in natural language processing
KS Kalyan, A Rajasekharan, S Sangeetha - arxiv preprint arxiv …, 2021 - arxiv.org
Transformer-based pretrained language models (T-PTLMs) have achieved great success in
almost every NLP task. The evolution of these models started with GPT and BERT. These …
almost every NLP task. The evolution of these models started with GPT and BERT. These …
C-pack: Packed resources for general chinese embeddings
We introduce C-Pack, a package of resources that significantly advances the field of general
text embeddings for Chinese. C-Pack includes three critical resources. 1) C-MTP is a …
text embeddings for Chinese. C-Pack includes three critical resources. 1) C-MTP is a …
Glm-130b: An open bilingual pre-trained model
We introduce GLM-130B, a bilingual (English and Chinese) pre-trained language model
with 130 billion parameters. It is an attempt to open-source a 100B-scale model at least as …
with 130 billion parameters. It is an attempt to open-source a 100B-scale model at least as …
The bigscience roots corpus: A 1.6 tb composite multilingual dataset
As language models grow ever larger, the need for large-scale high-quality text datasets has
never been more pressing, especially in multilingual settings. The BigScience workshop, a 1 …
never been more pressing, especially in multilingual settings. The BigScience workshop, a 1 …
Cogview: Mastering text-to-image generation via transformers
Text-to-Image generation in the general domain has long been an open problem, which
requires both a powerful generative model and cross-modal understanding. We propose …
requires both a powerful generative model and cross-modal understanding. We propose …
Ppt: Pre-trained prompt tuning for few-shot learning
Prompts for pre-trained language models (PLMs) have shown remarkable performance by
bridging the gap between pre-training tasks and various downstream tasks. Among these …
bridging the gap between pre-training tasks and various downstream tasks. Among these …
Continual learning of large language models: A comprehensive survey
The recent success of large language models (LLMs) trained on static, pre-collected,
general datasets has sparked numerous research directions and applications. One such …
general datasets has sparked numerous research directions and applications. One such …
NusaCrowd: Open source initiative for Indonesian NLP resources
We present NusaCrowd, a collaborative initiative to collect and unify existing resources for
Indonesian languages, including opening access to previously non-public resources …
Indonesian languages, including opening access to previously non-public resources …
The all-seeing project: Towards panoptic visual recognition and understanding of the open world
We present the All-Seeing (AS) project: a large-scale data and model for recognizing and
understanding everything in the open world. Using a scalable data engine that incorporates …
understanding everything in the open world. Using a scalable data engine that incorporates …