A comprehensive overview of large language models

H Naveed, AU Khan, S Qiu, M Saqib, S Anwar… - arxiv preprint arxiv …, 2023 - arxiv.org
Large Language Models (LLMs) have recently demonstrated remarkable capabilities in
natural language processing tasks and beyond. This success of LLMs has led to a large …

Datasets for large language models: A comprehensive survey

Y Liu, J Cao, C Liu, K Ding, L ** - arxiv preprint arxiv:2402.18041, 2024 - arxiv.org
This paper embarks on an exploration into the Large Language Model (LLM) datasets,
which play a crucial role in the remarkable advancements of LLMs. The datasets serve as …

Ppt: Pre-trained prompt tuning for few-shot learning

Y Gu, X Han, Z Liu, M Huang - arxiv preprint arxiv:2109.04332, 2021 - arxiv.org
Prompts for pre-trained language models (PLMs) have shown remarkable performance by
bridging the gap between pre-training tasks and various downstream tasks. Among these …

Ernie 3.0: Large-scale knowledge enhanced pre-training for language understanding and generation

Y Sun, S Wang, S Feng, S Ding, C Pang… - arxiv preprint arxiv …, 2021 - arxiv.org
Pre-trained models have achieved state-of-the-art results in various Natural Language
Processing (NLP) tasks. Recent works such as T5 and GPT-3 have shown that scaling up …

Graph neural networks for natural language processing: A survey

L Wu, Y Chen, K Shen, X Guo, H Gao… - … and Trends® in …, 2023 - nowpublishers.com
Deep learning has become the dominant approach in addressing various tasks in Natural
Language Processing (NLP). Although text inputs are typically represented as a sequence …

Revisiting pre-trained models for Chinese natural language processing

Y Cui, W Che, T Liu, B Qin, S Wang, G Hu - arxiv preprint arxiv …, 2020 - arxiv.org
Bidirectional Encoder Representations from Transformers (BERT) has shown marvelous
improvements across various NLP tasks, and consecutive variants have been proposed to …

K-bert: Enabling language representation with knowledge graph

W Liu, P Zhou, Z Zhao, Z Wang, Q Ju, H Deng… - Proceedings of the …, 2020 - ojs.aaai.org
Pre-trained language representation models, such as BERT, capture a general language
representation from large-scale corpora, but lack domain-specific knowledge. When reading …

Pre-training with whole word masking for chinese bert

Y Cui, W Che, T Liu, B Qin… - IEEE/ACM Transactions on …, 2021 - ieeexplore.ieee.org
Bidirectional Encoder Representations from Transformers (BERT) has shown marvelous
improvements across various NLP tasks, and its consecutive variants have been proposed …

Ernie 2.0: A continual pre-training framework for language understanding

Y Sun, S Wang, Y Li, S Feng, H Tian, H Wu… - Proceedings of the AAAI …, 2020 - ojs.aaai.org
Recently pre-trained models have achieved state-of-the-art results in various language
understanding tasks. Current pre-training procedures usually focus on training the model …

Ernie: Enhanced representation through knowledge integration

Y Sun, S Wang, Y Li, S Feng, X Chen, H Zhang… - arxiv preprint arxiv …, 2019 - arxiv.org
We present a novel language representation model enhanced by knowledge called ERNIE
(Enhanced Representation through kNowledge IntEgration). Inspired by the masking …