[HTML][HTML] Pre-trained language models and their applications

H Wang, J Li, H Wu, E Hovy, Y Sun - Engineering, 2023 - Elsevier
Pre-trained language models have achieved striking success in natural language
processing (NLP), leading to a paradigm shift from supervised learning to pre-training …

Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing

P Liu, W Yuan, J Fu, Z Jiang, H Hayashi… - ACM computing …, 2023 - dl.acm.org
This article surveys and organizes research works in a new paradigm in natural language
processing, which we dub “prompt-based learning.” Unlike traditional supervised learning …

[HTML][HTML] ChatGPT: Vision and challenges

SS Gill, R Kaur - Internet of Things and Cyber-Physical Systems, 2023 - Elsevier
Artificial intelligence (AI) and machine learning have changed the nature of scientific inquiry
in recent years. Of these, the development of virtual assistants has accelerated greatly in the …

Compute trends across three eras of machine learning

J Sevilla, L Heim, A Ho, T Besiroglu… - … Joint Conference on …, 2022 - ieeexplore.ieee.org
Compute, data, and algorithmic advances are the three fundamental factors that drive
progress in modern Machine Learning (ML). In this paper we study trends in the most readily …

Ernie 3.0: Large-scale knowledge enhanced pre-training for language understanding and generation

Y Sun, S Wang, S Feng, S Ding, C Pang… - arxiv preprint arxiv …, 2021 - arxiv.org
Pre-trained models have achieved state-of-the-art results in various Natural Language
Processing (NLP) tasks. Recent works such as T5 and GPT-3 have shown that scaling up …

Pre-trained language models for text generation: A survey

J Li, T Tang, WX Zhao, JY Nie, JR Wen - ACM Computing Surveys, 2024 - dl.acm.org
Text Generation aims to produce plausible and readable text in human language from input
data. The resurgence of deep learning has greatly advanced this field, in particular, with the …

In conversation with artificial intelligence: aligning language models with human values

A Kasirzadeh, I Gabriel - Philosophy & Technology, 2023 - Springer
Large-scale language technologies are increasingly used in various forms of
communication with humans across different contexts. One particular use case for these …

PanGu-: Large-scale Autoregressive Pretrained Chinese Language Models with Auto-parallel Computation

W Zeng, X Ren, T Su, H Wang, Y Liao, Z Wang… - arxiv preprint arxiv …, 2021 - arxiv.org
Large-scale Pretrained Language Models (PLMs) have become the new paradigm for
Natural Language Processing (NLP). PLMs with hundreds of billions parameters such as …

Chinesebert: Chinese pretraining enhanced by glyph and pinyin information

Z Sun, X Li, X Sun, Y Meng, X Ao, Q He, F Wu… - arxiv preprint arxiv …, 2021 - arxiv.org
Recent pretraining models in Chinese neglect two important aspects specific to the Chinese
language: glyph and pinyin, which carry significant syntax and semantic information for …

Cpt: A pre-trained unbalanced transformer for both chinese language understanding and generation

Y Shao, Z Geng, Y Liu, J Dai, H Yan, F Yang… - Science China …, 2024 - Springer
In this paper, we take the advantage of previous pre-trained models (PTMs) and propose a
novel Chinese pre-trained unbalanced transformer (CPT). Different from previous Chinese …