One small step for generative ai, one giant leap for agi: A complete survey on chatgpt in aigc era

C Zhang, C Zhang, C Li, Y Qiao, S Zheng… - arxiv preprint arxiv …, 2023 - arxiv.org
OpenAI has recently released GPT-4 (aka ChatGPT plus), which is demonstrated to be one
small step for generative AI (GAI), but one giant leap for artificial general intelligence (AGI) …

Deep learning--based text classification: a comprehensive review

S Minaee, N Kalchbrenner, E Cambria… - ACM computing …, 2021 - dl.acm.org
Deep learning--based models have surpassed classical machine learning--based
approaches in various text classification tasks, including sentiment analysis, news …

Beyond the imitation game: Quantifying and extrapolating the capabilities of language models

A Srivastava, A Rastogi, A Rao, AAM Shoeb… - arxiv preprint arxiv …, 2022 - arxiv.org
Language models demonstrate both quantitative improvement and new qualitative
capabilities with increasing scale. Despite their potentially transformative impact, these new …

Byt5: Towards a token-free future with pre-trained byte-to-byte models

L Xue, A Barua, N Constant, R Al-Rfou… - Transactions of the …, 2022 - direct.mit.edu
Most widely used pre-trained language models operate on sequences of tokens
corresponding to word or subword units. By comparison, token-free models that operate …

Restoring and attributing ancient texts using deep neural networks

Y Assael, T Sommerschield, B Shillingford, M Bordbar… - Nature, 2022 - nature.com
Ancient history relies on disciplines such as epigraphy—the study of inscribed texts known
as inscriptions—for evidence of the thought, language, society and history of past …

Pre-trained models for natural language processing: A survey

X Qiu, T Sun, Y Xu, Y Shao, N Dai, X Huang - Science China …, 2020 - Springer
Recently, the emergence of pre-trained models (PTMs) has brought natural language
processing (NLP) to a new era. In this survey, we provide a comprehensive review of PTMs …

Curriculum learning: A survey

P Soviany, RT Ionescu, P Rota, N Sebe - International Journal of …, 2022 - Springer
Training machine learning models in a meaningful order, from the easy samples to the hard
ones, using curriculum learning can provide performance improvements over the standard …

Biological structure and function emerge from scaling unsupervised learning to 250 million protein sequences

A Rives, J Meier, T Sercu, S Goyal… - Proceedings of the …, 2021 - National Acad Sciences
In the field of artificial intelligence, a combination of scale in data and model capacity
enabled by unsupervised learning has led to major advances in representation learning and …

Federated learning for mobile keyboard prediction

A Hard, K Rao, R Mathews, S Ramaswamy… - arxiv preprint arxiv …, 2018 - arxiv.org
We train a recurrent neural network language model using a distributed, on-device learning
framework called federated learning for the purpose of next-word prediction in a virtual …

A survey of the usages of deep learning for natural language processing

DW Otter, JR Medina, JK Kalita - IEEE transactions on neural …, 2020 - ieeexplore.ieee.org
Over the last several years, the field of natural language processing has been propelled
forward by an explosion in the use of deep learning models. This article provides a brief …