Recent advances in natural language processing via large pre-trained language models: A survey

B Min, H Ross, E Sulem, APB Veyseh… - ACM Computing …, 2023 - dl.acm.org
Large, pre-trained language models (PLMs) such as BERT and GPT have drastically
changed the Natural Language Processing (NLP) field. For numerous NLP tasks …

Linguistically inspired roadmap for building biologically reliable protein language models

MH Vu, R Akbar, PA Robert, B Swiatczak… - Nature Machine …, 2023 - nature.com
Deep neural-network-based language models (LMs) are increasingly applied to large-scale
protein sequence data to predict protein function. However, being largely black-box models …

Testing the predictions of surprisal theory in 11 languages

EG Wilcox, T Pimentel, C Meister, R Cotterell… - Transactions of the …, 2023 - direct.mit.edu
Surprisal theory posits that less-predictable words should take more time to process, with
word predictability quantified as surprisal, ie, negative log probability in context. While …

A primer on pretrained multilingual language models

S Doddapaneni, G Ramesh, MM Khapra… - arxiv preprint arxiv …, 2021 - arxiv.org
Multilingual Language Models (\MLLMs) such as mBERT, XLM, XLM-R,\textit {etc.} have
emerged as a viable option for bringing the power of pretraining to a large number of …

Parsing with multilingual BERT, a small corpus, and a small treebank

EC Chau, LH Lin, NA Smith - arxiv preprint arxiv:2009.14124, 2020 - arxiv.org
Pretrained multilingual contextual representations have shown great success, but due to the
limits of their pretraining data, their benefits do not apply equally to all language varieties …

Fine-tuning BERT for low-resource natural language understanding via active learning

D Grießhaber, J Maucher, NT Vu - arxiv preprint arxiv:2012.02462, 2020 - arxiv.org
Recently, leveraging pre-trained Transformer based language models in down stream, task
specific models has advanced state of the art results in natural language understanding …

Lost in translation: large language models in non-English content analysis

G Nicholas, A Bhatia - arxiv preprint arxiv:2306.07377, 2023 - arxiv.org
In recent years, large language models (eg, Open AI's GPT-4, Meta's LLaMa, Google's
PaLM) have become the dominant approach for building AI systems to analyze and …

A Warm Start and a Clean Crawled Corpus--A Recipe for Good Language Models

V Snæbjarnarson, HB Símonarson… - arxiv preprint arxiv …, 2022 - arxiv.org
We train several language models for Icelandic, including IceBERT, that achieve state-of-the-
art performance in a variety of downstream tasks, including part-of-speech tagging, named …

Are multilingual models the best choice for moderately under-resourced languages? A comprehensive assessment for Catalan

J Armengol-Estapé, CP Carrino… - arxiv preprint arxiv …, 2021 - arxiv.org
Multilingual language models have been a crucial breakthrough as they considerably
reduce the need of data for under-resourced languages. Nevertheless, the superiority of …

Multidimensional affective analysis for low-resource languages: A use case with guarani-spanish code-switching language

MM Agüero-Torales, AG López-Herrera… - Cognitive Computation, 2023 - Springer
This paper focuses on text-based affective computing for Jopara, a code-switching language
that combines Guarani and Spanish. First, we collected a dataset of tweets primarily written …