Recent advances in natural language processing via large pre-trained language models: A survey
Large, pre-trained language models (PLMs) such as BERT and GPT have drastically
changed the Natural Language Processing (NLP) field. For numerous NLP tasks …
changed the Natural Language Processing (NLP) field. For numerous NLP tasks …
Linguistically inspired roadmap for building biologically reliable protein language models
Deep neural-network-based language models (LMs) are increasingly applied to large-scale
protein sequence data to predict protein function. However, being largely black-box models …
protein sequence data to predict protein function. However, being largely black-box models …
Testing the predictions of surprisal theory in 11 languages
Surprisal theory posits that less-predictable words should take more time to process, with
word predictability quantified as surprisal, ie, negative log probability in context. While …
word predictability quantified as surprisal, ie, negative log probability in context. While …
A primer on pretrained multilingual language models
Multilingual Language Models (\MLLMs) such as mBERT, XLM, XLM-R,\textit {etc.} have
emerged as a viable option for bringing the power of pretraining to a large number of …
emerged as a viable option for bringing the power of pretraining to a large number of …
Parsing with multilingual BERT, a small corpus, and a small treebank
Pretrained multilingual contextual representations have shown great success, but due to the
limits of their pretraining data, their benefits do not apply equally to all language varieties …
limits of their pretraining data, their benefits do not apply equally to all language varieties …
Fine-tuning BERT for low-resource natural language understanding via active learning
Recently, leveraging pre-trained Transformer based language models in down stream, task
specific models has advanced state of the art results in natural language understanding …
specific models has advanced state of the art results in natural language understanding …
Lost in translation: large language models in non-English content analysis
In recent years, large language models (eg, Open AI's GPT-4, Meta's LLaMa, Google's
PaLM) have become the dominant approach for building AI systems to analyze and …
PaLM) have become the dominant approach for building AI systems to analyze and …
A Warm Start and a Clean Crawled Corpus--A Recipe for Good Language Models
V Snæbjarnarson, HB Símonarson… - arxiv preprint arxiv …, 2022 - arxiv.org
We train several language models for Icelandic, including IceBERT, that achieve state-of-the-
art performance in a variety of downstream tasks, including part-of-speech tagging, named …
art performance in a variety of downstream tasks, including part-of-speech tagging, named …
Are multilingual models the best choice for moderately under-resourced languages? A comprehensive assessment for Catalan
Multilingual language models have been a crucial breakthrough as they considerably
reduce the need of data for under-resourced languages. Nevertheless, the superiority of …
reduce the need of data for under-resourced languages. Nevertheless, the superiority of …
Multidimensional affective analysis for low-resource languages: A use case with guarani-spanish code-switching language
This paper focuses on text-based affective computing for Jopara, a code-switching language
that combines Guarani and Spanish. First, we collected a dataset of tweets primarily written …
that combines Guarani and Spanish. First, we collected a dataset of tweets primarily written …