Ammus: A survey of transformer-based pretrained models in natural language processing

KS Kalyan, A Rajasekharan, S Sangeetha - arxiv preprint arxiv …, 2021 - arxiv.org
Transformer-based pretrained language models (T-PTLMs) have achieved great success in
almost every NLP task. The evolution of these models started with GPT and BERT. These …

[PDF][PDF] Language model behavior: A comprehensive survey

TA Chang, BK Bergen - Computational Linguistics, 2024 - direct.mit.edu
Transformer language models have received widespread public attention, yet their
generated text is often surprising even to NLP researchers. In this survey, we discuss over …

A systematic review of transformer-based pre-trained language models through self-supervised learning

E Kotei, R Thirunavukarasu - Information, 2023 - mdpi.com
Transfer learning is a technique utilized in deep learning applications to transmit learned
inference to a different target domain. The approach is mainly to solve the problem of a few …

Towards tracing trustworthiness dynamics: Revisiting pre-training period of large language models

C Qian, J Zhang, W Yao, D Liu, Z Yin, Y Qiao… - arxiv preprint arxiv …, 2024 - arxiv.org
Ensuring the trustworthiness of large language models (LLMs) is crucial. Most studies
concentrate on fully pre-trained LLMs to better understand and improve LLMs' …

Semantic structure in deep learning

E Pavlick - Annual Review of Linguistics, 2022 - annualreviews.org
Deep learning has recently come to dominate computational linguistics, leading to claims of
human-level performance in a range of language processing tasks. Like much previous …

A closer look at how fine-tuning changes BERT

Y Zhou, V Srikumar - arxiv preprint arxiv:2106.14282, 2021 - arxiv.org
Given the prevalence of pre-trained contextualized representations in today's NLP, there
have been many efforts to understand what information they contain, and why they seem to …

How transfer learning impacts linguistic knowledge in deep NLP models?

N Durrani, H Sajjad, F Dalvi - arxiv preprint arxiv:2105.15179, 2021 - arxiv.org
Transfer learning from pre-trained neural language models towards downstream tasks has
been a predominant theme in NLP recently. Several researchers have shown that deep NLP …

Towards trustworthy and aligned machine learning: A data-centric survey with causality perspectives

H Liu, M Chaudhary, H Wang - arxiv preprint arxiv:2307.16851, 2023 - arxiv.org
The trustworthiness of machine learning has emerged as a critical topic in the field,
encompassing various applications and research areas such as robustness, security …

Estimating knowledge in large language models without generating a single token

D Gottesman, M Geva - arxiv preprint arxiv:2406.12673, 2024 - arxiv.org
To evaluate knowledge in large language models (LLMs), current methods query the model
and then evaluate its generated responses. In this work, we ask whether evaluation can be …

TopoBERT: Exploring the topology of fine-tuned word representations

A Rathore, Y Zhou, V Srikumar… - Information …, 2023 - journals.sagepub.com
Transformer-based language models such as BERT and its variants have found widespread
use in natural language processing (NLP). A common way of using these models is to fine …