Recent advances in natural language processing via large pre-trained language models: A survey
Large, pre-trained language models (PLMs) such as BERT and GPT have drastically
changed the Natural Language Processing (NLP) field. For numerous NLP tasks …
changed the Natural Language Processing (NLP) field. For numerous NLP tasks …
A survey on deep neural network pruning: Taxonomy, comparison, analysis, and recommendations
Modern deep neural networks, particularly recent large language models, come with
massive model sizes that require significant computational and storage resources. To …
massive model sizes that require significant computational and storage resources. To …
AdaLoRA: Adaptive budget allocation for parameter-efficient fine-tuning
Fine-tuning large pre-trained language models on downstream tasks has become an
important paradigm in NLP. However, common practice fine-tunes all of the parameters in a …
important paradigm in NLP. However, common practice fine-tunes all of the parameters in a …
Losparse: Structured compression of large language models based on low-rank and sparse approximation
Transformer models have achieved remarkable results in various natural language tasks,
but they are often prohibitively large, requiring massive memories and computational …
but they are often prohibitively large, requiring massive memories and computational …
Revisiting out-of-distribution robustness in nlp: Benchmarks, analysis, and LLMs evaluations
This paper reexamines the research on out-of-distribution (OOD) robustness in the field of
NLP. We find that the distribution shift settings in previous studies commonly lack adequate …
NLP. We find that the distribution shift settings in previous studies commonly lack adequate …
Task-specific skill localization in fine-tuned language models
Pre-trained language models can be fine-tuned to solve diverse NLP tasks, including in few-
shot settings. Thus fine-tuning allows the model to quickly pick up task-specific" skills," but …
shot settings. Thus fine-tuning allows the model to quickly pick up task-specific" skills," but …
Platon: Pruning large transformer models with upper confidence bound of weight importance
Large Transformer-based models have exhibited superior performance in various natural
language processing and computer vision tasks. However, these models contain enormous …
language processing and computer vision tasks. However, these models contain enormous …
State-of-the-art generalisation research in NLP: a taxonomy and review
The ability to generalise well is one of the primary desiderata of natural language
processing (NLP). Yet, what'good generalisation'entails and how it should be evaluated is …
processing (NLP). Yet, what'good generalisation'entails and how it should be evaluated is …
Edge AI: A taxonomy, systematic review and future directions
Abstract Edge Artificial Intelligence (AI) incorporates a network of interconnected systems
and devices that receive, cache, process, and analyse data in close communication with the …
and devices that receive, cache, process, and analyse data in close communication with the …
Configurable foundation models: Building llms from a modular perspective
Advancements in LLMs have recently unveiled challenges tied to computational efficiency
and continual scalability due to their requirements of huge parameters, making the …
and continual scalability due to their requirements of huge parameters, making the …