A survey on deep neural network pruning: Taxonomy, comparison, analysis, and recommendations

H Cheng, M Zhang, JQ Shi - IEEE Transactions on Pattern …, 2024 - ieeexplore.ieee.org
Modern deep neural networks, particularly recent large language models, come with
massive model sizes that require significant computational and storage resources. To …

Continual learning for recurrent neural networks: an empirical evaluation

A Cossu, A Carta, V Lomonaco, D Bacciu - Neural Networks, 2021 - Elsevier
Learning continuously during all model lifetime is fundamental to deploy machine learning
solutions robust to drifts in the data distribution. Advances in Continual Learning (CL) with …

Sparsity in deep learning: Pruning and growth for efficient inference and training in neural networks

T Hoefler, D Alistarh, T Ben-Nun, N Dryden… - Journal of Machine …, 2021 - jmlr.org
The growing energy and performance costs of deep learning have driven the community to
reduce the size of neural networks by selectively pruning components. Similarly to their …

Analyzing multi-head self-attention: Specialized heads do the heavy lifting, the rest can be pruned

E Voita, D Talbot, F Moiseev, R Sennrich… - arxiv preprint arxiv …, 2019 - arxiv.org
Multi-head self-attention is a key component of the Transformer, a state-of-the-art
architecture for neural machine translation. In this work we evaluate the contribution made …

Mind the GAP: A balanced corpus of gendered ambiguous pronouns

K Webster, M Recasens, V Axelrod… - Transactions of the …, 2018 - direct.mit.edu
Coreference resolution is an important task for natural language understanding, and the
resolution of ambiguous pronouns a longstanding challenge. Nonetheless, existing corpora …

The flores evaluation datasets for low-resource machine translation: Nepali-english and sinhala-english

F Guzmán, PJ Chen, M Ott, J Pino, G Lample… - arxiv preprint arxiv …, 2019 - arxiv.org
For machine translation, a vast majority of language pairs in the world are considered low-
resource because they have little parallel data available. Besides the technical challenges …

Cross-lingual transfer learning for multilingual task oriented dialog

S Schuster, S Gupta, R Shah, M Lewis - arxiv preprint arxiv:1810.13327, 2018 - arxiv.org
One of the first steps in the utterance interpretation pipeline of many task-oriented
conversational AI systems is to identify user intents and the corresponding slots. Since data …

Context-aware neural machine translation learns anaphora resolution

E Voita, P Serdyukov, R Sennrich, I Titov - arxiv preprint arxiv:1805.10163, 2018 - arxiv.org
Standard machine translation systems process sentences in isolation and hence ignore
extra-sentential information, even though extended context can both prevent mistakes in …

When a good translation is wrong in context: Context-aware machine translation improves on deixis, ellipsis, and lexical cohesion

E Voita, R Sennrich, I Titov - arxiv preprint arxiv:1905.05979, 2019 - arxiv.org
Though machine translation errors caused by the lack of context beyond one sentence have
long been acknowledged, the development of context-aware NMT systems is hampered by …

Self-training improves pre-training for natural language understanding

J Du, E Grave, B Gunel, V Chaudhary, O Celebi… - arxiv preprint arxiv …, 2020 - arxiv.org
Unsupervised pre-training has led to much recent progress in natural language
understanding. In this paper, we study self-training as another way to leverage unlabeled …