Neural network approximation

R DeVore, B Hanin, G Petrova - Acta Numerica, 2021 - cambridge.org
Neural networks (NNs) are the method of choice for building learning algorithms. They are
now being investigated for other numerical tasks such as solving high-dimensional partial …

Foundational challenges in assuring alignment and safety of large language models

U Anwar, A Saparov, J Rando, D Paleka… - arxiv preprint arxiv …, 2024 - arxiv.org
This work identifies 18 foundational challenges in assuring the alignment and safety of large
language models (LLMs). These challenges are organized into three different categories …

The power of quantum neural networks

A Abbas, D Sutter, C Zoufal, A Lucchi, A Figalli… - Nature Computational …, 2021 - nature.com
It is unknown whether near-term quantum computers are advantageous for machine
learning tasks. In this work we address this question by trying to understand how powerful …

Intrinsic dimensionality explains the effectiveness of language model fine-tuning

A Aghajanyan, L Zettlemoyer, S Gupta - arxiv preprint arxiv:2012.13255, 2020 - arxiv.org
Although pretrained language models can be fine-tuned to produce state-of-the-art results
for a very wide range of language understanding tasks, the dynamics of this process are not …

Pruning neural networks without any data by iteratively conserving synaptic flow

H Tanaka, D Kunin, DL Yamins… - Advances in neural …, 2020 - proceedings.neurips.cc
Pruning the parameters of deep neural networks has generated intense interest due to
potential savings in time, memory and energy both during training and at test time. Recent …

Task-specific skill localization in fine-tuned language models

A Panigrahi, N Saunshi, H Zhao… - … on Machine Learning, 2023 - proceedings.mlr.press
Pre-trained language models can be fine-tuned to solve diverse NLP tasks, including in few-
shot settings. Thus fine-tuning allows the model to quickly pick up task-specific" skills," but …

A primer on Bayesian neural networks: review and debates

J Arbel, K Pitas, M Vladimirova, V Fortuin - arxiv preprint arxiv:2309.16314, 2023 - arxiv.org
Neural networks have achieved remarkable performance across various problem domains,
but their widespread applicability is hindered by inherent limitations such as overconfidence …

Learning imbalanced datasets with label-distribution-aware margin loss

K Cao, C Wei, A Gaidon… - Advances in neural …, 2019 - proceedings.neurips.cc
Deep learning algorithms can fare poorly when the training dataset suffers from heavy class-
imbalance but the testing criterion requires good generalization on less frequent classes …

Fantastic generalization measures and where to find them

Y Jiang, B Neyshabur, H Mobahi, D Krishnan… - arxiv preprint arxiv …, 2019 - arxiv.org
Generalization of deep networks has been of great interest in recent years, resulting in a
number of theoretically and empirically motivated complexity measures. However, most …

Fine-grained analysis of optimization and generalization for overparameterized two-layer neural networks

S Arora, S Du, W Hu, Z Li… - … conference on machine …, 2019 - proceedings.mlr.press
Recent works have cast some light on the mystery of why deep nets fit any data and
generalize despite being very overparametrized. This paper analyzes training and …