Provable repair of deep neural networks

M Sotoudeh, AV Thakur - Proceedings of the 42nd ACM SIGPLAN …, 2021 - dl.acm.org
Deep Neural Networks (DNNs) have grown in popularity over the past decade and are now
being used in safety-critical domains such as aircraft collision avoidance. This has motivated …

Learning parities with neural networks

A Daniely, E Malach - Advances in Neural Information …, 2020 - proceedings.neurips.cc
In recent years we see a rapidly growing line of research which shows learnability of various
models via common neural network algorithms. Yet, besides a very few outliers, these …

Unraveling attention via convex duality: Analysis and interpretations of vision transformers

A Sahiner, T Ergen, B Ozturkler… - International …, 2022 - proceedings.mlr.press
Vision transformers using self-attention or its proposed alternatives have demonstrated
promising results in many image related tasks. However, the underpinning inductive bias of …

Fast convex optimization for two-layer relu networks: Equivalent model classes and cone decompositions

A Mishkin, A Sahiner, M Pilanci - … Conference on Machine …, 2022 - proceedings.mlr.press
We develop fast algorithms and robust software for convex optimization of two-layer neural
networks with ReLU activation functions. Our work leverages a convex re-formulation of the …

The influence of learning rule on representation dynamics in wide neural networks

B Bordelon, C Pehlevan - The Eleventh International Conference on …, 2022 - openreview.net
It is unclear how changing the learning rule of a deep neural network alters its learning
dynamics and representations. To gain insight into the relationship between learned …

How does a kernel based on gradients of infinite-width neural networks come to be widely used: a review of the neural tangent kernel

Y Tan, H Liu - International Journal of Multimedia Information …, 2024 - Springer
The neural tangent kernel (NTK) was created in the context of using the limit idea to study
the theory of neural network. NTKs are defined from neural network models in the infinite …

Globally gated deep linear networks

Q Li, H Sompolinsky - Advances in Neural Information …, 2022 - proceedings.neurips.cc
Abstract Recently proposed Gated Linear Networks (GLNs) present a tractable nonlinear
network architecture, and exhibit interesting capabilities such as learning with local error …

Optimal sets and solution paths of ReLU networks

A Mishkin, M Pilanci - International Conference on Machine …, 2023 - proceedings.mlr.press
We develop an analytical framework to characterize the set of optimal ReLU neural networks
by reformulating the non-convex training problem as a convex program. We show that the …

[PDF][PDF] Correcting deep neural networks with small, generalizing patches

M Sotoudeh, A Thakur - … on safety and robustness in decision …, 2019 - thakur.cs.ucdavis.edu
We consider the problem of patching a deep neural network: applying a small change to the
network weights in order to produce a desired change in the classifications made by the …

Towards understanding learning in neural networks with linear teachers

R Sarussi, A Brutzkus… - … Conference on Machine …, 2021 - proceedings.mlr.press
Can a neural network minimizing cross-entropy learn linearly separable data? Despite
progress in the theory of deep learning, this question remains unsolved. Here we prove that …