Inverting gradients-how easy is it to break privacy in federated learning?

J Gei**, H Bauermeister, H Dröge… - Advances in neural …, 2020 - proceedings.neurips.cc
The idea of federated learning is to collaboratively train a neural network on a server. Each
user receives the current weights of the network and in turns sends parameter updates …

Finite versus infinite neural networks: an empirical study

J Lee, S Schoenholz, J Pennington… - Advances in …, 2020 - proceedings.neurips.cc
We perform a careful, thorough, and large scale empirical study of the correspondence
between wide neural networks and kernel methods. By doing so, we resolve a variety of …

Recent advances in deep learning theory

F He, D Tao - ar**, W Czaja, M Goldblum… - ar**, M Goldblum, PE Pope, M Moeller… - arxiv preprint arxiv …, 2021 - arxiv.org
It is widely believed that the implicit regularization of SGD is fundamental to the impressive
generalization behavior we observe in neural networks. In this work, we demonstrate that …

Unraveling meta-learning: Understanding feature representations for few-shot tasks

M Goldblum, S Reich, L Fowl, R Ni… - International …, 2020 - proceedings.mlr.press
Meta-learning algorithms produce feature extractors which achieve state-of-the-art
performance on few-shot classification. While the literature is rich with meta-learning …

TCT: Convexifying federated learning using bootstrapped neural tangent kernels

Y Yu, A Wei, SP Karimireddy, Y Ma… - Advances in Neural …, 2022 - proceedings.neurips.cc
State-of-the-art federated learning methods can perform far worse than their centralized
counterparts when clients have dissimilar data distributions. For neural networks, even when …

Security Considerations in AI-Robotics: A Survey of Current Methods, Challenges, and Opportunities

S Neupane, S Mitra, IA Fernandez, S Saha… - IEEE …, 2024 - ieeexplore.ieee.org
Robotics and Artificial Intelligence (AI) have been inextricably intertwined since their
inception. Today, AI-Robotics systems have become an integral part of our daily lives, from …

A linearized framework and a new benchmark for model selection for fine-tuning

A Deshpande, A Achille, A Ravichandran, H Li… - arxiv preprint arxiv …, 2021 - arxiv.org
Fine-tuning from a collection of models pre-trained on different domains (a" model zoo") is
emerging as a technique to improve test accuracy in the low-data regime. However, model …

Liberty or depth: Deep Bayesian neural nets do not need complex weight posterior approximations

S Farquhar, L Smith, Y Gal - Advances in Neural …, 2020 - proceedings.neurips.cc
We challenge the longstanding assumption that the mean-field approximation for variational
inference in Bayesian neural networks is severely restrictive, and show this is not the case in …