Problems and opportunities in training deep learning software systems: An analysis of variance

HV Pham, S Qian, J Wang, T Lutellier… - Proceedings of the 35th …, 2020 - dl.acm.org
Deep learning (DL) training algorithms utilize nondeterminism to improve models' accuracy
and training efficiency. Hence, multiple identical training runs (eg, identical training data …

Symbolic discovery of optimization algorithms

X Chen, C Liang, D Huang, E Real… - Advances in neural …, 2024 - proceedings.neurips.cc
We present a method to formulate algorithm discovery as program search, and apply it to
discover optimization algorithms for deep neural network training. We leverage efficient …

Dataset distillation using neural feature regression

Y Zhou, E Nezhadarya, J Ba - Advances in Neural …, 2022 - proceedings.neurips.cc
Dataset distillation aims to learn a small synthetic dataset that preserves most of the
information from the original dataset. Dataset distillation can be formulated as a bi-level …

A survey of meta-reinforcement learning

J Beck, R Vuorio, EZ Liu, Z **ong, L Zintgraf… - arxiv preprint arxiv …, 2023 - arxiv.org
While deep reinforcement learning (RL) has fueled multiple high-profile successes in
machine learning, it is held back from more widespread adoption by its often poor data …

Meta-learning in neural networks: A survey

T Hospedales, A Antoniou, P Micaelli… - IEEE transactions on …, 2021 - ieeexplore.ieee.org
The field of meta-learning, or learning-to-learn, has seen a dramatic rise in interest in recent
years. Contrary to conventional approaches to AI where tasks are solved from scratch using …

Brax--a differentiable physics engine for large scale rigid body simulation

CD Freeman, E Frey, A Raichuk, S Girgin… - arxiv preprint arxiv …, 2021 - arxiv.org
We present Brax, an open source library for rigid body simulation with a focus on
performance and parallelism on accelerators, written in JAX. We present results on a suite of …

Learning to optimize: A primer and a benchmark

T Chen, X Chen, W Chen, H Heaton, J Liu… - Journal of Machine …, 2022 - jmlr.org
Learning to optimize (L2O) is an emerging approach that leverages machine learning to
develop optimization methods, aiming at reducing the laborious iterations of hand …

Data distillation: A survey

N Sachdeva, J McAuley - arxiv preprint arxiv:2301.04272, 2023 - arxiv.org
The popularity of deep learning has led to the curation of a vast number of massive and
multifarious datasets. Despite having close-to-human performance on individual tasks …

Do differentiable simulators give better policy gradients?

HJ Suh, M Simchowitz, K Zhang… - … on Machine Learning, 2022 - proceedings.mlr.press
Differentiable simulators promise faster computation time for reinforcement learning by
replacing zeroth-order gradient estimates of a stochastic objective with an estimate based …

General-purpose in-context learning by meta-learning transformers

L Kirsch, J Harrison, J Sohl-Dickstein, L Metz - arxiv preprint arxiv …, 2022 - arxiv.org
Modern machine learning requires system designers to specify aspects of the learning
pipeline, such as losses, architectures, and optimizers. Meta-learning, or learning-to-learn …