Conditional gradient methods

G Braun, A Carderera, CW Combettes… - arxiv preprint arxiv …, 2022 - arxiv.org
The purpose of this survey is to serve both as a gentle introduction and a coherent overview
of state-of-the-art Frank--Wolfe algorithms, also called conditional gradient algorithms, for …

Asif: Coupled data turns unimodal models to multimodal without training

A Norelli, M Fumero, V Maiorca… - Advances in …, 2023 - proceedings.neurips.cc
CLIP proved that aligning visual and language spaces is key to solving many vision tasks
without explicit training, but required to train image and text encoders from scratch on a huge …

Cardinality minimization, constraints, and regularization: a survey

AM Tillmann, D Bienstock, A Lodi, A Schwartz - SIAM Review, 2024 - SIAM
We survey optimization problems that involve the cardinality of variable vectors in
constraints or the objective function. We provide a unified viewpoint on the general problem …

Training neural networks for and by interpolation

L Berrada, A Zisserman… - … conference on machine …, 2020 - proceedings.mlr.press
In modern supervised learning, many deep neural networks are able to interpolate the data:
the empirical loss can be driven to near zero on all samples simultaneously. In this work, we …

Breaking the linear iteration cost barrier for some well-known conditional gradient methods using maxip data-structures

Z Xu, Z Song, A Shrivastava - Advances in Neural …, 2021 - proceedings.neurips.cc
Conditional gradient methods (CGM) are widely used in modern machine learning. CGM's
overall running time usually consists of two parts: the number of iterations and the cost of …

Linearly convergent Frank-Wolfe with backtracking line-search

F Pedregosa, G Negiar, A Askari… - … conference on artificial …, 2020 - proceedings.mlr.press
Abstract Structured constraints in Machine Learning have recently brought the Frank-Wolfe
(FW) family of algorithms back in the spotlight. While the classical FW algorithm has poor …

Robust Gaussian Processes via Relevance Pursuit

S Ament, E Santorella, D Eriksson… - Advances in …, 2025 - proceedings.neurips.cc
Gaussian processes (GPs) are non-parametric probabilistic regression models that are
popular due to their flexibility, data efficiency, and well-calibrated uncertainty estimates …

Stochastic Frank-Wolfe for constrained finite-sum minimization

G Négiar, G Dresdner, A Tsai… - international …, 2020 - proceedings.mlr.press
We propose a novel Stochastic Frank-Wolfe (aka conditional gradient) algorithm for
constrained smooth finite-sum minimization with a generalized linear prediction/structure …

A conditional gradient framework for composite convex minimization with applications to semidefinite programming

A Yurtsever, O Fercoq, F Locatello… - … on machine learning, 2018 - proceedings.mlr.press
We propose a conditional gradient framework for a composite convex minimization template
with broad applications. Our approach combines smoothing and homotopy techniques …

Restarting frank-wolfe

T Kerdreux, A d'Aspremont… - The 22nd international …, 2019 - proceedings.mlr.press
Abstract Conditional Gradients (aka Frank-Wolfe algorithms) form a classical set of methods
for constrained smooth convex minimization due to their simplicity, the absence of projection …