A Damped Newton Method Achieves Global and Local Quadratic Convergence Rate

S Hanzely, D Kamzolov… - Advances in …, 2022 - proceedings.neurips.cc
In this paper, we present the first stepsize schedule for Newton method resulting in fast
global and local convergence guarantees. In particular, we a) prove an $\mathcal O\left …

[HTML][HTML] SHED: A Newton-type algorithm for federated learning based on incremental Hessian eigenvector sharing

N Dal Fabbro, S Dey, M Rossi, L Schenato - Automatica, 2024 - Elsevier
There is a growing interest in the distributed optimization framework that goes under the
name of Federated Learning (FL). In particular, much attention is being turned to FL …

Distributed adaptive greedy quasi-Newton methods with explicit non-asymptotic convergence bounds

Y Du, K You - Automatica, 2024 - Elsevier
Though quasi-Newton methods have been extensively studied in the literature, they either
suffer from local convergence or use a series of line searches for global convergence which …

FedZeN: Towards superlinear zeroth-order federated learning via incremental hessian estimation

A Maritan, S Dey, L Schenato - arxiv preprint arxiv:2309.17174, 2023 - arxiv.org
Federated learning is a distributed learning framework that allows a set of clients to
collaboratively train a model under the orchestration of a central server, without sharing raw …

[PDF][PDF] Accelerated adaptive cubic regularized quasi-newton methods

D Kamzolov, K Ziu, A Agafonov… - arxiv preprint arxiv …, 2023 - researchgate.net
In this paper, we propose Cubic Regularized Quasi-Newton Methods for (strongly)
starconvex and Accelerated Cubic Regularized Quasi-Newton for convex optimization. The …

FedZeN: Quadratic convergence in zeroth-order federated learning via incremental Hessian estimation

A Maritan, S Dey, L Schenato - 2024 European Control …, 2024 - ieeexplore.ieee.org
Federated learning is a distributed learning framework that allows a set of clients to
collaboratively train a model under the orchestration of a central server, without sharing raw …

Preconditioning meets biased compression for efficient distributed optimization

V Pirau, A Beznosikov, M Takáč, V Matyukhin… - Computational …, 2024 - Springer
Methods with preconditioned updates show up well in badly scaled and/or ill-conditioned
convex optimization problems. However, theoretical analysis of these methods in distributed …

OPTAMI: Global Superlinear Convergence of High-order Methods

D Kamzolov, D Pasechnyuk, A Agafonov… - arxiv preprint arxiv …, 2024 - arxiv.org
Second-order methods for convex optimization outperform first-order methods in terms of
theoretical iteration convergence, achieving rates up to $ O (k^{-5}) $ for highly-smooth …

FLECS-CGD: A Federated Learning Second-Order Framework via Compression and Sketching with Compressed Gradient Differences

A Agafonov, B Erraji, M Takáč - arxiv preprint arxiv:2210.09626, 2022 - arxiv.org
In the recent paper FLECS (Agafonov et al, FLECS: A Federated Learning Second-Order
Framework via Compression and Sketching), the second-order framework FLECS was …

Fed-Sophia: A Communication-Efficient Second-Order Federated Learning Algorithm

A Elbakary, CB Issaid, M Shehab, K Seddik… - arxiv preprint arxiv …, 2024 - arxiv.org
Federated learning is a machine learning approach where multiple devices collaboratively
learn with the help of a parameter server by sharing only their local updates. While gradient …