A Damped Newton Method Achieves Global and Local Quadratic Convergence Rate
In this paper, we present the first stepsize schedule for Newton method resulting in fast
global and local convergence guarantees. In particular, we a) prove an $\mathcal O\left …
global and local convergence guarantees. In particular, we a) prove an $\mathcal O\left …
[HTML][HTML] SHED: A Newton-type algorithm for federated learning based on incremental Hessian eigenvector sharing
There is a growing interest in the distributed optimization framework that goes under the
name of Federated Learning (FL). In particular, much attention is being turned to FL …
name of Federated Learning (FL). In particular, much attention is being turned to FL …
Distributed adaptive greedy quasi-Newton methods with explicit non-asymptotic convergence bounds
Y Du, K You - Automatica, 2024 - Elsevier
Though quasi-Newton methods have been extensively studied in the literature, they either
suffer from local convergence or use a series of line searches for global convergence which …
suffer from local convergence or use a series of line searches for global convergence which …
FedZeN: Towards superlinear zeroth-order federated learning via incremental hessian estimation
Federated learning is a distributed learning framework that allows a set of clients to
collaboratively train a model under the orchestration of a central server, without sharing raw …
collaboratively train a model under the orchestration of a central server, without sharing raw …
[PDF][PDF] Accelerated adaptive cubic regularized quasi-newton methods
In this paper, we propose Cubic Regularized Quasi-Newton Methods for (strongly)
starconvex and Accelerated Cubic Regularized Quasi-Newton for convex optimization. The …
starconvex and Accelerated Cubic Regularized Quasi-Newton for convex optimization. The …
FedZeN: Quadratic convergence in zeroth-order federated learning via incremental Hessian estimation
Federated learning is a distributed learning framework that allows a set of clients to
collaboratively train a model under the orchestration of a central server, without sharing raw …
collaboratively train a model under the orchestration of a central server, without sharing raw …
Preconditioning meets biased compression for efficient distributed optimization
Methods with preconditioned updates show up well in badly scaled and/or ill-conditioned
convex optimization problems. However, theoretical analysis of these methods in distributed …
convex optimization problems. However, theoretical analysis of these methods in distributed …
OPTAMI: Global Superlinear Convergence of High-order Methods
Second-order methods for convex optimization outperform first-order methods in terms of
theoretical iteration convergence, achieving rates up to $ O (k^{-5}) $ for highly-smooth …
theoretical iteration convergence, achieving rates up to $ O (k^{-5}) $ for highly-smooth …
FLECS-CGD: A Federated Learning Second-Order Framework via Compression and Sketching with Compressed Gradient Differences
In the recent paper FLECS (Agafonov et al, FLECS: A Federated Learning Second-Order
Framework via Compression and Sketching), the second-order framework FLECS was …
Framework via Compression and Sketching), the second-order framework FLECS was …
Fed-Sophia: A Communication-Efficient Second-Order Federated Learning Algorithm
Federated learning is a machine learning approach where multiple devices collaboratively
learn with the help of a parameter server by sharing only their local updates. While gradient …
learn with the help of a parameter server by sharing only their local updates. While gradient …