Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
A Damped Newton Method Achieves Global and Local Quadratic Convergence Rate
In this paper, we present the first stepsize schedule for Newton method resulting in fast
global and local convergence guarantees. In particular, we a) prove an $\mathcal O\left …
global and local convergence guarantees. In particular, we a) prove an $\mathcal O\left …
[HTML][HTML] SHED: A Newton-type algorithm for federated learning based on incremental Hessian eigenvector sharing
There is a growing interest in the distributed optimization framework that goes under the
name of Federated Learning (FL). In particular, much attention is being turned to FL …
name of Federated Learning (FL). In particular, much attention is being turned to FL …
Distributed adaptive greedy quasi-Newton methods with explicit non-asymptotic convergence bounds
Y Du, K You - Automatica, 2024 - Elsevier
Though quasi-Newton methods have been extensively studied in the literature, they either
suffer from local convergence or use a series of line searches for global convergence which …
suffer from local convergence or use a series of line searches for global convergence which …
FedZeN: Towards superlinear zeroth-order federated learning via incremental hessian estimation
Federated learning is a distributed learning framework that allows a set of clients to
collaboratively train a model under the orchestration of a central server, without sharing raw …
collaboratively train a model under the orchestration of a central server, without sharing raw …
[PDF][PDF] Accelerated adaptive cubic regularized quasi-newton methods
In this paper, we propose Cubic Regularized Quasi-Newton Methods for (strongly)
starconvex and Accelerated Cubic Regularized Quasi-Newton for convex optimization. The …
starconvex and Accelerated Cubic Regularized Quasi-Newton for convex optimization. The …
FedZeN: Quadratic convergence in zeroth-order federated learning via incremental Hessian estimation
Federated learning is a distributed learning framework that allows a set of clients to
collaboratively train a model under the orchestration of a central server, without sharing raw …
collaboratively train a model under the orchestration of a central server, without sharing raw …
Preconditioning meets biased compression for efficient distributed optimization
Methods with preconditioned updates show up well in badly scaled and/or ill-conditioned
convex optimization problems. However, theoretical analysis of these methods in distributed …
convex optimization problems. However, theoretical analysis of these methods in distributed …
FLECS-CGD: A Federated Learning Second-Order Framework via Compression and Sketching with Compressed Gradient Differences
In the recent paper FLECS (Agafonov et al, FLECS: A Federated Learning Second-Order
Framework via Compression and Sketching), the second-order framework FLECS was …
Framework via Compression and Sketching), the second-order framework FLECS was …
Fed-Sophia: A Communication-Efficient Second-Order Federated Learning Algorithm
Federated learning is a machine learning approach where multiple devices collaboratively
learn with the help of a parameter server by sharing only their local updates. While gradient …
learn with the help of a parameter server by sharing only their local updates. While gradient …
Novel Bounds for Incremental Hessian Estimation With Application to Zeroth-Order Federated Learning
The Hessian matrix conveys important information about the curvature, spectrum and partial
derivatives of a function, and is required in a variety of tasks. However, computing the exact …
derivatives of a function, and is required in a variety of tasks. However, computing the exact …