Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Variance reduced proxskip: Algorithm, theory and application to federated learning
We study distributed optimization methods based on the {\em local training (LT)} paradigm,
ie, methods which achieve communication efficiency by performing richer local gradient …
ie, methods which achieve communication efficiency by performing richer local gradient …
Proxskip: Yes! local gradient steps provably lead to communication acceleration! finally!
We introduce ProxSkip—a surprisingly simple and provably efficient method for minimizing
the sum of a smooth ($ f $) and an expensive nonsmooth proximable ($\psi $) function. The …
the sum of a smooth ($ f $) and an expensive nonsmooth proximable ($\psi $) function. The …
Adaptive personalized federated learning
Investigation of the degree of personalization in federated learning algorithms has shown
that only maximizing the performance of the global model will confine the capacity of the …
that only maximizing the performance of the global model will confine the capacity of the …
Federated learning of a mixture of global and local models
We propose a new optimization formulation for training federated learning models. The
standard formulation has the form of an empirical risk minimization problem constructed to …
standard formulation has the form of an empirical risk minimization problem constructed to …
Where to begin? on the impact of pre-training and initialization in federated learning
An oft-cited challenge of federated learning is the presence of heterogeneity.\emph {Data
heterogeneity} refers to the fact that data from different clients may follow very different …
heterogeneity} refers to the fact that data from different clients may follow very different …
Fedavg with fine tuning: Local updates lead to representation learning
Abstract The Federated Averaging (FedAvg) algorithm, which consists of alternating
between a few local stochastic gradient updates at client nodes, followed by a model …
between a few local stochastic gradient updates at client nodes, followed by a model …
Mime: Mimicking centralized stochastic algorithms in federated learning
Federated learning (FL) is a challenging setting for optimization due to the heterogeneity of
the data across different clients which gives rise to the client drift phenomenon. In fact …
the data across different clients which gives rise to the client drift phenomenon. In fact …
Communication-efficient and distributed learning over wireless networks: Principles and applications
Machine learning (ML) is a promising enabler for the fifth-generation (5G) communication
systems and beyond. By imbuing intelligence into the network edge, edge nodes can …
systems and beyond. By imbuing intelligence into the network edge, edge nodes can …
Cooperative SGD: A unified framework for the design and analysis of local-update SGD algorithms
When training machine learning models using stochastic gradient descent (SGD) with a
large number of nodes or massive edge devices, the communication cost of synchronizing …
large number of nodes or massive edge devices, the communication cost of synchronizing …
Distributionally robust federated averaging
In this paper, we study communication efficient distributed algorithms for distributionally
robust federated learning via periodic averaging with adaptive sampling. In contrast to …
robust federated learning via periodic averaging with adaptive sampling. In contrast to …