Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Optimization for deep learning: An overview
RY Sun - Journal of the Operations Research Society of China, 2020 - Springer
Optimization is a critical component in deep learning. We think optimization for neural
networks is an interesting topic for theoretical research due to various reasons. First, its …
networks is an interesting topic for theoretical research due to various reasons. First, its …
Scaffold: Stochastic controlled averaging for federated learning
Federated learning is a key scenario in modern large-scale machine learning where the
data remains distributed over a large number of clients and the task is to learn a centralized …
data remains distributed over a large number of clients and the task is to learn a centralized …
Federated learning: Challenges, methods, and future directions
Federated learning involves training statistical models over remote devices or siloed data
centers, such as mobile phones or hospitals, while kee** data localized. Training in …
centers, such as mobile phones or hospitals, while kee** data localized. Training in …
Federated optimization in heterogeneous networks
Federated Learning is a distributed learning paradigm with two key challenges that
differentiate it from traditional distributed optimization:(1) significant variability in terms of the …
differentiate it from traditional distributed optimization:(1) significant variability in terms of the …
Harmofl: Harmonizing local and global drifts in federated learning on heterogeneous medical images
Multiple medical institutions collaboratively training a model using federated learning (FL)
has become a promising solution for maximizing the potential of data-driven models, yet the …
has become a promising solution for maximizing the potential of data-driven models, yet the …
Adam can converge without any modification on update rules
Ever since\citet {reddi2019convergence} pointed out the divergence issue of Adam, many
new variants have been designed to obtain convergence. However, vanilla Adam remains …
new variants have been designed to obtain convergence. However, vanilla Adam remains …
SGD: General analysis and improved rates
We propose a general yet simple theorem describing the convergence of SGD under the
arbitrary sampling paradigm. Our theorem describes the convergence of an infinite array of …
arbitrary sampling paradigm. Our theorem describes the convergence of an infinite array of …
Loss landscapes and optimization in over-parameterized non-linear systems and neural networks
The success of deep learning is due, to a large extent, to the remarkable effectiveness of
gradient-based optimization methods applied to large neural networks. The purpose of this …
gradient-based optimization methods applied to large neural networks. The purpose of this …
Mime: Mimicking centralized stochastic algorithms in federated learning
Federated learning (FL) is a challenging setting for optimization due to the heterogeneity of
the data across different clients which gives rise to the client drift phenomenon. In fact …
the data across different clients which gives rise to the client drift phenomenon. In fact …
Reasonable effectiveness of random weighting: A litmus test for multi-task learning
Multi-Task Learning (MTL) has achieved success in various fields. However, how to balance
different tasks to achieve good performance is a key problem. To achieve the task balancing …
different tasks to achieve good performance is a key problem. To achieve the task balancing …