Sok: Model inversion attack landscape: Taxonomy, challenges, and future roadmap
SV Dibbo - 2023 IEEE 36th Computer Security Foundations …, 2023 - ieeexplore.ieee.org
A crucial module of the widely applied machine learning (ML) model is the model training
phase, which involves large-scale training data, often including sensitive private data. ML …
phase, which involves large-scale training data, often including sensitive private data. ML …
Advances in asynchronous parallel and distributed optimization
Motivated by large-scale optimization problems arising in the context of machine learning,
there have been several advances in the study of asynchronous parallel and distributed …
there have been several advances in the study of asynchronous parallel and distributed …
Cocktailsgd: Fine-tuning foundation models over 500mbps networks
Distributed training of foundation models, especially large language models (LLMs), is
communication-intensive and so has heavily relied on centralized data centers with fast …
communication-intensive and so has heavily relied on centralized data centers with fast …
Communication-efficient federated learning
Federated learning (FL) enables edge devices, such as Internet of Things devices (eg,
sensors), servers, and institutions (eg, hospitals), to collaboratively train a machine learning …
sensors), servers, and institutions (eg, hospitals), to collaboratively train a machine learning …
Decentralized federated averaging
Federated averaging (FedAvg) is a communication-efficient algorithm for distributed training
with an enormous number of clients. In FedAvg, clients keep their data locally for privacy …
with an enormous number of clients. In FedAvg, clients keep their data locally for privacy …
Sharper convergence guarantees for asynchronous SGD for distributed and federated learning
We study the asynchronous stochastic gradient descent algorithm, for distributed training
over $ n $ workers that might be heterogeneous. In this algorithm, workers compute …
over $ n $ workers that might be heterogeneous. In this algorithm, workers compute …
SGD: General analysis and improved rates
We propose a general yet simple theorem describing the convergence of SGD under the
arbitrary sampling paradigm. Our theorem describes the convergence of an infinite array of …
arbitrary sampling paradigm. Our theorem describes the convergence of an infinite array of …
Qsparse-local-SGD: Distributed SGD with quantization, sparsification and local computations
Communication bottleneck has been identified as a significant issue in distributed
optimization of large-scale learning models. Recently, several approaches to mitigate this …
optimization of large-scale learning models. Recently, several approaches to mitigate this …
Fedavg with fine tuning: Local updates lead to representation learning
Abstract The Federated Averaging (FedAvg) algorithm, which consists of alternating
between a few local stochastic gradient updates at client nodes, followed by a model …
between a few local stochastic gradient updates at client nodes, followed by a model …
Stochastic polyak step-size for sgd: An adaptive learning rate for fast convergence
We propose a stochastic variant of the classical Polyak step-size (Polyak, 1987) commonly
used in the subgradient method. Although computing the Polyak step-size requires …
used in the subgradient method. Although computing the Polyak step-size requires …