Communication-efficient federated learning via knowledge distillation
Federated learning is a privacy-preserving machine learning technique to train intelligent
models from decentralized data, which enables exploiting private data by communicating …
models from decentralized data, which enables exploiting private data by communicating …
Dense: Data-free one-shot federated learning
Abstract One-shot Federated Learning (FL) has recently emerged as a promising approach,
which allows the central server to learn a model in a single communication round. Despite …
which allows the central server to learn a model in a single communication round. Despite …
Fedpara: Low-rank hadamard product for communication-efficient federated learning
In this work, we propose a communication-efficient parameterization, FedPara, for federated
learning (FL) to overcome the burdens on frequent model uploads and downloads. Our …
learning (FL) to overcome the burdens on frequent model uploads and downloads. Our …
What kinds of functions do deep neural networks learn? Insights from variational spline theory
We develop a variational framework to understand the properties of functions learned by
fitting deep neural networks with rectified linear unit (ReLU) activations to data. We propose …
fitting deep neural networks with rectified linear unit (ReLU) activations to data. We propose …
Llm360: Towards fully transparent open-source llms
The recent surge in open-source Large Language Models (LLMs), such as LLaMA, Falcon,
and Mistral, provides diverse options for AI practitioners and researchers. However, most …
and Mistral, provides diverse options for AI practitioners and researchers. However, most …
Deep learning meets sparse regularization: A signal processing perspective
Deep learning (DL) has been wildly successful in practice, and most of the state-of-the-art
machine learning methods are based on neural networks (NNs). Lacking, however, is a …
machine learning methods are based on neural networks (NNs). Lacking, however, is a …
Low-rank lottery tickets: finding efficient low-rank neural networks via matrix differential equations
Neural networks have achieved tremendous success in a large variety of applications.
However, their memory footprint and computational demand can render them impractical in …
However, their memory footprint and computational demand can render them impractical in …
Optimus-CC: Efficient large NLP model training with 3D parallelism aware communication compression
In training of modern large natural language processing (NLP) models, it has become a
common practice to split models using 3D parallelism to multiple GPUs. Such technique …
common practice to split models using 3D parallelism to multiple GPUs. Such technique …
Layer-wise adaptive model aggregation for scalable federated learning
Abstract In Federated Learning (FL), a common approach for aggregating local solutions
across clients is periodic full model averaging. It is, however, known that different layers of …
across clients is periodic full model averaging. It is, however, known that different layers of …
Fedhm: Efficient federated learning for heterogeneous models via low-rank factorization
One underlying assumption of recent federated learning (FL) paradigms is that all local
models usually share the same network architecture and size, which becomes impractical …
models usually share the same network architecture and size, which becomes impractical …