A survey on deep neural network pruning: Taxonomy, comparison, analysis, and recommendations
Modern deep neural networks, particularly recent large language models, come with
massive model sizes that require significant computational and storage resources. To …
massive model sizes that require significant computational and storage resources. To …
Exploring the landscape of machine unlearning: A comprehensive survey and taxonomy
Machine unlearning (MU) is gaining increasing attention due to the need to remove or
modify predictions made by machine learning (ML) models. While training models have …
modify predictions made by machine learning (ML) models. While training models have …
Spvit: Enabling faster vision transformers via latency-aware soft token pruning
Abstract Recently, Vision Transformer (ViT) has continuously established new milestones in
the computer vision field, while the high computation and memory cost makes its …
the computer vision field, while the high computation and memory cost makes its …
Chex: Channel exploration for cnn model compression
Channel pruning has been broadly recognized as an effective technique to reduce the
computation and memory cost of deep convolutional neural networks. However …
computation and memory cost of deep convolutional neural networks. However …
Model sparsity can simplify machine unlearning
In response to recent data regulation requirements, machine unlearning (MU) has emerged
as a critical process to remove the influence of specific examples from a given model …
as a critical process to remove the influence of specific examples from a given model …
Federated dynamic sparse training: Computing less, communicating less, yet learning better
Federated learning (FL) enables distribution of machine learning workloads from the cloud
to resource-limited edge devices. Unfortunately, current deep networks remain not only too …
to resource-limited edge devices. Unfortunately, current deep networks remain not only too …
Advancing model pruning via bi-level optimization
The deployment constraints in practical applications necessitate the pruning of large-scale
deep learning models, ie, promoting their weight sparsity. As illustrated by the Lottery Ticket …
deep learning models, ie, promoting their weight sparsity. As illustrated by the Lottery Ticket …
Coarsening the granularity: Towards structurally sparse lottery tickets
The lottery ticket hypothesis (LTH) has shown that dense models contain highly sparse
subnetworks (ie, winning tickets) that can be trained in isolation to match full accuracy …
subnetworks (ie, winning tickets) that can be trained in isolation to match full accuracy …
An Introduction to Bilevel Optimization: Foundations and applications in signal processing and machine learning
Recently, bilevel optimization (BLO) has taken center stage in some very exciting
developments in the area of signal processing (SP) and machine learning (ML). Roughly …
developments in the area of signal processing (SP) and machine learning (ML). Roughly …
Rare gems: Finding lottery tickets at initialization
Large neural networks can be pruned to a small fraction of their original size, with little loss
in accuracy, by following a time-consuming" train, prune, re-train" approach. Frankle & …
in accuracy, by following a time-consuming" train, prune, re-train" approach. Frankle & …