Effortless Efficiency: Low-Cost Pruning of Diffusion Models

Y Zhang, E **, Y Dong, A Khakzar, P Torr… - arxiv preprint arxiv …, 2024 - arxiv.org
Diffusion models have achieved impressive advancements in various vision tasks. However,
these gains often rely on increasing model size, which escalates computational complexity …

BlockPruner: Fine-grained Pruning for Large Language Models

L Zhong, F Wan, R Chen, X Quan, L Li - arxiv preprint arxiv:2406.10594, 2024 - arxiv.org
With the rapid growth in the size and complexity of large language models (LLMs), the costs
associated with their training and inference have escalated significantly. Research indicates …

FASP: Fast and Accurate Structured Pruning of Large Language Models

H Hu, P Zhao, P Li, Y Zheng, Z Wang… - arxiv preprint arxiv …, 2025 - arxiv.org
The rapid increase in the size of large language models (LLMs) has significantly escalated
their computational and memory demands, posing challenges for efficient deployment …

From General to Expert: Custom Pruning LLMs Across Language, Domain, and Task

Y Zhao, W Zhang, G Chen, K Kawaguchi, L Bing - openreview.net
Large Language Models (LLMs) have transformed natural language processing, yet their
substantial model sizes often demand significant computational resources. To conserve …