A review of deep learning techniques for speech processing

A Mehrish, N Majumder, R Bharadwaj, R Mihalcea… - Information …, 2023 - Elsevier
The field of speech processing has undergone a transformative shift with the advent of deep
learning. The use of multiple processing layers has enabled the creation of models capable …

Survey on evolutionary deep learning: Principles, algorithms, applications, and open issues

N Li, L Ma, G Yu, B Xue, M Zhang, Y ** - ACM Computing Surveys, 2023 - dl.acm.org
Over recent years, there has been a rapid development of deep learning (DL) in both
industry and academia fields. However, finding the optimal hyperparameters of a DL model …

Efficient large language models: A survey

Z Wan, X Wang, C Liu, S Alam, Y Zheng, J Liu… - arxiv preprint arxiv …, 2023 - arxiv.org
Large Language Models (LLMs) have demonstrated remarkable capabilities in important
tasks such as natural language understanding and language generation, and thus have the …

Losparse: Structured compression of large language models based on low-rank and sparse approximation

Y Li, Y Yu, Q Zhang, C Liang, P He… - International …, 2023 - proceedings.mlr.press
Transformer models have achieved remarkable results in various natural language tasks,
but they are often prohibitively large, requiring massive memories and computational …

Assessing the brittleness of safety alignment via pruning and low-rank modifications

B Wei, K Huang, Y Huang, T **e, X Qi, M **a… - arxiv preprint arxiv …, 2024 - arxiv.org
Large language models (LLMs) show inherent brittleness in their safety mechanisms, as
evidenced by their susceptibility to jailbreaking and even non-malicious fine-tuning. This …

Asvd: Activation-aware singular value decomposition for compressing large language models

Z Yuan, Y Shang, Y Song, Q Wu, Y Yan… - arxiv preprint arxiv …, 2023 - arxiv.org
In this paper, we introduce a new post-training compression paradigm for Large Language
Models (LLMs) to facilitate their wider adoption. We delve into LLM weight low-rank …

Sparsity in transformers: A systematic literature review

M Farina, U Ahmad, A Taha, H Younes, Y Mesbah… - Neurocomputing, 2024 - Elsevier
Transformers have become the state-of-the-art architectures for various tasks in Natural
Language Processing (NLP) and Computer Vision (CV); however, their space and …

Lq-lora: Low-rank plus quantized matrix decomposition for efficient language model finetuning

H Guo, P Greengard, EP **ng, Y Kim - arxiv preprint arxiv:2311.12023, 2023 - arxiv.org
We propose a simple approach for memory-efficient adaptation of pretrained language
models. Our approach uses an iterative algorithm to decompose each pretrained matrix into …

Trojvit: Trojan insertion in vision transformers

M Zheng, Q Lou, L Jiang - … of the IEEE/CVF Conference on …, 2023 - openaccess.thecvf.com
Abstract Vision Transformers (ViTs) have demonstrated the state-of-the-art performance in
various vision-related tasks. The success of ViTs motivates adversaries to perform backdoor …

Svdqunat: Absorbing outliers by low-rank components for 4-bit diffusion models

M Li, Y Lin, Z Zhang, T Cai, X Li, J Guo, E **e… - arxiv preprint arxiv …, 2024 - arxiv.org
Diffusion models have been proven highly effective at generating high-quality images.
However, as these models grow larger, they require significantly more memory and suffer …