Efficientqat: Efficient quantization-aware training for large language models

M Chen, W Shao, P Xu, J Wang, P Gao… - arxiv preprint arxiv …, 2024 - arxiv.org
Large language models (LLMs) are crucial in modern natural language processing and
artificial intelligence. However, they face challenges in managing their significant memory …

A survey of low-bit large language models: Basics, systems, and algorithms

R Gong, Y Ding, Z Wang, C Lv, X Zheng, J Du… - arxiv preprint arxiv …, 2024 - arxiv.org
Large language models (LLMs) have achieved remarkable advancements in natural
language processing, showcasing exceptional performance across various tasks. However …

How good are low-bit quantized llama3 models? an empirical study

W Huang, X Ma, H Qin, X Zheng, C Lv, H Chen… - arxiv e …, 2024 - ui.adsabs.harvard.edu
Meta's LLaMA family has become one of the most powerful open-source Large Language
Model (LLM) series. Notably, LLaMA3 models have recently been released and achieve …

An empirical study of llama3 quantization: From llms to mllms

W Huang, X Zheng, X Ma, H Qin, C Lv, H Chen, J Luo… - Visual Intelligence, 2024 - Springer
The LLaMA family, a collection of foundation language models ranging from 7B to 65B
parameters, has become one of the most powerful open-source large language models …

PTQ4SAM: Post-Training Quantization for Segment Anything

C Lv, H Chen, J Guo, Y Ding… - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
Abstract Segment Anything Model (SAM) has achieved impressive performance in many
computer vision tasks. However as a large-scale model the immense memory and …

A Comprehensive Approach Towards Wheat Leaf Disease Identification Leveraging Transformer Models and Federated Learning

M Fahim-Ul-Islam, A Chakrabarty, ST Ahmed… - IEEE …, 2024 - ieeexplore.ieee.org
Wheat is one of the most extensively cultivated crops worldwide that contributes significantly
to global food caloric and protein production and is grown on millions of hectares yearly …

Stbllm: Breaking the 1-bit barrier with structured binary llms

P Dong, L Li, Y Zhong, D Du, R Fan, Y Chen… - arxiv preprint arxiv …, 2024 - arxiv.org
In this paper, we present the first structural binarization method for LLM compression to less
than 1-bit precision. Although LLMs have achieved remarkable performance, their memory …

A Hybrid Framework of Transformer Encoder and Residential Conventional for Cardiovascular Disease Recognition Using Heart Sounds

RM Al-Tam, AM Al-Hejri, E Naji, FA Hashim… - IEEE …, 2024 - ieeexplore.ieee.org
Valvular heart disease (VHD) is one of the primary causes of cardiovascular illnesses with
high mortality rates worldwide. Early detection of VHD enables optimal treatment and stops …

Knowledge Augmentation for Distillation: A General and Effective Approach to Enhance Knowledge Distillation

Y Tang, Z Guo, L Wang, B Fan, F Cao, K Gao… - Proceedings of the 1st …, 2024 - dl.acm.org
Knowledge Distillation (KD), which extracts knowledge from a well-performed large neural
network (aka teacher network) to guide the training of a small network (aka student network) …

QVD: Post-training Quantization for Video Diffusion Models

S Tian, H Chen, C Lv, Y Liu, J Guo, X Liu, S Li… - Proceedings of the …, 2024 - dl.acm.org
Recently, video diffusion models (VDMs) have garnered significant attention due to their
notable advancements in generating coherent and realistic video content. However …