Ptq4sam: Post-training quantization for segment anything

C Lv, H Chen, J Guo, Y Ding… - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
Abstract Segment Anything Model (SAM) has achieved impressive performance in many
computer vision tasks. However as a large-scale model the immense memory and …

Efficientqat: Efficient quantization-aware training for large language models

M Chen, W Shao, P Xu, J Wang, P Gao… - arxiv preprint arxiv …, 2024 - arxiv.org
Large language models (LLMs) are crucial in modern natural language processing and
artificial intelligence. However, they face challenges in managing their significant memory …

A survey of low-bit large language models: Basics, systems, and algorithms

R Gong, Y Ding, Z Wang, C Lv, X Zheng, J Du… - arxiv preprint arxiv …, 2024 - arxiv.org
Large language models (LLMs) have achieved remarkable advancements in natural
language processing, showcasing exceptional performance across various tasks. However …

An empirical study of llama3 quantization: From llms to mllms

W Huang, X Zheng, X Ma, H Qin, C Lv, H Chen, J Luo… - Visual Intelligence, 2024 - Springer
The LLaMA family, a collection of foundation language models ranging from 7B to 65B
parameters, has become one of the most powerful open-source large language models …

SliM-LLM: Salience-driven mixed-precision quantization for large language models

W Huang, H Qin, Y Liu, Y Li, X Liu, L Benini… - arxiv preprint arxiv …, 2024 - arxiv.org
Large language models (LLMs) achieve remarkable performance in natural language
understanding but require substantial computation and memory resources. Post-training …

Demystifying the compression of mixture-of-experts through a unified framework

S He, D Dong, L Ding, A Li - arxiv preprint arxiv:2406.02500, 2024 - arxiv.org
Scaling large language models has revolutionized the performance across diverse domains,
yet the continual growth in model size poses significant challenges for real-world …

A Comprehensive Approach Towards Wheat Leaf Disease Identification Leveraging Transformer Models and Federated Learning

M Fahim-Ul-Islam, A Chakrabarty, ST Ahmed… - IEEE …, 2024 - ieeexplore.ieee.org
Wheat is one of the most extensively cultivated crops worldwide that contributes significantly
to global food caloric and protein production and is grown on millions of hectares yearly …

Enhanced sentiment analysis and topic modeling during the pandemic using automated latent Dirichlet allocation

A Batool, YC Byun - IEEE Access, 2024 - ieeexplore.ieee.org
The COVID-19 pandemic has profoundly impacted human societies, resulting in the loss of
millions of lives and slowing economic growth worldwide. This devastating pandemic …

Stbllm: Breaking the 1-bit barrier with structured binary llms

P Dong, L Li, Y Zhong, D Du, R Fan, Y Chen… - arxiv preprint arxiv …, 2024 - arxiv.org
In this paper, we present the first structural binarization method for LLM compression to less
than 1-bit precision. Although LLMs have achieved remarkable performance, their memory …

Ternaryllm: Ternarized large language model

T Chen, Z Li, W Xu, Z Zhu, D Li, L Tian… - arxiv preprint arxiv …, 2024 - arxiv.org
Large language models (LLMs) have achieved remarkable performance on Natural
Language Processing (NLP) tasks, but they are hindered by high computational costs and …