ADFQ-ViT: Activation-Distribution-Friendly Post-Training Quantization for Vision Transformers

Y Jiang, N Sun, X **e, F Yang, T Li - Neural Networks, 2025 - Elsevier
Abstract Vision Transformers (ViTs) have exhibited exceptional performance across diverse
computer vision tasks, while their substantial parameter size incurs significantly increased …

Towards Accurate Post-Training Quantization of Vision Transformers via Error Reduction

Y Zhong, Y Huang, J Hu, Y Zhang… - IEEE Transactions on …, 2025 - ieeexplore.ieee.org
Post-training quantization (PTQ) for vision transformers (ViTs) has received increasing
attention from both academic and industrial communities due to its minimal data needs and …

Semantics Prompting Data-Free Quantization for Low-Bit Vision Transformers

Y Zhong, Y Zhou, Y Zhang, S Li, Y Li, F Chao… - arxiv preprint arxiv …, 2024 - arxiv.org
Data-free quantization (DFQ), which facilitates model quantization without real data to
address increasing concerns about data security, has garnered significant attention within …

Low-Bit Quantization Favors Undertrained LLMs: Scaling Laws for Quantized LLMs with 100T Training Tokens

X Ouyang, T Ge, T Hartvigsen, Z Zhang, H Mi… - arxiv preprint arxiv …, 2024 - arxiv.org
We reveal that low-bit quantization favors undertrained large language models (LLMs) by
observing that models with larger sizes or fewer training tokens experience less quantization …

Mixed Non-linear Quantization for Vision Transformers

G Kim, J Lee, S Park, Y Kwon, H Kim - arxiv preprint arxiv:2407.18437, 2024 - arxiv.org
The majority of quantization methods have been proposed to reduce the model size of
Vision Transformers, yet most of them have overlooked the quantization of non-linear …