Подписаться
Yuzhang Shang
Yuzhang Shang
Подтвержден адрес электронной почты в домене hawk.iit.edu - Главная страница
Название
Процитировано
Процитировано
Год
Post-training Quantization on Diffusion Models
Y Shang, Z Yuan, B Xie, B Wu, Y Yan
CVPR, 2023
1532023
Rptq: Reorder-based post-training quantization for large language models
Z Yuan, L Niu, J Liu, W Liu, X Wang, Y Shang, G Sun, Q Wu, J Wu, B Wu
arXiv preprint arXiv:2304.01089, 2023
812023
Llava-prumerge: Adaptive token reduction for efficient large multimodal models
Y Shang, M Cai, B Xu, YJ Lee, Y Yan
arXiv preprint arXiv:2403.15388, 2024
732024
LLM Inference Unveiled: Survey and Roofline Model Insights
Z Yuan*, Y Shang*, Y Zhou*, Z Dong, C Xue, B Wu, Z Li, Q Gu, YJ Lee, ...
arXiv preprint arXiv:2402.16363, 2024
612024
Pb-llm: Partially binarized large language models
Y Shang, Z Yuan, Q Wu, Z Dong
ICLR, 2024
502024
ASVD: Activation-aware Singular Value Decomposition for Compressing Large Language Models
Z Yuan*, Y Shang*, Y Song, Q Wu, Y Yan, G Sun
arXiv preprint arXiv:2312.05821, 2023
372023
Lipschitz Continuity Guided Knowledge Distillation
Y Shang, B Duan, Z Zong, L Nie, Y Yan
ICCV, 2021
362021
Network Binarization via Contrastive Learning
Y Shang, X Dan, Z Zong, L Nie, Y Yan
ECCV, 2022
312022
Lipschitz Continuity Retained Binary Neural Network
Y Shang, X Dan, B Duan, Z Zong, L Nie, Y Yan
ECCV, 2022
242022
Quest: Low-bit diffusion model quantization via efficient selective finetuning
H Wang, Y Shang, Z Yuan, J Wu, J Yan, Y Yan
arXiv preprint arXiv:2402.03666, 2024
162024
MIM4DD: Mutual Information Maximization for Dataset Distillation
Y Shang, Z Yuan, Y Yan
NeurIPS, 2023
142023
PTQ4DiT: Post-training Quantization for Diffusion Transformers
J Wu*, H Wang*, Y Shang, M Shah, Y Yan
NeurIPS, 2024
122024
Temporalbench: Benchmarking fine-grained temporal understanding for multimodal video models
M Cai, R Tan, J Zhang, B Zou, K Zhang, F Yao, F Zhu, J Gu, Y Zhong, ...
arXiv preprint arXiv:2410.10818, 2024
9*2024
Enhancing Post-training Quantization Calibration through Contrastive Learning
Y Shang, G Liu, RR Kompella, Y Yan
CVPR, 2024
92024
Interpolating video-llms: Toward longer-sequence lmms in a training-free manner
Y Shang, B Xu, W Kang, M Cai, Y Li, Z Wen, Z Dong, K Keutzer, YJ Lee, ...
arXiv preprint arXiv:2409.12963, 2024
52024
A closer look at time steps is worthy of triple speed-up for diffusion model training
K Wang, M Shi, Y Zhou, Z Li, Z Yuan, Y Shang, X Peng, H Zhang, Y You
arXiv preprint arXiv:2405.17403, 2024
52024
Dataset Quantization with Active Learning based Adaptive Sampling
Z Zhao, Y Shang, J Wu, Y Yan
ECCV, 2024
42024
Causal-DFQ: Causality Guided Data-free Network Quantization
Y Shang, B Xu, G Liu, R Kompella, Y Yan
ICCV, 2023
42023
Network specialization via feature-level knowledge distillation
G Liu, Y Shang, Y Yao, R Kompella
CVPR-W, 2023
42023
Adaptive cross-architecture mutual knowledge distillation
J Ni, H Tang, Y Shang, B Duan, Y Yan
2024 IEEE 18th International Conference on Automatic Face and Gesture …, 2024
32024
В данный момент система не может выполнить эту операцию. Повторите попытку позднее.
Статьи 1–20