Segui
Weilin Zhao
Titolo
Citata da
Citata da
Anno
Parameter-efficient fine-tuning of large-scale pre-trained language models
N Ding, Y Qin, G Yang, F Wei, Z Yang, Y Su, S Hu, Y Chen, CM Chan, ...
Nature Machine Intelligence 5 (3), 220-235, 2023
6792023
Ptr: Prompt tuning with rules for text classification
X Han, W Zhao, N Ding, Z Liu, M Sun
AI Open 3, 182-192, 2022
5262022
Openprompt: An open-source framework for prompt-learning
N Ding, S Hu, W Zhao, Y Chen, Z Liu, HT Zheng, M Sun
arXiv preprint arXiv:2111.01998, 2021
3172021
Tool learning with foundation models
Y Qin, S Hu, Y Lin, W Chen, N Ding, G Cui, Z Zeng, Y Huang, C Xiao, ...
arXiv preprint arXiv:2304.08354, 2023
2562023
Delta Tuning: A Comprehensive Study of Parameter Efficient Methods for Pre-trained Language Models
N Ding, Y Qin, G Yang, F Wei, Z Yang, Y Su, S Hu, Y Chen, CM Chan, ...
arXiv preprint arXiv:2203.06904, 2022
2482022
MiniCPM-V: A GPT-4V Level MLLM on Your Phone
Y Yao, T Yu, A Zhang, C Wang, J Cui, H Zhu, T Cai, H Li, W Zhao, Z He, ...
arXiv preprint arXiv:2408.01800, 2024
1892024
MiniCPM: Unveiling the Potential of Small Language Models with Scalable Training Strategies
S Hu, Y Tu, X Han, C He, G Cui, X Long, Z Zheng, Y Fang, Y Huang, ...
arXiv preprint arXiv:2404.06395, 2024
1832024
Moderate-fitting as a Natural Backdoor Defender for Pre-trained Language Models
B Zhu, Y Qin, G Cui, Y Chen, W Zhao, C Fu, Y Deng, Z Liu, J Wang, W Wu, ...
Advances in Neural Information Processing Systems 35, 1086-1099, 2022
322022
OpenDelta: A Plug-and-play Library for Parameter-efficient Adaptation of Pre-trained Models
S Hu, N Ding, W Zhao, X Lv, Z Zhang, Z Liu, M Sun
arXiv preprint arXiv:2307.03084, 2023
132023
Ouroboros: Generating Longer Drafts Phrase by Phrase for Faster Speculative Decoding
W Zhao, Y Huang, X Han, W Xu, C Xiao, X Zhang, Y Fang, K Zhang, Z Liu, ...
10*2024
Configurable Foundation Models: Building LLMs from a Modular Perspective
C Xiao, Z Zhang, C Song, D Jiang, F Yao, X Han, X Wang, S Wang, ...
arXiv preprint arXiv:2409.02877, 2024
92024
BMCook: A task-agnostic compression toolkit for big models
Z Zhang, B Gong, Y Chen, X Han, G Zeng, W Zhao, Y Chen, Z Liu, M Sun
Proceedings of the 2022 Conference on Empirical Methods in Natural Language …, 2022
92022
BMInf: An Efficient Toolkit for Big Model Inference and Tuning
X Han, G Zeng, W Zhao, Z Liu, Z Zhang, J Zhou, J Zhang, J Chao, M Sun
Proceedings of the 60th Annual Meeting of the Association for Computational …, 2022
82022
CA-LoRA: Adapting Existing LoRA for Compressed LLMs to Enable Efficient Multi-Tasking on Personal Devices
W Zhao, Y Huang, X Han, Z Liu, Z Zhang, K Li, C Chen, TAO YANG, ...
First Conference on Language Modeling, 2024
7*2024
Beyond the Turn-Based Game: Enabling Real-Time Conversations with Duplex Models
X Zhang, Y Chen, S Hu, X Han, Z Xu, Y Xu, W Zhao, M Sun, Z Liu
arXiv preprint arXiv:2406.15718, 2024
52024
BurstAttention: An Efficient Distributed Attention Framework for Extremely Long Sequences
S Ao, W Zhao, X Han, C Yang, Z Liu, C Shi, M Sun, S Wang, T Su
arXiv preprint arXiv:2403.09347, 2024
42024
Unlock Predictable Scaling from Emergent Abilities
S Hu, X Liu, X Han, X Zhang, C He, W Zhao, Y Lin, N Ding, Z Ou, G Zeng, ...
arXiv preprint arXiv:2310.03262, 2023
22023
Seq1F1B: Efficient Sequence-Level Pipeline Parallelism for Large Language Model Training
S Ao, W Zhao, X Han, C Yang, Z Liu, C Shi, M Sun
arXiv preprint arXiv:2406.03488, 2024
12024
Optimal RoPE extension via Bayesian Optimization for training-free length generalization
X Zhang, S Hu, W Zhao, H Wang, X Han, C He, G Zeng, Z Liu, M Sun
AI Open, 2025
2025
H3T: efficient integration of memory optimization and parallelism for high-throughput transformer training
Y Wang, X Han, W Zhao, G Zeng, Z Liu, M Sun
Proceedings of the 37th International Conference on Neural Information …, 2023
2023
Il sistema al momento non può eseguire l'operazione. Riprova più tardi.
Articoli 1–20