关注
Sourab Mangrulkar
Sourab Mangrulkar
Hugging Face
在 huggingface.co 的电子邮件经过验证
标题
引用次数
引用次数
年份
Peft: State-of-the-art parameter-efficient fine-tuning methods
S Mangrulkar, S Gugger, L Debut, Y Belkada, S Paul, B Bossan
URL: https://github. com/huggingface/peft, 2022
2952022
Lysandre Debut, Younes Belkada, Sayak Paul, and Benjamin Bossan. 2022. Peft: State-of-the-art parameterefficient fine-tuning methods
S Mangrulkar, S Gugger
github. com/huggingface/pe, 0
268
SantaCoder: don't reach for the stars!
LB Allal, R Li, D Kocetkov, C Mou, C Akiki, CM Ferrandis, N Muennighoff, ...
arXiv preprint arXiv:2301.03988, 2023
2222023
Accelerate: Training and inference at scale made simple, efficient and adaptable
S Gugger, L Debut, T Wolf, P Schmid, Z Mueller, S Mangrulkar, M Sun, ...
662022
PEFT: state-of-the-art parameter-efficient fine-tuning methods (2022)
S Mangrulkar, S Gugger, L Debut, Y Belkada, S Paul, B Bossan
URL https://github. com/huggingface/peft, 2023
302023
A context-aware convolutional natural language generation model for dialogue systems
S Mangrulkar, S Shrivastava, V Thenkanidiyoor, DA Dinesh
Proceedings of the 19th Annual SIGdial Meeting on Discourse and Dialogue …, 2018
112018
Accelerate: training and inference at scale made simple, efficient and adaptable (2022)
S Gugger, L Debut, T Wolf, P Schmid, Z Mueller, S Mangrulkar, M Sun, ...
URL https://github. com/huggingface/accelerate, 0
10
Lysandre Debut, Thomas Wolf, Philipp Schmid, Zachary Mueller, and Sourab Mangrulkar. Accelerate: Training and Inference at Scale Made Simple, Efficient and Adaptable
S Gugger
92022
Making llms even more accessible with bitsandbytes, 4-bit quantization and qlora
Y Belkada, T Dettmers, A Pagnoni, S Gugger, S Mangrulkar
82023
Lysandre Debut, Thomas Wolf, Philipp Schmid, Zachary Mueller, and Sourab Mangrulkar. 2022. Accelerate: Training and inference at scale made simple, efficient and adaptable
S Gugger
8
Be3r: Bert based early-exit using expert routing
S Mangrulkar, A MS, V Sembium
Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and …, 2022
72022
Mixture of experts explained
O Sanseviero, L Tunstall, P Schmid, S Mangrulkar, Y Belkada, P Cuenca
Hugging Face Blog, 2023
62023
Peft: State-ofthe-art parameter-efficient fine-tuning methods
S Mangrulkar, S Gugger, L Debut, Y Belkada, S Paul, B Bossan
62022
Multilingual semantic sourcing using product images for cross-lingual alignment
S Mangrulkar, A MS, V Sembium
Companion Proceedings of the Web Conference 2022, 41-51, 2022
52022
Peft: State-of-the-art parameter-efficient fine-tuning methods
LDYBS Paul, S Mangrulkar, S Gugger
52022
Mixture of experts explained, 2023
O Sanseviero, L Tunstall, P Schmid, S Mangrulkar, Y Belkada, P Cuenca
URL: https://huggingface. co/blog/moe, 0
5
Lysandre Debut. 2022. Accelerate: Training and inference at scale made simple, efficient and adaptable
TWPSZ Mueller, SMS Gugger
4
HISS: A novel hybrid inference architecture in embedding based product sourcing using knowledge distillation
MS Ankith, S Mangrulkar, V Sembium
32022
Lysandre Debut, Younes Belkada, and Sayak Paul. Peft: State-of-the-art parameter-efficient fine-tuning methods
S Mangrulkar, S Gugger
2022
系统目前无法执行此操作,请稍后再试。
文章 1–19