Peft: State-of-the-art parameter-efficient fine-tuning methods S Mangrulkar, S Gugger, L Debut, Y Belkada, S Paul, B Bossan URL: https://github. com/huggingface/peft, 2022 | 295 | 2022 |
Lysandre Debut, Younes Belkada, Sayak Paul, and Benjamin Bossan. 2022. Peft: State-of-the-art parameterefficient fine-tuning methods S Mangrulkar, S Gugger github. com/huggingface/pe, 0 | 268 | |
SantaCoder: don't reach for the stars! LB Allal, R Li, D Kocetkov, C Mou, C Akiki, CM Ferrandis, N Muennighoff, ... arXiv preprint arXiv:2301.03988, 2023 | 222 | 2023 |
Accelerate: Training and inference at scale made simple, efficient and adaptable S Gugger, L Debut, T Wolf, P Schmid, Z Mueller, S Mangrulkar, M Sun, ... | 66 | 2022 |
PEFT: state-of-the-art parameter-efficient fine-tuning methods (2022) S Mangrulkar, S Gugger, L Debut, Y Belkada, S Paul, B Bossan URL https://github. com/huggingface/peft, 2023 | 30 | 2023 |
A context-aware convolutional natural language generation model for dialogue systems S Mangrulkar, S Shrivastava, V Thenkanidiyoor, DA Dinesh Proceedings of the 19th Annual SIGdial Meeting on Discourse and Dialogue …, 2018 | 11 | 2018 |
Accelerate: training and inference at scale made simple, efficient and adaptable (2022) S Gugger, L Debut, T Wolf, P Schmid, Z Mueller, S Mangrulkar, M Sun, ... URL https://github. com/huggingface/accelerate, 0 | 10 | |
Lysandre Debut, Thomas Wolf, Philipp Schmid, Zachary Mueller, and Sourab Mangrulkar. Accelerate: Training and Inference at Scale Made Simple, Efficient and Adaptable S Gugger | 9 | 2022 |
Making llms even more accessible with bitsandbytes, 4-bit quantization and qlora Y Belkada, T Dettmers, A Pagnoni, S Gugger, S Mangrulkar | 8 | 2023 |
Lysandre Debut, Thomas Wolf, Philipp Schmid, Zachary Mueller, and Sourab Mangrulkar. 2022. Accelerate: Training and inference at scale made simple, efficient and adaptable S Gugger | 8 | |
Be3r: Bert based early-exit using expert routing S Mangrulkar, A MS, V Sembium Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and …, 2022 | 7 | 2022 |
Mixture of experts explained O Sanseviero, L Tunstall, P Schmid, S Mangrulkar, Y Belkada, P Cuenca Hugging Face Blog, 2023 | 6 | 2023 |
Peft: State-ofthe-art parameter-efficient fine-tuning methods S Mangrulkar, S Gugger, L Debut, Y Belkada, S Paul, B Bossan | 6 | 2022 |
Multilingual semantic sourcing using product images for cross-lingual alignment S Mangrulkar, A MS, V Sembium Companion Proceedings of the Web Conference 2022, 41-51, 2022 | 5 | 2022 |
Peft: State-of-the-art parameter-efficient fine-tuning methods LDYBS Paul, S Mangrulkar, S Gugger | 5 | 2022 |
Mixture of experts explained, 2023 O Sanseviero, L Tunstall, P Schmid, S Mangrulkar, Y Belkada, P Cuenca URL: https://huggingface. co/blog/moe, 0 | 5 | |
Lysandre Debut. 2022. Accelerate: Training and inference at scale made simple, efficient and adaptable TWPSZ Mueller, SMS Gugger | 4 | |
HISS: A novel hybrid inference architecture in embedding based product sourcing using knowledge distillation MS Ankith, S Mangrulkar, V Sembium | 3 | 2022 |
Lysandre Debut, Younes Belkada, and Sayak Paul. Peft: State-of-the-art parameter-efficient fine-tuning methods S Mangrulkar, S Gugger | | 2022 |