Opt: Open pre-trained transformer language models S Zhang, S Roller, N Goyal, M Artetxe, M Chen, S Chen, C Dewan, ... arXiv preprint arXiv:2205.01068, 2022 | 2859 | 2022 |
Xi Victoria Lin, Todor Mihaylov, Myle Ott, Sam Shleifer, Kurt Shuster, Daniel Simig, Punit Singh Koura, Anjali Sridhar, Tianlu Wang, and Luke Zettlemoyer. 2022 S Zhang, S Roller, N Goyal, M Artetxe, M Chen, S Chen, C Dewan, ... Opt: Open pretrained transformer language models 1, 2022 | 868 | 2022 |
Opt-iml: Scaling language model instruction meta learning through the lens of generalization S Iyer, XV Lin, R Pasunuru, T Mihaylov, D Simig, P Yu, K Shuster, T Wang, ... arXiv preprint arXiv:2212.12017, 2022 | 109 | 2022 |
Opt: Open pre-trained transformer language models. arXiv 2022 S Zhang, S Roller, N Goyal, M Artetxe, M Chen, S Chen, C Dewan, ... arXiv preprint arXiv:2205.01068, 2023 | 62 | 2023 |
Pytext: A seamless path from nlp research to production A Aly, K Lakhotia, S Zhao, M Mohit, B Oguz, A Arora, S Gupta, C Dewan, ... arXiv preprint arXiv:1812.08729, 2018 | 16 | 2018 |
Systems and methods for efficiently updating neural networks N Rotem, AU Diril, M Smelyanskiy, JS Park, C Dewan US Patent 10,699,190, 2020 | 8 | 2020 |
Systems and methods for efficiently updating neural networks N Rotem, AU Diril, M Smelyanskiy, JS Park, C Dewan US Patent 10,817,783, 2020 | 2 | 2020 |