Language models are few-shot learners T Brown, B Mann, N Ryder, M Subbiah, JD Kaplan, P Dhariwal, ... Advances in neural information processing systems 33, 1877-1901, 2020 | 39745 | 2020 |
Rewon Child, Aditya Ramesh, Daniel M. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray TB Brown, B Mann, N Ryder, M Subbiah, J Kaplan, P Dhariwal, ... Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford …, 2020 | 9477 | 2020 |
Gpt-4 technical report J Achiam, S Adler, S Agarwal, L Ahmad, I Akkaya, FL Aleman, D Almeida, ... arXiv preprint arXiv:2303.08774, 2023 | 7777 | 2023 |
Evaluating large language models trained on code M Chen, J Tworek, H Jun, Q Yuan, HPDO Pinto, J Kaplan, H Edwards, ... arXiv preprint arXiv:2107.03374, 2021 | 3817 | 2021 |
Scaling laws for autoregressive generative modeling T Henighan, J Kaplan, M Katz, M Chen, C Hesse, J Jackson, H Jun, ... arXiv preprint arXiv:2010.14701, 2020 | 378 | 2020 |
ChatGPT: Optimizing language models for dialogue J Schulman, B Zoph, C Kim, J Hilton, J Menick, J Weng, JFC Uribe, ... OpenAI blog 2 (4), 2022 | 322 | 2022 |
Language Models are Few-Shot Learners. 2020. doi: 10.48550 TB Brown, B Mann, N Ryder, M Subbiah, J Kaplan, P Dhariwal, ... arxiv, 5-7, 2005 | 260 | 2005 |
& Amodei, D.(2020) TB Brown, B Mann, N Ryder, M Subbiah, J Kaplan, P Dhariwal, ... Language models are few-shot learners, 2005 | 200 | 2005 |
Language models are few-shot learners B Mann, N Ryder, M Subbiah, J Kaplan, P Dhariwal, A Neelakantan, ... arXiv preprint arXiv:2005.14165 1, 2020 | 193 | 2020 |
Tensor programs v: Tuning large neural networks via zero-shot hyperparameter transfer G Yang, EJ Hu, I Babuschkin, S Sidor, X Liu, D Farhi, N Ryder, J Pachocki, ... arXiv preprint arXiv:2203.03466, 2022 | 131 | 2022 |
Language models are few-shot learners (arXiv: 2005.14165). arXiv TB Brown, B Mann, N Ryder, M Subbiah, J Kaplan, P Dhariwal, ... | 131 | 2005 |
Gpt-4o system card A Hurst, A Lerer, AP Goucher, A Perelman, A Ramesh, A Clark, AJ Ostrow, ... arXiv preprint arXiv:2410.21276, 2024 | 128 | 2024 |
Tuning large neural networks via zero-shot hyperparameter transfer G Yang, E Hu, I Babuschkin, S Sidor, X Liu, D Farhi, N Ryder, J Pachocki, ... Advances in Neural Information Processing Systems 34, 17084-17097, 2021 | 102 | 2021 |
Evaluating large language models trained on code. arXiv 2021 M Chen, J Tworek, H Jun, Q Yuan, HPO Pinto, J Kaplan, H Edwards, ... arXiv preprint arXiv:2107.03374 10, 2021 | 72 | 2021 |
Language models are few-shot learners. cite TB Brown, B Mann, N Ryder, M Subbiah, J Kaplan, P Dhariwal, ... arXiv preprint arxiv:2005.14165, 2020 | 64 | 2020 |
Alethea Power, Lukasz Kaiser, Mohammad Bavarian, Clemens Winter, Philippe Tillet, Felipe Petroski Such, David W M Chen, J Tworek, H Jun, Q Yuan, H Ponde, J Kaplan, H Edwards, ... | 56 | 2021 |
& Agarwal, S.(2020). Language models are few-shot learners TB Brown, B Mann, N Ryder, M Subbiah, J Kaplan, P Dhariwal arXiv preprint arXiv:2005.14165, 2005 | 55 | 2005 |
Introducing chatgpt J Schulman, B Zoph, C Kim, J Hilton, J Menick, J Weng, JFC Uribe, ... OpenAI Blog, 2022 | 48 | 2022 |
The geometry of rank decompositions of matrix multiplication II: 3× 3 matrices G Ballard, C Ikenmeyer, JM Landsberg, N Ryder Journal of Pure and Applied Algebra 223 (8), 3205-3224, 2019 | 30 | 2019 |
Openai o1 system card A Jaech, A Kalai, A Lerer, A Richardson, A El-Kishky, A Low, A Helyar, ... arXiv preprint arXiv:2412.16720, 2024 | 21 | 2024 |