دنبال کردن
Ofir Press
Ofir Press
ایمیل تأیید شده در princeton.edu - صفحهٔ اصلی
عنوان
نقل شده توسط
نقل شده توسط
سال
Bloom: A 176b-parameter open-access multilingual language model
T Le Scao, A Fan, C Akiki, E Pavlick, S Ilić, D Hesslow, R Castagné, ...
17682023
Using the Output Embedding to Improve Language Models
O Press, L Wolf
EACL 2017, 2017
8052017
Train Short, Test Long: Attention with Linear Biases Enables Input Length Extrapolation
O Press, NA Smith, M Lewis
ICLR 2022, 2021
6522021
Measuring and narrowing the compositionality gap in language models
O Press, M Zhang, S Min, L Schmidt, NA Smith, M Lewis
Findings of EMNLP 2023, 2022
462*2022
SWE-bench: Can Language Models Resolve Real-World GitHub Issues?
CE Jimenez, J Yang, A Wettig, S Yao, K Pei, O Press, K Narasimhan
ICLR 2024, 2023
3522023
How language model hallucinations can snowball
M Zhang, O Press, W Merrill, A Liu, NA Smith
ICML 2024, 2023
2662023
SWE-agent: Agent-Computer Interfaces Enable Automated Software Engineering
J Yang, CE Jimenez, A Wettig, K Lieret, S Yao, K Narasimhan, O Press
NeurIPS 2024, 2024
1462024
Language Generation with Recurrent Generative Adversarial Networks without Pre-training
O Press, A Bar, B Bogin, J Berant, L Wolf
1st Workshop on Learning to Generate Natural Language at ICML 2017, 2017
1382017
What Language Model to Train if You Have One Million GPU Hours?
T Le Scao, T Wang, D Hesslow, L Saulnier, S Bekman, MS Bari, ...
Findings of EMNLP 2022, 2022
1182022
Transformer Language Models without Positional Encodings Still Learn Positional Information
A Haviv, O Ram, O Press, P Izsak, O Levy
Findings of EMNLP 2022, 2022
1122022
Improving Transformer Models by Reordering their Sublayers
O Press, NA Smith, O Levy
ACL 2020, 2019
852019
Shortformer: Better Language Modeling using Shorter Inputs
O Press, NA Smith, M Lewis
ACL 2021, 2020
832020
You may not need attention
O Press, NA Smith
arXiv preprint arXiv:1810.13409, 2018
282018
SWE-bench Multimodal: Do AI Systems Generalize to Visual Software Domains?
J Yang, CE Jimenez, AL Zhang, K Lieret, J Yang, X Wu, O Press, ...
ICLR 2025, 2024
92024
Scicode: A research coding benchmark curated by scientists
M Tian, L Gao, SD Zhang, X Chen, C Fan, X Guo, R Haas, P Ji, ...
NeurIPS 2024, 2024
92024
Partially shuffling the training data to improve language models
O Press
arXiv preprint arXiv:1903.04167, 2019
52019
AssistantBench: Can Web Agents Solve Realistic and Time-Consuming Tasks?
O Yoran, SJ Amouyal, C Malaviya, B Bogin, O Press, J Berant
EMNLP 2024, 2024
42024
CiteME: Can Language Models Accurately Cite Scientific Claims?
O Press, A Hochlehnert, A Prabhu, V Udandarao, O Press, M Bethge
NeurIPS 2024, 2024
42024
EnIGMA: Enhanced Interactive Generative Model Agent for CTF Challenges
T Abramovich, M Udeshi, M Shao, K Lieret, H Xi, K Milner, S Jancheska, ...
arXiv preprint arXiv:2409.16165, 2024
22024
BLOOM: A 176B-Parameter Open-Access Multilingual Language Model
V Danchev, V Nikoulina, V Laippala, V Lepercq, V Prabhu, Z Alyafeai, ...
2023
سیستم در حال حاضر قادر به انجام عملکرد نیست. بعداً دوباره امتحان کنید.
مقاله‌ها 1–20