Stebėti
Iz Beltagy
Iz Beltagy
Allen Institute for Artificial Intelligence
Patvirtintas el. paštas beltagy.net - Pagrindinis puslapis
Pavadinimas
Cituota
Cituota
Metai
Longformer: The long-document transformer
I Beltagy, ME Peters, A Cohan
arXiv preprint arXiv:2004.05150, 2020
47082020
SciBERT: A Pretrained Language Model for Scientific Text
I Beltagy, K Lo, A Cohan
Proceedings of the 2019 Conference on Empirical Methods in Natural Language …, 2019
42132019
Don't stop pretraining: Adapt language models to domains and tasks
S Gururangan, A Marasović, S Swayamdipta, K Lo, I Beltagy, D Downey, ...
arXiv preprint arXiv:2004.10964, 2020
25402020
Bloom: A 176b-parameter open-access multilingual language model
T Le Scao, A Fan, C Akiki, E Pavlick, S Ilić, D Hesslow, R Castagné, ...
17752023
ScispaCy: fast and robust models for biomedical natural language processing
M Neumann, D King, I Beltagy, W Ammar
arXiv preprint arXiv:1902.07669, 2019
8902019
Specter: Document-level representation learning using citation-informed transformers
A Cohan, S Feldman, I Beltagy, D Downey, DS Weld
arXiv preprint arXiv:2004.07180, 2020
5882020
Construction of the literature graph in semantic scholar
W Ammar, D Groeneveld, C Bhagavatula, I Beltagy, M Crawford, ...
arXiv preprint arXiv:1805.02262, 2018
5192018
How far can camels go? exploring the state of instruction tuning on open resources
Y Wang, H Ivison, P Dasigi, J Hessel, T Khot, K Chandu, D Wadden, ...
Advances in Neural Information Processing Systems 36, 74764-74786, 2023
2902023
A dataset of information-seeking questions and answers anchored in research papers
P Dasigi, K Lo, I Beltagy, A Cohan, NA Smith, M Gardner
arXiv preprint arXiv:2105.03011, 2021
2632021
Longformer: The long-document transformer. arXiv 2020
I Beltagy, ME Peters, A Cohan
arXiv preprint arXiv:2004.05150, 2004
2302004
PRIMERA: Pyramid-based masked sentence pre-training for multi-document summarization
W Xiao, I Beltagy, G Carenini, A Cohan
arXiv preprint arXiv:2110.08499, 2021
1902021
What language model architecture and pretraining objective works best for zero-shot generalization?
T Wang, A Roberts, D Hesslow, T Le Scao, HW Chung, I Beltagy, ...
International Conference on Machine Learning, 22964-22984, 2022
1832022
Camels in a changing climate: Enhancing lm adaptation with tulu 2
H Ivison, Y Wang, V Pyatkin, N Lambert, M Peters, P Dasigi, J Jang, ...
arXiv preprint arXiv:2311.10702, 2023
1712023
SciREX: A challenge dataset for document-level information extraction
S Jain, M Van Zuylen, H Hajishirzi, I Beltagy
arXiv preprint arXiv:2005.00512, 2020
1702020
Machine learning for reliable mmwave systems: Blockage prediction and proactive handoff
A Alkhateeb, I Beltagy, S Alex
2018 IEEE Global conference on signal and information processing (GlobalSIP …, 2018
1622018
Pretrained language models for sequential sentence classification
A Cohan, I Beltagy, D King, B Dalvi, DS Weld
arXiv preprint arXiv:1909.04054, 2019
1602019
Olmo: Accelerating the science of language models
D Groeneveld, I Beltagy, P Walsh, A Bhagia, R Kinney, O Tafjord, AH Jha, ...
arXiv preprint arXiv:2402.00838, 2024
1592024
Dolma: An open corpus of three trillion tokens for language model pretraining research
L Soldaini, R Kinney, A Bhagia, D Schwenk, D Atkinson, R Authur, ...
arXiv preprint arXiv:2402.00159, 2024
1292024
Ms2: Multi-document summarization of medical studies
J DeYoung, I Beltagy, M van Zuylen, B Kuehl, LL Wang
arXiv preprint arXiv:2104.06486, 2021
1252021
Flex: Unifying evaluation for few-shot nlp
J Bragg, A Cohan, K Lo, I Beltagy
Advances in neural information processing systems 34, 15787-15800, 2021
1222021
Sistema negali atlikti operacijos. Bandykite vėliau dar kartą.
Straipsniai 1–20