Παρακολούθηση
Akshita Bhagia
Akshita Bhagia
Allen Institute for AI
Η διεύθυνση ηλεκτρονικού ταχυδρομείου έχει επαληθευτεί στον τομέα allenai.org - Αρχική σελίδα
Τίτλος
Παρατίθεται από
Παρατίθεται από
Έτος
Olmo: Accelerating the science of language models
D Groeneveld, I Beltagy, P Walsh, A Bhagia, R Kinney, O Tafjord, AH Jha, ...
ACL 2024 🏆 Best Theme Paper, 2024
254*2024
Dolma: An open corpus of three trillion tokens for language model pretraining research
L Soldaini, R Kinney, A Bhagia, D Schwenk, D Atkinson, R Authur, ...
ACL 2024 🏆 Best Resource Paper, 2024
176*2024
What's In My Big Data?
Y Elazar, A Bhagia, I Magnusson, A Ravichander, D Schwenk, A Suhr, ...
ICLR 2024 (Spotlight; 5% papers), 2023
832023
HINT: Hypernetwork Instruction Tuning for Efficient Zero-& Few-Shot Generalisation
H Ivison, A Bhagia, Y Wang, H Hajishirzi, M Peters
ACL 2023, 2022
272022
Paloma: A benchmark for evaluating language model fit
I Magnusson, A Bhagia, V Hofmann, L Soldaini, AH Jha, O Tafjord, ...
NeurIPS 2024, 2023
24*2023
Findings of the WMT’22 shared task on large-scale machine translation evaluation for African languages
D Adelani, MMI Alam, A Anastasopoulos, A Bhagia, MR Costa-jussà, ...
Proceedings of the Seventh Conference on Machine Translation (WMT), 773-800, 2022
142022
Olmoe: Open mixture-of-experts language models
N Muennighoff, L Soldaini, D Groeneveld, K Lo, J Morrison, S Min, W Shi, ...
arXiv preprint arXiv:2409.02060, 2024
42024
2 OLMo 2 Furious
T OLMo, P Walsh, L Soldaini, D Groeneveld, K Lo, S Arora, A Bhagia, ...
arXiv preprint arXiv:2501.00656, 2024
32024
Catwalk: A unified language model evaluation framework for many datasets
D Groeneveld, A Awadalla, I Beltagy, A Bhagia, I Magnusson, H Peng, ...
arXiv preprint arXiv:2312.10253, 2023
32023
Continued pretraining for better zero-and few-shot promptability
Z Wu, RL Logan IV, P Walsh, A Bhagia, D Groeneveld, S Singh, I Beltagy
EMNLP 2022, 2022
32022
On advances in text generation from images beyond captioning: A case study in self-rationalization
S Palaskar, A Bhagia, Y Bisk, F Metze, AW Black, A Marasović
Findings of EMNLP 2022, 2022
22022
Establishing Task Scaling Laws via Compute-Efficient Model Ladders
A Bhagia, J Liu, A Wettig, D Heineman, O Tafjord, AH Jha, L Soldaini, ...
arXiv preprint arXiv:2412.04403, 2024
2024
Robust Tooling and New Resources for Large Language Model Evaluation via Catwalk
K Richardson, I Magnusson, O Tafjord, A Bhagia, I Beltagy, A Cohan, ...
Δεν είναι δυνατή η εκτέλεση της ενέργειας από το σύστημα αυτή τη στιγμή. Προσπαθήστε ξανά αργότερα.
Άρθρα 1–13