Прати
Mike Lewis
Mike Lewis
Facebook AI Research
Верификована је имејл адреса на fb.com - Почетна страница
Наслов
Навело
Навело
Година
Roberta: A robustly optimized bert pretraining approach
Y Liu, M Ott, N Goyal, J Du, M Joshi, D Chen, O Levy, M Lewis, ...
arXiv preprint arXiv:1907.11692, 2019
31731*2019
Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension
M Lewis, Y Liu, N Goyal, M Ghazvininejad, A Mohamed, O Levy, ...
arXiv preprint arXiv:1910.13461, 2019
120962019
Retrieval-augmented generation for knowledge-intensive nlp tasks
P Lewis, E Perez, A Piktus, F Petroni, V Karpukhin, N Goyal, H Küttler, ...
Advances in neural information processing systems 33, 9459-9474, 2020
61992020
The llama 3 herd of models
A Dubey, A Jauhri, A Pandey, A Kadian, A Al-Dahle, A Letman, A Mathur, ...
arXiv preprint arXiv:2407.21783, 2024
3427*2024
Multilingual denoising pre-training for neural machine translation
Y Liu, J Gu, N Goyal, X Li, S Edunov, M Ghazvininejad, M Lewis, ...
Transactions of the Association for Computational Linguistics 8, 726-742, 2020
19662020
Hierarchical neural story generation
A Fan, M Lewis, Y Dauphin
arXiv preprint arXiv:1805.04833, 2018
18392018
Rethinking the role of demonstrations: What makes in-context learning work?
S Min, X Lyu, A Holtzman, M Artetxe, M Lewis, H Hajishirzi, L Zettlemoyer
arXiv preprint arXiv:2202.12837, 2022
12962022
End-to-end neural coreference resolution
K Lee, L He, M Lewis, L Zettlemoyer
arXiv preprint arXiv:1707.07045, 2017
12102017
Gpt3. int8 (): 8-bit matrix multiplication for transformers at scale
T Dettmers, M Lewis, Y Belkada, L Zettlemoyer
Advances in neural information processing systems 35, 30318-30332, 2022
10062022
Lima: Less is more for alignment
C Zhou, P Liu, P Xu, S Iyer, J Sun, Y Mao, X Ma, A Efrat, P Yu, L Yu, ...
Advances in Neural Information Processing Systems 36, 55006-55021, 2023
9732023
Generalization through memorization: Nearest neighbor language models
U Khandelwal, O Levy, D Jurafsky, L Zettlemoyer, M Lewis
arXiv preprint arXiv:1911.00172, 2019
8692019
Train short, test long: Attention with linear biases enables input length extrapolation
O Press, NA Smith, M Lewis
arXiv preprint arXiv:2108.12409, 2021
6542021
Incoder: A generative model for code infilling and synthesis
D Fried, A Aghajanyan, J Lin, S Wang, E Wallace, F Shi, R Zhong, W Yih, ...
arXiv preprint arXiv:2204.05999, 2022
6402022
Deep semantic role labeling: What works and what’s next
L He, K Lee, M Lewis, L Zettlemoyer
Proceedings of the 55th Annual Meeting of the Association for Computational …, 2017
5822017
Deal or no deal? end-to-end learning for negotiation dialogues
M Lewis, D Yarats, YN Dauphin, D Parikh, D Batra
arXiv preprint arXiv:1706.05125, 2017
5252017
Replug: Retrieval-augmented black-box language models
W Shi, S Min, M Yasunaga, M Seo, R James, M Lewis, L Zettlemoyer, ...
arXiv preprint arXiv:2301.12652, 2023
515*2023
Factscore: Fine-grained atomic evaluation of factual precision in long form text generation
S Min, K Krishna, X Lyu, M Lewis, W Yih, PW Koh, M Iyyer, L Zettlemoyer, ...
arXiv preprint arXiv:2305.14251, 2023
4912023
Asking and answering questions to evaluate the factual consistency of summaries
A Wang, K Cho, M Lewis
arXiv preprint arXiv:2004.04228, 2020
4662020
Measuring and narrowing the compositionality gap in language models
O Press, M Zhang, S Min, L Schmidt, NA Smith, M Lewis
arXiv preprint arXiv:2210.03350, 2022
463*2022
Efficient streaming language models with attention sinks
G Xiao, Y Tian, B Chen, S Han, M Lewis
arXiv preprint arXiv:2309.17453, 2023
4602023
Систем тренутно не може да изврши ову радњу. Пробајте поново касније.
Чланци 1–20