Folgen
Sachin Kumar
Sachin Kumar
Bestätigte E-Mail-Adresse bei osu.edu - Startseite
Titel
Zitiert von
Zitiert von
Jahr
Dolma: An open corpus of three trillion tokens for language model pretraining research
L Soldaini, R Kinney, A Bhagia, D Schwenk, D Atkinson, R Authur, ...
arXiv preprint arXiv:2402.00159, 2024
160*2024
Rewardbench: Evaluating reward models for language modeling
N Lambert, V Pyatkin, J Morrison, LJ Miranda, BY Lin, K Chandu, N Dziri, ...
arXiv preprint arXiv:2403.13787, 2024
150*2024
Earth mover's distance pooling over siamese LSTMs for automatic short answer grading
S Kumar, S Chakrabarti, S Roy
952017
Von mises-fisher loss for training sequence to sequence models with continuous outputs
S Kumar, Y Tsvetkov
arXiv preprint arXiv:1812.04616, 2018
852018
Language generation models can cause harm: So what can we do about it? an actionable survey
S Kumar, V Balachandran, L Njoo, A Anastasopoulos, Y Tsvetkov
arXiv preprint arXiv:2210.07700, 2022
822022
Controlled text generation as continuous optimization with multiple constraints
S Kumar, E Malmi, A Severyn, Y Tsvetkov
Advances in Neural Information Processing Systems 34, 14542-14554, 2021
792021
Do all languages cost the same? tokenization in the era of commercial language models
O Ahia, S Kumar, H Gonen, J Kasai, DR Mortensen, NA Smith, Y Tsvetkov
arXiv preprint arXiv:2305.13707, 2023
752023
Ssd-lm: Semi-autoregressive simplex-based diffusion language model for text generation and modular control
X Han, S Kumar, Y Tsvetkov
arXiv preprint arXiv:2210.17432, 2022
742022
Minding language models'(lack of) theory of mind: A plug-and-play multi-character belief tracker
M Sclar, S Kumar, P West, A Suhr, Y Choi, Y Tsvetkov
arXiv preprint arXiv:2306.00924, 2023
592023
Gradient-based constrained sampling from language models
S Kumar, B Paria, Y Tsvetkov
arXiv preprint arXiv:2205.12558, 2022
55*2022
Topics to avoid: Demoting latent confounds in text classification
S Kumar, S Wintner, NA Smith, Y Tsvetkov
arXiv preprint arXiv:1909.00453, 2019
402019
Assessing language model deployment with risk cards
L Derczynski, HR Kirk, V Balachandran, S Kumar, Y Tsvetkov, MR Leiser, ...
arXiv preprint arXiv:2303.18190, 2023
352023
On the blind spots of model-based evaluation metrics for text generation
T He, J Zhang, T Wang, S Kumar, K Cho, J Glass, Y Tsvetkov
arXiv preprint arXiv:2212.10020, 2022
342022
Machine translation into low-resource language varieties
S Kumar, A Anastasopoulos, S Wintner, Y Tsvetkov
arXiv preprint arXiv:2106.06797, 2021
342021
Neural abstractive summarization with structural attention
T Chowdhury, S Kumar, T Chakraborty
arXiv preprint arXiv:2004.09739, 2020
192020
WildTeaming at Scale: From In-the-Wild Jailbreaks to (Adversarially) Safer Language Models
L Jiang, K Rao, S Han, A Ettinger, F Brahman, S Kumar, N Mireshghallah, ...
arXiv preprint arXiv:2406.18510, 2024
17*2024
A deep reinforced model for zero-shot cross-lingual summarization with bilingual semantic similarity rewards
ZY Dou, S Kumar, Y Tsvetkov
arXiv preprint arXiv:2006.15454, 2020
172020
Referee: Reference-free sentence summarization with sharper controllability through symbolic knowledge distillation
M Sclar, P West, S Kumar, Y Tsvetkov, Y Choi
arXiv preprint arXiv:2210.13800, 2022
152022
The art of saying no: Contextual noncompliance in language models
F Brahman, S Kumar, V Balachandran, P Dasigi, V Pyatkin, ...
arXiv preprint arXiv:2407.12043, 2024
112024
Ssd-2: Scaling and inference-time fusion of diffusion language models
X Han, S Kumar, Y Tsvetkov, M Ghazvininejad
arXiv preprint arXiv:2305.14771, 2023
102023
Das System kann den Vorgang jetzt nicht ausführen. Versuchen Sie es später erneut.
Artikel 1–20