Neural machine translation for low-resource languages: A survey

S Ranathunga, ESA Lee, M Prifti Skenduli… - ACM Computing …, 2023 - dl.acm.org
Neural Machine Translation (NMT) has seen tremendous growth in the last ten years since
the early 2000s and has already entered a mature phase. While considered the most widely …

[HTML][HTML] A survey on recent named entity recognition and relationship extraction techniques on clinical texts

P Bose, S Srinivasan, WC Sleeman IV, J Palta… - Applied Sciences, 2021 - mdpi.com
Significant growth in Electronic Health Records (EHR) over the last decade has provided an
abundance of clinical text that is mostly unstructured and untapped. This huge amount of …

Beyond the imitation game: Quantifying and extrapolating the capabilities of language models

A Srivastava, A Rastogi, A Rao, AAM Shoeb… - arxiv preprint arxiv …, 2022 - arxiv.org
Language models demonstrate both quantitative improvement and new qualitative
capabilities with increasing scale. Despite their potentially transformative impact, these new …

Refiner: Reasoning feedback on intermediate representations

D Paul, M Ismayilzada, M Peyrard, B Borges… - arxiv preprint arxiv …, 2023 - arxiv.org
Language models (LMs) have recently shown remarkable performance on reasoning tasks
by explicitly generating intermediate inferences, eg, chain-of-thought prompting. However …

mgpt: Few-shot learners go multilingual

O Shliazhko, A Fenogenova, M Tikhonova… - Transactions of the …, 2024 - direct.mit.edu
This paper introduces mGPT, a multilingual variant of GPT-3, pretrained on 61 languages
from 25 linguistically diverse language families using Wikipedia and the C4 Corpus. We …

Comparison of text preprocessing methods

CP Chai - Natural Language Engineering, 2023 - cambridge.org
Text preprocessing is not only an essential step to prepare the corpus for modeling but also
a key area that directly affects the natural language processing (NLP) application results. For …

Data augmentation using llms: Data perspectives, learning paradigms and challenges

B Ding, C Qin, R Zhao, T Luo, X Li… - Findings of the …, 2024 - aclanthology.org
In the rapidly evolving field of large language models (LLMs), data augmentation (DA) has
emerged as a pivotal technique for enhancing model performance by diversifying training …

An empirical survey of data augmentation for limited data learning in NLP

J Chen, D Tam, C Raffel, M Bansal… - Transactions of the …, 2023 - direct.mit.edu
NLP has achieved great progress in the past decade through the use of neural models and
large labeled datasets. The dependence on abundant data prevents NLP models from being …

Walklm: A uniform language model fine-tuning framework for attributed graph embedding

Y Tan, Z Zhou, H Lv, W Liu… - Advances in neural …, 2023 - proceedings.neurips.cc
Graphs are widely used to model interconnected entities and improve downstream
predictions in various real-world applications. However, real-world graphs nowadays are …

mmarco: A multilingual version of the ms marco passage ranking dataset

L Bonifacio, V Jeronymo, HQ Abonizio… - arxiv preprint arxiv …, 2021 - arxiv.org
The MS MARCO ranking dataset has been widely used for training deep learning models for
IR tasks, achieving considerable effectiveness on diverse zero-shot scenarios. However, this …