Neural machine translation for low-resource languages: A survey

S Ranathunga, ESA Lee, M Prifti Skenduli… - ACM Computing …, 2023 - dl.acm.org
Neural Machine Translation (NMT) has seen tremendous growth in the last ten years since
the early 2000s and has already entered a mature phase. While considered the most widely …

Semantic memory: A review of methods, models, and current challenges

AA Kumar - Psychonomic bulletin & review, 2021 - Springer
Adult semantic memory has been traditionally conceptualized as a relatively static memory
system that consists of knowledge about the world, concepts, and symbols. Considerable …

Language models are multilingual chain-of-thought reasoners

F Shi, M Suzgun, M Freitag, X Wang, S Srivats… - arxiv preprint arxiv …, 2022 - arxiv.org
We evaluate the reasoning abilities of large language models in multilingual settings. We
introduce the Multilingual Grade School Math (MGSM) benchmark, by manually translating …

Mind the gap: Understanding the modality gap in multi-modal contrastive representation learning

VW Liang, Y Zhang, Y Kwon… - Advances in Neural …, 2022 - proceedings.neurips.cc
We present modality gap, an intriguing geometric phenomenon of the representation space
of multi-modal models. Specifically, we show that different data modalities (eg images and …

The linear representation hypothesis and the geometry of large language models

K Park, YJ Choe, V Veitch - arxiv preprint arxiv:2311.03658, 2023 - arxiv.org
Informally, the'linear representation hypothesis' is the idea that high-level concepts are
represented linearly as directions in some representation space. In this paper, we address …

Multilingual denoising pre-training for neural machine translation

Y Liu, J Gu, N Goyal, X Li, S Edunov… - Transactions of the …, 2020 - direct.mit.edu
This paper demonstrates that multilingual denoising pre-training produces significant
performance gains across a wide variety of machine translation (MT) tasks. We present …

Are all languages created equal in multilingual BERT?

S Wu, M Dredze - arxiv preprint arxiv:2005.09093, 2020 - arxiv.org
Multilingual BERT (mBERT) trained on 104 languages has shown surprisingly good cross-
lingual performance on several NLP tasks, even without explicit cross-lingual signals …

Unsupervised translation of programming languages

B Roziere, MA Lachaux… - Advances in neural …, 2020 - proceedings.neurips.cc
A transcompiler, also known as source-to-source translator, is a system that converts source
code from a high-level programming language (such as C++ or Python) to another …

Billion-scale similarity search with GPUs

J Johnson, M Douze, H Jégou - IEEE Transactions on Big Data, 2019 - ieeexplore.ieee.org
Similarity search finds application in database systems handling complex data such as
images or videos, which are typically represented by high-dimensional features and require …

Z-score normalization, hubness, and few-shot learning

N Fei, Y Gao, Z Lu, T **ang - Proceedings of the IEEE/CVF …, 2021 - openaccess.thecvf.com
The goal of few-shot learning (FSL) is to recognize a set of novel classes with only few
labeled samples by exploiting a large set of abundant base class samples. Adopting a meta …