XNLI: Evaluating cross-lingual sentence representations
State-of-the-art natural language processing systems rely on supervision in the form of
annotated data to learn competent models. These models are generally trained on data in a …
annotated data to learn competent models. These models are generally trained on data in a …
A deep convolutional coupling network for change detection based on heterogeneous optical and radar images
We propose an unsupervised deep convolutional coupling network for change detection
based on two heterogeneous images acquired by optical sensors and radars on different …
based on two heterogeneous images acquired by optical sensors and radars on different …
Multilingual models for compositional distributed semantics
We present a novel technique for learning semantic representations, which extends the
distributional hypothesis to multilingual data and joint-space embeddings. Our models …
distributional hypothesis to multilingual data and joint-space embeddings. Our models …
Learning joint multilingual sentence representations with neural machine translation
In this paper, we use the framework of neural machine translation to learn joint sentence
representations across six very different languages. Our aim is that a representation which is …
representations across six very different languages. Our aim is that a representation which is …
Correlational neural networks
Common representation learning (CRL), wherein different descriptions (or views) of the data
are embedded in a common subspace, has been receiving a lot of attention recently. Two …
are embedded in a common subspace, has been receiving a lot of attention recently. Two …
Filtering and mining parallel data in a joint multilingual space
H Schwenk - arxiv preprint arxiv:1805.09822, 2018 - arxiv.org
We learn a joint multilingual sentence embedding and use the distance between sentences
in different languages to filter noisy parallel data and to mine for parallel data in large news …
in different languages to filter noisy parallel data and to mine for parallel data in large news …
Concatenated power mean word embeddings as universal cross-lingual sentence representations
Average word embeddings are a common baseline for more sophisticated sentence
embedding techniques. However, they typically fall short of the performances of more …
embedding techniques. However, they typically fall short of the performances of more …
Multilingual distributed representations without word alignment
Distributed representations of meaning are a natural way to encode covariance
relationships between words and phrases in NLP. By overcoming data sparsity problems, as …
relationships between words and phrases in NLP. By overcoming data sparsity problems, as …
[PDF][PDF] Learning bilingual sentiment word embeddings for cross-language sentiment classification
H Zhou, L Chen, F Shi, D Huang - … of the 53rd Annual Meeting of …, 2015 - aclanthology.org
The sentiment classification performance relies on high-quality sentiment resources.
However, these resources are imbalanced in different languages. Cross-language …
However, these resources are imbalanced in different languages. Cross-language …
Learning bilingual word representations by marginalizing alignments
We present a probabilistic model that simultaneously learns alignments and distributed
representations for bilingual data. By marginalizing over word alignments the model …
representations for bilingual data. By marginalizing over word alignments the model …