Preserving integrity in online social networks
Preserving integrity in online social networks Page 1 92 COMMUNICATIONS OF THE ACM |
FEBRUARY 2022 | VOL. 65 | NO. 2 review articles THE GOAL OF online social networks is to …
FEBRUARY 2022 | VOL. 65 | NO. 2 review articles THE GOAL OF online social networks is to …
Do vision transformers see like convolutional neural networks?
Convolutional neural networks (CNNs) have so far been the de-facto model for visual data.
Recent work has shown that (Vision) Transformer models (ViT) can achieve comparable or …
Recent work has shown that (Vision) Transformer models (ViT) can achieve comparable or …
On the cross-lingual transferability of monolingual representations
State-of-the-art unsupervised multilingual models (eg, multilingual BERT) have been shown
to generalize in a zero-shot cross-lingual setting. This generalization ability has been …
to generalize in a zero-shot cross-lingual setting. This generalization ability has been …
[PDF][PDF] Unsupervised cross-lingual representation learning at scale
A Conneau - arxiv preprint arxiv:1911.02116, 2019 - fq.pkwyx.com
This paper shows that pretraining multilingual language models at scale leads to significant
performance gains for a wide range of cross-lingual transfer tasks. We train a Transformer …
performance gains for a wide range of cross-lingual transfer tasks. We train a Transformer …
InfoXLM: An information-theoretic framework for cross-lingual language model pre-training
In this work, we present an information-theoretic framework that formulates cross-lingual
language model pre-training as maximizing mutual information between multilingual-multi …
language model pre-training as maximizing mutual information between multilingual-multi …
Do wide and deep networks learn the same things? uncovering how neural network representations vary with width and depth
A key factor in the success of deep neural networks is the ability to scale models to improve
performance by varying the architecture depth and width. This simple property of neural …
performance by varying the architecture depth and width. This simple property of neural …
Probing pretrained language models for lexical semantics
The success of large pretrained language models (LMs) such as BERT and RoBERTa has
sparked interest in probing their representations, in order to unveil what types of knowledge …
sparked interest in probing their representations, in order to unveil what types of knowledge …
Are all languages created equal in multilingual BERT?
Multilingual BERT (mBERT) trained on 104 languages has shown surprisingly good cross-
lingual performance on several NLP tasks, even without explicit cross-lingual signals …
lingual performance on several NLP tasks, even without explicit cross-lingual signals …
From zero to hero: On the limitations of zero-shot cross-lingual transfer with multilingual transformers
Massively multilingual transformers pretrained with language modeling objectives (eg,
mBERT, XLM-R) have become a de facto default transfer paradigm for zero-shot cross …
mBERT, XLM-R) have become a de facto default transfer paradigm for zero-shot cross …
mslam: Massively multilingual joint pre-training for speech and text
We present mSLAM, a multilingual Speech and LAnguage Model that learns cross-lingual
cross-modal representations of speech and text by pre-training jointly on large amounts of …
cross-modal representations of speech and text by pre-training jointly on large amounts of …