BERT: a review of applications in natural language processing and understanding
MV Koroteev - arxiv preprint arxiv:2103.11943, 2021 - arxiv.org
In this review, we describe the application of one of the most popular deep learning-based
language models-BERT. The paper describes the mechanism of operation of this model, the …
language models-BERT. The paper describes the mechanism of operation of this model, the …
Bertscore: Evaluating text generation with bert
We propose BERTScore, an automatic evaluation metric for text generation. Analogously to
common metrics, BERTScore computes a similarity score for each token in the candidate …
common metrics, BERTScore computes a similarity score for each token in the candidate …
YiSi-a unified semantic MT quality evaluation and estimation metric for languages with different levels of available resources
C Lo - Proceedings of the Fourth Conference on Machine …, 2019 - aclanthology.org
We present YiSi, a unified automatic semantic machine translation quality evaluation and
estimation metric for languages with different levels of available resources. Underneath the …
estimation metric for languages with different levels of available resources. Underneath the …
Findings of the WMT 2018 shared task on parallel corpus filtering
We posed the shared task of assigning sentence-level quality scores for a very noisy corpus
of sentence pairs crawled from the web, with the goal of sub-selecting 1% and 10% of high …
of sentence pairs crawled from the web, with the goal of sub-selecting 1% and 10% of high …
Automatic text evaluation through the lens of Wasserstein barycenters
A new metric\texttt {BaryScore} to evaluate text generation based on deep contextualized
embeddings eg, BERT, Roberta, ELMo) is introduced. This metric is motivated by a new …
embeddings eg, BERT, Roberta, ELMo) is introduced. This metric is motivated by a new …
Results of the WMT18 metrics shared task: Both characters and embeddings achieve good performance
This paper presents the results of the WMT18 Metrics Shared Task. We asked participants of
this task to score the outputs of the MT systems involved in the WMT18 News Translation …
this task to score the outputs of the MT systems involved in the WMT18 News Translation …
Parallel corpus filtering via pre-trained language models
Web-crawled data provides a good source of parallel corpora for training machine
translation models. It is automatically obtained, but extremely noisy, and recent work shows …
translation models. It is automatically obtained, but extremely noisy, and recent work shows …
Towards reference-free text simplification evaluation with a BERT siamese network architecture
Text simplification (TS) aims to modify sentences to make their both content and structure
easier to understand. Traditional n-gram matching-based TS evaluation metrics heavily rely …
easier to understand. Traditional n-gram matching-based TS evaluation metrics heavily rely …
MSLC24 submissions to the general machine translation task
Abstract The MSLC (Metric Score Landscape Challenge) submissions for English-German,
English-Spanish, and Japanese-Chinese are constrained systems built using Transformer …
English-Spanish, and Japanese-Chinese are constrained systems built using Transformer …
Fully unsupervised crosslingual semantic textual similarity metric based on BERT for identifying parallel data
We present a fully unsupervised crosslingual semantic textual similarity (STS) metric, based
on contextual embeddings extracted from BERT–Bidirectional Encoder Representations …
on contextual embeddings extracted from BERT–Bidirectional Encoder Representations …