Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Bringing order into the realm of Transformer-based language models for artificial intelligence and law
Transformer-based language models (TLMs) have widely been recognized to be a cutting-
edge technology for the successful development of deep-learning-based solutions to …
edge technology for the successful development of deep-learning-based solutions to …
[PDF][PDF] Retrieval-augmented generation for large language models: A survey
Y Gao, Y **ong, X Gao, K Jia, J Pan, Y Bi… - arxiv preprint arxiv …, 2023 - simg.baai.ac.cn
Large language models (LLMs) demonstrate powerful capabilities, but they still face
challenges in practical applications, such as hallucinations, slow knowledge updates, and …
challenges in practical applications, such as hallucinations, slow knowledge updates, and …
Lift yourself up: Retrieval-augmented text generation with self-memory
With direct access to human-written reference as memory, retrieval-augmented generation
has achieved much progress in a wide range of text generation tasks. Since better memory …
has achieved much progress in a wide range of text generation tasks. Since better memory …
Findings of the 2021 conference on machine translation (WMT21)
F Akhbardeh, A Arkhangorodsky, M Biesialska… - Proceedings of the sixth …, 2021 - cris.fbk.eu
This paper presents the results of the news translation task, the multilingual low-resource
translation for Indo-European languages, the triangular translation task, and the automatic …
translation for Indo-European languages, the triangular translation task, and the automatic …
Lexically constrained decoding for sequence generation using grid beam search
We present Grid Beam Search (GBS), an algorithm which extends beam search to allow the
inclusion of pre-specified lexical constraints. The algorithm can be used with any model that …
inclusion of pre-specified lexical constraints. The algorithm can be used with any model that …
Understanding and detecting hallucinations in neural machine translation via model introspection
Neural sequence generation models are known to “hallucinate”, by producing outputs that
are unrelated to the source text. These hallucinations are potentially harmful, yet it remains …
are unrelated to the source text. These hallucinations are potentially harmful, yet it remains …
The united nations parallel corpus v1. 0
M Ziemski, M Junczys-Dowmunt… - Proceedings of the …, 2016 - aclanthology.org
This paper describes the creation process and statistics of the official United Nations Parallel
Corpus, the first parallel corpus composed from United Nations documents published by the …
Corpus, the first parallel corpus composed from United Nations documents published by the …
Improving neural machine translation models with monolingual data
Neural Machine Translation (NMT) has obtained state-of-the art performance for several
language pairs, while only using parallel data for training. Target-side monolingual data …
language pairs, while only using parallel data for training. Target-side monolingual data …
[PDF][PDF] PPDB: The paraphrase database
We present the 1.0 release of our paraphrase database, PPDB. Its English portion, PPDB:
Eng, contains over 220 million paraphrase pairs, consisting of 73 million phrasal and 8 …
Eng, contains over 220 million paraphrase pairs, consisting of 73 million phrasal and 8 …
Automatic language identification in texts: A survey
Language identification (" LI") is the problem of determining the natural language that a
document or part thereof is written in. Automatic LI has been extensively researched for over …
document or part thereof is written in. Automatic LI has been extensively researched for over …