Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Survey of hallucination in natural language generation
Natural Language Generation (NLG) has improved exponentially in recent years thanks to
the development of sequence-to-sequence deep learning technologies such as Transformer …
the development of sequence-to-sequence deep learning technologies such as Transformer …
Recent advances in deep learning based dialogue systems: A systematic survey
Dialogue systems are a popular natural language processing (NLP) task as it is promising in
real-life applications. It is also a complicated task since many NLP tasks deserving study are …
real-life applications. It is also a complicated task since many NLP tasks deserving study are …
Memory-based model editing at scale
Even the largest neural networks make errors, and once-correct predictions can become
invalid as the world changes. Model editors make local updates to the behavior of base (pre …
invalid as the world changes. Model editors make local updates to the behavior of base (pre …
Red teaming language models with language models
Language Models (LMs) often cannot be deployed because of their potential to harm users
in hard-to-predict ways. Prior work identifies harmful behaviors before deployment by using …
in hard-to-predict ways. Prior work identifies harmful behaviors before deployment by using …
Orpo: Monolithic preference optimization without reference model
While recent preference alignment algorithms for language models have demonstrated
promising results, supervised fine-tuning (SFT) remains imperative for achieving successful …
promising results, supervised fine-tuning (SFT) remains imperative for achieving successful …
Recipes for building an open-domain chatbot
Building open-domain chatbots is a challenging area for machine learning research. While
prior work has shown that scaling neural models in the number of parameters and the size of …
prior work has shown that scaling neural models in the number of parameters and the size of …
Quark: Controllable text generation with reinforced unlearning
Large-scale language models often learn behaviors that are misaligned with user
expectations. Generated text may contain offensive or toxic language, contain significant …
expectations. Generated text may contain offensive or toxic language, contain significant …
FEQA: A question answering evaluation framework for faithfulness assessment in abstractive summarization
Neural abstractive summarization models are prone to generate content inconsistent with
the source document, ie unfaithful. Existing automatic metrics do not capture such mistakes …
the source document, ie unfaithful. Existing automatic metrics do not capture such mistakes …
Locally typical sampling
Today's probabilistic language generators fall short when it comes to producing coherent
and fluent text despite the fact that the underlying models perform well under standard …
and fluent text despite the fact that the underlying models perform well under standard …
How much do language models copy from their training data? evaluating linguistic novelty in text generation using raven
Current language models can generate high-quality text. Are they simply copying text they
have seen before, or have they learned generalizable linguistic abstractions? To tease apart …
have seen before, or have they learned generalizable linguistic abstractions? To tease apart …