Paraphrasing evades detectors of ai-generated text, but retrieval is an effective defense

K Krishna, Y Song, M Karpinska… - Advances in Neural …, 2023 - proceedings.neurips.cc
The rise in malicious usage of large language models, such as fake content creation and
academic plagiarism, has motivated the development of approaches that identify AI …

[HTML][HTML] Question answering models for human–machine interaction in the manufacturing industry

E Ruiz, MI Torres, A del Pozo - Computers in Industry, 2023 - Elsevier
This paper presents a question answering (QA) system that will enable workers from the
manufacturing industry to'hands-free'request information. This kind of systems, that are …

Paraphrase Types for Generation and Detection

JP Wahle, B Gipp, T Ruas - arxiv preprint arxiv:2310.14863, 2023 - arxiv.org
Current approaches in paraphrase generation and detection heavily rely on a single general
similarity score, ignoring the intricate linguistic properties of language. This paper introduces …

Measuring reliability of large language models through semantic consistency

H Raj, D Rosati, S Majumdar - arxiv preprint arxiv:2211.05853, 2022 - arxiv.org
While large pretrained language models (PLMs) demonstrate incredible fluency and
performance on many natural language tasks, recent work has shown that well-performing …

Predicting question-answering performance of large language models through semantic consistency

E Rabinovich, S Ackerman, O Raz, E Farchi… - arxiv preprint arxiv …, 2023 - arxiv.org
Semantic consistency of a language model is broadly defined as the model's ability to
produce semantically-equivalent outputs, given semantically-equivalent inputs. We address …

Comparative analysis of paraphrasing performance of ChatGPT, GPT‐3, and T5 language models using a new ChatGPT generated dataset: ParaGPT

M Kurt Pehlivanoğlu, RT Gobosho, MA Syakura… - Expert …, 2024 - Wiley Online Library
Paraphrase generation is a fundamental natural language processing (NLP) task that refers
to the process of generating a well‐formed and coherent output sentence that exhibits both …

Mcpg: A flexible multi-level controllable framework for unsupervised paraphrase generation

Y Chen, H Jiang, L Liu, R Wang, S Shi… - Findings of the …, 2022 - aclanthology.org
We present MCPG: a simple and effectiveapproach for controllable unsupervised
paraphrase generation, which is also flexible toadapt to specific domains without extra …

Explicit syntactic guidance for neural text generation

Y Li, L Cui, J Yan, Y Yin, W Bi, S Shi… - arxiv preprint arxiv …, 2023 - arxiv.org
Most existing text generation models follow the sequence-to-sequence paradigm.
Generative Grammar suggests that humans generate natural language texts by learning …

HU at SemEval-2024 Task 8A: Can Contrastive Learning Learn Embeddings to Detect Machine-Generated Text?

SR Dipta, S Shahriar - arxiv preprint arxiv:2402.11815, 2024 - arxiv.org
This paper describes our system developed for SemEval-2024 Task 8," Multigenerator,
Multidomain, and Multilingual Black-Box Machine-Generated Text Detection." Machine …

Lexical generalization improves with larger models and longer training

E Bandel, Y Goldberg, Y Elazar - arxiv preprint arxiv:2210.12673, 2022 - arxiv.org
While fine-tuned language models perform well on many tasks, they were also shown to rely
on superficial surface features such as lexical overlap. Excessive utilization of such …