Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Factcheck-bench: Fine-grained evaluation benchmark for automatic fact-checkers
The increased use of large language models (LLMs) across a variety of real-world
applications calls for mechanisms to verify the factual accuracy of their outputs. In this work …
applications calls for mechanisms to verify the factual accuracy of their outputs. In this work …
DEFAME: Dynamic Evidence-based FAct-checking with Multimodal Experts
The proliferation of disinformation presents a growing threat to societal trust and democracy,
necessitating robust and scalable Fact-Checking systems. In this work, we present Dynamic …
necessitating robust and scalable Fact-Checking systems. In this work, we present Dynamic …
OpenFactCheck: A Unified Framework for Factuality Evaluation of LLMs
The increased use of large language models (LLMs) across a variety of real-world
applications calls for automatic tools to check the factual accuracy of their outputs, as LLMs …
applications calls for automatic tools to check the factual accuracy of their outputs, as LLMs …
Loki: An open-source tool for fact verification
We introduce Loki, an open-source tool designed to address the growing problem of
misinformation. Loki adopts a human-centered approach, striking a balance between the …
misinformation. Loki adopts a human-centered approach, striking a balance between the …
LlamaLens: Specialized Multilingual LLM for Analyzing News and Social Media Content
Large Language Models (LLMs) have demonstrated remarkable success as general-
purpose task solvers across various fields, including NLP, healthcare, finance, and law …
purpose task solvers across various fields, including NLP, healthcare, finance, and law …
Chinese SafetyQA: A Safety Short-form Factuality Benchmark for Large Language Models
With the rapid advancement of Large Language Models (LLMs), significant safety concerns
have emerged. Fundamentally, the safety of large language models is closely linked to the …
have emerged. Fundamentally, the safety of large language models is closely linked to the …
[PDF][PDF] CORE: Robust Factual Precision Scoring with Informative Sub-Claim Identification
ZJJZN Weir, SEMWK Sanders… - arxiv preprint arxiv …, 2024 - katesanders9.github.io
Hallucinations—the generation of untrue claims—pose a challenge to the application of
large language models (LLMs)[12] thereby motivating the development of metrics to …
large language models (LLMs)[12] thereby motivating the development of metrics to …
Automatic Fact-Checking with Frame-Semantics
We propose a novel paradigm for automatic fact-checking that leverages frame semantics to
enhance the structured understanding of claims, addressing the challenges posed by …
enhance the structured understanding of claims, addressing the challenges posed by …
Decomposition Dilemmas: Does Claim Decomposition Boost or Burden Fact-Checking Performance?
Fact-checking pipelines increasingly adopt the Decompose-Then-Verify paradigm, where
texts are broken down into smaller claims for individual verification and subsequently …
texts are broken down into smaller claims for individual verification and subsequently …
Model Selection for HERITAGE-AI: Evaluating LLMs for Contextual Data Analysis of Maryland's Domestic Traffic Ads (1824–1864)
The HERITAGE-AI (Harnessing Enhanced Research and Instructional Technologies for
Archival Generative Exploration using AI), as part of the IMLS grant initiative, GenAI-4 …
Archival Generative Exploration using AI), as part of the IMLS grant initiative, GenAI-4 …