Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
A survey of confidence estimation and calibration in large language models
Large language models (LLMs) have demonstrated remarkable capabilities across a wide
range of tasks in various domains. Despite their impressive performance, they can be …
range of tasks in various domains. Despite their impressive performance, they can be …
Harnessing the power of llms in practice: A survey on chatgpt and beyond
This article presents a comprehensive and practical guide for practitioners and end-users
working with Large Language Models (LLMs) in their downstream Natural Language …
working with Large Language Models (LLMs) in their downstream Natural Language …
Trustworthy llms: a survey and guideline for evaluating large language models' alignment
Ensuring alignment, which refers to making models behave in accordance with human
intentions [1, 2], has become a critical task before deploying large language models (LLMs) …
intentions [1, 2], has become a critical task before deploying large language models (LLMs) …
Active prompting with chain-of-thought for large language models
The increasing scale of large language models (LLMs) brings emergent abilities to various
complex tasks requiring reasoning, such as arithmetic and commonsense reasoning. It is …
complex tasks requiring reasoning, such as arithmetic and commonsense reasoning. It is …
The'Problem'of Human Label Variation: On Ground Truth in Data, Modeling and Evaluation
B Plank - arxiv preprint arxiv:2211.02570, 2022 - arxiv.org
Human variation in labeling is often considered noise. Annotation projects for machine
learning (ML) aim at minimizing human label variation, with the assumption to maximize …
learning (ML) aim at minimizing human label variation, with the assumption to maximize …
Teaching models to express their uncertainty in words
We show that a GPT-3 model can learn to express uncertainty about its own answers in
natural language--without use of model logits. When given a question, the model generates …
natural language--without use of model logits. When given a question, the model generates …
A survey of mix-based data augmentation: Taxonomy, methods, applications, and explainability
Data augmentation (DA) is indispensable in modern machine learning and deep neural
networks. The basic idea of DA is to construct new training data to improve the model's …
networks. The basic idea of DA is to construct new training data to improve the model's …
Knowledgeable prompt-tuning: Incorporating knowledge into prompt verbalizer for text classification
Tuning pre-trained language models (PLMs) with task-specific prompts has been a
promising approach for text classification. Particularly, previous studies suggest that prompt …
promising approach for text classification. Particularly, previous studies suggest that prompt …
Navigating the grey area: How expressions of uncertainty and overconfidence affect language models
The increased deployment of LMs for real-world tasks involving knowledge and facts makes
it important to understand model epistemology: what LMs think they know, and how their …
it important to understand model epistemology: what LMs think they know, and how their …
How Can We Know When Language Models Know? On the Calibration of Language Models for Question Answering
Recent works have shown that language models (LM) capture different types of knowledge
regarding facts or common sense. However, because no model is perfect, they still fail to …
regarding facts or common sense. However, because no model is perfect, they still fail to …