Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
On reinforcement learning and distribution matching for fine-tuning language models with no catastrophic forgetting
The availability of large pre-trained models is changing the landscape of Machine Learning
research and practice, moving from a" training from scratch" to a" fine-tuning''paradigm …
research and practice, moving from a" training from scratch" to a" fine-tuning''paradigm …
Oops i took a gradient: Scalable sampling for discrete distributions
We propose a general and scalable approximate sampling strategy for probabilistic models
with discrete variables. Our approach uses gradients of the likelihood function with respect …
with discrete variables. Our approach uses gradients of the likelihood function with respect …
On the calibration of pre-trained language models using mixup guided by area under the margin and saliency
A well-calibrated neural model produces confidence (probability outputs) closely
approximated by the expected accuracy. While prior studies have shown that mixup training …
approximated by the expected accuracy. While prior studies have shown that mixup training …
Building minimal and reusable causal state abstractions for reinforcement learning
Two desiderata of reinforcement learning (RL) algorithms are the ability to learn from
relatively little experience and the ability to learn policies that generalize to a range of …
relatively little experience and the ability to learn policies that generalize to a range of …
On the Calibration of Multilingual Question Answering LLMs
Multilingual pre-trained Large Language Models (LLMs) are incredibly effective at Question
Answering (QA), a core task in Natural Language Understanding, achieving high accuracies …
Answering (QA), a core task in Natural Language Understanding, achieving high accuracies …
Triple-Hybrid Energy-based Model Makes Better Calibrated Natural Language Understanding Models
Though pre-trained language models achieve notable success in many applications, it's
usually controversial for over-confident predictions. Specifically, the in-distribution (ID) …
usually controversial for over-confident predictions. Specifically, the in-distribution (ID) …
Energy-based models with applications to speech and language processing
Z Ou - Foundations and Trends® in Signal Processing, 2024 - nowpublishers.com
Abstract Energy-Based Models (EBMs) are an important class of probabilistic models, also
known as random fields and undirected graphical models. EBMs are un-normalized and …
known as random fields and undirected graphical models. EBMs are un-normalized and …
Consistent and efficient long document understanding
Q Zeng - 2023 - ideals.illinois.edu
In the age of information overload, people's information needs from long documents are
rapidly emerging, while people's patience for careful reading and reasoning is gradually …
rapidly emerging, while people's patience for careful reading and reasoning is gradually …
Improving NMT Models by Retrofitting Quality Estimators into Trainable Energy Loss
G Yoo, JY Lee - Proceedings of the 31st International Conference …, 2025 - aclanthology.org
Reinforcement learning has shown great promise in aligning language models with human
preferences in a variety of text generation tasks, including machine translation. For …
preferences in a variety of text generation tasks, including machine translation. For …
Consistent training via energy-based gflownets for modeling discrete joint distributions
Generative Flow Networks (GFlowNets) have demonstrated significant performance
improvements for generating diverse discrete objects $ x $ given a reward function $ R (x) …
improvements for generating diverse discrete objects $ x $ given a reward function $ R (x) …