Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Weak-to-strong generalization: Eliciting strong capabilities with weak supervision
Widely used alignment techniques, such as reinforcement learning from human feedback
(RLHF), rely on the ability of humans to supervise model behavior-for example, to evaluate …
(RLHF), rely on the ability of humans to supervise model behavior-for example, to evaluate …
Freematch: Self-adaptive thresholding for semi-supervised learning
Pseudo labeling and consistency regularization approaches with confidence-based
thresholding have made great progress in semi-supervised learning (SSL). In this paper, we …
thresholding have made great progress in semi-supervised learning (SSL). In this paper, we …
Causal inference in natural language processing: Estimation, prediction, interpretation and beyond
A fundamental goal of scientific research is to learn about causal relationships. However,
despite its critical role in the life and social sciences, causality has not had the same …
despite its critical role in the life and social sciences, causality has not had the same …
[HTML][HTML] Self-training: A survey
Self-training methods have gained significant attention in recent years due to their
effectiveness in leveraging small labeled datasets and large unlabeled observations for …
effectiveness in leveraging small labeled datasets and large unlabeled observations for …
Cycle self-training for domain adaptation
Mainstream approaches for unsupervised domain adaptation (UDA) learn domain-invariant
representations to narrow the domain shift, which are empirically effective but theoretically …
representations to narrow the domain shift, which are empirically effective but theoretically …
Test time adaptation via conjugate pseudo-labels
Test-time adaptation (TTA) refers to adapting neural networks to distribution shifts,
specifically with just access to unlabeled test samples from the new domain at test-time …
specifically with just access to unlabeled test samples from the new domain at test-time …
Theoretical analysis of self-training with deep networks on unlabeled data
Self-training algorithms, which train a model to fit pseudolabels predicted by another
previously-learned model, have been very successful for learning with unlabeled data using …
previously-learned model, have been very successful for learning with unlabeled data using …
Theoretical analysis of weak-to-strong generalization
Strong student models can learn from weaker teachers: when trained on the predictions of a
weaker model, a strong pretrained student can learn to correct the weak model's errors and …
weaker model, a strong pretrained student can learn to correct the weak model's errors and …
Robust learning with progressive data expansion against spurious correlation
While deep learning models have shown remarkable performance in various tasks, they are
susceptible to learning non-generalizable _spurious features_ rather than the core features …
susceptible to learning non-generalizable _spurious features_ rather than the core features …
Masktune: Mitigating spurious correlations by forcing to explore
A fundamental challenge of over-parameterized deep learning models is learning
meaningful data representations that yield good performance on a downstream task without …
meaningful data representations that yield good performance on a downstream task without …