Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
A survey on data selection for language models
A major factor in the recent success of large language models is the use of enormous and
ever-growing text datasets for unsupervised pre-training. However, naively training a model …
ever-growing text datasets for unsupervised pre-training. However, naively training a model …
Robust learning with progressive data expansion against spurious correlation
While deep learning models have shown remarkable performance in various tasks, they are
susceptible to learning non-generalizable _spurious features_ rather than the core features …
susceptible to learning non-generalizable _spurious features_ rather than the core features …
The limits and potentials of local sgd for distributed heterogeneous learning with intermittent communication
Local SGD is a popular optimization method in distributed learning, often outperforming mini-
batch SGD. Despite this practical success, proving the efficiency of local SGD has been …
batch SGD. Despite this practical success, proving the efficiency of local SGD has been …
Optimal multi-distribution learning
Abstract Multi-distribution learning (MDL), which seeks to learn a shared model that
minimizes the worst-case risk across $ k $ distinct data distributions, has emerged as a …
minimizes the worst-case risk across $ k $ distinct data distributions, has emerged as a …
A unifying perspective on multi-calibration: Game dynamics for multi-objective learning
We provide a unifying framework for the design and analysis of multi-calibrated predictors.
By placing the multi-calibration problem in the general setting of multi-objective learning …
By placing the multi-calibration problem in the general setting of multi-objective learning …
Stochastic approximation approaches to group distributionally robust optimization
This paper investigates group distributionally robust optimization (GDRO), with the purpose
to learn a model that performs well over $ m $ different distributions. First, we formulate …
to learn a model that performs well over $ m $ different distributions. First, we formulate …
The sample complexity of multi-distribution learning
B Peng - The Thirty Seventh Annual Conference on Learning …, 2024 - proceedings.mlr.press
Multi-distribution learning generalizes the classic PAC learning to handle data coming from
multiple distributions. Given a set of $ k $ data distributions and a hypothesis class of VC …
multiple distributions. Given a set of $ k $ data distributions and a hypothesis class of VC …
Why does throwing away data improve worst-group error?
When facing data with imbalanced classes or groups, practitioners follow an intriguing
strategy to achieve best results. They throw away examples until the classes or groups are …
strategy to achieve best results. They throw away examples until the classes or groups are …
Open problem: The sample complexity of multi-distribution learning for VC classes
Multi-distribution learning is a natural generalization of PAC learning to settings with multiple
data distributions. There remains a significant gap between the known upper and lower …
data distributions. There remains a significant gap between the known upper and lower …
Derandomizing Multi-Distribution Learning
Multi-distribution or collaborative learning involves learning a single predictor that works
well across multiple data distributions, using samples from each during training. Recent …
well across multiple data distributions, using samples from each during training. Recent …