Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Variance-reduced methods for machine learning
Stochastic optimization lies at the heart of machine learning, and its cornerstone is
stochastic gradient descent (SGD), a method introduced over 60 years ago. The last eight …
stochastic gradient descent (SGD), a method introduced over 60 years ago. The last eight …
Closing the gap: Tighter analysis of alternating stochastic gradient methods for bilevel problems
Stochastic nested optimization, including stochastic compositional, min-max, and bilevel
optimization, is gaining popularity in many machine learning applications. While the three …
optimization, is gaining popularity in many machine learning applications. While the three …
Decentralized gossip-based stochastic bilevel optimization over communication networks
Bilevel optimization have gained growing interests, with numerous applications found in
meta learning, minimax games, reinforcement learning, and nested composition …
meta learning, minimax games, reinforcement learning, and nested composition …
Multi-agent reinforcement learning via double averaging primal-dual optimization
Despite the success of single-agent reinforcement learning, multi-agent reinforcement
learning (MARL) remains challenging due to complex interactions between agents …
learning (MARL) remains challenging due to complex interactions between agents …
Stochastic optimization of areas under precision-recall curves with provable convergence
Areas under ROC (AUROC) and precision-recall curves (AUPRC) are common metrics for
evaluating classification performance for imbalanced problems. Compared with AUROC …
evaluating classification performance for imbalanced problems. Compared with AUROC …
Stochastic variance reduction methods for policy evaluation
Policy evaluation is concerned with estimating the value function that predicts long-term
values of states under a given policy. It is a crucial step in many reinforcement-learning …
values of states under a given policy. It is a crucial step in many reinforcement-learning …
A single timescale stochastic approximation method for nested stochastic optimization
We study constrained nested stochastic optimization problems in which the objective
function is a composition of two smooth functions whose exact values and derivatives are …
function is a composition of two smooth functions whose exact values and derivatives are …
A single-timescale method for stochastic bilevel optimization
Stochastic bilevel optimization generalizes the classic stochastic optimization from the
minimization of a single objective to the minimization of an objective function that depends …
minimization of a single objective to the minimization of an objective function that depends …
On the convergence and sample efficiency of variance-reduced policy gradient method
Policy gradient (PG) gives rise to a rich class of reinforcement learning (RL) methods.
Recently, there has been an emerging trend to augment the existing PG methods such as …
Recently, there has been an emerging trend to augment the existing PG methods such as …
A single-timescale method for stochastic bilevel optimization
Stochastic bilevel optimization generalizes the classic stochastic optimization from the
minimization of a single objective to the minimization of an objective function that depends …
minimization of a single objective to the minimization of an objective function that depends …