Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
[HTML][HTML] A machine learning approach for air quality prediction: Model regularization and optimization
In this paper, we tackle air quality forecasting by using machine learning approaches to
predict the hourly concentration of air pollutants (eg, ozone, particle matter (PM 2.5) and …
predict the hourly concentration of air pollutants (eg, ozone, particle matter (PM 2.5) and …
Adjusting learning depth in nonnegative latent factorization of tensors for accurately modeling temporal patterns in dynamic QoS data
A nonnegative latent factorization of tensors (NLFT) model precisely represents the temporal
patterns hidden in multichannel data emerging from various applications. It often adopts a …
patterns hidden in multichannel data emerging from various applications. It often adopts a …
Online primal-dual mirror descent under stochastic constraints
We consider online convex optimization with stochastic constraints where the objective
functions are arbitrarily time-varying and the constraint functions are independent and …
functions are arbitrarily time-varying and the constraint functions are independent and …
Tfpnp: Tuning-free plug-and-play proximal algorithms with applications to inverse imaging problems
Plug-and-Play (PnP) is a non-convex optimization framework that combines proximal
algorithms, for example, the alternating direction method of multipliers (ADMM), with …
algorithms, for example, the alternating direction method of multipliers (ADMM), with …
Nonconvex zeroth-order stochastic admm methods with lower function query complexity
F Huang, S Gao, J Pei, H Huang - IEEE Transactions on Pattern …, 2024 - ieeexplore.ieee.org
Zeroth-order (aka, derivative-free) methods are a class of effective optimization methods for
solving complex machine learning problems, where gradients of the objective functions are …
solving complex machine learning problems, where gradients of the objective functions are …
[HTML][HTML] FedADMM-InSa: An inexact and self-adaptive ADMM for federated learning
Y Song, Z Wang, E Zuazua - Neural Networks, 2025 - Elsevier
Federated learning (FL) is a promising framework for learning from distributed data while
maintaining privacy. The development of efficient FL algorithms encounters various …
maintaining privacy. The development of efficient FL algorithms encounters various …
Robust subspace tracking with missing data and outliers: Novel algorithm with convergence guarantee
In this article, we propose a novel algorithm, namely PETRELS-ADMM, to deal with
subspace tracking in the presence of outliers and missing data. The proposed approach …
subspace tracking in the presence of outliers and missing data. The proposed approach …
Universal stagewise learning for non-convex problems with convergence on averaged solutions
Although stochastic gradient descent (SGD) method and its variants (eg, stochastic
momentum methods, AdaGrad) are the choice of algorithms for solving non-convex …
momentum methods, AdaGrad) are the choice of algorithms for solving non-convex …
Accelerated variance reduction stochastic ADMM for large-scale machine learning
Recently, many stochastic variance reduced alternating direction methods of multipliers
(ADMMs)(eg, SAG-ADMM and SVRG-ADMM) have made exciting progress such as linear …
(ADMMs)(eg, SAG-ADMM and SVRG-ADMM) have made exciting progress such as linear …
Faster stochastic alternating direction method of multipliers for nonconvex optimization
In this paper, we propose a faster stochastic alternating direction method of multipliers
(ADMM) for nonconvex optimization by using a new stochastic path-integrated differential …
(ADMM) for nonconvex optimization by using a new stochastic path-integrated differential …