Derivative-free reinforcement learning: A review

H Qian, Y Yu - Frontiers of Computer Science, 2021 - Springer
Reinforcement learning is about learning agent models that make the best sequential
decisions in unknown environments. In an unknown environment, the agent needs to …

Harvnet: resource-optimized operation of multi-exit deep neural networks on energy harvesting devices

S Jeon, Y Choi, Y Cho, H Cha - Proceedings of the 21st Annual …, 2023 - dl.acm.org
Optimizing deep neural networks (DNNs) running on resource-constrained devices, such as
energy harvesting sensor devices, poses unique challenges due to the limited memory and …

Zeroth-order algorithms for stochastic distributed nonconvex optimization

X Yi, S Zhang, T Yang, KH Johansson - Automatica, 2022 - Elsevier
In this paper, we consider a stochastic distributed nonconvex optimization problem with the
cost function being distributed over n agents having access only to zeroth-order (ZO) …

L2ight: Enabling on-chip learning for optical neural networks via efficient in-situ subspace optimization

J Gu, H Zhu, C Feng, Z Jiang… - Advances in Neural …, 2021 - proceedings.neurips.cc
Silicon-photonics-based optical neural network (ONN) is a promising hardware platform that
could represent a paradigm shift in efficient AI with its CMOS-compatibility, flexibility, ultra …

Efficient on-chip learning for optical neural networks through power-aware sparse zeroth-order optimization

J Gu, C Feng, Z Zhao, Z Ying, RT Chen… - Proceedings of the AAAI …, 2021 - ojs.aaai.org
Optical neural networks (ONNs) have demonstrated record-breaking potential in high-
performance neuromorphic computing due to their ultra-high execution speed and low …

Tensor-compressed back-propagation-free training for (physics-informed) neural networks

Y Zhao, X Yu, Z Chen, Z Liu, S Liu, Z Zhang - arxiv preprint arxiv …, 2023 - arxiv.org
Backward propagation (BP) is widely used to compute the gradients in neural network
training. However, it is hard to implement BP on edge devices due to the lack of hardware …

Curvature-aware derivative-free optimization

B Kim, HQ Cai, D McKenzie, W Yin - arxiv preprint arxiv:2109.13391, 2021 - arxiv.org
The paper discusses derivative-free optimization (DFO), which involves minimizing a
function without access to gradients or directional derivatives, only function evaluations …

Minibatch stochastic three points method for unconstrained smooth minimization

S Boucherouite, G Malinovsky, P Richtárik… - Proceedings of the …, 2024 - ojs.aaai.org
We present a new zero-order optimization method called Minibatch Stochastic Three Points
(MiSTP), specifically designed to solve stochastic unconstrained minimization problems …

Derivative-Free Optimization with Transformed Objective Functions (DFOTO) and the Algorithm Based on the Least Frobenius Norm Updating Quadratic Model

P **e, YX Yuan - arxiv preprint arxiv:2302.12021, 2023 - arxiv.org
Derivative-free optimization problems are optimization problems where derivative
information is unavailable. The least Frobenius norm updating quadratic interpolation model …

Randomized directional search for nonconvex optimization

Y Zhang, W **ng - arxiv preprint arxiv:2501.00469, 2024 - arxiv.org
Direct search methods are a class of popular global optimization algorithms for general
nonconvex programs. In this paper, we propose a randomized directional search algorithm …