A modern introduction to online learning
F Orabona - arxiv preprint arxiv:1912.13213, 2019 - arxiv.org
In this monograph, I introduce the basic concepts of Online Learning through a modern view
of Online Convex Optimization. Here, online learning refers to the framework of regret …
of Online Convex Optimization. Here, online learning refers to the framework of regret …
Data-dependent bounds for online portfolio selection without Lipschitzness and smoothness
This work introduces the first small-loss and gradual-variation regret bounds for online
portfolio selection, marking the first instances of data-dependent bounds for online convex …
portfolio selection, marking the first instances of data-dependent bounds for online convex …
Effective online portfolio selection for the long-short market using mirror gradient descent
W Zhang, X Li, Y Chen, N Ye… - IEEE Signal Processing …, 2023 - ieeexplore.ieee.org
Online portfolio selection has been actively studied to maximise overall returns by selecting
the optimal portfolio weights using online algorithms. However, most work has focused on …
the optimal portfolio weights using online algorithms. However, most work has focused on …
A Bregman proximal perspective on classical and quantum Blahut-Arimoto algorithms
The Blahut-Arimoto algorithm is a well-known method to compute classical channel
capacities and rate-distortion functions. Recent works have extended this algorithm to …
capacities and rate-distortion functions. Recent works have extended this algorithm to …
Fast Minimization of Expected Logarithmic Loss via Stochastic Dual Averaging
Consider the problem of minimizing an expected logarithmic loss over either the probability
simplex or the set of quantum density matrices. This problem includes tasks such as solving …
simplex or the set of quantum density matrices. This problem includes tasks such as solving …
Fast asymptotically optimal algorithms for non-parametric stochastic bandits
We consider the problem of regret minimization in non-parametric stochastic bandits. When
the rewards are known to be bounded from above, there exists asymptotically optimal …
the rewards are known to be bounded from above, there exists asymptotically optimal …
Online Learning Quantum States with the Logarithmic Loss via VB-FTRL
WF Tseng, KC Chen, ZH **ao, YH Li - arxiv preprint arxiv:2311.04237, 2023 - arxiv.org
Online learning quantum states with the logarithmic loss (LL-OLQS) is a quantum
generalization of online portfolio selection, a classic open problem in the field of online …
generalization of online portfolio selection, a classic open problem in the field of online …
Faster Stochastic First-Order Method for Maximum-Likelihood Quantum State Tomography
In maximum-likelihood quantum state tomography, both the sample size and dimension
grow exponentially with the number of qubits. It is therefore desirable to develop a stochastic …
grow exponentially with the number of qubits. It is therefore desirable to develop a stochastic …
Generalized Exponentiated Gradient Algorithms and Their Application to On-Line Portfolio Selection
This paper introduces a novel family of generalized exponentiated gradient (EG) updates
derived from an Alpha-Beta divergence regularization function. Collectively referred to as …
derived from an Alpha-Beta divergence regularization function. Collectively referred to as …
Data-Dependent Regret Bounds for Adversarial Multi-Armed Bandits and Online Portfolio Selection
SR Putta - 2024 - search.proquest.com
This dissertation studies\textit {Data-Dependent} regret bounds for two online learning
problems. As opposed to worst-case regret bounds, data-dependent bounds are able to …
problems. As opposed to worst-case regret bounds, data-dependent bounds are able to …