A survey on causal inference
Causal inference is a critical research topic across many domains, such as statistics,
computer science, education, public policy, and economics, for decades. Nowadays …
computer science, education, public policy, and economics, for decades. Nowadays …
AI and personalization
This chapter reviews the recent developments at the intersection of personalization and AI in
marketing and related fields. We provide a formal definition of personalized policy and …
marketing and related fields. We provide a formal definition of personalized policy and …
Introduction to multi-armed bandits
A Slivkins - Foundations and Trends® in Machine Learning, 2019 - nowpublishers.com
Multi-armed bandits a simple but very powerful framework for algorithms that make
decisions over time under uncertainty. An enormous body of work has accumulated over the …
decisions over time under uncertainty. An enormous body of work has accumulated over the …
Beyond ucb: Optimal and efficient contextual bandits with regression oracles
A fundamental challenge in contextual bandits is to develop flexible, general-purpose
algorithms with computational requirements no worse than classical supervised learning …
algorithms with computational requirements no worse than classical supervised learning …
Balanced linear contextual bandits
Contextual bandit algorithms are sensitive to the estimation method of the outcome model as
well as the exploration method used, particularly in the presence of rich heterogeneity or …
well as the exploration method used, particularly in the presence of rich heterogeneity or …
Estimation considerations in contextual bandits
Contextual bandit algorithms are sensitive to the estimation method of the outcome model as
well as the exploration method used, particularly in the presence of rich heterogeneity or …
well as the exploration method used, particularly in the presence of rich heterogeneity or …
Bypassing the monster: A faster and simpler optimal algorithm for contextual bandits under realizability
We consider the general (stochastic) contextual bandit problem under the realizability
assumption, that is, the expected reward, as a function of contexts and actions, belongs to a …
assumption, that is, the expected reward, as a function of contexts and actions, belongs to a …
Mostly exploration-free algorithms for contextual bandits
The contextual bandit literature has traditionally focused on algorithms that address the
exploration–exploitation tradeoff. In particular, greedy algorithms that exploit current …
exploration–exploitation tradeoff. In particular, greedy algorithms that exploit current …
Instance-dependent complexity of contextual bandits and reinforcement learning: A disagreement-based perspective
In the classical multi-armed bandit problem, instance-dependent algorithms attain improved
performance on" easy" problems with a gap between the best and second-best arm. Are …
performance on" easy" problems with a gap between the best and second-best arm. Are …
Contextual bandits with large action spaces: Made practical
A central problem in sequential decision making is to develop algorithms that are practical
and computationally efficient, yet support the use of flexible, general-purpose models …
and computationally efficient, yet support the use of flexible, general-purpose models …