Contextual combinatorial bandits with probabilistically triggered arms
We study contextual combinatorial bandits with probabilistically triggered arms (C $^ 2$
MAB-T) under a variety of smoothness conditions that capture a wide range of applications …
MAB-T) under a variety of smoothness conditions that capture a wide range of applications …
Batch-size independent regret bounds for combinatorial semi-bandits with probabilistically triggered arms or independent arms
In this paper, we study the combinatorial semi-bandits (CMAB) and focus on reducing the
dependency of the batch-size $ K $ in the regret bound, where $ K $ is the total number of …
dependency of the batch-size $ K $ in the regret bound, where $ K $ is the total number of …
Bandit multi-linear DR-submodular maximization and its applications on adversarial submodular bandits
We investigate the online bandit learning of the monotone multi-linear DR-submodular
functions, designing the algorithm $\mathtt {BanditMLSM} $ that attains $ O (T^{2/3}\log T) …
functions, designing the algorithm $\mathtt {BanditMLSM} $ that attains $ O (T^{2/3}\log T) …
A survey on location-driven influence maximization
Influence Maximization (IM), which aims to select a set of users from a social network to
maximize the expected number of influenced users, is an evergreen hot research topic. Its …
maximize the expected number of influenced users, is an evergreen hot research topic. Its …
Dynamic influence maximization
B Peng - Advances in Neural Information Processing …, 2021 - proceedings.neurips.cc
We initiate a systematic study on {\em dynamic influence maximization}(DIM). In the DIM
problem, one maintains a seed set $ S $ of at most $ k $ nodes in a dynamically involving …
problem, one maintains a seed set $ S $ of at most $ k $ nodes in a dynamically involving …
An explore-then-commit algorithm for submodular maximization under full-bandit feedback
We investigate the problem of combinatorial multi-armed bandits with stochastic submodular
(in expectation) rewards and full-bandit feedback, where no extra information other than the …
(in expectation) rewards and full-bandit feedback, where no extra information other than the …
Combinatorial causal bandits
In combinatorial causal bandits (CCB), the learning agent chooses at most K variables in
each round to intervene, collects feedback from the observed variables, with the goal of …
each round to intervene, collects feedback from the observed variables, with the goal of …
Randomized greedy learning for non-monotone stochastic submodular maximization under full-bandit feedback
We investigate the problem of unconstrained combinatorial multi-armed bandits with full-
bandit feedback and stochastic rewards for submodular maximization. Previous works …
bandit feedback and stochastic rewards for submodular maximization. Previous works …
A community-aware framework for social influence maximization
We consider the problem of Influence Maximization (IM), the task of selecting seed nodes in
a social network such that the expected number of nodes influenced is maximized. We …
a social network such that the expected number of nodes influenced is maximized. We …
Combinatorial stochastic-greedy bandit
We propose a novel combinatorial stochastic-greedy bandit (SGB) algorithm for
combinatorial multi-armed bandit problems when no extra information other than the joint …
combinatorial multi-armed bandit problems when no extra information other than the joint …