Introduction to multi-armed bandits
A Slivkins - Foundations and Trends® in Machine Learning, 2019 - nowpublishers.com
Multi-armed bandits a simple but very powerful framework for algorithms that make
decisions over time under uncertainty. An enormous body of work has accumulated over the …
decisions over time under uncertainty. An enormous body of work has accumulated over the …
Introduction to online convex optimization
E Hazan - Foundations and Trends® in Optimization, 2016 - nowpublishers.com
This monograph portrays optimization as a process. In many practical applications the
environment is so complex that it is infeasible to lay out a comprehensive theoretical model …
environment is so complex that it is infeasible to lay out a comprehensive theoretical model …
Regret analysis of stochastic and nonstochastic multi-armed bandit problems
S Bubeck, N Cesa-Bianchi - Foundations and Trends® in …, 2012 - nowpublishers.com
Multi-armed bandit problems are the most basic examples of sequential decision problems
with an exploration-exploitation trade-off. This is the balance between staying with the option …
with an exploration-exploitation trade-off. This is the balance between staying with the option …
Online learning and online convex optimization
S Shalev-Shwartz - Foundations and Trends® in Machine …, 2012 - nowpublishers.com
Online learning is a well established learning paradigm which has both theoretical and
practical appeals. The goal of online learning is to make a sequence of accurate predictions …
practical appeals. The goal of online learning is to make a sequence of accurate predictions …
Gaussian process optimization in the bandit setting: No regret and experimental design
Many applications require optimizing an unknown, noisy function that is expensive to
evaluate. We formalize this task as a multi-armed bandit problem, where the payoff function …
evaluate. We formalize this task as a multi-armed bandit problem, where the payoff function …
An information-theoretic analysis of thompson sampling
We provide an information-theoretic analysis of Thompson sampling that applies across a
broad range of online optimization problems in which a decision-maker must learn from …
broad range of online optimization problems in which a decision-maker must learn from …
Contextual gaussian process bandit optimization
How should we design experiments to maximize performance of a complex system, taking
into account uncontrollable environmental conditions? How should we select relevant …
into account uncontrollable environmental conditions? How should we select relevant …
[PDF][PDF] Stochastic Linear Optimization under Bandit Feedback.
In the classical stochastic k-armed bandit problem, in each of a sequence of T rounds, a
decision maker chooses one of k arms and incurs a cost chosen from an unknown …
decision maker chooses one of k arms and incurs a cost chosen from an unknown …
Off-policy evaluation for slate recommendation
A Swaminathan, A Krishnamurthy… - Advances in …, 2017 - proceedings.neurips.cc
This paper studies the evaluation of policies that recommend an ordered set of items (eg, a
ranking) based on some context---a common scenario in web search, ads, and …
ranking) based on some context---a common scenario in web search, ads, and …
Contextual bandits with similarity information
A Slivkins - Proceedings of the 24th annual Conference On …, 2011 - proceedings.mlr.press
In a multi-armed bandit (MAB) problem, an online algorithm makes a sequence of choices.
In each round it chooses from a time-invariant set of alternatives and receives the payoff …
In each round it chooses from a time-invariant set of alternatives and receives the payoff …