\( \newcommand{\argmax}{\operatorname{arg\,max}\limits} \) \( \newcommand{\P}[1]{\mathbf{P} \left\{ #1\right\}} \) \( \newcommand{\E}{\mathbf{E}} \) \( \newcommand{\R}{\mathbb{R}} \) \( \newcommand{\set}[1]{\left\{#1\right\}} \) \( \newcommand{\floor}[1]{\left \lfloor {#1} \right\rfloor} \) \( \newcommand{\ceil}[1]{\left \lceil {#1} \right\rceil} \) \( \newcommand{\logp}{\log_{+}\!} \) \( \let\epsilon\varepsilon\)

About me

I was an assistant professor at the Department of Computer Science at Indiana University in Bloomington. Currently I am visiting the Australian National University until late March before joining Google DeepMind in London.

My work is mostly on the theory of sequential decision making, including multi-armed bandits and reinforcement learning.

You can contact me at firstname.lastname@gmail.com.


  • February 2017: The paper with Csaba on the asymptotic regret for linear bandits was accepted at AISTATS.
  • October 2016: Csaba and I have just released a new paper on finite-armed linear bandits to the arxiv. The main result is a characterisation of the optimal achievable asymptotic regret in the sense of Lai & Robbins, including an (extraordinarily naive) algorithm that attains the optimal rate. An interesting corollary of the analysis is that strategies based on optimism or Thompson sampling can be arbitrarily far from optimal, even in very simple cases.
  • September 2016: Csaba Szepesvari and I are writing a series of blog posts on bandit algorithms. The posts coincide with a course on bandits this semester at the UofA, and the same material will be used in a course at IU next semester. Eventually we plan to use the material (and the aid of the crowd) to write a book. The first post is up and you can expect about one new post per week.
  • September 2016: Goodbye UoA. Hello IU! I am in Bloomington now.
  • August 2016: I will be going to NIPS (four papers accepted).
  • June 2016: Finnian Lattimore (who happens to be my sister), Mark Reid and I have just released a new bandit paper to the arxiv. The idea is to show that causal information can sometimes be exploited when minimising the simple regret (pure exploration).
  • June 2016: The paper with Jan, Marcus, Laurent and I won best student paper at UAI. Congratulations Jan!
  • June 2016: Bandit strategies based on distinct phases of exploration followed by exploitation are pretty natural. Aurélien Garivier, Emilie Kaufmann and I show that such approaches are doomed to sub-optimality in the illustrative two-armed Gaussian bandit with known horizon.
  • May 2016: There are missing lower bounds for adversarial finite-armed bandits. Sébastien Gerchinovitz and I have now added some (preprint). Specifically: high probability bounds, first-order bounds (in terms of the loss of the best arm) and second-order bounds (in terms of the quadratic variation). There are also some impossibility results. For example, the presence of an arm that is optimal in every round cannot help, and neither can losses that lie in a small range. The latter results are in contrast to the full-information setting where these things lead to smaller regret.
  • May 2016: There is a new near-near-optimal algorithm for finite-armed subgaussian bandits. See the preprint or a brief description.
  • May 2016: Together with Jan Leike, Laurent Orseau and Marcus Hutter, I have a paper on asymptotic results for a version of Thompson sampling in general reinforcement learning environments accepted at UAI (http://arxiv.org/abs/1602.07905).
  • April 2016: I have papers accepted at COLT (regret analysis of the Gittins index, http://arxiv.org/abs/1511.06014) and ICML (on “conservative” bandits to guarantee revenue while exploring, http://arxiv.org/abs/1602.04282).