CS 201: Relaxing the I.I.D. Assumption: Adaptively Minimax Optimal Regret via Root-Entropic Regularization, DANIEL ROY – BLAIR BILODEAU, University of Toronto

Speaker: Daniel Roy - Blair Bilodeau
Affiliation: University of Toronto

ABSTRACT:

We consider sequential prediction with expert advice when data are generated from distributions varying arbitrarily within an unknown constraint set. We quantify relaxations of the classical i.i.d. assumption in terms of these constraint sets, with i.i.d. sequences at one extreme and adversarial mechanisms at the other. The Hedge algorithm, long known to be minimax optimal in the adversarial regime, was recently shown to be minimax optimal for i.i.d. data. We show that Hedge with deterministic learning rates is suboptimal between these extremes, and present a new algorithm that adaptively achieves the minimax optimal rate of regret with respect to our relaxations of the i.i.d. assumption, and does so without knowledge of the underlying constraint set. We analyze our algorithm using the follow-the-regularized-leader framework, and prove it corresponds to Hedge with an adaptive learning rate that implicitly scales as the square root of the entropy of the current predictive distribution, rather than the entropy of the initial predictive distribution.

Joint work with Blair Bilodeau and Jeffrey Negrea. https://arxiv.org/abs/2007.06552

BIO:

Daniel Roy is Canada CIFAR AI Chair at the Vector Institute and associate professor in the Departments of Statistical Sciences, Computer Science, Electrical and Computer Engineering, and Computer and Mathematical Sciences. Roy’s research spans machine learning, mathematical statistics, and theoretical computer science.

BIO:

Blair Bilodeau is a PhD candidate in statistics at the University of Toronto, supported by the Vector Institute and an NSERC Doctoral Canada Graduate Scholarship. Blair received his BSc in financial mathematics from Western University in 2018. Blair’s research area is broadly statistical machine learning, with a focus on theoretical performance guarantees for sequential decision making.

Hosted by Professor Quanquan Gu

Date/Time:
Date(s) - Mar 30, 2021
4:00 pm - 5:45 pm

Location:
Zoom Webinar
404 Westwood Plaza Los Angeles
Map Unavailable