Abstract
Algorithms for full-information online learning are classically tuned to minimize their worst-case regret. Modern algorithms additionally provide tighter guarantees outside the adversarial regime, most notably in the form of constant pseudoregret bounds under statistical margin assumptions. We investigate the multiscale extension of the problem where the loss ranges of the experts are vastly different. Here, the regret with respect to each expert needs to scale with its range, instead of the maximum overall range. We develop new multiscale algorithms, tuning schemes and analysis techniques to show that worst-case robustness and adaptation to easy data can be combined at a negligible cost. We further develop an extension with optimism and apply it to solve multiscale two-player zero-sum games. We demonstrate experimentally the superior performance of our scale-adaptive algorithm and discuss the subtle relationship of our results to Freund's 2016 open problem.
Original language | English |
---|---|
Title of host publication | Proceedings of the 36th Conference on Neural Information Processing Systems, NeurIPS 2022 |
Editors | S. Koyejo, S. Mohamed, A. Agarwal, D. Belgrave, K. Cho, A. Oh |
Publisher | Neural information processing systems foundation |
Number of pages | 11 |
ISBN (Electronic) | 9781713871088 |
Publication status | Published - 2023 |
Externally published | Yes |
Event | 36th Conference on Neural Information Processing Systems, NeurIPS 2022 - Hybrid, New Orleans, United States Duration: 28 Nov 2022 → 9 Dec 2022 Conference number: 36 |
Publication series
Name | Advances in Neural Information Processing Systems |
---|---|
Volume | 35 |
Conference
Conference | 36th Conference on Neural Information Processing Systems, NeurIPS 2022 |
---|---|
Abbreviated title | NeurIPS 2022 |
Country/Territory | United States |
City | New Orleans |
Period | 28/11/22 → 9/12/22 |