[C51] - Online and stochastic optimization beyond Lipschitz continuity: A Riemannian approach

K. Antonakopoulos, E. V. Belmega, and P. Mertikopoulos. In ICLR '20: Proceedings of the 2020 International Conference on Learning Representations, 2020.


Motivated by applications to machine learning and imaging science, we study a class of online and stochastic optimization problems with loss functions that are not Lipschitz continuous; in particular, the loss functions encountered by the optimizer could exhibit gradient singularities or be singular themselves. Drawing on tools and techniques from Riemannian geometry, we examine a Riemann–Lipschitz (RL) continuity condition which is tailored to the singularity landscape of the problem’s loss functions. In this way, we are able to tackle cases beyond the Lipschitz framework provided by a global norm, and we derive optimal regret bounds and last iterate convergence results through the use of regularized learning methods (such as online mirror descent). These results are subsequently validated in a class of stochastic Poisson inverse problems that arise in imaging science.

[Spotlight talk at ICLR 2020]

Nifty tech tag lists from Wouter Beeftink