Skip to yearly menu bar Skip to main content


Poster

Optimal Sample Complexity Bounds for Non-convex Optimization under Kurdyka-Lojasiewicz Condition

Qian Yu · Yining Wang · Baihe Huang · Qi Lei · Jason Lee

Auditorium 1 Foyer 101

Abstract:

Optimization of smooth reward functions under bandit feedback is a long-standing problem in online learning. This paper approaches this problem by studying the convergence under smoothness and Kurdyka-Lojasiewicz conditions. We designed a search-based algorithm that achieves an improved rate compared to the standard gradient-based method. In conjunction with a matching lower bound, this algorithm shows optimality in the dependence on precision for the low-dimension regime.

Live content is unavailable. Log in and register to view live content