|Time:||星期四, 10月 14, 2021, 10:30am|
|Title:||Classical and quantum algorithms for nonconvex optimization|
|Speaker:||Chenyi Zhang (Tsinghua)|
Nonconvex optimization is a central research topic in machine learning, mainly because the loss functions in many machine learning models (including neural networks) are typically nonconvex. However, finding a global optimum of a nonconvex function is very hard in general (the running time may be exponentially large). Instead, many theoretical works focus on finding local optima.
The main obstacle comes from ubiquitous saddle points where the optimization algorithms may get stuck. We propose a simple gradient-descent based algorithm such that for a smooth function, it outputs an approximate local minimum of the function in O(log n) gradient descent iterations. Compared to the previous state-of-the-art algorithm using O( log^4 n) iterations, our algorithm is polynomially better in terms of (log n). Technically, our main contribution is an idea of implementing a Hessian power method using only gradients, which can find negative curvature near saddle points that indicates feasible directions to escape from them.
For the same problem, we also propose a quantum algorithm that outputs an approximate local minimum using O(log^2 n) queries to a quantum oracle providing no gradient information but only function value..
Host: Dandan Xu