资源论文Efficiently escaping saddle points on manifolds

Efficiently escaping saddle points on manifolds

2020-02-20 | |  50 |   38 |   0

Abstract

Smooth, non-convex optimization problems on Riemannian manifolds occur in machine learning as a result of orthonormality, rank or positivity constraints. Firstand second-order necessary optimality conditions state that the Riemannian gradient must be zero, and the Riemannian Hessian must be positive semidefinite. Generalizing Jin et al.’s recent work on perturbed gradient descent (PGD) for optimization on linear spaces [How to Escape Saddle Points Efficiently (2017) [17], Stochastic Gradient Descent Escapes Saddle Points Efficiently (2019) [18]], we propose a version of perturbed Riemannian gradient descent (PRGD) to show that necessary optimality conditions can be met approximately with high probability, without evaluating the Hessian. Specifically, for an arbitrary Riemannian manifold M of dimension d, a sufficiently smooth (possibly non-convex) objective function f , and under weak conditions on the retraction chosen to move on the manifold, with high probability, our version p of PRGD produces a point with gradient smaller than 图片.png and Hessian within 图片.png of being positive semidefinite in 图片.png gradient queries. This matches the complexity of PGD in the Euclidean case. Crucially, the dependence on dimension is low. This matters for large-scale applications including PCA and low-rank matrix completion, which both admit natural formulations on manifolds. The key technical idea is to generalize PRGD with a distinction between two types of gradient steps: “steps on the manifold” and “perturbed steps in a tangent space of the manifold.” Ultimately, this distinction makes it possible to extend Jin et al.’s analysis seamlessly.

上一篇:Information Competing Process for Learning Diversified Representations

下一篇:Learning Perceptual Inference by Contrasting

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...