Abstract
In this paper, we propose an accelerated first-order method for geodesically convex optimization, which is the generalization of the standard Nesterov’s accelerated method from Euclidean space to nonlinear Riemannian space. We first derive two equations and obtain two nonlinear operators for geodesically convex optimization instead of the linear extrapolation step in Euclidean space. In particular, we analyze the global convergence properties of our accelerated method for geodesically strongly-convex problems, which show p that our method improves the convergence rate from . Moreover, our method also improves the global convergence rate on geodesically general convex problems from to . Finally, we give a specific iterative scheme for matrix Karcher mean problems, and validate our theoretical results with experiments.