资源论文AS TOCHASTIC DERIVATIVE FREE OPTIMIZATIONM ETHOD WITH MOMENTUM

AS TOCHASTIC DERIVATIVE FREE OPTIMIZATIONM ETHOD WITH MOMENTUM

2020-01-02 | |  58 |   48 |   0

Abstract

We consider the problem of unconstrained minimization of a smooth objective function in Rd in setting where only function evaluations are possible. We propose and analyze stochastic zeroth-order method with heavy ball momentum. In particular, we propose, SMTP, a momentum version of the stochastic three-point method (STP) Bergou et al. (2019). We show new complexity results for non-convex, convex and strongly convex functions. We test our method on a collection of learning to continuous control tasks on several MuJoCo Todorov et al. (2012) environments with varying difficulty and compare against STP, other state-of-the-art derivative-free optimization algorithms and against policy gradient methods. SMTP significantly outperforms STP and all other methods that we considered in our numerical experiments. Our second contribution is SMTP with importance sampling which we call SMTP_IS. We provide convergence analysis of this method for non-convex, convex and strongly convex objectives.

上一篇:VARIATIONAL RECURRENT MODELS FOR SOLVINGPARTIALLY OBSERVABLE CONTROL TASKS

下一篇:THE SHAPE OF DATA :I NTRINSIC DISTANCE FOR DATA DISTRIBUTIONS

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...