资源论文Variants of RMSProp and Adagrad with Logarithmic Regret Bounds

Variants of RMSProp and Adagrad with Logarithmic Regret Bounds

2020-03-10 | |  132 |   87 |   0

Abstract

Adaptive gradient methods have become recently very popular, in particular as they have been shown to be useful in the training of deep neural networks. In this paper we have analyzed RMSProp, originally proposed for the training of deep neural networks, in the context of on? line convex optimization and show T -type regret bounds. Moreover, we propose two variants SC-Adagrad and SC-RMSProp for which we show logarithmic regret bounds for strongly convex functions. Finally, we demonstrate in the experiments that these new variants outperform other adaptive gradient techniques or stochastic gradient descent in the optimization of strongly convex functions as well as in training of deep neural networks.

上一篇:Uncovering Causality from Multivariate Hawkes Integrated Cumulants

下一篇:When can Multi-Site Datasets be Pooled for Regression? Hypothesis Tests, `2 -consistency and Neuroscience Applications

用户评价
全部评价

热门资源

  • Deep Cross-media ...

    Cross-media retrieval is a research hotspot in ...

  • Regularizing RNNs...

    Recently, caption generation with an encoder-de...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Visual Reinforcem...

    For an autonomous agent to fulfill a wide range...

  • Joint Pose and Ex...

    Facial expression recognition (FER) is a challe...