资源论文PROX -SGD: TRAINING STRUCTURED NEURAL NET-WORKS UNDER REGULARIZATION AND CONSTRAINTS

PROX -SGD: TRAINING STRUCTURED NEURAL NET-WORKS UNDER REGULARIZATION AND CONSTRAINTS

2019-12-30 | |  88 |   44 |   0

Abstract

In this paper, we consider the problem of training structured neural networks (NN) with nonsmooth regularization (e.g. `1 -norm) and constraints (e.g. interval constraints). We formulate training as a constrained nonsmooth nonconvex optimization problem, and propose a convergent proximal-type stochastic gradient descent (Prox-SGD) algorithm. We show that under properly selected learning rates, with probability 1, every limit point of the sequence generated by the proposed ProxSGD algorithm is a stationary point. Finally, to support the theoretical analysis and demonstrate the flexibility of Prox-SGD, we show by extensive numerical tests how Prox-SGD can be used to train either sparse or binary neural networks through an adequate selection of the regularization function and constraint set.

上一篇:SIZE -FREE GENERALIZATION BOUNDSFOR CONVOLUTIONAL NEURAL NETWORKS

下一篇:REGULARIZING ACTIVATIONS IN NEURAL NETWORKSVIA DISTRIBUTION MATCHING WITH THE WASSER -STEIN METRIC

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...