资源论文An Optimal Control Approach to Deep Learning and Applications to Discrete-Weight Neural Networks

An Optimal Control Approach to Deep Learning and Applications to Discrete-Weight Neural Networks

2020-03-16 | |  60 |   42 |   0

Abstract

Deep learning is formulated as a discrete-time optimal control problem. This allows one to characterize necessary conditions for optimality and develop training algorithms that do not rely on g dients with respect to the trainable parameters. particular, we introduce the discrete-time method of successive approximations (MSA), which is based on the Pontryagin’s maximum principle, for training neural networks. A rigorous error es timate for the discrete MSA is obtained, which sheds light on its dynamics and the means to stabilize the algorithm. The developed methods are applied to train, in a rather principled way, neu networks with weights that are constrained to tak values in a discrete set. We obtain competitive p formance and interestingly, very sparse weights i the case of ternary networks, which may be useful in model deployment in low-memory devices.

上一篇:Multi-Fidelity Black-Box Optimization with Hierarchical Partitions

下一篇:Mixed batches and symmetric discriminators for GAN training

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...