资源论文Constrained Optimization Based Low-Rank Approximation of Deep Neural Networks

Constrained Optimization Based Low-Rank Approximation of Deep Neural Networks

2019-10-23 | |  69 |   42 |   0

Abstract. We present COBLA—Constrained Optimization Based Lowrank Approximation—a systematic method of fifinding an optimal lowrank approximation of a trained convolutional neural network, subject to constraints in the number of multiply-accumulate (MAC) operations and the memory footprint. COBLA optimally allocates the constrained computation resources into each layer of the approximated network. The singular value decomposition of the network weight is computed, then a binary masking variable is introduced to denote whether a particular singular value and the corresponding singular vectors are used in low-rank approximation. With this formulation, the number of the MAC operations and the memory footprint are represented as linear constraints in terms of the binary masking variables. The resulted 0-1 integer programming problem is approximately solved by sequential quadratic programming. COBLA does not introduce any hyperparameter. We empirically demonstrate that COBLA outperforms prior art using the SqueezeNet and VGG-16 architecture on the ImageNet dataset

上一篇:Escaping from Collapsing Modes in a Constrained Space

下一篇:A Scalable Exemplar-based Subspace ClusteringAlgorithm for Class-Imbalanced Data

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...