资源论文Blockout: Dynamic Model Selection for Hierarchical Deep Networks

Blockout: Dynamic Model Selection for Hierarchical Deep Networks

2019-12-27 | |  87 |   46 |   0

Abstract

Most deep architectures for image classification–eventhose that are trained to classify a large number of diversecategories–learn shared image representations with a sin-gle model. Intuitively, however, categories that are more similar should share more information than those that are very different. While hierarchical deep networks address this problem by learning separate features for subsets of related categories, current implementations require simplified models using fixed architectures specified via heuristic clustering methods. Instead, we propose Blockout, a method for regularization and model selection that simultaneously learns both the model architecture and parameters. A generalization of Dropout, our approach gives a novel parametrization of hierarchical architectures that allows for structure learning via back-propagation. To demon-strate its utility, we evaluate Blockout on the CIFAR and ImageNet datasets, demonstrating improved classification accuracy, better regularization performance, faster training, and the clear emergence of hierarchical network structures.

上一篇:Gradual DropIn of Layers to Train Very Deep Neural Networks

下一篇:Learning Structured Inference Neural Networks with Label Relations

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...