资源论文Large-Scale Distributed Second-Order Optimization Using Kronecker-FactoredApproximate Curvature for Deep Convolutional Neural Networks

Large-Scale Distributed Second-Order Optimization Using Kronecker-FactoredApproximate Curvature for Deep Convolutional Neural Networks

2019-09-17 | |  90 |   46 |   0 0 0
Abstract Large-scale distributed training of deep neural networks suffer from the generalization gap caused by the increase in the effective mini-batch size. Previous approaches try to solve this problem by varying the learning rate and batch size over epochs and layers, or some ad hoc modification of the batch normalization. We propose an alternative approach using a second-order optimization method that shows similar generalization capability to first-order methods, but converges faster and can handle larger minibatches. To test our method on a benchmark where highly optimized first-order methods are available as references, we train ResNet-50 on ImageNet. We converged to 75% Top-1 validation accuracy in 35 epochs for mini-batch sizes under 16,384, and achieved 75% even with a mini-batch size of 131,072, which took only 978 iterations

上一篇:Knowledge-Embedded Routing Network for Scene Graph Generation

下一篇:LaSO: Label-Set Operations networks for multi-label few-shot learning

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • Learning to learn...

    The move from hand-designed features to learned...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...