资源论文Generalization of Two-layer Neural Net-works: An Asymptotic Viewpoint

Generalization of Two-layer Neural Net-works: An Asymptotic Viewpoint

2019-12-30 | |  66 |   38 |   0

Abstract

This paper investigates the generalization properties of two-layer neural networks in high-dimensions, i.e. when the number of samples n, features d, and neurons h tend to infinity at the same rate. Specifically, we derive the exact population risk of the unregularized least squares regression problem with two-layer neural networks when either the first or the second layer is trained using a gradient flow under different initialization setups. When only the second layer coefficients are optimized, we recover the double descent phenomenon: a cusp in the population risk appears at h 图片.png n and further overparameterization decreases the risk. In contrast, when the first layer weights are optimized, we highlight how different scales initialization lead to different inductive bias, and show that the resulting risk is independent of overparameterization. Our theoretical and experimental results suggest that previously studied model setups that provably give rise to double descent phenomenon might not translate to optimizing two-layer neural networks.

上一篇:EDITABLE NEURAL NETWORKS

下一篇:STABLE RANK NORMALIZATION FOR IMPROVED GEN -ERALIZATION IN NEURAL NETWORKS AND GAN S

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...