资源论文On the number of variables to use in principal component regression

On the number of variables to use in principal component regression

2020-02-19 | |  68 |   47 |   0

Abstract

We study least squares linear regression over N uncorrelated Gaussian features that are selected in order of decreasing variance. When the number of selected features p is at most the sample size n, the estimator under consideration coincides with the principal component regression estimator; when p > n, the estimator is the least 图片.png2 norm solution over the selected features. We give an average-case analysis of the out-of-sample prediction error as p, n, N 图片.png with p/N 图片.png and n/N 图片.png , for some constants 图片.png and 图片.png. In this average-case setting, the prediction error exhibits a “double descent” shape as a function of p. We also establish conditions under which the minimum risk is achieved in the interpolating (p > n) regime.

上一篇:An Adaptive Empirical Bayesian Method for Sparse Deep Learning

下一篇:Modelling the Dynamics of Multiagent Q-Learning in Repeated Symmetric Games: a Mean Field Theoretic Approach

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...