资源论文New Subsampling Algorithms for Fast Least Squares Regression

New Subsampling Algorithms for Fast Least Squares Regression

2020-01-16 | |  95 |   55 |   0

Abstract

We address the problem of fast estimation of ordinary least squares (OLS) from large amounts of data 图片.png We propose three methods which solve the big data problem by subsampling the covariance matrix using either a single or two stage estimation. All three run in the order ofp size of input i.e. O(np) and our best method, Uluru, gives an error bound of 图片.pngwhich is independent of the amount of subsampling as long as it is above a threshold. We provide theoretical bounds for our algorithms in the fixed design (with Randomized Hadamard preconditioning) as well as sub-Gaussian random design setting. We also compare the performance of our methods on synthetic and real-world datasets and show that if observations are i.i.d., sub-Gaussian then one can directly subsample without the expensive Randomized Hadamard preconditioning without loss of accuracy.

上一篇:Deep content-based music recommendation

下一篇:Stochastic Majorization-Minimization Algorithms for Large-Scale Optimization

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • Learning to learn...

    The move from hand-designed features to learned...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...