资源论文Diffusion Approximations for Online Principal Component Estimation and Global Convergence

Diffusion Approximations for Online Principal Component Estimation and Global Convergence

2020-02-10 | |  82 |   47 |   0

Abstract 

In this paper, we propose to adopt the diffusion approximation tools to study the dynamics of Oja’s iteration which is an online stochastic gradient descent method for the principal component analysis. Oja’s iteration maintains a running estimate of the true principal component from streaming data and enjoys less temporal and spatial complexities. We show that the Oja’s iteration for the top eigenvector generates a continuous-state discrete-time Markov chain over the unit sphere. We characterize the Oja’s iteration in three phases using diffusion approximation and weak convergence tools. Our three-phase analysis further provides a finite-sample error bound for the running estimate, which matches the minimax information lower bound for principal component analysis under the additional assumption of bounded samples.

上一篇:Unsupervised Sequence Classification using Sequential Output Statistics

下一篇:Is the Bellman residual a bad proxy?

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...