资源论文More data speeds up training time in learning halfspaces over sparse vectors

More data speeds up training time in learning halfspaces over sparse vectors

2020-01-16 | |  66 |   43 |   0

Abstract

The increased availability of data in recent years has led several authors to ask whether it is possible to use data as a computational resource. That is, if more data is available, beyond the sample complexity limit, is it possible to use the extra examples to speed up the computation time required to perform the learning task? We give the first positive answer to this question for a natural supervised learning problem — we consider agnostic PAC learning of halfspaces over 3-sparse  vectors in 图片.png This class is inefficiently learnable using O n/2 examples. Our main contribution is a novel, non-cryptographic, methodology for establishing computational-statistical gaps, which allows us to show that, under a widely believed assumption that refuting random 3CNFformulas is hard, it is impossible to efficiently learn this class using only 图片.png examples.  We further show that under stronger hardness assumptions, even 图片.png examples do not suffice. On the other  hand, we show a new algorithm that learns this class efficiently using 图片.png examples. This formally establishes the tradeoff between sample and computational complexity for a natural supervised learning problem.

上一篇:Data-driven Distributionally Robust Polynomial Optimization

下一篇:On the Representational Efficiency of Restricted Boltzmann Machines

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...