资源论文Large Linear Classification When Data Cannot Fit in Memory

Large Linear Classification When Data Cannot Fit in Memory

2019-11-12 | |  57 |   45 |   0

Abstract

Linear classification is a useful tool for dealing with large-scale data in applications such as doc-ument classification and natural language process-ing. Recent developments of linear classification have shown that the training process can be effi-ciently conducted. However, when the data size ex-ceeds the memory capacity, most training methods suffer from very slow convergencedue to the severe disk swapping. Although some methods have at-tempted to handle such a situation, they are usually too complicated to support some important func-tions such as parameter selection. In this paper, we introduce a block minimization framework for data larger than memory. Under the framework, a solver splits data into blocks and stores them into separate files. Then, at each time, the solver trains a data block loaded from disk. Although the framework is simple, the experimental results show that it ef-fectively handles a data set 20 times larger than the memory capacity.


上一篇:A Framework for Longitudinal Influence Measurement between Communication Content and Social Networks

下一篇:Decision Support through Argumentation-Based Practical Reasoning

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...