资源论文CRAFTML, an Efficient Clustering-based Random Forest for Extreme Multi-label Learning

CRAFTML, an Efficient Clustering-based Random Forest for Extreme Multi-label Learning

2020-03-16 | |  57 |   45 |   0

Abstract

Extreme Multi-label Learning (XML) considers large sets of items described by a number of labels that can exceed one million. Tree-based methods, which hierarchically partition the problem into small scale sub-problems, are particularly promising in this context to reduce the lear ing/prediction complexity and to open the way to parallelization. However, the current best approaches do not exploit tree randomization which has shown its efficiency in random forests and they resort to complex partitioning strategies. To overcome these limits, we here introduce a new random forest based algorithm with a very fast partitioning approach called CRAFTML. Experimental comparisons on nine datasets from the XML literature show that it outperforms the other tree-based approaches. Moreover with a parallelized implementation reduced to five cores, it i competitive with the best state-of-the-art methods which run on one hundred-core machines.

上一篇:The Power of Interpolation: Understanding the Effectiveness of SGD in Modern Over-parametrized Learning†

下一篇:Coordinated Exploration in Concurrent Reinforcement Learning

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...