资源论文A P300 BCI for the Masses: Prior Information Enables Instant Unsupervised Spelling

A P300 BCI for the Masses: Prior Information Enables Instant Unsupervised Spelling

2020-01-16 | |  76 |   47 |   0

Abstract
The usability of Brain Computer Interfaces (BCI) based on the P300 speller is severely hindered by the need for long training times and many repetitions of the same stimulus. In this contribution we introduce a set of unsupervised hierarchical probabilistic models that tackle both problems simultaneously by incorporating prior knowledge from two sources: information from other training subjects (through transfer learning) and information about the words being spelled (through language models). We show, that due to this prior knowledge, the performance of the unsupervised models parallels and in some cases even surpasses that of supervised models, while eliminating the tedious training session.

上一篇:Emergence of Object-Selective Features in Unsupervised Feature Learning

下一篇:Nonparametric Bayesian Inverse Reinforcement Learning for Multiple Reward Functionss

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...