Abstract
The ability to recognize human actions in video has many
potential applications. Human action recognition, however,
is tremendously challenging for computers due to the complexity of video data and the subtlety of human actions.
Most current recognition systems flounder on the inability
to separate human actions from co-occurring factors that
usually dominate subtle human actions.
In this paper, we propose a novel approach for training
a human action recognizer, one that can: (1) explicitly factorize human actions from the co-occurring factors; (2) deliberately build a model for human actions and a separate
model for all correlated contextual elements; and (3) effectively combine the models for human action recognition.
Our approach exploits the benefits of conjugate samples of
human actions, which are video clips that are contextually similar to human action samples, but do not contain
the action. Experiments on ActionThread, PASCAL VOC,
UCF101, and Hollywood2 datasets demonstrate the ability
to separate action from context of the proposed approach