Abstract
This paper focuses on transcription generation in the form of sub ject, verb, ob ject (SVO) triplets for videos in the wild, given off- the-shelf visual concept detectors. This problem is challenging due to the availability of sentence only annotations, the unreliability of con- cept detectors, and the lack of training samples for many words. Fac- ing these challenges, we propose a Semantic Aware Transcription (SAT) framework based on Random Forest classifiers. It takes concept detec- tion results as input, and outputs a distribution of English words. SAT uses video, sentence pairs for training. It hierarchically learns node splits by grouping semantically similar words, measured by a continuous skip- gram language model. This not only addresses the sparsity of training samples per word, but also yields semantically reasonable errors during transcription. SAT provides a systematic way to measure the related- ness of a concept detector to real words, which helps us understand the relationship between current visual detectors and words in a semantic space. Experiments on a large video dataset with 1,970 clips and 85,550 sentences are used to demonstrate our idea.