资源论文Multimodal Abstractive Summarization for How2 Videos

Multimodal Abstractive Summarization for How2 Videos

2019-09-23 | |  182 |   121 |   0 0 0
Abstract In this paper, we study abstractive summarization for open-domain videos. Unlike the traditional text news summarization, the goal is less to “compress” text information but rather to provide a fluent textual summary of information that has been collected and fused from different source modalities, in our case video and audio transcripts (or text). We show how a multi-source sequence-to-sequence model with hierarchical attention can integrate information from different modalities into a coherent output, compare various models trained with different modalities and present pilot experiments on the How2 corpus of instructional videos. We also propose a new evaluation metric (Content F1) for abstractive summarization task that measures semantic adequacy rather than fluency of the summaries, which is covered by metrics like ROUGE and BLEU.

上一篇:Keep Meeting Summaries on Topic: Abstractive Multi-Modal Meeting Summarization

下一篇:Multi-News: a Large-Scale Multi-Document Summarization Dataset and Abstractive Hierarchical Model

用户评价
全部评价

热门资源

  • Deep Cross-media ...

    Cross-media retrieval is a research hotspot in ...

  • Regularizing RNNs...

    Recently, caption generation with an encoder-de...

  • Supervised Descen...

    Many computer vision problems (e.

  • Attributed Graph ...

    Graph clustering is a fundamental task which di...

  • Hierarchical Task...

    We extend hierarchical task network planning wi...