arXiv Analytics

Sign in

arXiv:1609.08758 [cs.CV]AbstractReferencesReviewsResources

Video Summarization using Deep Semantic Features

Mayu Otani, Yuta Nakashima, Esa Rahtu, Janne Heikkilä, Naokazu Yokoya

Published 2016-09-28Version 1

This paper presents a video summarization technique for an Internet video to provide a quick way to overview its content. This is a challenging problem because finding important or informative parts of the original video requires to understand its content. Furthermore the content of Internet videos is very diverse, ranging from home videos to documentaries, which makes video summarization much more tough as prior knowledge is almost not available. To tackle this problem, we propose to use deep video features that can encode various levels of content semantics, including objects, actions, and scenes, improving the efficiency of standard video summarization techniques. For this, we design a deep neural network that maps videos as well as descriptions to a common semantic space and jointly trained it with associated pairs of videos and descriptions. To generate a video summary, we extract the deep features from each segment of the original video and apply a clustering-based summarization technique to them. We evaluate our video summaries using the SumMe dataset as well as baseline approaches. The results demonstrated the advantages of incorporating our deep semantic features in a video summarization technique.

Comments: 16 pages, the 13th Asian Conference on Computer Vision (ACCV'16)
Categories: cs.CV
Related articles: Most relevant | Search more
arXiv:1702.01528 [cs.CV] (Published 2017-02-06)
Textually Customized Video Summaries
arXiv:1401.3590 [cs.CV] (Published 2014-01-14, updated 2016-04-19)
An Enhanced Method For Evaluating Automatic Video Summaries
arXiv:2002.03740 [cs.CV] (Published 2020-01-31)
Convolutional Hierarchical Attention Network for Query-Focused Video Summarization