Video summarization using deep semantic features
Otani, Mayu; Nakashima, Yuta; Rahtu, Esa; Heikkilä, Janne; Yokoya, Naokazu (2017-03-11)
Otani M., Nakashima Y., Rahtu E., Heikkilä J., Yokoya N. (2017) Video Summarization Using Deep Semantic Features. In: Lai SH., Lepetit V., Nishino K., Sato Y. (eds) Computer Vision – ACCV 2016. ACCV 2016. Lecture Notes in Computer Science, vol 10115. Springer, Cham
© Springer International Publishing AG 2017. This is a post-peer-review, pre-copyedit version of an article published in ACCV 2016: Computer Vision – ACCV 2016. The final authenticated version is available online at: https://doi.org/10.1007/978-3-319-54193-8_23.
https://rightsstatements.org/vocab/InC/1.0/
https://urn.fi/URN:NBN:fi-fe2019060318262
Tiivistelmä
Abstract
This paper presents a video summarization technique for an Internet video to provide a quick way to overview its content. This is a challenging problem because finding important or informative parts of the original video requires to understand its content. Furthermore the content of Internet videos is very diverse, ranging from home videos to documentaries, which makes video summarization much more tough as prior knowledge is almost not available. To tackle this problem, we propose to use deep video features that can encode various levels of content semantics, including objects, actions, and scenes, improving the efficiency of standard video summarization techniques. For this, we design a deep neural network that maps videos as well as descriptions to a common semantic space and jointly trained it with associated pairs of videos and descriptions. To generate a video summary, we extract the deep features from each segment of the original video and apply a clustering-based summarization technique to them. We evaluate our video summaries using the SumMe dataset as well as baseline approaches. The results demonstrated the advantages of incorporating our deep semantic features in a video summarization technique.
Kokoelmat
- Avoin saatavuus [37125]