Cargando…
Self-Supervised Learning to Detect Key Frames in Videos
Detecting key frames in videos is a common problem in many applications such as video classification, action recognition and video summarization. These tasks can be performed more efficiently using only a handful of key frames rather than the full video. Existing key frame detection approaches are m...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7731244/ https://www.ncbi.nlm.nih.gov/pubmed/33291759 http://dx.doi.org/10.3390/s20236941 |
_version_ | 1783621861490294784 |
---|---|
author | Yan, Xiang Gilani, Syed Zulqarnain Feng, Mingtao Zhang, Liang Qin, Hanlin Mian, Ajmal |
author_facet | Yan, Xiang Gilani, Syed Zulqarnain Feng, Mingtao Zhang, Liang Qin, Hanlin Mian, Ajmal |
author_sort | Yan, Xiang |
collection | PubMed |
description | Detecting key frames in videos is a common problem in many applications such as video classification, action recognition and video summarization. These tasks can be performed more efficiently using only a handful of key frames rather than the full video. Existing key frame detection approaches are mostly designed for supervised learning and require manual labelling of key frames in a large corpus of training data to train the models. Labelling requires human annotators from different backgrounds to annotate key frames in videos which is not only expensive and time consuming but also prone to subjective errors and inconsistencies between the labelers. To overcome these problems, we propose an automatic self-supervised method for detecting key frames in a video. Our method comprises a two-stream ConvNet and a novel automatic annotation architecture able to reliably annotate key frames in a video for self-supervised learning of the ConvNet. The proposed ConvNet learns deep appearance and motion features to detect frames that are unique. The trained network is then able to detect key frames in test videos. Extensive experiments on UCF101 human action and video summarization VSUMM datasets demonstrates the effectiveness of our proposed method. |
format | Online Article Text |
id | pubmed-7731244 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-77312442020-12-12 Self-Supervised Learning to Detect Key Frames in Videos Yan, Xiang Gilani, Syed Zulqarnain Feng, Mingtao Zhang, Liang Qin, Hanlin Mian, Ajmal Sensors (Basel) Article Detecting key frames in videos is a common problem in many applications such as video classification, action recognition and video summarization. These tasks can be performed more efficiently using only a handful of key frames rather than the full video. Existing key frame detection approaches are mostly designed for supervised learning and require manual labelling of key frames in a large corpus of training data to train the models. Labelling requires human annotators from different backgrounds to annotate key frames in videos which is not only expensive and time consuming but also prone to subjective errors and inconsistencies between the labelers. To overcome these problems, we propose an automatic self-supervised method for detecting key frames in a video. Our method comprises a two-stream ConvNet and a novel automatic annotation architecture able to reliably annotate key frames in a video for self-supervised learning of the ConvNet. The proposed ConvNet learns deep appearance and motion features to detect frames that are unique. The trained network is then able to detect key frames in test videos. Extensive experiments on UCF101 human action and video summarization VSUMM datasets demonstrates the effectiveness of our proposed method. MDPI 2020-12-04 /pmc/articles/PMC7731244/ /pubmed/33291759 http://dx.doi.org/10.3390/s20236941 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Yan, Xiang Gilani, Syed Zulqarnain Feng, Mingtao Zhang, Liang Qin, Hanlin Mian, Ajmal Self-Supervised Learning to Detect Key Frames in Videos |
title | Self-Supervised Learning to Detect Key Frames in Videos |
title_full | Self-Supervised Learning to Detect Key Frames in Videos |
title_fullStr | Self-Supervised Learning to Detect Key Frames in Videos |
title_full_unstemmed | Self-Supervised Learning to Detect Key Frames in Videos |
title_short | Self-Supervised Learning to Detect Key Frames in Videos |
title_sort | self-supervised learning to detect key frames in videos |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7731244/ https://www.ncbi.nlm.nih.gov/pubmed/33291759 http://dx.doi.org/10.3390/s20236941 |
work_keys_str_mv | AT yanxiang selfsupervisedlearningtodetectkeyframesinvideos AT gilanisyedzulqarnain selfsupervisedlearningtodetectkeyframesinvideos AT fengmingtao selfsupervisedlearningtodetectkeyframesinvideos AT zhangliang selfsupervisedlearningtodetectkeyframesinvideos AT qinhanlin selfsupervisedlearningtodetectkeyframesinvideos AT mianajmal selfsupervisedlearningtodetectkeyframesinvideos |