Cargando…
Prediction of Head Movement in 360-Degree Videos Using Attention Model
In this paper, we propose a prediction algorithm, the combination of Long Short-Term Memory (LSTM) and attention model, based on machine learning models to predict the vision coordinates when watching 360-degree videos in a Virtual Reality (VR) or Augmented Reality (AR) system. Predicting the vision...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8198419/ https://www.ncbi.nlm.nih.gov/pubmed/34070560 http://dx.doi.org/10.3390/s21113678 |
_version_ | 1783707134024744960 |
---|---|
author | Lee, Dongwon Choi, Minji Lee, Joohyun |
author_facet | Lee, Dongwon Choi, Minji Lee, Joohyun |
author_sort | Lee, Dongwon |
collection | PubMed |
description | In this paper, we propose a prediction algorithm, the combination of Long Short-Term Memory (LSTM) and attention model, based on machine learning models to predict the vision coordinates when watching 360-degree videos in a Virtual Reality (VR) or Augmented Reality (AR) system. Predicting the vision coordinates while video streaming is important when the network condition is degraded. However, the traditional prediction models such as Moving Average (MA) and Autoregression Moving Average (ARMA) are linear so they cannot consider the nonlinear relationship. Therefore, machine learning models based on deep learning are recently used for nonlinear predictions. We use the Long Short-Term Memory (LSTM) and Gated Recurrent Unit (GRU) neural network methods, originated in Recurrent Neural Networks (RNN), and predict the head position in the 360-degree videos. Therefore, we adopt the attention model to LSTM to make more accurate results. We also compare the performance of the proposed model with the other machine learning models such as Multi-Layer Perceptron (MLP) and RNN using the root mean squared error (RMSE) of predicted and real coordinates. We demonstrate that our model can predict the vision coordinates more accurately than the other models in various videos. |
format | Online Article Text |
id | pubmed-8198419 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-81984192021-06-14 Prediction of Head Movement in 360-Degree Videos Using Attention Model Lee, Dongwon Choi, Minji Lee, Joohyun Sensors (Basel) Article In this paper, we propose a prediction algorithm, the combination of Long Short-Term Memory (LSTM) and attention model, based on machine learning models to predict the vision coordinates when watching 360-degree videos in a Virtual Reality (VR) or Augmented Reality (AR) system. Predicting the vision coordinates while video streaming is important when the network condition is degraded. However, the traditional prediction models such as Moving Average (MA) and Autoregression Moving Average (ARMA) are linear so they cannot consider the nonlinear relationship. Therefore, machine learning models based on deep learning are recently used for nonlinear predictions. We use the Long Short-Term Memory (LSTM) and Gated Recurrent Unit (GRU) neural network methods, originated in Recurrent Neural Networks (RNN), and predict the head position in the 360-degree videos. Therefore, we adopt the attention model to LSTM to make more accurate results. We also compare the performance of the proposed model with the other machine learning models such as Multi-Layer Perceptron (MLP) and RNN using the root mean squared error (RMSE) of predicted and real coordinates. We demonstrate that our model can predict the vision coordinates more accurately than the other models in various videos. MDPI 2021-05-25 /pmc/articles/PMC8198419/ /pubmed/34070560 http://dx.doi.org/10.3390/s21113678 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Lee, Dongwon Choi, Minji Lee, Joohyun Prediction of Head Movement in 360-Degree Videos Using Attention Model |
title | Prediction of Head Movement in 360-Degree Videos Using Attention Model |
title_full | Prediction of Head Movement in 360-Degree Videos Using Attention Model |
title_fullStr | Prediction of Head Movement in 360-Degree Videos Using Attention Model |
title_full_unstemmed | Prediction of Head Movement in 360-Degree Videos Using Attention Model |
title_short | Prediction of Head Movement in 360-Degree Videos Using Attention Model |
title_sort | prediction of head movement in 360-degree videos using attention model |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8198419/ https://www.ncbi.nlm.nih.gov/pubmed/34070560 http://dx.doi.org/10.3390/s21113678 |
work_keys_str_mv | AT leedongwon predictionofheadmovementin360degreevideosusingattentionmodel AT choiminji predictionofheadmovementin360degreevideosusingattentionmodel AT leejoohyun predictionofheadmovementin360degreevideosusingattentionmodel |