Cargando…

Cross-Modality Interaction Network for Equine Activity Recognition Using Imbalanced Multi-Modal Data †

With the recent advances in deep learning, wearable sensors have increasingly been used in automated animal activity recognition. However, there are two major challenges in improving recognition performance—multi-modal feature fusion and imbalanced data modeling. In this study, to improve classifica...

Descripción completa

Detalles Bibliográficos
Autores principales: Mao, Axiu, Huang, Endai, Gan, Haiming, Parkes, Rebecca S. V., Xu, Weitao, Liu, Kai
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8434387/
https://www.ncbi.nlm.nih.gov/pubmed/34502709
http://dx.doi.org/10.3390/s21175818
_version_ 1783751586664677376
author Mao, Axiu
Huang, Endai
Gan, Haiming
Parkes, Rebecca S. V.
Xu, Weitao
Liu, Kai
author_facet Mao, Axiu
Huang, Endai
Gan, Haiming
Parkes, Rebecca S. V.
Xu, Weitao
Liu, Kai
author_sort Mao, Axiu
collection PubMed
description With the recent advances in deep learning, wearable sensors have increasingly been used in automated animal activity recognition. However, there are two major challenges in improving recognition performance—multi-modal feature fusion and imbalanced data modeling. In this study, to improve classification performance for equine activities while tackling these two challenges, we developed a cross-modality interaction network (CMI-Net) involving a dual convolution neural network architecture and a cross-modality interaction module (CMIM). The CMIM adaptively recalibrated the temporal- and axis-wise features in each modality by leveraging multi-modal information to achieve deep intermodality interaction. A class-balanced (CB) focal loss was adopted to supervise the training of CMI-Net to alleviate the class imbalance problem. Motion data was acquired from six neck-attached inertial measurement units from six horses. The CMI-Net was trained and verified with leave-one-out cross-validation. The results demonstrated that our CMI-Net outperformed the existing algorithms with high precision (79.74%), recall (79.57%), F1-score (79.02%), and accuracy (93.37%). The adoption of CB focal loss improved the performance of CMI-Net, with increases of 2.76%, 4.16%, and 3.92% in precision, recall, and F1-score, respectively. In conclusion, CMI-Net and CB focal loss effectively enhanced the equine activity classification performance using imbalanced multi-modal sensor data.
format Online
Article
Text
id pubmed-8434387
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-84343872021-09-12 Cross-Modality Interaction Network for Equine Activity Recognition Using Imbalanced Multi-Modal Data † Mao, Axiu Huang, Endai Gan, Haiming Parkes, Rebecca S. V. Xu, Weitao Liu, Kai Sensors (Basel) Article With the recent advances in deep learning, wearable sensors have increasingly been used in automated animal activity recognition. However, there are two major challenges in improving recognition performance—multi-modal feature fusion and imbalanced data modeling. In this study, to improve classification performance for equine activities while tackling these two challenges, we developed a cross-modality interaction network (CMI-Net) involving a dual convolution neural network architecture and a cross-modality interaction module (CMIM). The CMIM adaptively recalibrated the temporal- and axis-wise features in each modality by leveraging multi-modal information to achieve deep intermodality interaction. A class-balanced (CB) focal loss was adopted to supervise the training of CMI-Net to alleviate the class imbalance problem. Motion data was acquired from six neck-attached inertial measurement units from six horses. The CMI-Net was trained and verified with leave-one-out cross-validation. The results demonstrated that our CMI-Net outperformed the existing algorithms with high precision (79.74%), recall (79.57%), F1-score (79.02%), and accuracy (93.37%). The adoption of CB focal loss improved the performance of CMI-Net, with increases of 2.76%, 4.16%, and 3.92% in precision, recall, and F1-score, respectively. In conclusion, CMI-Net and CB focal loss effectively enhanced the equine activity classification performance using imbalanced multi-modal sensor data. MDPI 2021-08-29 /pmc/articles/PMC8434387/ /pubmed/34502709 http://dx.doi.org/10.3390/s21175818 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Mao, Axiu
Huang, Endai
Gan, Haiming
Parkes, Rebecca S. V.
Xu, Weitao
Liu, Kai
Cross-Modality Interaction Network for Equine Activity Recognition Using Imbalanced Multi-Modal Data †
title Cross-Modality Interaction Network for Equine Activity Recognition Using Imbalanced Multi-Modal Data †
title_full Cross-Modality Interaction Network for Equine Activity Recognition Using Imbalanced Multi-Modal Data †
title_fullStr Cross-Modality Interaction Network for Equine Activity Recognition Using Imbalanced Multi-Modal Data †
title_full_unstemmed Cross-Modality Interaction Network for Equine Activity Recognition Using Imbalanced Multi-Modal Data †
title_short Cross-Modality Interaction Network for Equine Activity Recognition Using Imbalanced Multi-Modal Data †
title_sort cross-modality interaction network for equine activity recognition using imbalanced multi-modal data †
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8434387/
https://www.ncbi.nlm.nih.gov/pubmed/34502709
http://dx.doi.org/10.3390/s21175818
work_keys_str_mv AT maoaxiu crossmodalityinteractionnetworkforequineactivityrecognitionusingimbalancedmultimodaldata
AT huangendai crossmodalityinteractionnetworkforequineactivityrecognitionusingimbalancedmultimodaldata
AT ganhaiming crossmodalityinteractionnetworkforequineactivityrecognitionusingimbalancedmultimodaldata
AT parkesrebeccasv crossmodalityinteractionnetworkforequineactivityrecognitionusingimbalancedmultimodaldata
AT xuweitao crossmodalityinteractionnetworkforequineactivityrecognitionusingimbalancedmultimodaldata
AT liukai crossmodalityinteractionnetworkforequineactivityrecognitionusingimbalancedmultimodaldata