Cargando…
Cross-Modality Interaction Network for Equine Activity Recognition Using Imbalanced Multi-Modal Data †
With the recent advances in deep learning, wearable sensors have increasingly been used in automated animal activity recognition. However, there are two major challenges in improving recognition performance—multi-modal feature fusion and imbalanced data modeling. In this study, to improve classifica...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8434387/ https://www.ncbi.nlm.nih.gov/pubmed/34502709 http://dx.doi.org/10.3390/s21175818 |
_version_ | 1783751586664677376 |
---|---|
author | Mao, Axiu Huang, Endai Gan, Haiming Parkes, Rebecca S. V. Xu, Weitao Liu, Kai |
author_facet | Mao, Axiu Huang, Endai Gan, Haiming Parkes, Rebecca S. V. Xu, Weitao Liu, Kai |
author_sort | Mao, Axiu |
collection | PubMed |
description | With the recent advances in deep learning, wearable sensors have increasingly been used in automated animal activity recognition. However, there are two major challenges in improving recognition performance—multi-modal feature fusion and imbalanced data modeling. In this study, to improve classification performance for equine activities while tackling these two challenges, we developed a cross-modality interaction network (CMI-Net) involving a dual convolution neural network architecture and a cross-modality interaction module (CMIM). The CMIM adaptively recalibrated the temporal- and axis-wise features in each modality by leveraging multi-modal information to achieve deep intermodality interaction. A class-balanced (CB) focal loss was adopted to supervise the training of CMI-Net to alleviate the class imbalance problem. Motion data was acquired from six neck-attached inertial measurement units from six horses. The CMI-Net was trained and verified with leave-one-out cross-validation. The results demonstrated that our CMI-Net outperformed the existing algorithms with high precision (79.74%), recall (79.57%), F1-score (79.02%), and accuracy (93.37%). The adoption of CB focal loss improved the performance of CMI-Net, with increases of 2.76%, 4.16%, and 3.92% in precision, recall, and F1-score, respectively. In conclusion, CMI-Net and CB focal loss effectively enhanced the equine activity classification performance using imbalanced multi-modal sensor data. |
format | Online Article Text |
id | pubmed-8434387 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-84343872021-09-12 Cross-Modality Interaction Network for Equine Activity Recognition Using Imbalanced Multi-Modal Data † Mao, Axiu Huang, Endai Gan, Haiming Parkes, Rebecca S. V. Xu, Weitao Liu, Kai Sensors (Basel) Article With the recent advances in deep learning, wearable sensors have increasingly been used in automated animal activity recognition. However, there are two major challenges in improving recognition performance—multi-modal feature fusion and imbalanced data modeling. In this study, to improve classification performance for equine activities while tackling these two challenges, we developed a cross-modality interaction network (CMI-Net) involving a dual convolution neural network architecture and a cross-modality interaction module (CMIM). The CMIM adaptively recalibrated the temporal- and axis-wise features in each modality by leveraging multi-modal information to achieve deep intermodality interaction. A class-balanced (CB) focal loss was adopted to supervise the training of CMI-Net to alleviate the class imbalance problem. Motion data was acquired from six neck-attached inertial measurement units from six horses. The CMI-Net was trained and verified with leave-one-out cross-validation. The results demonstrated that our CMI-Net outperformed the existing algorithms with high precision (79.74%), recall (79.57%), F1-score (79.02%), and accuracy (93.37%). The adoption of CB focal loss improved the performance of CMI-Net, with increases of 2.76%, 4.16%, and 3.92% in precision, recall, and F1-score, respectively. In conclusion, CMI-Net and CB focal loss effectively enhanced the equine activity classification performance using imbalanced multi-modal sensor data. MDPI 2021-08-29 /pmc/articles/PMC8434387/ /pubmed/34502709 http://dx.doi.org/10.3390/s21175818 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Mao, Axiu Huang, Endai Gan, Haiming Parkes, Rebecca S. V. Xu, Weitao Liu, Kai Cross-Modality Interaction Network for Equine Activity Recognition Using Imbalanced Multi-Modal Data † |
title | Cross-Modality Interaction Network for Equine Activity Recognition Using Imbalanced Multi-Modal Data † |
title_full | Cross-Modality Interaction Network for Equine Activity Recognition Using Imbalanced Multi-Modal Data † |
title_fullStr | Cross-Modality Interaction Network for Equine Activity Recognition Using Imbalanced Multi-Modal Data † |
title_full_unstemmed | Cross-Modality Interaction Network for Equine Activity Recognition Using Imbalanced Multi-Modal Data † |
title_short | Cross-Modality Interaction Network for Equine Activity Recognition Using Imbalanced Multi-Modal Data † |
title_sort | cross-modality interaction network for equine activity recognition using imbalanced multi-modal data † |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8434387/ https://www.ncbi.nlm.nih.gov/pubmed/34502709 http://dx.doi.org/10.3390/s21175818 |
work_keys_str_mv | AT maoaxiu crossmodalityinteractionnetworkforequineactivityrecognitionusingimbalancedmultimodaldata AT huangendai crossmodalityinteractionnetworkforequineactivityrecognitionusingimbalancedmultimodaldata AT ganhaiming crossmodalityinteractionnetworkforequineactivityrecognitionusingimbalancedmultimodaldata AT parkesrebeccasv crossmodalityinteractionnetworkforequineactivityrecognitionusingimbalancedmultimodaldata AT xuweitao crossmodalityinteractionnetworkforequineactivityrecognitionusingimbalancedmultimodaldata AT liukai crossmodalityinteractionnetworkforequineactivityrecognitionusingimbalancedmultimodaldata |