Cargando…

Motion Artifact Quantification and Sensor Fusion for Unobtrusive Health Monitoring

Sensors integrated into objects of everyday life potentially allow unobtrusive health monitoring at home. However, since the coupling of sensors and subject is not as well-defined as compared to a clinical setting, the signal quality is much more variable and can be disturbed significantly by motion...

Descripción completa

Detalles Bibliográficos
Autores principales: Hoog Antink, Christoph, Schulz, Florian, Leonhardt, Steffen, Walter, Marian
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2017
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5795602/
https://www.ncbi.nlm.nih.gov/pubmed/29295594
http://dx.doi.org/10.3390/s18010038
_version_ 1783297329000873984
author Hoog Antink, Christoph
Schulz, Florian
Leonhardt, Steffen
Walter, Marian
author_facet Hoog Antink, Christoph
Schulz, Florian
Leonhardt, Steffen
Walter, Marian
author_sort Hoog Antink, Christoph
collection PubMed
description Sensors integrated into objects of everyday life potentially allow unobtrusive health monitoring at home. However, since the coupling of sensors and subject is not as well-defined as compared to a clinical setting, the signal quality is much more variable and can be disturbed significantly by motion artifacts. One way of tackling this challenge is the combined evaluation of multiple channels via sensor fusion. For robust and accurate sensor fusion, analyzing the influence of motion on different modalities is crucial. In this work, a multimodal sensor setup integrated into an armchair is presented that combines capacitively coupled electrocardiography, reflective photoplethysmography, two high-frequency impedance sensors and two types of ballistocardiography sensors. To quantify motion artifacts, a motion protocol performed by healthy volunteers is recorded with a motion capture system, and reference sensors perform cardiorespiratory monitoring. The shape-based signal-to-noise ratio SNR [Formula: see text] is introduced and used to quantify the effect on motion on different sensing modalities. Based on this analysis, an optimal combination of sensors and fusion methodology is developed and evaluated. Using the proposed approach, beat-to-beat heart-rate is estimated with a coverage of 99.5% and a mean absolute error of 7.9 ms on 425 min of data from seven volunteers in a proof-of-concept measurement scenario.
format Online
Article
Text
id pubmed-5795602
institution National Center for Biotechnology Information
language English
publishDate 2017
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-57956022018-02-13 Motion Artifact Quantification and Sensor Fusion for Unobtrusive Health Monitoring Hoog Antink, Christoph Schulz, Florian Leonhardt, Steffen Walter, Marian Sensors (Basel) Article Sensors integrated into objects of everyday life potentially allow unobtrusive health monitoring at home. However, since the coupling of sensors and subject is not as well-defined as compared to a clinical setting, the signal quality is much more variable and can be disturbed significantly by motion artifacts. One way of tackling this challenge is the combined evaluation of multiple channels via sensor fusion. For robust and accurate sensor fusion, analyzing the influence of motion on different modalities is crucial. In this work, a multimodal sensor setup integrated into an armchair is presented that combines capacitively coupled electrocardiography, reflective photoplethysmography, two high-frequency impedance sensors and two types of ballistocardiography sensors. To quantify motion artifacts, a motion protocol performed by healthy volunteers is recorded with a motion capture system, and reference sensors perform cardiorespiratory monitoring. The shape-based signal-to-noise ratio SNR [Formula: see text] is introduced and used to quantify the effect on motion on different sensing modalities. Based on this analysis, an optimal combination of sensors and fusion methodology is developed and evaluated. Using the proposed approach, beat-to-beat heart-rate is estimated with a coverage of 99.5% and a mean absolute error of 7.9 ms on 425 min of data from seven volunteers in a proof-of-concept measurement scenario. MDPI 2017-12-25 /pmc/articles/PMC5795602/ /pubmed/29295594 http://dx.doi.org/10.3390/s18010038 Text en © 2017 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Hoog Antink, Christoph
Schulz, Florian
Leonhardt, Steffen
Walter, Marian
Motion Artifact Quantification and Sensor Fusion for Unobtrusive Health Monitoring
title Motion Artifact Quantification and Sensor Fusion for Unobtrusive Health Monitoring
title_full Motion Artifact Quantification and Sensor Fusion for Unobtrusive Health Monitoring
title_fullStr Motion Artifact Quantification and Sensor Fusion for Unobtrusive Health Monitoring
title_full_unstemmed Motion Artifact Quantification and Sensor Fusion for Unobtrusive Health Monitoring
title_short Motion Artifact Quantification and Sensor Fusion for Unobtrusive Health Monitoring
title_sort motion artifact quantification and sensor fusion for unobtrusive health monitoring
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5795602/
https://www.ncbi.nlm.nih.gov/pubmed/29295594
http://dx.doi.org/10.3390/s18010038
work_keys_str_mv AT hoogantinkchristoph motionartifactquantificationandsensorfusionforunobtrusivehealthmonitoring
AT schulzflorian motionartifactquantificationandsensorfusionforunobtrusivehealthmonitoring
AT leonhardtsteffen motionartifactquantificationandsensorfusionforunobtrusivehealthmonitoring
AT waltermarian motionartifactquantificationandsensorfusionforunobtrusivehealthmonitoring