Cargando…
Hybrid fNIRS-EEG based classification of auditory and visual perception processes
For multimodal Human-Computer Interaction (HCI), it is very useful to identify the modalities on which the user is currently processing information. This would enable a system to select complementary output modalities to reduce the user's workload. In this paper, we develop a hybrid Brain-Compu...
Autores principales: | , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2014
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4235375/ https://www.ncbi.nlm.nih.gov/pubmed/25477777 http://dx.doi.org/10.3389/fnins.2014.00373 |
_version_ | 1782345015570202624 |
---|---|
author | Putze, Felix Hesslinger, Sebastian Tse, Chun-Yu Huang, YunYing Herff, Christian Guan, Cuntai Schultz, Tanja |
author_facet | Putze, Felix Hesslinger, Sebastian Tse, Chun-Yu Huang, YunYing Herff, Christian Guan, Cuntai Schultz, Tanja |
author_sort | Putze, Felix |
collection | PubMed |
description | For multimodal Human-Computer Interaction (HCI), it is very useful to identify the modalities on which the user is currently processing information. This would enable a system to select complementary output modalities to reduce the user's workload. In this paper, we develop a hybrid Brain-Computer Interface (BCI) which uses Electroencephalography (EEG) and functional Near Infrared Spectroscopy (fNIRS) to discriminate and detect visual and auditory stimulus processing. We describe the experimental setup we used for collection of our data corpus with 12 subjects. On this data, we performed cross-validation evaluation, of which we report accuracy for different classification conditions. The results show that the subject-dependent systems achieved a classification accuracy of 97.8% for discriminating visual and auditory perception processes from each other and a classification accuracy of up to 94.8% for detecting modality-specific processes independently of other cognitive activity. The same classification conditions could also be discriminated in a subject-independent fashion with accuracy of up to 94.6 and 86.7%, respectively. We also look at the contributions of the two signal types and show that the fusion of classifiers using different features significantly increases accuracy. |
format | Online Article Text |
id | pubmed-4235375 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2014 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-42353752014-12-04 Hybrid fNIRS-EEG based classification of auditory and visual perception processes Putze, Felix Hesslinger, Sebastian Tse, Chun-Yu Huang, YunYing Herff, Christian Guan, Cuntai Schultz, Tanja Front Neurosci Neuroscience For multimodal Human-Computer Interaction (HCI), it is very useful to identify the modalities on which the user is currently processing information. This would enable a system to select complementary output modalities to reduce the user's workload. In this paper, we develop a hybrid Brain-Computer Interface (BCI) which uses Electroencephalography (EEG) and functional Near Infrared Spectroscopy (fNIRS) to discriminate and detect visual and auditory stimulus processing. We describe the experimental setup we used for collection of our data corpus with 12 subjects. On this data, we performed cross-validation evaluation, of which we report accuracy for different classification conditions. The results show that the subject-dependent systems achieved a classification accuracy of 97.8% for discriminating visual and auditory perception processes from each other and a classification accuracy of up to 94.8% for detecting modality-specific processes independently of other cognitive activity. The same classification conditions could also be discriminated in a subject-independent fashion with accuracy of up to 94.6 and 86.7%, respectively. We also look at the contributions of the two signal types and show that the fusion of classifiers using different features significantly increases accuracy. Frontiers Media S.A. 2014-11-18 /pmc/articles/PMC4235375/ /pubmed/25477777 http://dx.doi.org/10.3389/fnins.2014.00373 Text en Copyright © 2014 Putze, Hesslinger, Tse, Huang, Herff, Guan and Schultz. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Neuroscience Putze, Felix Hesslinger, Sebastian Tse, Chun-Yu Huang, YunYing Herff, Christian Guan, Cuntai Schultz, Tanja Hybrid fNIRS-EEG based classification of auditory and visual perception processes |
title | Hybrid fNIRS-EEG based classification of auditory and visual perception processes |
title_full | Hybrid fNIRS-EEG based classification of auditory and visual perception processes |
title_fullStr | Hybrid fNIRS-EEG based classification of auditory and visual perception processes |
title_full_unstemmed | Hybrid fNIRS-EEG based classification of auditory and visual perception processes |
title_short | Hybrid fNIRS-EEG based classification of auditory and visual perception processes |
title_sort | hybrid fnirs-eeg based classification of auditory and visual perception processes |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4235375/ https://www.ncbi.nlm.nih.gov/pubmed/25477777 http://dx.doi.org/10.3389/fnins.2014.00373 |
work_keys_str_mv | AT putzefelix hybridfnirseegbasedclassificationofauditoryandvisualperceptionprocesses AT hesslingersebastian hybridfnirseegbasedclassificationofauditoryandvisualperceptionprocesses AT tsechunyu hybridfnirseegbasedclassificationofauditoryandvisualperceptionprocesses AT huangyunying hybridfnirseegbasedclassificationofauditoryandvisualperceptionprocesses AT herffchristian hybridfnirseegbasedclassificationofauditoryandvisualperceptionprocesses AT guancuntai hybridfnirseegbasedclassificationofauditoryandvisualperceptionprocesses AT schultztanja hybridfnirseegbasedclassificationofauditoryandvisualperceptionprocesses |