Cargando…

Emognition dataset: emotion recognition with self-reports, facial expressions, and physiology using wearables

The Emognition dataset is dedicated to testing methods for emotion recognition (ER) from physiological responses and facial expressions. We collected data from 43 participants who watched short film clips eliciting nine discrete emotions: amusement, awe, enthusiasm, liking, surprise, anger, disgust,...

Descripción completa

Detalles Bibliográficos
Autores principales: Saganowski, Stanisław, Komoszyńska, Joanna, Behnke, Maciej, Perz, Bartosz, Kunc, Dominika, Klich, Bartłomiej, Kaczmarek, Łukasz D., Kazienko, Przemysław
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8989970/
https://www.ncbi.nlm.nih.gov/pubmed/35393434
http://dx.doi.org/10.1038/s41597-022-01262-0
Descripción
Sumario:The Emognition dataset is dedicated to testing methods for emotion recognition (ER) from physiological responses and facial expressions. We collected data from 43 participants who watched short film clips eliciting nine discrete emotions: amusement, awe, enthusiasm, liking, surprise, anger, disgust, fear, and sadness. Three wearables were used to record physiological data: EEG, BVP (2x), HR, EDA, SKT, ACC (3x), and GYRO (2x); in parallel with the upper-body videos. After each film clip, participants completed two types of self-reports: (1) related to nine discrete emotions and (2) three affective dimensions: valence, arousal, and motivation. The obtained data facilitates various ER approaches, e.g., multimodal ER, EEG- vs. cardiovascular-based ER, discrete to dimensional representation transitions. The technical validation indicated that watching film clips elicited the targeted emotions. It also supported signals’ high quality.