Cargando…

Self-Supervised Contrastive Learning for Medical Time Series: A Systematic Review

Medical time series are sequential data collected over time that measures health-related signals, such as electroencephalography (EEG), electrocardiography (ECG), and intensive care unit (ICU) readings. Analyzing medical time series and identifying the latent patterns and trends that lead to uncover...

Descripción completa

Detalles Bibliográficos
Autores principales: Liu, Ziyu, Alavi, Azadeh, Li, Minyi, Zhang, Xiang
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10181273/
https://www.ncbi.nlm.nih.gov/pubmed/37177423
http://dx.doi.org/10.3390/s23094221
_version_ 1785041534890016768
author Liu, Ziyu
Alavi, Azadeh
Li, Minyi
Zhang, Xiang
author_facet Liu, Ziyu
Alavi, Azadeh
Li, Minyi
Zhang, Xiang
author_sort Liu, Ziyu
collection PubMed
description Medical time series are sequential data collected over time that measures health-related signals, such as electroencephalography (EEG), electrocardiography (ECG), and intensive care unit (ICU) readings. Analyzing medical time series and identifying the latent patterns and trends that lead to uncovering highly valuable insights for enhancing diagnosis, treatment, risk assessment, and disease progression. However, data mining in medical time series is heavily limited by the sample annotation which is time-consuming and labor-intensive, and expert-depending. To mitigate this challenge, the emerging self-supervised contrastive learning, which has shown great success since 2020, is a promising solution. Contrastive learning aims to learn representative embeddings by contrasting positive and negative samples without the requirement for explicit labels. Here, we conducted a systematic review of how contrastive learning alleviates the label scarcity in medical time series based on PRISMA standards. We searched the studies in five scientific databases (IEEE, ACM, Scopus, Google Scholar, and PubMed) and retrieved 1908 papers based on the inclusion criteria. After applying excluding criteria, and screening at title, abstract, and full text levels, we carefully reviewed 43 papers in this area. Specifically, this paper outlines the pipeline of contrastive learning, including pre-training, fine-tuning, and testing. We provide a comprehensive summary of the various augmentations applied to medical time series data, the architectures of pre-training encoders, the types of fine-tuning classifiers and clusters, and the popular contrastive loss functions. Moreover, we present an overview of the different data types used in medical time series, highlight the medical applications of interest, and provide a comprehensive table of 51 public datasets that have been utilized in this field. In addition, this paper will provide a discussion on the promising future scopes such as providing guidance for effective augmentation design, developing a unified framework for analyzing hierarchical time series, and investigating methods for processing multimodal data. Despite being in its early stages, self-supervised contrastive learning has shown great potential in overcoming the need for expert-created annotations in the research of medical time series.
format Online
Article
Text
id pubmed-10181273
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-101812732023-05-13 Self-Supervised Contrastive Learning for Medical Time Series: A Systematic Review Liu, Ziyu Alavi, Azadeh Li, Minyi Zhang, Xiang Sensors (Basel) Systematic Review Medical time series are sequential data collected over time that measures health-related signals, such as electroencephalography (EEG), electrocardiography (ECG), and intensive care unit (ICU) readings. Analyzing medical time series and identifying the latent patterns and trends that lead to uncovering highly valuable insights for enhancing diagnosis, treatment, risk assessment, and disease progression. However, data mining in medical time series is heavily limited by the sample annotation which is time-consuming and labor-intensive, and expert-depending. To mitigate this challenge, the emerging self-supervised contrastive learning, which has shown great success since 2020, is a promising solution. Contrastive learning aims to learn representative embeddings by contrasting positive and negative samples without the requirement for explicit labels. Here, we conducted a systematic review of how contrastive learning alleviates the label scarcity in medical time series based on PRISMA standards. We searched the studies in five scientific databases (IEEE, ACM, Scopus, Google Scholar, and PubMed) and retrieved 1908 papers based on the inclusion criteria. After applying excluding criteria, and screening at title, abstract, and full text levels, we carefully reviewed 43 papers in this area. Specifically, this paper outlines the pipeline of contrastive learning, including pre-training, fine-tuning, and testing. We provide a comprehensive summary of the various augmentations applied to medical time series data, the architectures of pre-training encoders, the types of fine-tuning classifiers and clusters, and the popular contrastive loss functions. Moreover, we present an overview of the different data types used in medical time series, highlight the medical applications of interest, and provide a comprehensive table of 51 public datasets that have been utilized in this field. In addition, this paper will provide a discussion on the promising future scopes such as providing guidance for effective augmentation design, developing a unified framework for analyzing hierarchical time series, and investigating methods for processing multimodal data. Despite being in its early stages, self-supervised contrastive learning has shown great potential in overcoming the need for expert-created annotations in the research of medical time series. MDPI 2023-04-23 /pmc/articles/PMC10181273/ /pubmed/37177423 http://dx.doi.org/10.3390/s23094221 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Systematic Review
Liu, Ziyu
Alavi, Azadeh
Li, Minyi
Zhang, Xiang
Self-Supervised Contrastive Learning for Medical Time Series: A Systematic Review
title Self-Supervised Contrastive Learning for Medical Time Series: A Systematic Review
title_full Self-Supervised Contrastive Learning for Medical Time Series: A Systematic Review
title_fullStr Self-Supervised Contrastive Learning for Medical Time Series: A Systematic Review
title_full_unstemmed Self-Supervised Contrastive Learning for Medical Time Series: A Systematic Review
title_short Self-Supervised Contrastive Learning for Medical Time Series: A Systematic Review
title_sort self-supervised contrastive learning for medical time series: a systematic review
topic Systematic Review
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10181273/
https://www.ncbi.nlm.nih.gov/pubmed/37177423
http://dx.doi.org/10.3390/s23094221
work_keys_str_mv AT liuziyu selfsupervisedcontrastivelearningformedicaltimeseriesasystematicreview
AT alaviazadeh selfsupervisedcontrastivelearningformedicaltimeseriesasystematicreview
AT liminyi selfsupervisedcontrastivelearningformedicaltimeseriesasystematicreview
AT zhangxiang selfsupervisedcontrastivelearningformedicaltimeseriesasystematicreview