Cargando…

Qualitative analysis of manual annotations of clinical text with SNOMED CT

SNOMED CT provides about 300,000 codes with fine-grained concept definitions to support interoperability of health data. Coding clinical texts with medical terminologies it is not a trivial task and is prone to disagreements between coders. We conducted a qualitative analysis to identify sources of...

Descripción completa

Detalles Bibliográficos
Autores principales: Miñarro-Giménez, Jose Antonio, Martínez-Costa, Catalina, Karlsson, Daniel, Schulz, Stefan, Gøeg, Kirstine Rosenbeck
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2018
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6307753/
https://www.ncbi.nlm.nih.gov/pubmed/30589855
http://dx.doi.org/10.1371/journal.pone.0209547
_version_ 1783383060393230336
author Miñarro-Giménez, Jose Antonio
Martínez-Costa, Catalina
Karlsson, Daniel
Schulz, Stefan
Gøeg, Kirstine Rosenbeck
author_facet Miñarro-Giménez, Jose Antonio
Martínez-Costa, Catalina
Karlsson, Daniel
Schulz, Stefan
Gøeg, Kirstine Rosenbeck
author_sort Miñarro-Giménez, Jose Antonio
collection PubMed
description SNOMED CT provides about 300,000 codes with fine-grained concept definitions to support interoperability of health data. Coding clinical texts with medical terminologies it is not a trivial task and is prone to disagreements between coders. We conducted a qualitative analysis to identify sources of disagreements on an annotation experiment which used a subset of SNOMED CT with some restrictions. A corpus of 20 English clinical text fragments from diverse origins and languages was annotated independently by two domain medically trained annotators following a specific annotation guideline. By following this guideline, the annotators had to assign sets of SNOMED CT codes to noun phrases, together with concept and term coverage ratings. Then, the annotations were manually examined against a reference standard to determine sources of disagreements. Five categories were identified. In our results, the most frequent cause of inter-annotator disagreement was related to human issues. In several cases disagreements revealed gaps in the annotation guidelines and lack of training of annotators. The reminder issues can be influenced by some SNOMED CT features.
format Online
Article
Text
id pubmed-6307753
institution National Center for Biotechnology Information
language English
publishDate 2018
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-63077532019-01-08 Qualitative analysis of manual annotations of clinical text with SNOMED CT Miñarro-Giménez, Jose Antonio Martínez-Costa, Catalina Karlsson, Daniel Schulz, Stefan Gøeg, Kirstine Rosenbeck PLoS One Research Article SNOMED CT provides about 300,000 codes with fine-grained concept definitions to support interoperability of health data. Coding clinical texts with medical terminologies it is not a trivial task and is prone to disagreements between coders. We conducted a qualitative analysis to identify sources of disagreements on an annotation experiment which used a subset of SNOMED CT with some restrictions. A corpus of 20 English clinical text fragments from diverse origins and languages was annotated independently by two domain medically trained annotators following a specific annotation guideline. By following this guideline, the annotators had to assign sets of SNOMED CT codes to noun phrases, together with concept and term coverage ratings. Then, the annotations were manually examined against a reference standard to determine sources of disagreements. Five categories were identified. In our results, the most frequent cause of inter-annotator disagreement was related to human issues. In several cases disagreements revealed gaps in the annotation guidelines and lack of training of annotators. The reminder issues can be influenced by some SNOMED CT features. Public Library of Science 2018-12-27 /pmc/articles/PMC6307753/ /pubmed/30589855 http://dx.doi.org/10.1371/journal.pone.0209547 Text en © 2018 Miñarro-Giménez et al http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Miñarro-Giménez, Jose Antonio
Martínez-Costa, Catalina
Karlsson, Daniel
Schulz, Stefan
Gøeg, Kirstine Rosenbeck
Qualitative analysis of manual annotations of clinical text with SNOMED CT
title Qualitative analysis of manual annotations of clinical text with SNOMED CT
title_full Qualitative analysis of manual annotations of clinical text with SNOMED CT
title_fullStr Qualitative analysis of manual annotations of clinical text with SNOMED CT
title_full_unstemmed Qualitative analysis of manual annotations of clinical text with SNOMED CT
title_short Qualitative analysis of manual annotations of clinical text with SNOMED CT
title_sort qualitative analysis of manual annotations of clinical text with snomed ct
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6307753/
https://www.ncbi.nlm.nih.gov/pubmed/30589855
http://dx.doi.org/10.1371/journal.pone.0209547
work_keys_str_mv AT minarrogimenezjoseantonio qualitativeanalysisofmanualannotationsofclinicaltextwithsnomedct
AT martinezcostacatalina qualitativeanalysisofmanualannotationsofclinicaltextwithsnomedct
AT karlssondaniel qualitativeanalysisofmanualannotationsofclinicaltextwithsnomedct
AT schulzstefan qualitativeanalysisofmanualannotationsofclinicaltextwithsnomedct
AT gøegkirstinerosenbeck qualitativeanalysisofmanualannotationsofclinicaltextwithsnomedct