Cargando…

Why Cohen’s Kappa should be avoided as performance measure in classification

We show that Cohen’s Kappa and Matthews Correlation Coefficient (MCC), both extended and contrasted measures of performance in multi-class classification, are correlated in most situations, albeit can differ in others. Indeed, although in the symmetric case both match, we consider different unbalanc...

Descripción completa

Detalles Bibliográficos
Autores principales: Delgado, Rosario, Tibau, Xavier-Andoni
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6762152/
https://www.ncbi.nlm.nih.gov/pubmed/31557204
http://dx.doi.org/10.1371/journal.pone.0222916
_version_ 1783454158008877056
author Delgado, Rosario
Tibau, Xavier-Andoni
author_facet Delgado, Rosario
Tibau, Xavier-Andoni
author_sort Delgado, Rosario
collection PubMed
description We show that Cohen’s Kappa and Matthews Correlation Coefficient (MCC), both extended and contrasted measures of performance in multi-class classification, are correlated in most situations, albeit can differ in others. Indeed, although in the symmetric case both match, we consider different unbalanced situations in which Kappa exhibits an undesired behaviour, i.e. a worse classifier gets higher Kappa score, differing qualitatively from that of MCC. The debate about the incoherence in the behaviour of Kappa revolves around the convenience, or not, of using a relative metric, which makes the interpretation of its values difficult. We extend these concerns by showing that its pitfalls can go even further. Through experimentation, we present a novel approach to this topic. We carry on a comprehensive study that identifies an scenario in which the contradictory behaviour among MCC and Kappa emerges. Specifically, we find out that when there is a decrease to zero of the entropy of the elements out of the diagonal of the confusion matrix associated to a classifier, the discrepancy between Kappa and MCC rise, pointing to an anomalous performance of the former. We believe that this finding disables Kappa to be used in general as a performance measure to compare classifiers.
format Online
Article
Text
id pubmed-6762152
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-67621522019-10-13 Why Cohen’s Kappa should be avoided as performance measure in classification Delgado, Rosario Tibau, Xavier-Andoni PLoS One Research Article We show that Cohen’s Kappa and Matthews Correlation Coefficient (MCC), both extended and contrasted measures of performance in multi-class classification, are correlated in most situations, albeit can differ in others. Indeed, although in the symmetric case both match, we consider different unbalanced situations in which Kappa exhibits an undesired behaviour, i.e. a worse classifier gets higher Kappa score, differing qualitatively from that of MCC. The debate about the incoherence in the behaviour of Kappa revolves around the convenience, or not, of using a relative metric, which makes the interpretation of its values difficult. We extend these concerns by showing that its pitfalls can go even further. Through experimentation, we present a novel approach to this topic. We carry on a comprehensive study that identifies an scenario in which the contradictory behaviour among MCC and Kappa emerges. Specifically, we find out that when there is a decrease to zero of the entropy of the elements out of the diagonal of the confusion matrix associated to a classifier, the discrepancy between Kappa and MCC rise, pointing to an anomalous performance of the former. We believe that this finding disables Kappa to be used in general as a performance measure to compare classifiers. Public Library of Science 2019-09-26 /pmc/articles/PMC6762152/ /pubmed/31557204 http://dx.doi.org/10.1371/journal.pone.0222916 Text en © 2019 Delgado, Tibau http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Delgado, Rosario
Tibau, Xavier-Andoni
Why Cohen’s Kappa should be avoided as performance measure in classification
title Why Cohen’s Kappa should be avoided as performance measure in classification
title_full Why Cohen’s Kappa should be avoided as performance measure in classification
title_fullStr Why Cohen’s Kappa should be avoided as performance measure in classification
title_full_unstemmed Why Cohen’s Kappa should be avoided as performance measure in classification
title_short Why Cohen’s Kappa should be avoided as performance measure in classification
title_sort why cohen’s kappa should be avoided as performance measure in classification
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6762152/
https://www.ncbi.nlm.nih.gov/pubmed/31557204
http://dx.doi.org/10.1371/journal.pone.0222916
work_keys_str_mv AT delgadorosario whycohenskappashouldbeavoidedasperformancemeasureinclassification
AT tibauxavierandoni whycohenskappashouldbeavoidedasperformancemeasureinclassification