Cargando…

Fisher information metrics for binary classifier evaluation and training

<!--HTML-->Different evaluation metrics for binary classifiers are appropriate to different scientific domains and even to different problems within the same domain. This presentation focuses on the optimisation of event selection to minimise statistical errors in HEP parameter estimation, a p...

Descripción completa

Detalles Bibliográficos
Autor principal: Valassi, Andrea
Lenguaje:eng
Publicado: 2018
Materias:
Acceso en línea:http://cds.cern.ch/record/2312462
_version_ 1780957974429171712
author Valassi, Andrea
author_facet Valassi, Andrea
author_sort Valassi, Andrea
collection CERN
description <!--HTML-->Different evaluation metrics for binary classifiers are appropriate to different scientific domains and even to different problems within the same domain. This presentation focuses on the optimisation of event selection to minimise statistical errors in HEP parameter estimation, a problem that is best analysed in terms of the maximisation of Fisher information about the measured parameters. After describing a general formalism to derive evaluation metrics based on Fisher information, three more specific metrics are introduced for the measurements of signal cross sections in counting experiments (FIP1) or distribution fits (FIP2) and for the measurements of other parameters from distribution fits (FIP3). The FIP2 metric is particularly interesting because it can be derived from any ROC curve, provided that prevalence is also known. In addition to its relation to measurement errors when used as an evaluation criterion (which makes it more interesting that the ROC AUC), a further advantage of the FIP2 metric is that it can also be directly used for training decision trees (instead of the Shannon entropy or Gini coefficient). Preliminary results based on the Python sklearn framework are presented. The problem of overtraining for these classifiers is also briefly discussed, in terms of the difference of the FIP2 metric on the validation and training set, and of their difference from the theoretical limit. Finally, the expected Fisher information gain from completely random branch splits in the decision tree and its possible relevance in reducing overtraining is analysed.
id cern-2312462
institution Organización Europea para la Investigación Nuclear
language eng
publishDate 2018
record_format invenio
spelling cern-23124622022-11-02T22:34:03Zhttp://cds.cern.ch/record/2312462engValassi, AndreaFisher information metrics for binary classifier evaluation and training2nd IML Machine Learning WorkshopMachine Learning<!--HTML-->Different evaluation metrics for binary classifiers are appropriate to different scientific domains and even to different problems within the same domain. This presentation focuses on the optimisation of event selection to minimise statistical errors in HEP parameter estimation, a problem that is best analysed in terms of the maximisation of Fisher information about the measured parameters. After describing a general formalism to derive evaluation metrics based on Fisher information, three more specific metrics are introduced for the measurements of signal cross sections in counting experiments (FIP1) or distribution fits (FIP2) and for the measurements of other parameters from distribution fits (FIP3). The FIP2 metric is particularly interesting because it can be derived from any ROC curve, provided that prevalence is also known. In addition to its relation to measurement errors when used as an evaluation criterion (which makes it more interesting that the ROC AUC), a further advantage of the FIP2 metric is that it can also be directly used for training decision trees (instead of the Shannon entropy or Gini coefficient). Preliminary results based on the Python sklearn framework are presented. The problem of overtraining for these classifiers is also briefly discussed, in terms of the difference of the FIP2 metric on the validation and training set, and of their difference from the theoretical limit. Finally, the expected Fisher information gain from completely random branch splits in the decision tree and its possible relevance in reducing overtraining is analysed.oai:cds.cern.ch:23124622018
spellingShingle Machine Learning
Valassi, Andrea
Fisher information metrics for binary classifier evaluation and training
title Fisher information metrics for binary classifier evaluation and training
title_full Fisher information metrics for binary classifier evaluation and training
title_fullStr Fisher information metrics for binary classifier evaluation and training
title_full_unstemmed Fisher information metrics for binary classifier evaluation and training
title_short Fisher information metrics for binary classifier evaluation and training
title_sort fisher information metrics for binary classifier evaluation and training
topic Machine Learning
url http://cds.cern.ch/record/2312462
work_keys_str_mv AT valassiandrea fisherinformationmetricsforbinaryclassifierevaluationandtraining
AT valassiandrea 2ndimlmachinelearningworkshop