Cargando…

Evaluating generative models in high energy physics

There has been a recent explosion in research into machine-learning-based generative modeling to tackle computational challenges for simulations in high energy physics (HEP). In order to use such alternative simulators in practice, we need well-defined metrics to compare different generative models...

Descripción completa

Detalles Bibliográficos
Autores principales: Kansal, Raghav, Li, Anni, Duarte, Javier, Chernyavskaya, Nadezda, Pierini, Maurizio, Orzari, Breno, Tomei, Thiago
Lenguaje:eng
Publicado: 2022
Materias:
Acceso en línea:https://dx.doi.org/10.1103/PhysRevD.107.076017
http://cds.cern.ch/record/2841787
_version_ 1780976201467166720
author Kansal, Raghav
Li, Anni
Duarte, Javier
Chernyavskaya, Nadezda
Pierini, Maurizio
Orzari, Breno
Tomei, Thiago
author_facet Kansal, Raghav
Li, Anni
Duarte, Javier
Chernyavskaya, Nadezda
Pierini, Maurizio
Orzari, Breno
Tomei, Thiago
author_sort Kansal, Raghav
collection CERN
description There has been a recent explosion in research into machine-learning-based generative modeling to tackle computational challenges for simulations in high energy physics (HEP). In order to use such alternative simulators in practice, we need well-defined metrics to compare different generative models and evaluate their discrepancy from the true distributions. We present the first systematic review and investigation into evaluation metrics and their sensitivity to failure modes of generative models, using the framework of two-sample goodness-of-fit testing, and their relevance and viability for HEP. Inspired by previous work in both physics and computer vision, we propose two new metrics, the Fréchet and kernel physics distances (FPD and KPD, respectively) and perform a variety of experiments measuring their performance on simple Gaussian-distributed and simulated high energy jet datasets. We find FPD, in particular, to be the most sensitive metric to all alternative jet distributions tested and recommend its adoption, along with the KPD and Wasserstein distances between individual feature distributions, for evaluating generative models in HEP. We finally demonstrate the efficacy of these proposed metrics in evaluating and comparing a novel attention-based generative adversarial particle transformer to the state-of-the-art message-passing generative adversarial network jet simulation model. The code for our proposed metrics is provided in the open source jetnet python library.
id cern-2841787
institution Organización Europea para la Investigación Nuclear
language eng
publishDate 2022
record_format invenio
spelling cern-28417872023-05-26T02:21:17Zdoi:10.1103/PhysRevD.107.076017http://cds.cern.ch/record/2841787engKansal, RaghavLi, AnniDuarte, JavierChernyavskaya, NadezdaPierini, MaurizioOrzari, BrenoTomei, ThiagoEvaluating generative models in high energy physicsstat.APMathematical Physics and Mathematicscs.LGComputing and Computershep-exParticle Physics - ExperimentThere has been a recent explosion in research into machine-learning-based generative modeling to tackle computational challenges for simulations in high energy physics (HEP). In order to use such alternative simulators in practice, we need well-defined metrics to compare different generative models and evaluate their discrepancy from the true distributions. We present the first systematic review and investigation into evaluation metrics and their sensitivity to failure modes of generative models, using the framework of two-sample goodness-of-fit testing, and their relevance and viability for HEP. Inspired by previous work in both physics and computer vision, we propose two new metrics, the Fréchet and kernel physics distances (FPD and KPD, respectively) and perform a variety of experiments measuring their performance on simple Gaussian-distributed and simulated high energy jet datasets. We find FPD, in particular, to be the most sensitive metric to all alternative jet distributions tested and recommend its adoption, along with the KPD and Wasserstein distances between individual feature distributions, for evaluating generative models in HEP. We finally demonstrate the efficacy of these proposed metrics in evaluating and comparing a novel attention-based generative adversarial particle transformer to the state-of-the-art message-passing generative adversarial network jet simulation model. The code for our proposed metrics is provided in the open source jetnet python library.There has been a recent explosion in research into machine-learning-based generative modeling to tackle computational challenges for simulations in high energy physics (HEP). In order to use such alternative simulators in practice, we need well-defined metrics to compare different generative models and evaluate their discrepancy from the true distributions. We present the first systematic review and investigation into evaluation metrics and their sensitivity to failure modes of generative models, using the framework of two-sample goodness-of-fit testing, and their relevance and viability for HEP. Inspired by previous work in both physics and computer vision, we propose two new metrics, the Fréchet and kernel physics distances (FPD and KPD, respectively), and perform a variety of experiments measuring their performance on simple Gaussian-distributed, and simulated high energy jet datasets. We find FPD, in particular, to be the most sensitive metric to all alternative jet distributions tested and recommend its adoption, along with the KPD and Wasserstein distances between individual feature distributions, for evaluating generative models in HEP. We finally demonstrate the efficacy of these proposed metrics in evaluating and comparing a novel attention-based generative adversarial particle transformer to the state-of-the-art message-passing generative adversarial network jet simulation model. The code for our proposed metrics is provided in the open source JetNet Python library.arXiv:2211.10295FERMILAB-PUB-22-872-CMS-PPDoai:cds.cern.ch:28417872022-11-18
spellingShingle stat.AP
Mathematical Physics and Mathematics
cs.LG
Computing and Computers
hep-ex
Particle Physics - Experiment
Kansal, Raghav
Li, Anni
Duarte, Javier
Chernyavskaya, Nadezda
Pierini, Maurizio
Orzari, Breno
Tomei, Thiago
Evaluating generative models in high energy physics
title Evaluating generative models in high energy physics
title_full Evaluating generative models in high energy physics
title_fullStr Evaluating generative models in high energy physics
title_full_unstemmed Evaluating generative models in high energy physics
title_short Evaluating generative models in high energy physics
title_sort evaluating generative models in high energy physics
topic stat.AP
Mathematical Physics and Mathematics
cs.LG
Computing and Computers
hep-ex
Particle Physics - Experiment
url https://dx.doi.org/10.1103/PhysRevD.107.076017
http://cds.cern.ch/record/2841787
work_keys_str_mv AT kansalraghav evaluatinggenerativemodelsinhighenergyphysics
AT lianni evaluatinggenerativemodelsinhighenergyphysics
AT duartejavier evaluatinggenerativemodelsinhighenergyphysics
AT chernyavskayanadezda evaluatinggenerativemodelsinhighenergyphysics
AT pierinimaurizio evaluatinggenerativemodelsinhighenergyphysics
AT orzaribreno evaluatinggenerativemodelsinhighenergyphysics
AT tomeithiago evaluatinggenerativemodelsinhighenergyphysics