Cargando…
Evaluating generative models in high energy physics
There has been a recent explosion in research into machine-learning-based generative modeling to tackle computational challenges for simulations in high energy physics (HEP). In order to use such alternative simulators in practice, we need well-defined metrics to compare different generative models...
Autores principales: | , , , , , , |
---|---|
Lenguaje: | eng |
Publicado: |
2022
|
Materias: | |
Acceso en línea: | https://dx.doi.org/10.1103/PhysRevD.107.076017 http://cds.cern.ch/record/2841787 |
_version_ | 1780976201467166720 |
---|---|
author | Kansal, Raghav Li, Anni Duarte, Javier Chernyavskaya, Nadezda Pierini, Maurizio Orzari, Breno Tomei, Thiago |
author_facet | Kansal, Raghav Li, Anni Duarte, Javier Chernyavskaya, Nadezda Pierini, Maurizio Orzari, Breno Tomei, Thiago |
author_sort | Kansal, Raghav |
collection | CERN |
description | There has been a recent explosion in research into machine-learning-based generative modeling to tackle computational challenges for simulations in high energy physics (HEP). In order to use such alternative simulators in practice, we need well-defined metrics to compare different generative models and evaluate their discrepancy from the true distributions. We present the first systematic review and investigation into evaluation metrics and their sensitivity to failure modes of generative models, using the framework of two-sample goodness-of-fit testing, and their relevance and viability for HEP. Inspired by previous work in both physics and computer vision, we propose two new metrics, the Fréchet and kernel physics distances (FPD and KPD, respectively) and perform a variety of experiments measuring their performance on simple Gaussian-distributed and simulated high energy jet datasets. We find FPD, in particular, to be the most sensitive metric to all alternative jet distributions tested and recommend its adoption, along with the KPD and Wasserstein distances between individual feature distributions, for evaluating generative models in HEP. We finally demonstrate the efficacy of these proposed metrics in evaluating and comparing a novel attention-based generative adversarial particle transformer to the state-of-the-art message-passing generative adversarial network jet simulation model. The code for our proposed metrics is provided in the open source jetnet python library. |
id | cern-2841787 |
institution | Organización Europea para la Investigación Nuclear |
language | eng |
publishDate | 2022 |
record_format | invenio |
spelling | cern-28417872023-05-26T02:21:17Zdoi:10.1103/PhysRevD.107.076017http://cds.cern.ch/record/2841787engKansal, RaghavLi, AnniDuarte, JavierChernyavskaya, NadezdaPierini, MaurizioOrzari, BrenoTomei, ThiagoEvaluating generative models in high energy physicsstat.APMathematical Physics and Mathematicscs.LGComputing and Computershep-exParticle Physics - ExperimentThere has been a recent explosion in research into machine-learning-based generative modeling to tackle computational challenges for simulations in high energy physics (HEP). In order to use such alternative simulators in practice, we need well-defined metrics to compare different generative models and evaluate their discrepancy from the true distributions. We present the first systematic review and investigation into evaluation metrics and their sensitivity to failure modes of generative models, using the framework of two-sample goodness-of-fit testing, and their relevance and viability for HEP. Inspired by previous work in both physics and computer vision, we propose two new metrics, the Fréchet and kernel physics distances (FPD and KPD, respectively) and perform a variety of experiments measuring their performance on simple Gaussian-distributed and simulated high energy jet datasets. We find FPD, in particular, to be the most sensitive metric to all alternative jet distributions tested and recommend its adoption, along with the KPD and Wasserstein distances between individual feature distributions, for evaluating generative models in HEP. We finally demonstrate the efficacy of these proposed metrics in evaluating and comparing a novel attention-based generative adversarial particle transformer to the state-of-the-art message-passing generative adversarial network jet simulation model. The code for our proposed metrics is provided in the open source jetnet python library.There has been a recent explosion in research into machine-learning-based generative modeling to tackle computational challenges for simulations in high energy physics (HEP). In order to use such alternative simulators in practice, we need well-defined metrics to compare different generative models and evaluate their discrepancy from the true distributions. We present the first systematic review and investigation into evaluation metrics and their sensitivity to failure modes of generative models, using the framework of two-sample goodness-of-fit testing, and their relevance and viability for HEP. Inspired by previous work in both physics and computer vision, we propose two new metrics, the Fréchet and kernel physics distances (FPD and KPD, respectively), and perform a variety of experiments measuring their performance on simple Gaussian-distributed, and simulated high energy jet datasets. We find FPD, in particular, to be the most sensitive metric to all alternative jet distributions tested and recommend its adoption, along with the KPD and Wasserstein distances between individual feature distributions, for evaluating generative models in HEP. We finally demonstrate the efficacy of these proposed metrics in evaluating and comparing a novel attention-based generative adversarial particle transformer to the state-of-the-art message-passing generative adversarial network jet simulation model. The code for our proposed metrics is provided in the open source JetNet Python library.arXiv:2211.10295FERMILAB-PUB-22-872-CMS-PPDoai:cds.cern.ch:28417872022-11-18 |
spellingShingle | stat.AP Mathematical Physics and Mathematics cs.LG Computing and Computers hep-ex Particle Physics - Experiment Kansal, Raghav Li, Anni Duarte, Javier Chernyavskaya, Nadezda Pierini, Maurizio Orzari, Breno Tomei, Thiago Evaluating generative models in high energy physics |
title | Evaluating generative models in high energy physics |
title_full | Evaluating generative models in high energy physics |
title_fullStr | Evaluating generative models in high energy physics |
title_full_unstemmed | Evaluating generative models in high energy physics |
title_short | Evaluating generative models in high energy physics |
title_sort | evaluating generative models in high energy physics |
topic | stat.AP Mathematical Physics and Mathematics cs.LG Computing and Computers hep-ex Particle Physics - Experiment |
url | https://dx.doi.org/10.1103/PhysRevD.107.076017 http://cds.cern.ch/record/2841787 |
work_keys_str_mv | AT kansalraghav evaluatinggenerativemodelsinhighenergyphysics AT lianni evaluatinggenerativemodelsinhighenergyphysics AT duartejavier evaluatinggenerativemodelsinhighenergyphysics AT chernyavskayanadezda evaluatinggenerativemodelsinhighenergyphysics AT pierinimaurizio evaluatinggenerativemodelsinhighenergyphysics AT orzaribreno evaluatinggenerativemodelsinhighenergyphysics AT tomeithiago evaluatinggenerativemodelsinhighenergyphysics |