Cargando…
Transfer learning with LundNet
In this work we explore the use of transfer learning in the context of the study of the structure of jets in hadron collisions. We adopt the tool LundNet, a graph neural network for jet tagging, and we investigate the performance of the tagger in different transfer learning scenarios in comparison to...
Autor principal: | |
---|---|
Lenguaje: | eng |
Publicado: |
2021
|
Materias: | |
Acceso en línea: | http://cds.cern.ch/record/2784478 |
_version_ | 1780972105529032704 |
---|---|
author | Grabarczyk, Radoslaw Piotr |
author_facet | Grabarczyk, Radoslaw Piotr |
author_sort | Grabarczyk, Radoslaw Piotr |
collection | CERN |
description | In this work we explore the use of transfer learning in the context of the study of the structure of jets in hadron collisions. We adopt the tool LundNet, a graph neural network for jet tagging, and we investigate the performance of the tagger in different transfer learning scenarios in comparison to that of fully trained models. We find that the transfer of weights from early layers of a different pretrained tagger allows one to train new taggers with significantly reduced computational resources and time. We study in detail the performance of the resulting tagger, which nearly reproduces that of a model trained from scratch. This offers a promising avenue to facilitate the use of such tools in the context of collider physics experiments. |
id | cern-2784478 |
institution | Organización Europea para la Investigación Nuclear |
language | eng |
publishDate | 2021 |
record_format | invenio |
spelling | cern-27844782021-10-18T21:50:16Zhttp://cds.cern.ch/record/2784478engGrabarczyk, Radoslaw PiotrTransfer learning with LundNetComputing and ComputersIn this work we explore the use of transfer learning in the context of the study of the structure of jets in hadron collisions. We adopt the tool LundNet, a graph neural network for jet tagging, and we investigate the performance of the tagger in different transfer learning scenarios in comparison to that of fully trained models. We find that the transfer of weights from early layers of a different pretrained tagger allows one to train new taggers with significantly reduced computational resources and time. We study in detail the performance of the resulting tagger, which nearly reproduces that of a model trained from scratch. This offers a promising avenue to facilitate the use of such tools in the context of collider physics experiments.CERN-STUDENTS-Note-2021-211oai:cds.cern.ch:27844782021-10-18 |
spellingShingle | Computing and Computers Grabarczyk, Radoslaw Piotr Transfer learning with LundNet |
title | Transfer learning with LundNet |
title_full | Transfer learning with LundNet |
title_fullStr | Transfer learning with LundNet |
title_full_unstemmed | Transfer learning with LundNet |
title_short | Transfer learning with LundNet |
title_sort | transfer learning with lundnet |
topic | Computing and Computers |
url | http://cds.cern.ch/record/2784478 |
work_keys_str_mv | AT grabarczykradoslawpiotr transferlearningwithlundnet |