Cargando…

Automated Intensity Optimisation Using Reinforcement Learning at LEIR

High intensities in the CERN Low Energy Ion Ring (LEIR) are achieved by stacking up to seven consecutive multi-turn injections from Linac3. Two inclined septa combined with a collapsing horizontal orbit bump allow a 6-D phase space painting via a linearly ramped mean momentum along the Linac3 pulse...

Descripción completa

Detalles Bibliográficos
Autores principales: Madysa, Nico, Alemany-Fernández, Reyes, Biancacci, Nicolo, Goddard, Brennan, Kain, Verena, Velotti, Francesco
Lenguaje:eng
Publicado: 2022
Materias:
Acceso en línea:https://dx.doi.org/10.18429/JACoW-IPAC2022-TUPOST040
http://cds.cern.ch/record/2845859
_version_ 1780976593648222208
author Madysa, Nico
Alemany-Fernández, Reyes
Biancacci, Nicolo
Goddard, Brennan
Kain, Verena
Velotti, Francesco
author_facet Madysa, Nico
Alemany-Fernández, Reyes
Biancacci, Nicolo
Goddard, Brennan
Kain, Verena
Velotti, Francesco
author_sort Madysa, Nico
collection CERN
description High intensities in the CERN Low Energy Ion Ring (LEIR) are achieved by stacking up to seven consecutive multi-turn injections from Linac3. Two inclined septa combined with a collapsing horizontal orbit bump allow a 6-D phase space painting via a linearly ramped mean momentum along the Linac3 pulse and injection at high dispersion. The beam is cooled and dragged longitudinally via electron cooling (e-cooling) into a stacking momentum. For optimal accumulation, the electron energy and trajectory need to match the ion energy and orbit at the e-cooler section. In this paper, a reinforcement learning (RL) agent is trained to adjust various e-cooler and Linac3 parameters to maximise the intensity at the end of the injection plateau. Variational Auto-Encoders (VAE) are used to compress longitudinal Schottky spectra into a compact representation as input for the RL agent. The RL agent is pre-trained on a surrogate model of the LEIR e-cooling dynamics, which in turn is learned from the data collected for the training of the VAE. The performance of the VAE, the surrogate model, and the RL agent is investigated in this paper. An overview of planned tests in the upcoming LEIR runs is given.
id cern-2845859
institution Organización Europea para la Investigación Nuclear
language eng
publishDate 2022
record_format invenio
spelling cern-28458592023-01-11T21:31:26Zdoi:10.18429/JACoW-IPAC2022-TUPOST040http://cds.cern.ch/record/2845859engMadysa, NicoAlemany-Fernández, ReyesBiancacci, NicoloGoddard, BrennanKain, VerenaVelotti, FrancescoAutomated Intensity Optimisation Using Reinforcement Learning at LEIRAccelerators and Storage RingsHigh intensities in the CERN Low Energy Ion Ring (LEIR) are achieved by stacking up to seven consecutive multi-turn injections from Linac3. Two inclined septa combined with a collapsing horizontal orbit bump allow a 6-D phase space painting via a linearly ramped mean momentum along the Linac3 pulse and injection at high dispersion. The beam is cooled and dragged longitudinally via electron cooling (e-cooling) into a stacking momentum. For optimal accumulation, the electron energy and trajectory need to match the ion energy and orbit at the e-cooler section. In this paper, a reinforcement learning (RL) agent is trained to adjust various e-cooler and Linac3 parameters to maximise the intensity at the end of the injection plateau. Variational Auto-Encoders (VAE) are used to compress longitudinal Schottky spectra into a compact representation as input for the RL agent. The RL agent is pre-trained on a surrogate model of the LEIR e-cooling dynamics, which in turn is learned from the data collected for the training of the VAE. The performance of the VAE, the surrogate model, and the RL agent is investigated in this paper. An overview of planned tests in the upcoming LEIR runs is given.oai:cds.cern.ch:28458592022
spellingShingle Accelerators and Storage Rings
Madysa, Nico
Alemany-Fernández, Reyes
Biancacci, Nicolo
Goddard, Brennan
Kain, Verena
Velotti, Francesco
Automated Intensity Optimisation Using Reinforcement Learning at LEIR
title Automated Intensity Optimisation Using Reinforcement Learning at LEIR
title_full Automated Intensity Optimisation Using Reinforcement Learning at LEIR
title_fullStr Automated Intensity Optimisation Using Reinforcement Learning at LEIR
title_full_unstemmed Automated Intensity Optimisation Using Reinforcement Learning at LEIR
title_short Automated Intensity Optimisation Using Reinforcement Learning at LEIR
title_sort automated intensity optimisation using reinforcement learning at leir
topic Accelerators and Storage Rings
url https://dx.doi.org/10.18429/JACoW-IPAC2022-TUPOST040
http://cds.cern.ch/record/2845859
work_keys_str_mv AT madysanico automatedintensityoptimisationusingreinforcementlearningatleir
AT alemanyfernandezreyes automatedintensityoptimisationusingreinforcementlearningatleir
AT biancaccinicolo automatedintensityoptimisationusingreinforcementlearningatleir
AT goddardbrennan automatedintensityoptimisationusingreinforcementlearningatleir
AT kainverena automatedintensityoptimisationusingreinforcementlearningatleir
AT velottifrancesco automatedintensityoptimisationusingreinforcementlearningatleir