Cargando…

ATLAS Distributed Computing in LHC Run2

The ATLAS Distributed Computing infrastructure has evolved after the first period of LHC data taking in order to cope with the challenges of the upcoming LHC Run2. An increased data rate and computing demands of the Monte-Carlo simulation, as well as new approaches to ATLAS analysis, dictated a more...

Descripción completa

Detalles Bibliográficos
Autor principal: Campana, Simone
Lenguaje:eng
Publicado: 2015
Materias:
Acceso en línea:http://cds.cern.ch/record/2002420
_version_ 1780946050532507648
author Campana, Simone
author_facet Campana, Simone
author_sort Campana, Simone
collection CERN
description The ATLAS Distributed Computing infrastructure has evolved after the first period of LHC data taking in order to cope with the challenges of the upcoming LHC Run2. An increased data rate and computing demands of the Monte-Carlo simulation, as well as new approaches to ATLAS analysis, dictated a more dynamic workload management system (ProdSys2) and data management system (Rucio), overcoming the boundaries imposed by the design of the old computing model. In particular, the commissioning of new central computing system components was the core part of the migration toward the flexible computing model. The flexible computing utilization exploring the opportunistic resources such as HPC, cloud, and volunteer computing is embedded in the new computing model, the data access mechanisms have been enhanced with the remote access, and the network topology and performance is deeply integrated into the core of the system. Moreover a new data management strategy, based on defined lifetime for each dataset, has been defined to better manage the lifecycle of the data. In this note, the overview of the operational experience of the new system and its evolution is presented.
id cern-2002420
institution Organización Europea para la Investigación Nuclear
language eng
publishDate 2015
record_format invenio
spelling cern-20024202019-09-30T06:29:59Zhttp://cds.cern.ch/record/2002420engCampana, SimoneATLAS Distributed Computing in LHC Run2Particle Physics - ExperimentThe ATLAS Distributed Computing infrastructure has evolved after the first period of LHC data taking in order to cope with the challenges of the upcoming LHC Run2. An increased data rate and computing demands of the Monte-Carlo simulation, as well as new approaches to ATLAS analysis, dictated a more dynamic workload management system (ProdSys2) and data management system (Rucio), overcoming the boundaries imposed by the design of the old computing model. In particular, the commissioning of new central computing system components was the core part of the migration toward the flexible computing model. The flexible computing utilization exploring the opportunistic resources such as HPC, cloud, and volunteer computing is embedded in the new computing model, the data access mechanisms have been enhanced with the remote access, and the network topology and performance is deeply integrated into the core of the system. Moreover a new data management strategy, based on defined lifetime for each dataset, has been defined to better manage the lifecycle of the data. In this note, the overview of the operational experience of the new system and its evolution is presented.ATL-SOFT-SLIDE-2015-077oai:cds.cern.ch:20024202015-03-19
spellingShingle Particle Physics - Experiment
Campana, Simone
ATLAS Distributed Computing in LHC Run2
title ATLAS Distributed Computing in LHC Run2
title_full ATLAS Distributed Computing in LHC Run2
title_fullStr ATLAS Distributed Computing in LHC Run2
title_full_unstemmed ATLAS Distributed Computing in LHC Run2
title_short ATLAS Distributed Computing in LHC Run2
title_sort atlas distributed computing in lhc run2
topic Particle Physics - Experiment
url http://cds.cern.ch/record/2002420
work_keys_str_mv AT campanasimone atlasdistributedcomputinginlhcrun2