Cargando…
Evolution of the ATLAS Distributed Computing during the LHC long shutdown
The ATLAS Distributed Computing project (ADC) was established in 2007 to develop and operate a framework, following the ATLAS computing model, to enable data storage, processing and bookkeeping on top of the WLCG distributed infrastructure. ADC development has always been driven by operations and th...
Autor principal: | |
---|---|
Lenguaje: | eng |
Publicado: |
2013
|
Materias: | |
Acceso en línea: | http://cds.cern.ch/record/1606536 |
_version_ | 1780931687571521536 |
---|---|
author | Campana, S |
author_facet | Campana, S |
author_sort | Campana, S |
collection | CERN |
description | The ATLAS Distributed Computing project (ADC) was established in 2007 to develop and operate a framework, following the ATLAS computing model, to enable data storage, processing and bookkeeping on top of the WLCG distributed infrastructure. ADC development has always been driven by operations and this contributed to its success. The system has fulfilled the demanding requirements of ATLAS, daily consolidating worldwide up to 1PB of data and running more than 1.5 million payloads distributed globally, supporting almost one thousand concurrent distributed analysis users. Comprehensive automation and monitoring minimized the operational manpower required. The flexibility of the system to adjust to operational needs has been important to the success of the ATLAS physics program. The LHC shutdown in 2013-2015 affords an opportunity to improve the system in light of operational experience and scale it to cope with the demanding requirements of 2015 and beyond, most notably a much higher trigger rate and event pileup. We will describe the evolution of the ADC software foreseen during this period. This includes consolidating the existing Production and Distributed Analysis framework (PanDA) and ATLAS Grid Information System (AGIS), together with the development and commissioning of next generation systems for distributed data management (DDM/Rucio) and production (PRODSYS2). We will explain how new technologies such as Cloud Computing and NoSQL databases, which ATLAS investigated as R&D projects in past years, will be integrated in production. Finally, we will describe more fundamental developments such as breaking job-to-data locality by exploiting storage federations and caches, and event level (rather than file or dataset level) workload engines. |
id | cern-1606536 |
institution | Organización Europea para la Investigación Nuclear |
language | eng |
publishDate | 2013 |
record_format | invenio |
spelling | cern-16065362019-09-30T06:29:59Zhttp://cds.cern.ch/record/1606536engCampana, SEvolution of the ATLAS Distributed Computing during the LHC long shutdownDetectors and Experimental TechniquesThe ATLAS Distributed Computing project (ADC) was established in 2007 to develop and operate a framework, following the ATLAS computing model, to enable data storage, processing and bookkeeping on top of the WLCG distributed infrastructure. ADC development has always been driven by operations and this contributed to its success. The system has fulfilled the demanding requirements of ATLAS, daily consolidating worldwide up to 1PB of data and running more than 1.5 million payloads distributed globally, supporting almost one thousand concurrent distributed analysis users. Comprehensive automation and monitoring minimized the operational manpower required. The flexibility of the system to adjust to operational needs has been important to the success of the ATLAS physics program. The LHC shutdown in 2013-2015 affords an opportunity to improve the system in light of operational experience and scale it to cope with the demanding requirements of 2015 and beyond, most notably a much higher trigger rate and event pileup. We will describe the evolution of the ADC software foreseen during this period. This includes consolidating the existing Production and Distributed Analysis framework (PanDA) and ATLAS Grid Information System (AGIS), together with the development and commissioning of next generation systems for distributed data management (DDM/Rucio) and production (PRODSYS2). We will explain how new technologies such as Cloud Computing and NoSQL databases, which ATLAS investigated as R&D projects in past years, will be integrated in production. Finally, we will describe more fundamental developments such as breaking job-to-data locality by exploiting storage federations and caches, and event level (rather than file or dataset level) workload engines.ATL-SOFT-SLIDE-2013-804oai:cds.cern.ch:16065362013-10-07 |
spellingShingle | Detectors and Experimental Techniques Campana, S Evolution of the ATLAS Distributed Computing during the LHC long shutdown |
title | Evolution of the ATLAS Distributed Computing during the LHC long shutdown |
title_full | Evolution of the ATLAS Distributed Computing during the LHC long shutdown |
title_fullStr | Evolution of the ATLAS Distributed Computing during the LHC long shutdown |
title_full_unstemmed | Evolution of the ATLAS Distributed Computing during the LHC long shutdown |
title_short | Evolution of the ATLAS Distributed Computing during the LHC long shutdown |
title_sort | evolution of the atlas distributed computing during the lhc long shutdown |
topic | Detectors and Experimental Techniques |
url | http://cds.cern.ch/record/1606536 |
work_keys_str_mv | AT campanas evolutionoftheatlasdistributedcomputingduringthelhclongshutdown |