Cargando…

Performance evaluation of distributed file systems for the phase-II upgrade of the ATLAS experiment at CERN

Over the next few years, the LHC will prepare for the upcoming High-Luminosity upgrade in which it is expected to deliver ten times more p-p collisions. This will create a harsher radiation environment and higher detector occupancy. In this context, the ATLAS experiment, one of the general purpose e...

Descripción completa

Detalles Bibliográficos
Autores principales: Abed Abud, Adam, Le Goff, Fabrice, Avolio, Giuseppe
Lenguaje:eng
Publicado: 2019
Materias:
Acceso en línea:http://cds.cern.ch/record/2667379
_version_ 1780962063251668992
author Abed Abud, Adam
Le Goff, Fabrice
Avolio, Giuseppe
author_facet Abed Abud, Adam
Le Goff, Fabrice
Avolio, Giuseppe
author_sort Abed Abud, Adam
collection CERN
description Over the next few years, the LHC will prepare for the upcoming High-Luminosity upgrade in which it is expected to deliver ten times more p-p collisions. This will create a harsher radiation environment and higher detector occupancy. In this context, the ATLAS experiment, one of the general purpose experiments at the LHC, plans substantial upgrades to the detectors and to the trigger system in order to efficiently select events. Similarly, the Data Acquisition System (DAQ) will have to redesign the data-flow architecture to accommodate for the large increase in event and data rates. The Phase-II DAQ design involves a large distributed storage system that buffers data read out from the detector, while a computing farm (Event Filter) analyzes and selects the most interesting events. This system will have to handle 5.2 TB/s of input data for an event rate of 1 MHz and provide access to 3 TB/s of these data to the filtering farm. A possible implementation for such a design is based on distributed file systems (DFS) which are becoming unavoidable among the big data industry. Features of DFS such as replication strategies and smart placement policies match the distributed nature and the requirements of the new data-flow system. This paper presents an up-to-date performance evaluation of some of the DFS currently available: GlusterFS, HadoopFS and CephFS. After characterization of the future data-flow system’s workload, we report on small-scale raw performance and scalability studies. Finally, we conclude on the suitability of such systems to the tight constraints expected for the ATLAS experiment in phase-II and, in general, what the HEP community can profit from these storage technologies.
id cern-2667379
institution Organización Europea para la Investigación Nuclear
language eng
publishDate 2019
record_format invenio
spelling cern-26673792020-07-29T14:54:28Zhttp://cds.cern.ch/record/2667379engAbed Abud, AdamLe Goff, FabriceAvolio, GiuseppePerformance evaluation of distributed file systems for the phase-II upgrade of the ATLAS experiment at CERNParticle Physics - ExperimentOver the next few years, the LHC will prepare for the upcoming High-Luminosity upgrade in which it is expected to deliver ten times more p-p collisions. This will create a harsher radiation environment and higher detector occupancy. In this context, the ATLAS experiment, one of the general purpose experiments at the LHC, plans substantial upgrades to the detectors and to the trigger system in order to efficiently select events. Similarly, the Data Acquisition System (DAQ) will have to redesign the data-flow architecture to accommodate for the large increase in event and data rates. The Phase-II DAQ design involves a large distributed storage system that buffers data read out from the detector, while a computing farm (Event Filter) analyzes and selects the most interesting events. This system will have to handle 5.2 TB/s of input data for an event rate of 1 MHz and provide access to 3 TB/s of these data to the filtering farm. A possible implementation for such a design is based on distributed file systems (DFS) which are becoming unavoidable among the big data industry. Features of DFS such as replication strategies and smart placement policies match the distributed nature and the requirements of the new data-flow system. This paper presents an up-to-date performance evaluation of some of the DFS currently available: GlusterFS, HadoopFS and CephFS. After characterization of the future data-flow system’s workload, we report on small-scale raw performance and scalability studies. Finally, we conclude on the suitability of such systems to the tight constraints expected for the ATLAS experiment in phase-II and, in general, what the HEP community can profit from these storage technologies.ATL-DAQ-SLIDE-2019-080oai:cds.cern.ch:26673792019-03-18
spellingShingle Particle Physics - Experiment
Abed Abud, Adam
Le Goff, Fabrice
Avolio, Giuseppe
Performance evaluation of distributed file systems for the phase-II upgrade of the ATLAS experiment at CERN
title Performance evaluation of distributed file systems for the phase-II upgrade of the ATLAS experiment at CERN
title_full Performance evaluation of distributed file systems for the phase-II upgrade of the ATLAS experiment at CERN
title_fullStr Performance evaluation of distributed file systems for the phase-II upgrade of the ATLAS experiment at CERN
title_full_unstemmed Performance evaluation of distributed file systems for the phase-II upgrade of the ATLAS experiment at CERN
title_short Performance evaluation of distributed file systems for the phase-II upgrade of the ATLAS experiment at CERN
title_sort performance evaluation of distributed file systems for the phase-ii upgrade of the atlas experiment at cern
topic Particle Physics - Experiment
url http://cds.cern.ch/record/2667379
work_keys_str_mv AT abedabudadam performanceevaluationofdistributedfilesystemsforthephaseiiupgradeoftheatlasexperimentatcern
AT legofffabrice performanceevaluationofdistributedfilesystemsforthephaseiiupgradeoftheatlasexperimentatcern
AT avoliogiuseppe performanceevaluationofdistributedfilesystemsforthephaseiiupgradeoftheatlasexperimentatcern