Cargando…
Delivering LHC software to HPC compute elements
In recent years, there was a growing interest in improving the utilization of supercomputers by running applications of experiments at the Large Hadron Collider (LHC) at CERN when idle cores cannot be assigned to traditional HPC jobs. At the same time, the upcoming LHC machine and detector upgrades...
Autores principales: | , , , |
---|---|
Lenguaje: | eng |
Publicado: |
2017
|
Materias: | |
Acceso en línea: | https://dx.doi.org/10.1007/978-3-319-67630-2_52 http://cds.cern.ch/record/2302591 |
_version_ | 1780957289258156032 |
---|---|
author | Blomer, Jakob Ganis, Gerardo Hardi, Nikola Popescu, Radu |
author_facet | Blomer, Jakob Ganis, Gerardo Hardi, Nikola Popescu, Radu |
author_sort | Blomer, Jakob |
collection | CERN |
description | In recent years, there was a growing interest in improving the utilization of supercomputers by running applications of experiments at the Large Hadron Collider (LHC) at CERN when idle cores cannot be assigned to traditional HPC jobs. At the same time, the upcoming LHC machine and detector upgrades will produce some 60 times higher data rates and challenge LHC experiments to use so far untapped compute resources. LHC experiment applications are tailored to run on high-throughput computing resources and they have a different anatomy than HPC applications. LHC applications comprise a core framework that allows hundreds of researchers to plug in their specific algorithms. The software stacks easily accumulate to many gigabytes for a single release. New releases are often produced on a daily basis. To facilitate the distribution of these software stacks to world-wide distributed computing resources, LHC experiments use a purpose-built, global, POSIX file system, the CernVM File System. CernVM-FS pre-processes data into content-addressed, digitally signed Merkle trees and it uses web caches and proxies for data distribution. Fuse-mounted files system clients on the compute nodes load and cache on demand only the small fraction of files needed at any given moment. In this paper, we report on problems and lessons learned in the deployment of CernVM-FS on supercomputers such as the supercomputers at NERSC in Berkeley, at LRZ in Munich, and at CSCS in Lugano. We compare CernVM-FS to a shared software area on a traditional HPC storage system and to container-based systems. |
id | cern-2302591 |
institution | Organización Europea para la Investigación Nuclear |
language | eng |
publishDate | 2017 |
record_format | invenio |
spelling | cern-23025912019-09-30T06:29:59Zdoi:10.1007/978-3-319-67630-2_52http://cds.cern.ch/record/2302591engBlomer, JakobGanis, GerardoHardi, NikolaPopescu, RaduDelivering LHC software to HPC compute elementsComputing and ComputersIn recent years, there was a growing interest in improving the utilization of supercomputers by running applications of experiments at the Large Hadron Collider (LHC) at CERN when idle cores cannot be assigned to traditional HPC jobs. At the same time, the upcoming LHC machine and detector upgrades will produce some 60 times higher data rates and challenge LHC experiments to use so far untapped compute resources. LHC experiment applications are tailored to run on high-throughput computing resources and they have a different anatomy than HPC applications. LHC applications comprise a core framework that allows hundreds of researchers to plug in their specific algorithms. The software stacks easily accumulate to many gigabytes for a single release. New releases are often produced on a daily basis. To facilitate the distribution of these software stacks to world-wide distributed computing resources, LHC experiments use a purpose-built, global, POSIX file system, the CernVM File System. CernVM-FS pre-processes data into content-addressed, digitally signed Merkle trees and it uses web caches and proxies for data distribution. Fuse-mounted files system clients on the compute nodes load and cache on demand only the small fraction of files needed at any given moment. In this paper, we report on problems and lessons learned in the deployment of CernVM-FS on supercomputers such as the supercomputers at NERSC in Berkeley, at LRZ in Munich, and at CSCS in Lugano. We compare CernVM-FS to a shared software area on a traditional HPC storage system and to container-based systems.oai:cds.cern.ch:23025912017 |
spellingShingle | Computing and Computers Blomer, Jakob Ganis, Gerardo Hardi, Nikola Popescu, Radu Delivering LHC software to HPC compute elements |
title | Delivering LHC software to HPC compute elements |
title_full | Delivering LHC software to HPC compute elements |
title_fullStr | Delivering LHC software to HPC compute elements |
title_full_unstemmed | Delivering LHC software to HPC compute elements |
title_short | Delivering LHC software to HPC compute elements |
title_sort | delivering lhc software to hpc compute elements |
topic | Computing and Computers |
url | https://dx.doi.org/10.1007/978-3-319-67630-2_52 http://cds.cern.ch/record/2302591 |
work_keys_str_mv | AT blomerjakob deliveringlhcsoftwaretohpccomputeelements AT ganisgerardo deliveringlhcsoftwaretohpccomputeelements AT hardinikola deliveringlhcsoftwaretohpccomputeelements AT popescuradu deliveringlhcsoftwaretohpccomputeelements |