Cargando…

Delivering LHC software to HPC compute elements

In recent years, there was a growing interest in improving the utilization of supercomputers by running applications of experiments at the Large Hadron Collider (LHC) at CERN when idle cores cannot be assigned to traditional HPC jobs. At the same time, the upcoming LHC machine and detector upgrades...

Descripción completa

Detalles Bibliográficos
Autores principales: Blomer, Jakob, Ganis, Gerardo, Hardi, Nikola, Popescu, Radu
Lenguaje:eng
Publicado: 2017
Materias:
Acceso en línea:https://dx.doi.org/10.1007/978-3-319-67630-2_52
http://cds.cern.ch/record/2302591
_version_ 1780957289258156032
author Blomer, Jakob
Ganis, Gerardo
Hardi, Nikola
Popescu, Radu
author_facet Blomer, Jakob
Ganis, Gerardo
Hardi, Nikola
Popescu, Radu
author_sort Blomer, Jakob
collection CERN
description In recent years, there was a growing interest in improving the utilization of supercomputers by running applications of experiments at the Large Hadron Collider (LHC) at CERN when idle cores cannot be assigned to traditional HPC jobs. At the same time, the upcoming LHC machine and detector upgrades will produce some 60 times higher data rates and challenge LHC experiments to use so far untapped compute resources. LHC experiment applications are tailored to run on high-throughput computing resources and they have a different anatomy than HPC applications. LHC applications comprise a core framework that allows hundreds of researchers to plug in their specific algorithms. The software stacks easily accumulate to many gigabytes for a single release. New releases are often produced on a daily basis. To facilitate the distribution of these software stacks to world-wide distributed computing resources, LHC experiments use a purpose-built, global, POSIX file system, the CernVM File System. CernVM-FS pre-processes data into content-addressed, digitally signed Merkle trees and it uses web caches and proxies for data distribution. Fuse-mounted files system clients on the compute nodes load and cache on demand only the small fraction of files needed at any given moment. In this paper, we report on problems and lessons learned in the deployment of CernVM-FS on supercomputers such as the supercomputers at NERSC in Berkeley, at LRZ in Munich, and at CSCS in Lugano. We compare CernVM-FS to a shared software area on a traditional HPC storage system and to container-based systems.
id cern-2302591
institution Organización Europea para la Investigación Nuclear
language eng
publishDate 2017
record_format invenio
spelling cern-23025912019-09-30T06:29:59Zdoi:10.1007/978-3-319-67630-2_52http://cds.cern.ch/record/2302591engBlomer, JakobGanis, GerardoHardi, NikolaPopescu, RaduDelivering LHC software to HPC compute elementsComputing and ComputersIn recent years, there was a growing interest in improving the utilization of supercomputers by running applications of experiments at the Large Hadron Collider (LHC) at CERN when idle cores cannot be assigned to traditional HPC jobs. At the same time, the upcoming LHC machine and detector upgrades will produce some 60 times higher data rates and challenge LHC experiments to use so far untapped compute resources. LHC experiment applications are tailored to run on high-throughput computing resources and they have a different anatomy than HPC applications. LHC applications comprise a core framework that allows hundreds of researchers to plug in their specific algorithms. The software stacks easily accumulate to many gigabytes for a single release. New releases are often produced on a daily basis. To facilitate the distribution of these software stacks to world-wide distributed computing resources, LHC experiments use a purpose-built, global, POSIX file system, the CernVM File System. CernVM-FS pre-processes data into content-addressed, digitally signed Merkle trees and it uses web caches and proxies for data distribution. Fuse-mounted files system clients on the compute nodes load and cache on demand only the small fraction of files needed at any given moment. In this paper, we report on problems and lessons learned in the deployment of CernVM-FS on supercomputers such as the supercomputers at NERSC in Berkeley, at LRZ in Munich, and at CSCS in Lugano. We compare CernVM-FS to a shared software area on a traditional HPC storage system and to container-based systems.oai:cds.cern.ch:23025912017
spellingShingle Computing and Computers
Blomer, Jakob
Ganis, Gerardo
Hardi, Nikola
Popescu, Radu
Delivering LHC software to HPC compute elements
title Delivering LHC software to HPC compute elements
title_full Delivering LHC software to HPC compute elements
title_fullStr Delivering LHC software to HPC compute elements
title_full_unstemmed Delivering LHC software to HPC compute elements
title_short Delivering LHC software to HPC compute elements
title_sort delivering lhc software to hpc compute elements
topic Computing and Computers
url https://dx.doi.org/10.1007/978-3-319-67630-2_52
http://cds.cern.ch/record/2302591
work_keys_str_mv AT blomerjakob deliveringlhcsoftwaretohpccomputeelements
AT ganisgerardo deliveringlhcsoftwaretohpccomputeelements
AT hardinikola deliveringlhcsoftwaretohpccomputeelements
AT popescuradu deliveringlhcsoftwaretohpccomputeelements