Cargando…

An I/O Analysis of HPC Workloads on CephFS and Lustre

In this contribution we compare the performance of the Input/Output load (I/O) of a High-Performance Computing (HPC) application on two different File Systems: CephFS and Lustre; our goal is to assess whether CephFS could be considered a valid choice for intense HPC applications. We perform our anal...

Descripción completa

Detalles Bibliográficos
Autores principales: Chiusole, Alberto, Cozzini, Stefano, van der Ster, Daniel, Lamanna, Massimo, Giuliani, Graziano
Lenguaje:eng
Publicado: 2019
Materias:
Acceso en línea:https://dx.doi.org/10.1007/978-3-030-34356-9_24
http://cds.cern.ch/record/2806103
_version_ 1780972971140055040
author Chiusole, Alberto
Cozzini, Stefano
van der Ster, Daniel
Lamanna, Massimo
Giuliani, Graziano
author_facet Chiusole, Alberto
Cozzini, Stefano
van der Ster, Daniel
Lamanna, Massimo
Giuliani, Graziano
author_sort Chiusole, Alberto
collection CERN
description In this contribution we compare the performance of the Input/Output load (I/O) of a High-Performance Computing (HPC) application on two different File Systems: CephFS and Lustre; our goal is to assess whether CephFS could be considered a valid choice for intense HPC applications. We perform our analysis using a real HPC workload, namely RegCM, a climate simulation application, and IOR, a synthetic benchmark application, to simulate several I/O patterns using different I/O parallel libraries (MPI-IO, HDF5, PnetCDF). We compare writing performance for the two different I/O approaches that RegCM implements: the so-called spokesperson or serial, and a truly parallel one. The small difference registered between the serial I/O approach and the parallel one motivates us to explore in detail how the software stack interacts with the underlying File Systems. For this reason, we use IOR and MPIIO hints related to Collective Buffering and Data Sieving to analyze several I/O patterns on the two different File Systems. Finally we investigate Lazy I/O, a unique feature of CephFS, which disables file coherency locks introduced by the File System; this allows Ceph to buffer writes and to fully exploit its parallel and distributed architecture. Two clusters were set up for these benchmarks, one at CNRIOM and a second one at Pawsey Supercomputing Centre; we performed similar tests on both installations, and we recorded a four-times I/O performance improvement with Lazy I/O enabled. Preliminary results collected so far are quite promising and further actions and new possible I/O optimizations are presented and discussed.
id cern-2806103
institution Organización Europea para la Investigación Nuclear
language eng
publishDate 2019
record_format invenio
spelling cern-28061032022-04-11T08:12:51Zdoi:10.1007/978-3-030-34356-9_24http://cds.cern.ch/record/2806103engChiusole, AlbertoCozzini, Stefanovan der Ster, DanielLamanna, MassimoGiuliani, GrazianoAn I/O Analysis of HPC Workloads on CephFS and LustreComputing and ComputersIn this contribution we compare the performance of the Input/Output load (I/O) of a High-Performance Computing (HPC) application on two different File Systems: CephFS and Lustre; our goal is to assess whether CephFS could be considered a valid choice for intense HPC applications. We perform our analysis using a real HPC workload, namely RegCM, a climate simulation application, and IOR, a synthetic benchmark application, to simulate several I/O patterns using different I/O parallel libraries (MPI-IO, HDF5, PnetCDF). We compare writing performance for the two different I/O approaches that RegCM implements: the so-called spokesperson or serial, and a truly parallel one. The small difference registered between the serial I/O approach and the parallel one motivates us to explore in detail how the software stack interacts with the underlying File Systems. For this reason, we use IOR and MPIIO hints related to Collective Buffering and Data Sieving to analyze several I/O patterns on the two different File Systems. Finally we investigate Lazy I/O, a unique feature of CephFS, which disables file coherency locks introduced by the File System; this allows Ceph to buffer writes and to fully exploit its parallel and distributed architecture. Two clusters were set up for these benchmarks, one at CNRIOM and a second one at Pawsey Supercomputing Centre; we performed similar tests on both installations, and we recorded a four-times I/O performance improvement with Lazy I/O enabled. Preliminary results collected so far are quite promising and further actions and new possible I/O optimizations are presented and discussed.oai:cds.cern.ch:28061032019
spellingShingle Computing and Computers
Chiusole, Alberto
Cozzini, Stefano
van der Ster, Daniel
Lamanna, Massimo
Giuliani, Graziano
An I/O Analysis of HPC Workloads on CephFS and Lustre
title An I/O Analysis of HPC Workloads on CephFS and Lustre
title_full An I/O Analysis of HPC Workloads on CephFS and Lustre
title_fullStr An I/O Analysis of HPC Workloads on CephFS and Lustre
title_full_unstemmed An I/O Analysis of HPC Workloads on CephFS and Lustre
title_short An I/O Analysis of HPC Workloads on CephFS and Lustre
title_sort i/o analysis of hpc workloads on cephfs and lustre
topic Computing and Computers
url https://dx.doi.org/10.1007/978-3-030-34356-9_24
http://cds.cern.ch/record/2806103
work_keys_str_mv AT chiusolealberto anioanalysisofhpcworkloadsoncephfsandlustre
AT cozzinistefano anioanalysisofhpcworkloadsoncephfsandlustre
AT vandersterdaniel anioanalysisofhpcworkloadsoncephfsandlustre
AT lamannamassimo anioanalysisofhpcworkloadsoncephfsandlustre
AT giulianigraziano anioanalysisofhpcworkloadsoncephfsandlustre
AT chiusolealberto ioanalysisofhpcworkloadsoncephfsandlustre
AT cozzinistefano ioanalysisofhpcworkloadsoncephfsandlustre
AT vandersterdaniel ioanalysisofhpcworkloadsoncephfsandlustre
AT lamannamassimo ioanalysisofhpcworkloadsoncephfsandlustre
AT giulianigraziano ioanalysisofhpcworkloadsoncephfsandlustre