Cargando…

CERN Disk Storage Services: Report from last data taking, evolution and future outlook towards Exabyte-scale storage

The CERN IT Storage group operates multiple distributed storage systems to support all CERN data storage requirements: the physics data generated by LHC and non-LHC experiments; object and file storage for infrastructure services; block storage for the CERN cloud system; filesystems for general use...

Descripción completa

Detalles Bibliográficos
Autores principales: Mascetti, Luca, Rios, Maria Arsuaga, Bocchi, Enrico, Vicente, Joao Calado, Cheong, Belinda Chan Kwok, Castro, Diogo, Collet, Julien, Contescu, Cristian, Labrador, Hugo Gonzalez, Iven, Jan, Lamanna, Massimo, Lo Presti, Giuseppe, Mouratidis, Theofilos, Mościcki, Jakub T, Musset, Paul, Pelletier, Remy, Cameselle, Roberto Valverde, van der Ster, Daniel
Lenguaje:eng
Publicado: 2020
Materias:
Acceso en línea:https://dx.doi.org/10.1051/epjconf/202024504038
http://cds.cern.ch/record/2757341
_version_ 1780969976198332416
author Mascetti, Luca
Rios, Maria Arsuaga
Bocchi, Enrico
Vicente, Joao Calado
Cheong, Belinda Chan Kwok
Castro, Diogo
Collet, Julien
Contescu, Cristian
Labrador, Hugo Gonzalez
Iven, Jan
Lamanna, Massimo
Lo Presti, Giuseppe
Mouratidis, Theofilos
Mościcki, Jakub T
Musset, Paul
Pelletier, Remy
Cameselle, Roberto Valverde
van der Ster, Daniel
author_facet Mascetti, Luca
Rios, Maria Arsuaga
Bocchi, Enrico
Vicente, Joao Calado
Cheong, Belinda Chan Kwok
Castro, Diogo
Collet, Julien
Contescu, Cristian
Labrador, Hugo Gonzalez
Iven, Jan
Lamanna, Massimo
Lo Presti, Giuseppe
Mouratidis, Theofilos
Mościcki, Jakub T
Musset, Paul
Pelletier, Remy
Cameselle, Roberto Valverde
van der Ster, Daniel
author_sort Mascetti, Luca
collection CERN
description The CERN IT Storage group operates multiple distributed storage systems to support all CERN data storage requirements: the physics data generated by LHC and non-LHC experiments; object and file storage for infrastructure services; block storage for the CERN cloud system; filesystems for general use and specialized HPC clusters; content distribution filesystem for software distribution and condition databases; and sync&share; cloud storage for end-user files. The total integrated capacity of these systems exceeds 0.6 Exabyte.Large-scale experiment data taking has been supported by EOS and CASTOR for the last 10+ years. Particular highlights for 2018 include the special HeavyIon run which was the last part of the LHC Run2 Programme: the IT storage systems sustained over 10GB/s to flawlessly collect and archive more than 13 PB of data in a single month. While the tape archival continues to be handled by CASTOR, the effort to migrate the current experiment workflows to the new CERN Tape Archive system (CTA) is underway.Ceph infrastructure has operated for more than 5 years to provide block storage to CERN IT private OpenStack cloud, a shared filesystem (CephFS) to HPC clusters and NFS storage to replace commercial Filers. S3 service was introduced in 2018, following increased user requirements for S3-compatible object storage from physics experiments and IT use-cases.Since its introduction in 2014N, CERNBox has become a ubiquitous cloud storage interface for all CERN user groups: physicists, engineers and administration. CERNBox provides easy access to multi-petabyte data stores from a multitude of mobile and desktop devices and all mainstream, modern operating systems (Linux, Windows, macOS, Android, iOS). CERNBox provides synchronized storage for end-user’s devices as well as easy sharing for individual users and e-groups. CERNBox has also become a storage platform to host online applications to process the data such as SWAN (Service for Web-based Analysis) as well as file editors such as Collabora Online, Only Office, Draw.IO and more. An increasing number of online applications in the Windows infrastructure uses CIFS/SMB access to CERNBox files.CVMFS provides software repositories for all experiments across the WLCG infrastructure and has recently been optimized to efficiently handle nightlybuilds. While AFS continues to provide general-purpose filesystem for internal CERN users, especially as $HOME login area on central computing infrastructure, the migration of project and web spaces has significantly advanced.In this paper, we report on the experiences from the last year of LHC RUN2 data taking and evolution of our services in the past year.. We will highlight upcoming changes and future improvements and challenges.
id oai-inspirehep.net-1831566
institution Organización Europea para la Investigación Nuclear
language eng
publishDate 2020
record_format invenio
spelling oai-inspirehep.net-18315662021-03-22T22:08:56Zdoi:10.1051/epjconf/202024504038http://cds.cern.ch/record/2757341engMascetti, LucaRios, Maria ArsuagaBocchi, EnricoVicente, Joao CaladoCheong, Belinda Chan KwokCastro, DiogoCollet, JulienContescu, CristianLabrador, Hugo GonzalezIven, JanLamanna, MassimoLo Presti, GiuseppeMouratidis, TheofilosMościcki, Jakub TMusset, PaulPelletier, RemyCameselle, Roberto Valverdevan der Ster, DanielCERN Disk Storage Services: Report from last data taking, evolution and future outlook towards Exabyte-scale storageComputing and ComputersThe CERN IT Storage group operates multiple distributed storage systems to support all CERN data storage requirements: the physics data generated by LHC and non-LHC experiments; object and file storage for infrastructure services; block storage for the CERN cloud system; filesystems for general use and specialized HPC clusters; content distribution filesystem for software distribution and condition databases; and sync&share; cloud storage for end-user files. The total integrated capacity of these systems exceeds 0.6 Exabyte.Large-scale experiment data taking has been supported by EOS and CASTOR for the last 10+ years. Particular highlights for 2018 include the special HeavyIon run which was the last part of the LHC Run2 Programme: the IT storage systems sustained over 10GB/s to flawlessly collect and archive more than 13 PB of data in a single month. While the tape archival continues to be handled by CASTOR, the effort to migrate the current experiment workflows to the new CERN Tape Archive system (CTA) is underway.Ceph infrastructure has operated for more than 5 years to provide block storage to CERN IT private OpenStack cloud, a shared filesystem (CephFS) to HPC clusters and NFS storage to replace commercial Filers. S3 service was introduced in 2018, following increased user requirements for S3-compatible object storage from physics experiments and IT use-cases.Since its introduction in 2014N, CERNBox has become a ubiquitous cloud storage interface for all CERN user groups: physicists, engineers and administration. CERNBox provides easy access to multi-petabyte data stores from a multitude of mobile and desktop devices and all mainstream, modern operating systems (Linux, Windows, macOS, Android, iOS). CERNBox provides synchronized storage for end-user’s devices as well as easy sharing for individual users and e-groups. CERNBox has also become a storage platform to host online applications to process the data such as SWAN (Service for Web-based Analysis) as well as file editors such as Collabora Online, Only Office, Draw.IO and more. An increasing number of online applications in the Windows infrastructure uses CIFS/SMB access to CERNBox files.CVMFS provides software repositories for all experiments across the WLCG infrastructure and has recently been optimized to efficiently handle nightlybuilds. While AFS continues to provide general-purpose filesystem for internal CERN users, especially as $HOME login area on central computing infrastructure, the migration of project and web spaces has significantly advanced.In this paper, we report on the experiences from the last year of LHC RUN2 data taking and evolution of our services in the past year.. We will highlight upcoming changes and future improvements and challenges.oai:inspirehep.net:18315662020
spellingShingle Computing and Computers
Mascetti, Luca
Rios, Maria Arsuaga
Bocchi, Enrico
Vicente, Joao Calado
Cheong, Belinda Chan Kwok
Castro, Diogo
Collet, Julien
Contescu, Cristian
Labrador, Hugo Gonzalez
Iven, Jan
Lamanna, Massimo
Lo Presti, Giuseppe
Mouratidis, Theofilos
Mościcki, Jakub T
Musset, Paul
Pelletier, Remy
Cameselle, Roberto Valverde
van der Ster, Daniel
CERN Disk Storage Services: Report from last data taking, evolution and future outlook towards Exabyte-scale storage
title CERN Disk Storage Services: Report from last data taking, evolution and future outlook towards Exabyte-scale storage
title_full CERN Disk Storage Services: Report from last data taking, evolution and future outlook towards Exabyte-scale storage
title_fullStr CERN Disk Storage Services: Report from last data taking, evolution and future outlook towards Exabyte-scale storage
title_full_unstemmed CERN Disk Storage Services: Report from last data taking, evolution and future outlook towards Exabyte-scale storage
title_short CERN Disk Storage Services: Report from last data taking, evolution and future outlook towards Exabyte-scale storage
title_sort cern disk storage services: report from last data taking, evolution and future outlook towards exabyte-scale storage
topic Computing and Computers
url https://dx.doi.org/10.1051/epjconf/202024504038
http://cds.cern.ch/record/2757341
work_keys_str_mv AT mascettiluca cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage
AT riosmariaarsuaga cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage
AT bocchienrico cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage
AT vicentejoaocalado cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage
AT cheongbelindachankwok cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage
AT castrodiogo cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage
AT colletjulien cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage
AT contescucristian cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage
AT labradorhugogonzalez cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage
AT ivenjan cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage
AT lamannamassimo cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage
AT loprestigiuseppe cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage
AT mouratidistheofilos cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage
AT moscickijakubt cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage
AT mussetpaul cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage
AT pelletierremy cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage
AT camesellerobertovalverde cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage
AT vandersterdaniel cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage