Cargando…
CERN Disk Storage Services: Report from last data taking, evolution and future outlook towards Exabyte-scale storage
The CERN IT Storage group operates multiple distributed storage systems to support all CERN data storage requirements: the physics data generated by LHC and non-LHC experiments; object and file storage for infrastructure services; block storage for the CERN cloud system; filesystems for general use...
Autores principales: | , , , , , , , , , , , , , , , , , |
---|---|
Lenguaje: | eng |
Publicado: |
2020
|
Materias: | |
Acceso en línea: | https://dx.doi.org/10.1051/epjconf/202024504038 http://cds.cern.ch/record/2757341 |
_version_ | 1780969976198332416 |
---|---|
author | Mascetti, Luca Rios, Maria Arsuaga Bocchi, Enrico Vicente, Joao Calado Cheong, Belinda Chan Kwok Castro, Diogo Collet, Julien Contescu, Cristian Labrador, Hugo Gonzalez Iven, Jan Lamanna, Massimo Lo Presti, Giuseppe Mouratidis, Theofilos Mościcki, Jakub T Musset, Paul Pelletier, Remy Cameselle, Roberto Valverde van der Ster, Daniel |
author_facet | Mascetti, Luca Rios, Maria Arsuaga Bocchi, Enrico Vicente, Joao Calado Cheong, Belinda Chan Kwok Castro, Diogo Collet, Julien Contescu, Cristian Labrador, Hugo Gonzalez Iven, Jan Lamanna, Massimo Lo Presti, Giuseppe Mouratidis, Theofilos Mościcki, Jakub T Musset, Paul Pelletier, Remy Cameselle, Roberto Valverde van der Ster, Daniel |
author_sort | Mascetti, Luca |
collection | CERN |
description | The CERN IT Storage group operates multiple distributed storage systems to support all CERN data storage requirements: the physics data generated by LHC and non-LHC experiments; object and file storage for infrastructure services; block storage for the CERN cloud system; filesystems for general use and specialized HPC clusters; content distribution filesystem for software distribution and condition databases; and sync&share; cloud storage for end-user files. The total integrated capacity of these systems exceeds 0.6 Exabyte.Large-scale experiment data taking has been supported by EOS and CASTOR for the last 10+ years. Particular highlights for 2018 include the special HeavyIon run which was the last part of the LHC Run2 Programme: the IT storage systems sustained over 10GB/s to flawlessly collect and archive more than 13 PB of data in a single month. While the tape archival continues to be handled by CASTOR, the effort to migrate the current experiment workflows to the new CERN Tape Archive system (CTA) is underway.Ceph infrastructure has operated for more than 5 years to provide block storage to CERN IT private OpenStack cloud, a shared filesystem (CephFS) to HPC clusters and NFS storage to replace commercial Filers. S3 service was introduced in 2018, following increased user requirements for S3-compatible object storage from physics experiments and IT use-cases.Since its introduction in 2014N, CERNBox has become a ubiquitous cloud storage interface for all CERN user groups: physicists, engineers and administration. CERNBox provides easy access to multi-petabyte data stores from a multitude of mobile and desktop devices and all mainstream, modern operating systems (Linux, Windows, macOS, Android, iOS). CERNBox provides synchronized storage for end-user’s devices as well as easy sharing for individual users and e-groups. CERNBox has also become a storage platform to host online applications to process the data such as SWAN (Service for Web-based Analysis) as well as file editors such as Collabora Online, Only Office, Draw.IO and more. An increasing number of online applications in the Windows infrastructure uses CIFS/SMB access to CERNBox files.CVMFS provides software repositories for all experiments across the WLCG infrastructure and has recently been optimized to efficiently handle nightlybuilds. While AFS continues to provide general-purpose filesystem for internal CERN users, especially as $HOME login area on central computing infrastructure, the migration of project and web spaces has significantly advanced.In this paper, we report on the experiences from the last year of LHC RUN2 data taking and evolution of our services in the past year.. We will highlight upcoming changes and future improvements and challenges. |
id | oai-inspirehep.net-1831566 |
institution | Organización Europea para la Investigación Nuclear |
language | eng |
publishDate | 2020 |
record_format | invenio |
spelling | oai-inspirehep.net-18315662021-03-22T22:08:56Zdoi:10.1051/epjconf/202024504038http://cds.cern.ch/record/2757341engMascetti, LucaRios, Maria ArsuagaBocchi, EnricoVicente, Joao CaladoCheong, Belinda Chan KwokCastro, DiogoCollet, JulienContescu, CristianLabrador, Hugo GonzalezIven, JanLamanna, MassimoLo Presti, GiuseppeMouratidis, TheofilosMościcki, Jakub TMusset, PaulPelletier, RemyCameselle, Roberto Valverdevan der Ster, DanielCERN Disk Storage Services: Report from last data taking, evolution and future outlook towards Exabyte-scale storageComputing and ComputersThe CERN IT Storage group operates multiple distributed storage systems to support all CERN data storage requirements: the physics data generated by LHC and non-LHC experiments; object and file storage for infrastructure services; block storage for the CERN cloud system; filesystems for general use and specialized HPC clusters; content distribution filesystem for software distribution and condition databases; and sync&share; cloud storage for end-user files. The total integrated capacity of these systems exceeds 0.6 Exabyte.Large-scale experiment data taking has been supported by EOS and CASTOR for the last 10+ years. Particular highlights for 2018 include the special HeavyIon run which was the last part of the LHC Run2 Programme: the IT storage systems sustained over 10GB/s to flawlessly collect and archive more than 13 PB of data in a single month. While the tape archival continues to be handled by CASTOR, the effort to migrate the current experiment workflows to the new CERN Tape Archive system (CTA) is underway.Ceph infrastructure has operated for more than 5 years to provide block storage to CERN IT private OpenStack cloud, a shared filesystem (CephFS) to HPC clusters and NFS storage to replace commercial Filers. S3 service was introduced in 2018, following increased user requirements for S3-compatible object storage from physics experiments and IT use-cases.Since its introduction in 2014N, CERNBox has become a ubiquitous cloud storage interface for all CERN user groups: physicists, engineers and administration. CERNBox provides easy access to multi-petabyte data stores from a multitude of mobile and desktop devices and all mainstream, modern operating systems (Linux, Windows, macOS, Android, iOS). CERNBox provides synchronized storage for end-user’s devices as well as easy sharing for individual users and e-groups. CERNBox has also become a storage platform to host online applications to process the data such as SWAN (Service for Web-based Analysis) as well as file editors such as Collabora Online, Only Office, Draw.IO and more. An increasing number of online applications in the Windows infrastructure uses CIFS/SMB access to CERNBox files.CVMFS provides software repositories for all experiments across the WLCG infrastructure and has recently been optimized to efficiently handle nightlybuilds. While AFS continues to provide general-purpose filesystem for internal CERN users, especially as $HOME login area on central computing infrastructure, the migration of project and web spaces has significantly advanced.In this paper, we report on the experiences from the last year of LHC RUN2 data taking and evolution of our services in the past year.. We will highlight upcoming changes and future improvements and challenges.oai:inspirehep.net:18315662020 |
spellingShingle | Computing and Computers Mascetti, Luca Rios, Maria Arsuaga Bocchi, Enrico Vicente, Joao Calado Cheong, Belinda Chan Kwok Castro, Diogo Collet, Julien Contescu, Cristian Labrador, Hugo Gonzalez Iven, Jan Lamanna, Massimo Lo Presti, Giuseppe Mouratidis, Theofilos Mościcki, Jakub T Musset, Paul Pelletier, Remy Cameselle, Roberto Valverde van der Ster, Daniel CERN Disk Storage Services: Report from last data taking, evolution and future outlook towards Exabyte-scale storage |
title | CERN Disk Storage Services: Report from last data taking, evolution and future outlook towards Exabyte-scale storage |
title_full | CERN Disk Storage Services: Report from last data taking, evolution and future outlook towards Exabyte-scale storage |
title_fullStr | CERN Disk Storage Services: Report from last data taking, evolution and future outlook towards Exabyte-scale storage |
title_full_unstemmed | CERN Disk Storage Services: Report from last data taking, evolution and future outlook towards Exabyte-scale storage |
title_short | CERN Disk Storage Services: Report from last data taking, evolution and future outlook towards Exabyte-scale storage |
title_sort | cern disk storage services: report from last data taking, evolution and future outlook towards exabyte-scale storage |
topic | Computing and Computers |
url | https://dx.doi.org/10.1051/epjconf/202024504038 http://cds.cern.ch/record/2757341 |
work_keys_str_mv | AT mascettiluca cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage AT riosmariaarsuaga cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage AT bocchienrico cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage AT vicentejoaocalado cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage AT cheongbelindachankwok cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage AT castrodiogo cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage AT colletjulien cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage AT contescucristian cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage AT labradorhugogonzalez cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage AT ivenjan cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage AT lamannamassimo cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage AT loprestigiuseppe cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage AT mouratidistheofilos cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage AT moscickijakubt cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage AT mussetpaul cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage AT pelletierremy cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage AT camesellerobertovalverde cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage AT vandersterdaniel cerndiskstorageservicesreportfromlastdatatakingevolutionandfutureoutlooktowardsexabytescalestorage |