Cargando…

Quantitative Analysis of Data Caching for the HL-LHC Data-lake

Given the development of the High Luminosity Large Hadron Collider ( HL-LHC ), the Worldwide LHC Computing Grid ( WLCG ) will face unprecedented computing challenges. The amount and complexity of the data generated from the different experiments at CERN will increase, and so the WLCG has developed a...

Descripción completa

Detalles Bibliográficos
Autor principal: Umana Chacon, Irvin Jadurier
Lenguaje:eng
Publicado: 2018
Materias:
Acceso en línea:http://cds.cern.ch/record/2642553
Descripción
Sumario:Given the development of the High Luminosity Large Hadron Collider ( HL-LHC ), the Worldwide LHC Computing Grid ( WLCG ) will face unprecedented computing challenges. The amount and complexity of the data generated from the different experiments at CERN will increase, and so the WLCG has developed a strategy to handle this through caching and the creation of data-lakes. This paper focuses on a quantitative analysis of this caching by exploring the relationship between cache size and hit-rate. This, through a simulation based on the interactions of a Tier 2 site and the Data Center. The results that show that, at least for these two sites, increasing the cache’s capacity above 0.45 petabytes will not increase above a 41% hit-rate. These results are based on the Data Center’s interactions registered in log files from May 31 to June 25, 2018.