Cargando…

Big Data Archiving From Oracle to Hadoop

The CERN Accelerator Logging Service (CALS) is used to persist data of around 2 million predefined signals coming from heterogeneous sources such as the electricity infrastructure, industrial controls like cryogenics and vacuum, or beam related data. This old Oracle based logging system will be phas...

Descripción completa

Detalles Bibliográficos
Autores principales: Prieto Barreiro, Ivan, Sobieszek, Marcin
Lenguaje:eng
Publicado: 2020
Materias:
Acceso en línea:https://dx.doi.org/10.18429/JACoW-ICALEPCS2019-MOPHA117
http://cds.cern.ch/record/2778522
_version_ 1780971751686012928
author Prieto Barreiro, Ivan
Sobieszek, Marcin
author_facet Prieto Barreiro, Ivan
Sobieszek, Marcin
author_sort Prieto Barreiro, Ivan
collection CERN
description The CERN Accelerator Logging Service (CALS) is used to persist data of around 2 million predefined signals coming from heterogeneous sources such as the electricity infrastructure, industrial controls like cryogenics and vacuum, or beam related data. This old Oracle based logging system will be phased out at the end of the LHC’s Long Shut-down 2 (LS2) and will be replaced by the Next CERN Accelerator Logging Service (NXCALS) which is based on Hadoop. As a consequence, the different data sources must be adapted to persist the data in the new logging system. This paper describes the solution implemented to archive into NXCALS the data produced by QPS (Quench Protection System) and SCADAR (Supervisory Control And Data Acquisition Relational database) systems, which generate a total of around 175, 000 values per second. To cope with such a volume of data the new service has to be extremely robust, scalable and fail-safe with guaranteed data delivery and no data loss. The paper also explains how to recover from different failure scenarios like e.g. network disruption and how to manage and monitor this highly distributed service.
id cern-2778522
institution Organización Europea para la Investigación Nuclear
language eng
publishDate 2020
record_format invenio
spelling cern-27785222022-01-14T14:55:34Zdoi:10.18429/JACoW-ICALEPCS2019-MOPHA117http://cds.cern.ch/record/2778522engPrieto Barreiro, IvanSobieszek, MarcinBig Data Archiving From Oracle to HadoopAccelerators and Storage RingsThe CERN Accelerator Logging Service (CALS) is used to persist data of around 2 million predefined signals coming from heterogeneous sources such as the electricity infrastructure, industrial controls like cryogenics and vacuum, or beam related data. This old Oracle based logging system will be phased out at the end of the LHC’s Long Shut-down 2 (LS2) and will be replaced by the Next CERN Accelerator Logging Service (NXCALS) which is based on Hadoop. As a consequence, the different data sources must be adapted to persist the data in the new logging system. This paper describes the solution implemented to archive into NXCALS the data produced by QPS (Quench Protection System) and SCADAR (Supervisory Control And Data Acquisition Relational database) systems, which generate a total of around 175, 000 values per second. To cope with such a volume of data the new service has to be extremely robust, scalable and fail-safe with guaranteed data delivery and no data loss. The paper also explains how to recover from different failure scenarios like e.g. network disruption and how to manage and monitor this highly distributed service.oai:cds.cern.ch:27785222020
spellingShingle Accelerators and Storage Rings
Prieto Barreiro, Ivan
Sobieszek, Marcin
Big Data Archiving From Oracle to Hadoop
title Big Data Archiving From Oracle to Hadoop
title_full Big Data Archiving From Oracle to Hadoop
title_fullStr Big Data Archiving From Oracle to Hadoop
title_full_unstemmed Big Data Archiving From Oracle to Hadoop
title_short Big Data Archiving From Oracle to Hadoop
title_sort big data archiving from oracle to hadoop
topic Accelerators and Storage Rings
url https://dx.doi.org/10.18429/JACoW-ICALEPCS2019-MOPHA117
http://cds.cern.ch/record/2778522
work_keys_str_mv AT prietobarreiroivan bigdataarchivingfromoracletohadoop
AT sobieszekmarcin bigdataarchivingfromoracletohadoop