Cargando…
Big Data Archiving From Oracle to Hadoop
The CERN Accelerator Logging Service (CALS) is used to persist data of around 2 million predefined signals coming from heterogeneous sources such as the electricity infrastructure, industrial controls like cryogenics and vacuum, or beam related data. This old Oracle based logging system will be phas...
Autores principales: | , |
---|---|
Lenguaje: | eng |
Publicado: |
2020
|
Materias: | |
Acceso en línea: | https://dx.doi.org/10.18429/JACoW-ICALEPCS2019-MOPHA117 http://cds.cern.ch/record/2778522 |
_version_ | 1780971751686012928 |
---|---|
author | Prieto Barreiro, Ivan Sobieszek, Marcin |
author_facet | Prieto Barreiro, Ivan Sobieszek, Marcin |
author_sort | Prieto Barreiro, Ivan |
collection | CERN |
description | The CERN Accelerator Logging Service (CALS) is used to persist data of around 2 million predefined signals coming from heterogeneous sources such as the electricity infrastructure, industrial controls like cryogenics and vacuum, or beam related data. This old Oracle based logging system will be phased out at the end of the LHC’s Long Shut-down 2 (LS2) and will be replaced by the Next CERN Accelerator Logging Service (NXCALS) which is based on Hadoop. As a consequence, the different data sources must be adapted to persist the data in the new logging system. This paper describes the solution implemented to archive into NXCALS the data produced by QPS (Quench Protection System) and SCADAR (Supervisory Control And Data Acquisition Relational database) systems, which generate a total of around 175, 000 values per second. To cope with such a volume of data the new service has to be extremely robust, scalable and fail-safe with guaranteed data delivery and no data loss. The paper also explains how to recover from different failure scenarios like e.g. network disruption and how to manage and monitor this highly distributed service. |
id | cern-2778522 |
institution | Organización Europea para la Investigación Nuclear |
language | eng |
publishDate | 2020 |
record_format | invenio |
spelling | cern-27785222022-01-14T14:55:34Zdoi:10.18429/JACoW-ICALEPCS2019-MOPHA117http://cds.cern.ch/record/2778522engPrieto Barreiro, IvanSobieszek, MarcinBig Data Archiving From Oracle to HadoopAccelerators and Storage RingsThe CERN Accelerator Logging Service (CALS) is used to persist data of around 2 million predefined signals coming from heterogeneous sources such as the electricity infrastructure, industrial controls like cryogenics and vacuum, or beam related data. This old Oracle based logging system will be phased out at the end of the LHC’s Long Shut-down 2 (LS2) and will be replaced by the Next CERN Accelerator Logging Service (NXCALS) which is based on Hadoop. As a consequence, the different data sources must be adapted to persist the data in the new logging system. This paper describes the solution implemented to archive into NXCALS the data produced by QPS (Quench Protection System) and SCADAR (Supervisory Control And Data Acquisition Relational database) systems, which generate a total of around 175, 000 values per second. To cope with such a volume of data the new service has to be extremely robust, scalable and fail-safe with guaranteed data delivery and no data loss. The paper also explains how to recover from different failure scenarios like e.g. network disruption and how to manage and monitor this highly distributed service.oai:cds.cern.ch:27785222020 |
spellingShingle | Accelerators and Storage Rings Prieto Barreiro, Ivan Sobieszek, Marcin Big Data Archiving From Oracle to Hadoop |
title | Big Data Archiving From Oracle to Hadoop |
title_full | Big Data Archiving From Oracle to Hadoop |
title_fullStr | Big Data Archiving From Oracle to Hadoop |
title_full_unstemmed | Big Data Archiving From Oracle to Hadoop |
title_short | Big Data Archiving From Oracle to Hadoop |
title_sort | big data archiving from oracle to hadoop |
topic | Accelerators and Storage Rings |
url | https://dx.doi.org/10.18429/JACoW-ICALEPCS2019-MOPHA117 http://cds.cern.ch/record/2778522 |
work_keys_str_mv | AT prietobarreiroivan bigdataarchivingfromoracletohadoop AT sobieszekmarcin bigdataarchivingfromoracletohadoop |