Cargando…
High-throughput and low-latency network communication with NetIO
HPC network technologies like Infiniband, TrueScale or OmniPath provide low-latency and high-throughput communication between hosts, which makes them attractive options for data-acquisition systems in large-scale high-energy physics experiments. Like HPC networks, DAQ networks are local and include...
Autores principales: | , , |
---|---|
Lenguaje: | eng |
Publicado: |
2017
|
Materias: | |
Acceso en línea: | https://dx.doi.org/10.1088/1742-6596/898/8/082003 http://cds.cern.ch/record/2260396 |
_version_ | 1780953999547039744 |
---|---|
author | Schumacher, Jorn Plessl, Christian Vandelli, Wainer |
author_facet | Schumacher, Jorn Plessl, Christian Vandelli, Wainer |
author_sort | Schumacher, Jorn |
collection | CERN |
description | HPC network technologies like Infiniband, TrueScale or OmniPath provide low-latency and high-throughput communication between hosts, which makes them attractive options for data-acquisition systems in large-scale high-energy physics experiments. Like HPC networks, DAQ networks are local and include a well specified number of systems. Unfortunately traditional network communication APIs for HPC clusters like MPI or PGAS target exclusively the HPC community and are not suited well for DAQ applications. It is possible to build distributed DAQ applications using low-level system APIs like Infiniband Verbs, but it requires a non-negligible effort and expert knowledge. At the same time, message services like ZeroMQ have gained popularity in the HEP community. They allow building distributed applications with a high-level approach and provide good performance. Unfortunately their usage usually limits developers to TCP/IP-based networks. While it is possible to operate a TCP/IP stack on top of Infiniband and OmniPath, this approach may not be very efficient compared to a direct usage of native APIs. NetIO is a simple, novel asynchronous message service that can operate on Ethernet, Infiniband and similar network fabrics. In the publication the design and implementation of NetIO is presented and described, and its use is evaluated in comparison to other approaches and show performance studies. NetIO supports different high-level programming models and typical workloads of HEP applications. The ATLAS FELIX project successfully uses NetIO as its central communication platform. The architecture of NetIO is described in this paper, including the user-level API and the internal design of the data flow. The paper includes a performance evaluation of NetIO including throughput and latency measurements. The performance is compared against the state-of-the-art ZeroMQ message service. Performance measurements are performend in a lab environment with 40G Ethernet and FDR Infiniband networks. |
id | cern-2260396 |
institution | Organización Europea para la Investigación Nuclear |
language | eng |
publishDate | 2017 |
record_format | invenio |
spelling | cern-22603962019-10-15T15:18:49Zdoi:10.1088/1742-6596/898/8/082003http://cds.cern.ch/record/2260396engSchumacher, JornPlessl, ChristianVandelli, WainerHigh-throughput and low-latency network communication with NetIOParticle Physics - ExperimentHPC network technologies like Infiniband, TrueScale or OmniPath provide low-latency and high-throughput communication between hosts, which makes them attractive options for data-acquisition systems in large-scale high-energy physics experiments. Like HPC networks, DAQ networks are local and include a well specified number of systems. Unfortunately traditional network communication APIs for HPC clusters like MPI or PGAS target exclusively the HPC community and are not suited well for DAQ applications. It is possible to build distributed DAQ applications using low-level system APIs like Infiniband Verbs, but it requires a non-negligible effort and expert knowledge. At the same time, message services like ZeroMQ have gained popularity in the HEP community. They allow building distributed applications with a high-level approach and provide good performance. Unfortunately their usage usually limits developers to TCP/IP-based networks. While it is possible to operate a TCP/IP stack on top of Infiniband and OmniPath, this approach may not be very efficient compared to a direct usage of native APIs. NetIO is a simple, novel asynchronous message service that can operate on Ethernet, Infiniband and similar network fabrics. In the publication the design and implementation of NetIO is presented and described, and its use is evaluated in comparison to other approaches and show performance studies. NetIO supports different high-level programming models and typical workloads of HEP applications. The ATLAS FELIX project successfully uses NetIO as its central communication platform. The architecture of NetIO is described in this paper, including the user-level API and the internal design of the data flow. The paper includes a performance evaluation of NetIO including throughput and latency measurements. The performance is compared against the state-of-the-art ZeroMQ message service. Performance measurements are performend in a lab environment with 40G Ethernet and FDR Infiniband networks.HPC network technologies like Infiniband, TrueScale or OmniPath provide low- latency and high-throughput communication between hosts, which makes them attractive options for data-acquisition systems in large-scale high-energy physics experiments. Like HPC networks, DAQ networks are local and include a well specified number of systems. Unfortunately traditional network communication APIs for HPC clusters like MPI or PGAS exclusively target the HPC community and are not suited well for DAQ applications. It is possible to build distributed DAQ applications using low-level system APIs like Infiniband Verbs, but it requires a non-negligible effort and expert knowledge. At the same time, message services like ZeroMQ have gained popularity in the HEP community. They make it possible to build distributed applications with a high-level approach and provide good performance. Unfortunately, their usage usually limits developers to TCP/IP- based networks. While it is possible to operate a TCP/IP stack on top of Infiniband and OmniPath, this approach may not be very efficient compared to a direct use of native APIs. NetIO is a simple, novel asynchronous message service that can operate on Ethernet, Infiniband and similar network fabrics. In this paper the design and implementation of NetIO is presented and described, and its use is evaluated in comparison to other approaches. NetIO supports different high-level programming models and typical workloads of HEP applications. The ATLAS FELIX project [1] successfully uses NetIO as its central communication platform. The architecture of NetIO is described in this paper, including the user-level API and the internal data-flow design. The paper includes a performance evaluation of NetIO including throughput and latency measurements. The performance is compared against the state-of-the- art ZeroMQ message service. Performance measurements are performed in a lab environment with Ethernet and FDR Infiniband networks.ATL-DAQ-PROC-2017-010oai:cds.cern.ch:22603962017-04-19 |
spellingShingle | Particle Physics - Experiment Schumacher, Jorn Plessl, Christian Vandelli, Wainer High-throughput and low-latency network communication with NetIO |
title | High-throughput and low-latency network communication with NetIO |
title_full | High-throughput and low-latency network communication with NetIO |
title_fullStr | High-throughput and low-latency network communication with NetIO |
title_full_unstemmed | High-throughput and low-latency network communication with NetIO |
title_short | High-throughput and low-latency network communication with NetIO |
title_sort | high-throughput and low-latency network communication with netio |
topic | Particle Physics - Experiment |
url | https://dx.doi.org/10.1088/1742-6596/898/8/082003 http://cds.cern.ch/record/2260396 |
work_keys_str_mv | AT schumacherjorn highthroughputandlowlatencynetworkcommunicationwithnetio AT plesslchristian highthroughputandlowlatencynetworkcommunicationwithnetio AT vandelliwainer highthroughputandlowlatencynetworkcommunicationwithnetio |