Cargando…

Assessing the overhead of offloading compression tasks

Exploring compression is increasingly promising as trade-off between computations and data movement. There are two main reasons: First, the gap between processing speed and I/O continues to grow, and technology trends indicate a continuation of this. Second, performance is determined by energy effic...

Descripción completa

Detalles Bibliográficos
Autores principales: Promberger, Laura, Schwemmer, Rainer, Fröning, Holger
Lenguaje:eng
Publicado: 2020
Materias:
Acceso en línea:https://dx.doi.org/10.1145/3409390.3409405
http://cds.cern.ch/record/2751151
_version_ 1780969237991391232
author Promberger, Laura
Schwemmer, Rainer
Fröning, Holger
author_facet Promberger, Laura
Schwemmer, Rainer
Fröning, Holger
author_sort Promberger, Laura
collection CERN
description Exploring compression is increasingly promising as trade-off between computations and data movement. There are two main reasons: First, the gap between processing speed and I/O continues to grow, and technology trends indicate a continuation of this. Second, performance is determined by energy efficiency, and the overall power consumption is dominated by the consumption of data movements. For these reasons there is already a plethora of related works on compression from various domains. Most recently, a couple of accelerators have been introduced to offload compression tasks from the main processor, for instance by AHA, Intel and Microsoft. Yet, one lacks the understanding of the overhead of compression when offloading tasks. In particular, such offloading is most beneficial for overlap with other tasks, if the associated overhead on the main processor is negligible. This work evaluates the integration costs compared to a solely software-based solution considering multiple compression algorithms. Among others, High Energy Physics data are used as a prime example of big data sources. The results imply that on average the zlib implementation on the accelerator achieves a comparable compression ratio to zlib level 2 on a CPU, while having up to 17 times the throughput and utilizing over 80 % less CPU resources. These results suggest that, given the right orchestration of compression and data movement tasks, the overhead of offloading compression is limited but present. Considering that compression is only a single task of a larger data processing pipeline, this overhead cannot be neglected.
id oai-inspirehep.net-1844347
institution Organización Europea para la Investigación Nuclear
language eng
publishDate 2020
record_format invenio
spelling oai-inspirehep.net-18443472021-03-02T20:35:21Zdoi:10.1145/3409390.3409405http://cds.cern.ch/record/2751151engPromberger, LauraSchwemmer, RainerFröning, HolgerAssessing the overhead of offloading compression tasksComputing and ComputersExploring compression is increasingly promising as trade-off between computations and data movement. There are two main reasons: First, the gap between processing speed and I/O continues to grow, and technology trends indicate a continuation of this. Second, performance is determined by energy efficiency, and the overall power consumption is dominated by the consumption of data movements. For these reasons there is already a plethora of related works on compression from various domains. Most recently, a couple of accelerators have been introduced to offload compression tasks from the main processor, for instance by AHA, Intel and Microsoft. Yet, one lacks the understanding of the overhead of compression when offloading tasks. In particular, such offloading is most beneficial for overlap with other tasks, if the associated overhead on the main processor is negligible. This work evaluates the integration costs compared to a solely software-based solution considering multiple compression algorithms. Among others, High Energy Physics data are used as a prime example of big data sources. The results imply that on average the zlib implementation on the accelerator achieves a comparable compression ratio to zlib level 2 on a CPU, while having up to 17 times the throughput and utilizing over 80 % less CPU resources. These results suggest that, given the right orchestration of compression and data movement tasks, the overhead of offloading compression is limited but present. Considering that compression is only a single task of a larger data processing pipeline, this overhead cannot be neglected.oai:inspirehep.net:18443472020
spellingShingle Computing and Computers
Promberger, Laura
Schwemmer, Rainer
Fröning, Holger
Assessing the overhead of offloading compression tasks
title Assessing the overhead of offloading compression tasks
title_full Assessing the overhead of offloading compression tasks
title_fullStr Assessing the overhead of offloading compression tasks
title_full_unstemmed Assessing the overhead of offloading compression tasks
title_short Assessing the overhead of offloading compression tasks
title_sort assessing the overhead of offloading compression tasks
topic Computing and Computers
url https://dx.doi.org/10.1145/3409390.3409405
http://cds.cern.ch/record/2751151
work_keys_str_mv AT prombergerlaura assessingtheoverheadofoffloadingcompressiontasks
AT schwemmerrainer assessingtheoverheadofoffloadingcompressiontasks
AT froningholger assessingtheoverheadofoffloadingcompressiontasks