Cargando…
Straggler-Aware Distributed Learning: Communication–Computation Latency Trade-Off
When gradient descent (GD) is scaled to many parallel workers for large-scale machine learning applications, its per-iteration computation time is limited by straggling workers. Straggling workers can be tolerated by assigning redundant computations and/or coding across data and computations, but in...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7517046/ https://www.ncbi.nlm.nih.gov/pubmed/33286316 http://dx.doi.org/10.3390/e22050544 |
_version_ | 1783587139372449792 |
---|---|
author | Ozfatura, Emre Ulukus, Sennur Gündüz, Deniz |
author_facet | Ozfatura, Emre Ulukus, Sennur Gündüz, Deniz |
author_sort | Ozfatura, Emre |
collection | PubMed |
description | When gradient descent (GD) is scaled to many parallel workers for large-scale machine learning applications, its per-iteration computation time is limited by straggling workers. Straggling workers can be tolerated by assigning redundant computations and/or coding across data and computations, but in most existing schemes, each non-straggling worker transmits one message per iteration to the parameter server (PS) after completing all its computations. Imposing such a limitation results in two drawbacks: over-computation due to inaccurate prediction of the straggling behavior, and under-utilization due to discarding partial computations carried out by stragglers. To overcome these drawbacks, we consider multi-message communication (MMC) by allowing multiple computations to be conveyed from each worker per iteration, and propose novel straggler avoidance techniques for both coded computation and coded communication with MMC. We analyze how the proposed designs can be employed efficiently to seek a balance between the computation and communication latency. Furthermore, we identify the advantages and disadvantages of these designs in different settings through extensive simulations, both model-based and real implementation on Amazon EC2 servers, and demonstrate that proposed schemes with MMC can help improve upon existing straggler avoidance schemes. |
format | Online Article Text |
id | pubmed-7517046 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-75170462020-11-09 Straggler-Aware Distributed Learning: Communication–Computation Latency Trade-Off Ozfatura, Emre Ulukus, Sennur Gündüz, Deniz Entropy (Basel) Article When gradient descent (GD) is scaled to many parallel workers for large-scale machine learning applications, its per-iteration computation time is limited by straggling workers. Straggling workers can be tolerated by assigning redundant computations and/or coding across data and computations, but in most existing schemes, each non-straggling worker transmits one message per iteration to the parameter server (PS) after completing all its computations. Imposing such a limitation results in two drawbacks: over-computation due to inaccurate prediction of the straggling behavior, and under-utilization due to discarding partial computations carried out by stragglers. To overcome these drawbacks, we consider multi-message communication (MMC) by allowing multiple computations to be conveyed from each worker per iteration, and propose novel straggler avoidance techniques for both coded computation and coded communication with MMC. We analyze how the proposed designs can be employed efficiently to seek a balance between the computation and communication latency. Furthermore, we identify the advantages and disadvantages of these designs in different settings through extensive simulations, both model-based and real implementation on Amazon EC2 servers, and demonstrate that proposed schemes with MMC can help improve upon existing straggler avoidance schemes. MDPI 2020-05-13 /pmc/articles/PMC7517046/ /pubmed/33286316 http://dx.doi.org/10.3390/e22050544 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Ozfatura, Emre Ulukus, Sennur Gündüz, Deniz Straggler-Aware Distributed Learning: Communication–Computation Latency Trade-Off |
title | Straggler-Aware Distributed Learning: Communication–Computation Latency Trade-Off |
title_full | Straggler-Aware Distributed Learning: Communication–Computation Latency Trade-Off |
title_fullStr | Straggler-Aware Distributed Learning: Communication–Computation Latency Trade-Off |
title_full_unstemmed | Straggler-Aware Distributed Learning: Communication–Computation Latency Trade-Off |
title_short | Straggler-Aware Distributed Learning: Communication–Computation Latency Trade-Off |
title_sort | straggler-aware distributed learning: communication–computation latency trade-off |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7517046/ https://www.ncbi.nlm.nih.gov/pubmed/33286316 http://dx.doi.org/10.3390/e22050544 |
work_keys_str_mv | AT ozfaturaemre stragglerawaredistributedlearningcommunicationcomputationlatencytradeoff AT ulukussennur stragglerawaredistributedlearningcommunicationcomputationlatencytradeoff AT gunduzdeniz stragglerawaredistributedlearningcommunicationcomputationlatencytradeoff |