Cargando…
MemMAP: Compact and Generalizable Meta-LSTM Models for Memory Access Prediction
With the rise of Big Data, there has been a significant effort in increasing compute power through GPUs, TPUs, and heterogeneous architectures. As a result, many applications are memory bound, i.e., they are bottlenecked by the movement of data from main memory to compute units. One way to address t...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7206302/ http://dx.doi.org/10.1007/978-3-030-47436-2_5 |
_version_ | 1783530389913993216 |
---|---|
author | Srivastava, Ajitesh Wang, Ta-Yang Zhang, Pengmiao De Rose, Cesar Augusto F. Kannan, Rajgopal Prasanna, Viktor K. |
author_facet | Srivastava, Ajitesh Wang, Ta-Yang Zhang, Pengmiao De Rose, Cesar Augusto F. Kannan, Rajgopal Prasanna, Viktor K. |
author_sort | Srivastava, Ajitesh |
collection | PubMed |
description | With the rise of Big Data, there has been a significant effort in increasing compute power through GPUs, TPUs, and heterogeneous architectures. As a result, many applications are memory bound, i.e., they are bottlenecked by the movement of data from main memory to compute units. One way to address this issue is through data prefetching, which relies on accurate prediction of memory accesses. While recent deep learning models have performed well on sequence prediction problems, they are far too heavy in terms of model size and inference latency to be practical for data prefetching. Here, we propose extremely compact LSTM models that can predict the next memory access with high accuracy. Prior LSTM based work on access prediction has used orders of magnitude more parameters and developed one model for each application (trace). While one (specialized) model per application can result in more accuracy, it is not a scalable approach. In contrast, our models can predict for a class of applications by trading off specialization at the cost of few retraining steps at runtime, for a more generalizable compact meta-model. Our experiments on 13 benchmark applications demonstrate that three compact meta-models can obtain accuracy close to specialized models using few batches of retraining for majority of the applications. |
format | Online Article Text |
id | pubmed-7206302 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
record_format | MEDLINE/PubMed |
spelling | pubmed-72063022020-05-08 MemMAP: Compact and Generalizable Meta-LSTM Models for Memory Access Prediction Srivastava, Ajitesh Wang, Ta-Yang Zhang, Pengmiao De Rose, Cesar Augusto F. Kannan, Rajgopal Prasanna, Viktor K. Advances in Knowledge Discovery and Data Mining Article With the rise of Big Data, there has been a significant effort in increasing compute power through GPUs, TPUs, and heterogeneous architectures. As a result, many applications are memory bound, i.e., they are bottlenecked by the movement of data from main memory to compute units. One way to address this issue is through data prefetching, which relies on accurate prediction of memory accesses. While recent deep learning models have performed well on sequence prediction problems, they are far too heavy in terms of model size and inference latency to be practical for data prefetching. Here, we propose extremely compact LSTM models that can predict the next memory access with high accuracy. Prior LSTM based work on access prediction has used orders of magnitude more parameters and developed one model for each application (trace). While one (specialized) model per application can result in more accuracy, it is not a scalable approach. In contrast, our models can predict for a class of applications by trading off specialization at the cost of few retraining steps at runtime, for a more generalizable compact meta-model. Our experiments on 13 benchmark applications demonstrate that three compact meta-models can obtain accuracy close to specialized models using few batches of retraining for majority of the applications. 2020-04-17 /pmc/articles/PMC7206302/ http://dx.doi.org/10.1007/978-3-030-47436-2_5 Text en © Springer Nature Switzerland AG 2020 This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic. |
spellingShingle | Article Srivastava, Ajitesh Wang, Ta-Yang Zhang, Pengmiao De Rose, Cesar Augusto F. Kannan, Rajgopal Prasanna, Viktor K. MemMAP: Compact and Generalizable Meta-LSTM Models for Memory Access Prediction |
title | MemMAP: Compact and Generalizable Meta-LSTM Models for Memory Access Prediction |
title_full | MemMAP: Compact and Generalizable Meta-LSTM Models for Memory Access Prediction |
title_fullStr | MemMAP: Compact and Generalizable Meta-LSTM Models for Memory Access Prediction |
title_full_unstemmed | MemMAP: Compact and Generalizable Meta-LSTM Models for Memory Access Prediction |
title_short | MemMAP: Compact and Generalizable Meta-LSTM Models for Memory Access Prediction |
title_sort | memmap: compact and generalizable meta-lstm models for memory access prediction |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7206302/ http://dx.doi.org/10.1007/978-3-030-47436-2_5 |
work_keys_str_mv | AT srivastavaajitesh memmapcompactandgeneralizablemetalstmmodelsformemoryaccessprediction AT wangtayang memmapcompactandgeneralizablemetalstmmodelsformemoryaccessprediction AT zhangpengmiao memmapcompactandgeneralizablemetalstmmodelsformemoryaccessprediction AT derosecesaraugustof memmapcompactandgeneralizablemetalstmmodelsformemoryaccessprediction AT kannanrajgopal memmapcompactandgeneralizablemetalstmmodelsformemoryaccessprediction AT prasannaviktork memmapcompactandgeneralizablemetalstmmodelsformemoryaccessprediction |