Cargando…

MemMAP: Compact and Generalizable Meta-LSTM Models for Memory Access Prediction

With the rise of Big Data, there has been a significant effort in increasing compute power through GPUs, TPUs, and heterogeneous architectures. As a result, many applications are memory bound, i.e., they are bottlenecked by the movement of data from main memory to compute units. One way to address t...

Descripción completa

Detalles Bibliográficos
Autores principales: Srivastava, Ajitesh, Wang, Ta-Yang, Zhang, Pengmiao, De Rose, Cesar Augusto F., Kannan, Rajgopal, Prasanna, Viktor K.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7206302/
http://dx.doi.org/10.1007/978-3-030-47436-2_5
_version_ 1783530389913993216
author Srivastava, Ajitesh
Wang, Ta-Yang
Zhang, Pengmiao
De Rose, Cesar Augusto F.
Kannan, Rajgopal
Prasanna, Viktor K.
author_facet Srivastava, Ajitesh
Wang, Ta-Yang
Zhang, Pengmiao
De Rose, Cesar Augusto F.
Kannan, Rajgopal
Prasanna, Viktor K.
author_sort Srivastava, Ajitesh
collection PubMed
description With the rise of Big Data, there has been a significant effort in increasing compute power through GPUs, TPUs, and heterogeneous architectures. As a result, many applications are memory bound, i.e., they are bottlenecked by the movement of data from main memory to compute units. One way to address this issue is through data prefetching, which relies on accurate prediction of memory accesses. While recent deep learning models have performed well on sequence prediction problems, they are far too heavy in terms of model size and inference latency to be practical for data prefetching. Here, we propose extremely compact LSTM models that can predict the next memory access with high accuracy. Prior LSTM based work on access prediction has used orders of magnitude more parameters and developed one model for each application (trace). While one (specialized) model per application can result in more accuracy, it is not a scalable approach. In contrast, our models can predict for a class of applications by trading off specialization at the cost of few retraining steps at runtime, for a more generalizable compact meta-model. Our experiments on 13 benchmark applications demonstrate that three compact meta-models can obtain accuracy close to specialized models using few batches of retraining for majority of the applications.
format Online
Article
Text
id pubmed-7206302
institution National Center for Biotechnology Information
language English
publishDate 2020
record_format MEDLINE/PubMed
spelling pubmed-72063022020-05-08 MemMAP: Compact and Generalizable Meta-LSTM Models for Memory Access Prediction Srivastava, Ajitesh Wang, Ta-Yang Zhang, Pengmiao De Rose, Cesar Augusto F. Kannan, Rajgopal Prasanna, Viktor K. Advances in Knowledge Discovery and Data Mining Article With the rise of Big Data, there has been a significant effort in increasing compute power through GPUs, TPUs, and heterogeneous architectures. As a result, many applications are memory bound, i.e., they are bottlenecked by the movement of data from main memory to compute units. One way to address this issue is through data prefetching, which relies on accurate prediction of memory accesses. While recent deep learning models have performed well on sequence prediction problems, they are far too heavy in terms of model size and inference latency to be practical for data prefetching. Here, we propose extremely compact LSTM models that can predict the next memory access with high accuracy. Prior LSTM based work on access prediction has used orders of magnitude more parameters and developed one model for each application (trace). While one (specialized) model per application can result in more accuracy, it is not a scalable approach. In contrast, our models can predict for a class of applications by trading off specialization at the cost of few retraining steps at runtime, for a more generalizable compact meta-model. Our experiments on 13 benchmark applications demonstrate that three compact meta-models can obtain accuracy close to specialized models using few batches of retraining for majority of the applications. 2020-04-17 /pmc/articles/PMC7206302/ http://dx.doi.org/10.1007/978-3-030-47436-2_5 Text en © Springer Nature Switzerland AG 2020 This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic.
spellingShingle Article
Srivastava, Ajitesh
Wang, Ta-Yang
Zhang, Pengmiao
De Rose, Cesar Augusto F.
Kannan, Rajgopal
Prasanna, Viktor K.
MemMAP: Compact and Generalizable Meta-LSTM Models for Memory Access Prediction
title MemMAP: Compact and Generalizable Meta-LSTM Models for Memory Access Prediction
title_full MemMAP: Compact and Generalizable Meta-LSTM Models for Memory Access Prediction
title_fullStr MemMAP: Compact and Generalizable Meta-LSTM Models for Memory Access Prediction
title_full_unstemmed MemMAP: Compact and Generalizable Meta-LSTM Models for Memory Access Prediction
title_short MemMAP: Compact and Generalizable Meta-LSTM Models for Memory Access Prediction
title_sort memmap: compact and generalizable meta-lstm models for memory access prediction
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7206302/
http://dx.doi.org/10.1007/978-3-030-47436-2_5
work_keys_str_mv AT srivastavaajitesh memmapcompactandgeneralizablemetalstmmodelsformemoryaccessprediction
AT wangtayang memmapcompactandgeneralizablemetalstmmodelsformemoryaccessprediction
AT zhangpengmiao memmapcompactandgeneralizablemetalstmmodelsformemoryaccessprediction
AT derosecesaraugustof memmapcompactandgeneralizablemetalstmmodelsformemoryaccessprediction
AT kannanrajgopal memmapcompactandgeneralizablemetalstmmodelsformemoryaccessprediction
AT prasannaviktork memmapcompactandgeneralizablemetalstmmodelsformemoryaccessprediction