Cargando…

Adiabatic Quantum Computation Applied to Deep Learning Networks

Training deep learning networks is a difficult task due to computational complexity, and this is traditionally handled by simplifying network topology to enable parallel computation on graphical processing units (GPUs). However, the emergence of quantum devices allows reconsideration of complex topo...

Descripción completa

Detalles Bibliográficos
Autores principales: Liu, Jeremy, Spedalieri, Federico M., Yao, Ke-Thia, Potok, Thomas E., Schuman, Catherine, Young, Steven, Patton, Robert, Rose, Garrett S., Chamka, Gangotree
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2018
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7512898/
https://www.ncbi.nlm.nih.gov/pubmed/33265470
http://dx.doi.org/10.3390/e20050380
_version_ 1783586263468605440
author Liu, Jeremy
Spedalieri, Federico M.
Yao, Ke-Thia
Potok, Thomas E.
Schuman, Catherine
Young, Steven
Patton, Robert
Rose, Garrett S.
Chamka, Gangotree
author_facet Liu, Jeremy
Spedalieri, Federico M.
Yao, Ke-Thia
Potok, Thomas E.
Schuman, Catherine
Young, Steven
Patton, Robert
Rose, Garrett S.
Chamka, Gangotree
author_sort Liu, Jeremy
collection PubMed
description Training deep learning networks is a difficult task due to computational complexity, and this is traditionally handled by simplifying network topology to enable parallel computation on graphical processing units (GPUs). However, the emergence of quantum devices allows reconsideration of complex topologies. We illustrate a particular network topology that can be trained to classify MNIST data (an image dataset of handwritten digits) and neutrino detection data using a restricted form of adiabatic quantum computation known as quantum annealing performed by a D-Wave processor. We provide a brief description of the hardware and how it solves Ising models, how we translate our data into the corresponding Ising models, and how we use available expanded topology options to explore potential performance improvements. Although we focus on the application of quantum annealing in this article, the work discussed here is just one of three approaches we explored as part of a larger project that considers alternative means for training deep learning networks. The other approaches involve using a high performance computing (HPC) environment to automatically find network topologies with good performance and using neuromorphic computing to find a low-power solution for training deep learning networks. Our results show that our quantum approach can find good network parameters in a reasonable time despite increased network topology complexity; that HPC can find good parameters for traditional, simplified network topologies; and that neuromorphic computers can use low power memristive hardware to represent complex topologies and parameters derived from other architecture choices.
format Online
Article
Text
id pubmed-7512898
institution National Center for Biotechnology Information
language English
publishDate 2018
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-75128982020-11-09 Adiabatic Quantum Computation Applied to Deep Learning Networks Liu, Jeremy Spedalieri, Federico M. Yao, Ke-Thia Potok, Thomas E. Schuman, Catherine Young, Steven Patton, Robert Rose, Garrett S. Chamka, Gangotree Entropy (Basel) Article Training deep learning networks is a difficult task due to computational complexity, and this is traditionally handled by simplifying network topology to enable parallel computation on graphical processing units (GPUs). However, the emergence of quantum devices allows reconsideration of complex topologies. We illustrate a particular network topology that can be trained to classify MNIST data (an image dataset of handwritten digits) and neutrino detection data using a restricted form of adiabatic quantum computation known as quantum annealing performed by a D-Wave processor. We provide a brief description of the hardware and how it solves Ising models, how we translate our data into the corresponding Ising models, and how we use available expanded topology options to explore potential performance improvements. Although we focus on the application of quantum annealing in this article, the work discussed here is just one of three approaches we explored as part of a larger project that considers alternative means for training deep learning networks. The other approaches involve using a high performance computing (HPC) environment to automatically find network topologies with good performance and using neuromorphic computing to find a low-power solution for training deep learning networks. Our results show that our quantum approach can find good network parameters in a reasonable time despite increased network topology complexity; that HPC can find good parameters for traditional, simplified network topologies; and that neuromorphic computers can use low power memristive hardware to represent complex topologies and parameters derived from other architecture choices. MDPI 2018-05-18 /pmc/articles/PMC7512898/ /pubmed/33265470 http://dx.doi.org/10.3390/e20050380 Text en © 2018 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Liu, Jeremy
Spedalieri, Federico M.
Yao, Ke-Thia
Potok, Thomas E.
Schuman, Catherine
Young, Steven
Patton, Robert
Rose, Garrett S.
Chamka, Gangotree
Adiabatic Quantum Computation Applied to Deep Learning Networks
title Adiabatic Quantum Computation Applied to Deep Learning Networks
title_full Adiabatic Quantum Computation Applied to Deep Learning Networks
title_fullStr Adiabatic Quantum Computation Applied to Deep Learning Networks
title_full_unstemmed Adiabatic Quantum Computation Applied to Deep Learning Networks
title_short Adiabatic Quantum Computation Applied to Deep Learning Networks
title_sort adiabatic quantum computation applied to deep learning networks
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7512898/
https://www.ncbi.nlm.nih.gov/pubmed/33265470
http://dx.doi.org/10.3390/e20050380
work_keys_str_mv AT liujeremy adiabaticquantumcomputationappliedtodeeplearningnetworks
AT spedalierifedericom adiabaticquantumcomputationappliedtodeeplearningnetworks
AT yaokethia adiabaticquantumcomputationappliedtodeeplearningnetworks
AT potokthomase adiabaticquantumcomputationappliedtodeeplearningnetworks
AT schumancatherine adiabaticquantumcomputationappliedtodeeplearningnetworks
AT youngsteven adiabaticquantumcomputationappliedtodeeplearningnetworks
AT pattonrobert adiabaticquantumcomputationappliedtodeeplearningnetworks
AT rosegarretts adiabaticquantumcomputationappliedtodeeplearningnetworks
AT chamkagangotree adiabaticquantumcomputationappliedtodeeplearningnetworks