Cargando…

Intelligent Sensing in Dynamic Environments Using Markov Decision Process

In a network of low-powered wireless sensors, it is essential to capture as many environmental events as possible while still preserving the battery life of the sensor node. This paper focuses on a real-time learning algorithm to extend the lifetime of a sensor node to sense and transmit environment...

Descripción completa

Detalles Bibliográficos
Autores principales: Nanayakkara, Thrishantha, Halgamuge, Malka N., Sridhar, Prasanna, Madni, Asad M.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Molecular Diversity Preservation International (MDPI) 2011
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3274088/
https://www.ncbi.nlm.nih.gov/pubmed/22346624
http://dx.doi.org/10.3390/s110101229
_version_ 1782223007882674176
author Nanayakkara, Thrishantha
Halgamuge, Malka N.
Sridhar, Prasanna
Madni, Asad M.
author_facet Nanayakkara, Thrishantha
Halgamuge, Malka N.
Sridhar, Prasanna
Madni, Asad M.
author_sort Nanayakkara, Thrishantha
collection PubMed
description In a network of low-powered wireless sensors, it is essential to capture as many environmental events as possible while still preserving the battery life of the sensor node. This paper focuses on a real-time learning algorithm to extend the lifetime of a sensor node to sense and transmit environmental events. A common method that is generally adopted in ad-hoc sensor networks is to periodically put the sensor nodes to sleep. The purpose of the learning algorithm is to couple the sensor’s sleeping behavior to the natural statistics of the environment hence that it can be in optimal harmony with changes in the environment, the sensors can sleep when steady environment and stay awake when turbulent environment. This paper presents theoretical and experimental validation of a reward based learning algorithm that can be implemented on an embedded sensor. The key contribution of the proposed approach is the design and implementation of a reward function that satisfies a trade-off between the above two mutually contradicting objectives, and a linear critic function to approximate the discounted sum of future rewards in order to perform policy learning.
format Online
Article
Text
id pubmed-3274088
institution National Center for Biotechnology Information
language English
publishDate 2011
publisher Molecular Diversity Preservation International (MDPI)
record_format MEDLINE/PubMed
spelling pubmed-32740882012-02-15 Intelligent Sensing in Dynamic Environments Using Markov Decision Process Nanayakkara, Thrishantha Halgamuge, Malka N. Sridhar, Prasanna Madni, Asad M. Sensors (Basel) Article In a network of low-powered wireless sensors, it is essential to capture as many environmental events as possible while still preserving the battery life of the sensor node. This paper focuses on a real-time learning algorithm to extend the lifetime of a sensor node to sense and transmit environmental events. A common method that is generally adopted in ad-hoc sensor networks is to periodically put the sensor nodes to sleep. The purpose of the learning algorithm is to couple the sensor’s sleeping behavior to the natural statistics of the environment hence that it can be in optimal harmony with changes in the environment, the sensors can sleep when steady environment and stay awake when turbulent environment. This paper presents theoretical and experimental validation of a reward based learning algorithm that can be implemented on an embedded sensor. The key contribution of the proposed approach is the design and implementation of a reward function that satisfies a trade-off between the above two mutually contradicting objectives, and a linear critic function to approximate the discounted sum of future rewards in order to perform policy learning. Molecular Diversity Preservation International (MDPI) 2011-01-20 /pmc/articles/PMC3274088/ /pubmed/22346624 http://dx.doi.org/10.3390/s110101229 Text en © 2011 by the authors; licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution license (http://creativecommons.org/licenses/by/3.0/).
spellingShingle Article
Nanayakkara, Thrishantha
Halgamuge, Malka N.
Sridhar, Prasanna
Madni, Asad M.
Intelligent Sensing in Dynamic Environments Using Markov Decision Process
title Intelligent Sensing in Dynamic Environments Using Markov Decision Process
title_full Intelligent Sensing in Dynamic Environments Using Markov Decision Process
title_fullStr Intelligent Sensing in Dynamic Environments Using Markov Decision Process
title_full_unstemmed Intelligent Sensing in Dynamic Environments Using Markov Decision Process
title_short Intelligent Sensing in Dynamic Environments Using Markov Decision Process
title_sort intelligent sensing in dynamic environments using markov decision process
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3274088/
https://www.ncbi.nlm.nih.gov/pubmed/22346624
http://dx.doi.org/10.3390/s110101229
work_keys_str_mv AT nanayakkarathrishantha intelligentsensingindynamicenvironmentsusingmarkovdecisionprocess
AT halgamugemalkan intelligentsensingindynamicenvironmentsusingmarkovdecisionprocess
AT sridharprasanna intelligentsensingindynamicenvironmentsusingmarkovdecisionprocess
AT madniasadm intelligentsensingindynamicenvironmentsusingmarkovdecisionprocess