Cargando…
An Analysis of the Value of Information When Exploring Stochastic, Discrete Multi-Armed Bandits
In this paper, we propose an information-theoretic exploration strategy for stochastic, discrete multi-armed bandits that achieves optimal regret. Our strategy is based on the value of information criterion. This criterion measures the trade-off between policy information and obtainable rewards. Hig...
Autores principales: | Sledge, Isaac J., Príncipe, José C. |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2018
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7512671/ https://www.ncbi.nlm.nih.gov/pubmed/33265246 http://dx.doi.org/10.3390/e20030155 |
Ejemplares similares
-
Introduction to multi-armed bandits
por: Slivkins, Aleksandrs
Publicado: (2019) -
Multi-Armed Bandits in Brain-Computer Interfaces
por: Heskebeck, Frida, et al.
Publicado: (2022) -
Some performance considerations when using multi-armed bandit algorithms in the presence of missing data
por: Chen, Xijin, et al.
Publicado: (2022) -
The Perils of Misspecified Priors and Optional Stopping in Multi-Armed Bandits
por: Loecher, Markus
Publicado: (2021) -
Arm order recognition in multi-armed bandit problem with laser chaos time series
por: Narisawa, Naoki, et al.
Publicado: (2021)