Cargando…
Adversarial Thresholding Semi-Bandits
The classical multi-armed bandit is one of the most common examples of sequential decision-making, either by trading-off between exploiting and exploring arms to maximise some payoff or purely exploring arms until the optimal arm is identified. In particular, a bandit player wanting to only pull arm...
Autor principal: | Bower, Craig Steven |
---|---|
Lenguaje: | eng |
Publicado: |
2021
|
Materias: | |
Acceso en línea: | http://cds.cern.ch/record/2790271 |
Ejemplares similares
-
Introduction to multi-armed bandits
por: Slivkins, Aleksandrs
Publicado: (2019) -
Bandit Algorithms for Website Optimization
por: White, John Myles
Publicado: (2013) -
Benchmark of Generative Adversarial Networks for Fast HEP Calorimeter Simulations
por: Rehm, Florian, et al.
Publicado: (2021) -
Large-scale distributed training applied to generative adversarial networks for calorimeter simulation
por: Vlimant, Jean-Roch, et al.
Publicado: (2019) -
Adversarial bandit approach for RIS-aided OFDM communication
por: Ahmed Ouameur, Messaoud, et al.
Publicado: (2022)