Cargando…
Asynchronous Deep Double Dueling Q-learning for trading-signal execution in limit order book markets
We employ deep reinforcement learning (RL) to train an agent to successfully translate a high-frequency trading signal into a trading strategy that places individual limit orders. Based on the ABIDES limit order book simulator, we build a reinforcement learning OpenAI gym environment and utilize it...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10561243/ https://www.ncbi.nlm.nih.gov/pubmed/37818429 http://dx.doi.org/10.3389/frai.2023.1151003 |
_version_ | 1785117879484547072 |
---|---|
author | Nagy, Peer Calliess, Jan-Peter Zohren, Stefan |
author_facet | Nagy, Peer Calliess, Jan-Peter Zohren, Stefan |
author_sort | Nagy, Peer |
collection | PubMed |
description | We employ deep reinforcement learning (RL) to train an agent to successfully translate a high-frequency trading signal into a trading strategy that places individual limit orders. Based on the ABIDES limit order book simulator, we build a reinforcement learning OpenAI gym environment and utilize it to simulate a realistic trading environment for NASDAQ equities based on historic order book messages. To train a trading agent that learns to maximize its trading return in this environment, we use Deep Dueling Double Q-learning with the APEX (asynchronous prioritized experience replay) architecture. The agent observes the current limit order book state, its recent history, and a short-term directional forecast. To investigate the performance of RL for adaptive trading independently from a concrete forecasting algorithm, we study the performance of our approach utilizing synthetic alpha signals obtained by perturbing forward-looking returns with varying levels of noise. Here, we find that the RL agent learns an effective trading strategy for inventory management and order placing that outperforms a heuristic benchmark trading strategy having access to the same signal. |
format | Online Article Text |
id | pubmed-10561243 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-105612432023-10-10 Asynchronous Deep Double Dueling Q-learning for trading-signal execution in limit order book markets Nagy, Peer Calliess, Jan-Peter Zohren, Stefan Front Artif Intell Artificial Intelligence We employ deep reinforcement learning (RL) to train an agent to successfully translate a high-frequency trading signal into a trading strategy that places individual limit orders. Based on the ABIDES limit order book simulator, we build a reinforcement learning OpenAI gym environment and utilize it to simulate a realistic trading environment for NASDAQ equities based on historic order book messages. To train a trading agent that learns to maximize its trading return in this environment, we use Deep Dueling Double Q-learning with the APEX (asynchronous prioritized experience replay) architecture. The agent observes the current limit order book state, its recent history, and a short-term directional forecast. To investigate the performance of RL for adaptive trading independently from a concrete forecasting algorithm, we study the performance of our approach utilizing synthetic alpha signals obtained by perturbing forward-looking returns with varying levels of noise. Here, we find that the RL agent learns an effective trading strategy for inventory management and order placing that outperforms a heuristic benchmark trading strategy having access to the same signal. Frontiers Media S.A. 2023-09-25 /pmc/articles/PMC10561243/ /pubmed/37818429 http://dx.doi.org/10.3389/frai.2023.1151003 Text en Copyright © 2023 Nagy, Calliess and Zohren. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Artificial Intelligence Nagy, Peer Calliess, Jan-Peter Zohren, Stefan Asynchronous Deep Double Dueling Q-learning for trading-signal execution in limit order book markets |
title | Asynchronous Deep Double Dueling Q-learning for trading-signal execution in limit order book markets |
title_full | Asynchronous Deep Double Dueling Q-learning for trading-signal execution in limit order book markets |
title_fullStr | Asynchronous Deep Double Dueling Q-learning for trading-signal execution in limit order book markets |
title_full_unstemmed | Asynchronous Deep Double Dueling Q-learning for trading-signal execution in limit order book markets |
title_short | Asynchronous Deep Double Dueling Q-learning for trading-signal execution in limit order book markets |
title_sort | asynchronous deep double dueling q-learning for trading-signal execution in limit order book markets |
topic | Artificial Intelligence |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10561243/ https://www.ncbi.nlm.nih.gov/pubmed/37818429 http://dx.doi.org/10.3389/frai.2023.1151003 |
work_keys_str_mv | AT nagypeer asynchronousdeepdoubleduelingqlearningfortradingsignalexecutioninlimitorderbookmarkets AT calliessjanpeter asynchronousdeepdoubleduelingqlearningfortradingsignalexecutioninlimitorderbookmarkets AT zohrenstefan asynchronousdeepdoubleduelingqlearningfortradingsignalexecutioninlimitorderbookmarkets |