Cargando…

Asynchronous Deep Double Dueling Q-learning for trading-signal execution in limit order book markets

We employ deep reinforcement learning (RL) to train an agent to successfully translate a high-frequency trading signal into a trading strategy that places individual limit orders. Based on the ABIDES limit order book simulator, we build a reinforcement learning OpenAI gym environment and utilize it...

Descripción completa

Detalles Bibliográficos
Autores principales: Nagy, Peer, Calliess, Jan-Peter, Zohren, Stefan
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10561243/
https://www.ncbi.nlm.nih.gov/pubmed/37818429
http://dx.doi.org/10.3389/frai.2023.1151003
_version_ 1785117879484547072
author Nagy, Peer
Calliess, Jan-Peter
Zohren, Stefan
author_facet Nagy, Peer
Calliess, Jan-Peter
Zohren, Stefan
author_sort Nagy, Peer
collection PubMed
description We employ deep reinforcement learning (RL) to train an agent to successfully translate a high-frequency trading signal into a trading strategy that places individual limit orders. Based on the ABIDES limit order book simulator, we build a reinforcement learning OpenAI gym environment and utilize it to simulate a realistic trading environment for NASDAQ equities based on historic order book messages. To train a trading agent that learns to maximize its trading return in this environment, we use Deep Dueling Double Q-learning with the APEX (asynchronous prioritized experience replay) architecture. The agent observes the current limit order book state, its recent history, and a short-term directional forecast. To investigate the performance of RL for adaptive trading independently from a concrete forecasting algorithm, we study the performance of our approach utilizing synthetic alpha signals obtained by perturbing forward-looking returns with varying levels of noise. Here, we find that the RL agent learns an effective trading strategy for inventory management and order placing that outperforms a heuristic benchmark trading strategy having access to the same signal.
format Online
Article
Text
id pubmed-10561243
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-105612432023-10-10 Asynchronous Deep Double Dueling Q-learning for trading-signal execution in limit order book markets Nagy, Peer Calliess, Jan-Peter Zohren, Stefan Front Artif Intell Artificial Intelligence We employ deep reinforcement learning (RL) to train an agent to successfully translate a high-frequency trading signal into a trading strategy that places individual limit orders. Based on the ABIDES limit order book simulator, we build a reinforcement learning OpenAI gym environment and utilize it to simulate a realistic trading environment for NASDAQ equities based on historic order book messages. To train a trading agent that learns to maximize its trading return in this environment, we use Deep Dueling Double Q-learning with the APEX (asynchronous prioritized experience replay) architecture. The agent observes the current limit order book state, its recent history, and a short-term directional forecast. To investigate the performance of RL for adaptive trading independently from a concrete forecasting algorithm, we study the performance of our approach utilizing synthetic alpha signals obtained by perturbing forward-looking returns with varying levels of noise. Here, we find that the RL agent learns an effective trading strategy for inventory management and order placing that outperforms a heuristic benchmark trading strategy having access to the same signal. Frontiers Media S.A. 2023-09-25 /pmc/articles/PMC10561243/ /pubmed/37818429 http://dx.doi.org/10.3389/frai.2023.1151003 Text en Copyright © 2023 Nagy, Calliess and Zohren. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Artificial Intelligence
Nagy, Peer
Calliess, Jan-Peter
Zohren, Stefan
Asynchronous Deep Double Dueling Q-learning for trading-signal execution in limit order book markets
title Asynchronous Deep Double Dueling Q-learning for trading-signal execution in limit order book markets
title_full Asynchronous Deep Double Dueling Q-learning for trading-signal execution in limit order book markets
title_fullStr Asynchronous Deep Double Dueling Q-learning for trading-signal execution in limit order book markets
title_full_unstemmed Asynchronous Deep Double Dueling Q-learning for trading-signal execution in limit order book markets
title_short Asynchronous Deep Double Dueling Q-learning for trading-signal execution in limit order book markets
title_sort asynchronous deep double dueling q-learning for trading-signal execution in limit order book markets
topic Artificial Intelligence
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10561243/
https://www.ncbi.nlm.nih.gov/pubmed/37818429
http://dx.doi.org/10.3389/frai.2023.1151003
work_keys_str_mv AT nagypeer asynchronousdeepdoubleduelingqlearningfortradingsignalexecutioninlimitorderbookmarkets
AT calliessjanpeter asynchronousdeepdoubleduelingqlearningfortradingsignalexecutioninlimitorderbookmarkets
AT zohrenstefan asynchronousdeepdoubleduelingqlearningfortradingsignalexecutioninlimitorderbookmarkets