Cargando…
A reinforcement learning approach to improve the performance of the Avellaneda-Stoikov market-making algorithm
Market making is a high-frequency trading problem for which solutions based on reinforcement learning (RL) are being explored increasingly. This paper presents an approach to market making using deep reinforcement learning, with the novelty that, rather than to set the bid and ask prices directly, t...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9767337/ https://www.ncbi.nlm.nih.gov/pubmed/36538547 http://dx.doi.org/10.1371/journal.pone.0277042 |
_version_ | 1784853944226283520 |
---|---|
author | Falces Marin, Javier Díaz Pardo de Vera, David Lopez Gonzalo, Eduardo |
author_facet | Falces Marin, Javier Díaz Pardo de Vera, David Lopez Gonzalo, Eduardo |
author_sort | Falces Marin, Javier |
collection | PubMed |
description | Market making is a high-frequency trading problem for which solutions based on reinforcement learning (RL) are being explored increasingly. This paper presents an approach to market making using deep reinforcement learning, with the novelty that, rather than to set the bid and ask prices directly, the neural network output is used to tweak the risk aversion parameter and the output of the Avellaneda-Stoikov procedure to obtain bid and ask prices that minimise inventory risk. Two further contributions are, first, that the initial parameters for the Avellaneda-Stoikov equations are optimised with a genetic algorithm, which parameters are also used to create a baseline Avellaneda-Stoikov agent (Gen-AS); and second, that state-defining features forming the RL agent’s neural network input are selected based on their relative importance by means of a random forest. Two variants of the deep RL model (Alpha-AS-1 and Alpha-AS-2) were backtested on real data (L2 tick data from 30 days of bitcoin–dollar pair trading) alongside the Gen-AS model and two other baselines. The performance of the five models was recorded through four indicators (the Sharpe, Sortino and P&L-to-MAP ratios, and the maximum drawdown). Gen-AS outperformed the two other baseline models on all indicators, and in turn the two Alpha-AS models substantially outperformed Gen-AS on Sharpe, Sortino and P&L-to-MAP. Localised excessive risk-taking by the Alpha-AS models, as reflected in a few heavy dropdowns, is a source of concern for which possible solutions are discussed. |
format | Online Article Text |
id | pubmed-9767337 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-97673372022-12-21 A reinforcement learning approach to improve the performance of the Avellaneda-Stoikov market-making algorithm Falces Marin, Javier Díaz Pardo de Vera, David Lopez Gonzalo, Eduardo PLoS One Research Article Market making is a high-frequency trading problem for which solutions based on reinforcement learning (RL) are being explored increasingly. This paper presents an approach to market making using deep reinforcement learning, with the novelty that, rather than to set the bid and ask prices directly, the neural network output is used to tweak the risk aversion parameter and the output of the Avellaneda-Stoikov procedure to obtain bid and ask prices that minimise inventory risk. Two further contributions are, first, that the initial parameters for the Avellaneda-Stoikov equations are optimised with a genetic algorithm, which parameters are also used to create a baseline Avellaneda-Stoikov agent (Gen-AS); and second, that state-defining features forming the RL agent’s neural network input are selected based on their relative importance by means of a random forest. Two variants of the deep RL model (Alpha-AS-1 and Alpha-AS-2) were backtested on real data (L2 tick data from 30 days of bitcoin–dollar pair trading) alongside the Gen-AS model and two other baselines. The performance of the five models was recorded through four indicators (the Sharpe, Sortino and P&L-to-MAP ratios, and the maximum drawdown). Gen-AS outperformed the two other baseline models on all indicators, and in turn the two Alpha-AS models substantially outperformed Gen-AS on Sharpe, Sortino and P&L-to-MAP. Localised excessive risk-taking by the Alpha-AS models, as reflected in a few heavy dropdowns, is a source of concern for which possible solutions are discussed. Public Library of Science 2022-12-20 /pmc/articles/PMC9767337/ /pubmed/36538547 http://dx.doi.org/10.1371/journal.pone.0277042 Text en © 2022 Falces Marin et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. |
spellingShingle | Research Article Falces Marin, Javier Díaz Pardo de Vera, David Lopez Gonzalo, Eduardo A reinforcement learning approach to improve the performance of the Avellaneda-Stoikov market-making algorithm |
title | A reinforcement learning approach to improve the performance of the Avellaneda-Stoikov market-making algorithm |
title_full | A reinforcement learning approach to improve the performance of the Avellaneda-Stoikov market-making algorithm |
title_fullStr | A reinforcement learning approach to improve the performance of the Avellaneda-Stoikov market-making algorithm |
title_full_unstemmed | A reinforcement learning approach to improve the performance of the Avellaneda-Stoikov market-making algorithm |
title_short | A reinforcement learning approach to improve the performance of the Avellaneda-Stoikov market-making algorithm |
title_sort | reinforcement learning approach to improve the performance of the avellaneda-stoikov market-making algorithm |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9767337/ https://www.ncbi.nlm.nih.gov/pubmed/36538547 http://dx.doi.org/10.1371/journal.pone.0277042 |
work_keys_str_mv | AT falcesmarinjavier areinforcementlearningapproachtoimprovetheperformanceoftheavellanedastoikovmarketmakingalgorithm AT diazpardodeveradavid areinforcementlearningapproachtoimprovetheperformanceoftheavellanedastoikovmarketmakingalgorithm AT lopezgonzaloeduardo areinforcementlearningapproachtoimprovetheperformanceoftheavellanedastoikovmarketmakingalgorithm AT falcesmarinjavier reinforcementlearningapproachtoimprovetheperformanceoftheavellanedastoikovmarketmakingalgorithm AT diazpardodeveradavid reinforcementlearningapproachtoimprovetheperformanceoftheavellanedastoikovmarketmakingalgorithm AT lopezgonzaloeduardo reinforcementlearningapproachtoimprovetheperformanceoftheavellanedastoikovmarketmakingalgorithm |