Cargando…

A reinforcement learning approach to improve the performance of the Avellaneda-Stoikov market-making algorithm

Market making is a high-frequency trading problem for which solutions based on reinforcement learning (RL) are being explored increasingly. This paper presents an approach to market making using deep reinforcement learning, with the novelty that, rather than to set the bid and ask prices directly, t...

Descripción completa

Detalles Bibliográficos
Autores principales: Falces Marin, Javier, Díaz Pardo de Vera, David, Lopez Gonzalo, Eduardo
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9767337/
https://www.ncbi.nlm.nih.gov/pubmed/36538547
http://dx.doi.org/10.1371/journal.pone.0277042
_version_ 1784853944226283520
author Falces Marin, Javier
Díaz Pardo de Vera, David
Lopez Gonzalo, Eduardo
author_facet Falces Marin, Javier
Díaz Pardo de Vera, David
Lopez Gonzalo, Eduardo
author_sort Falces Marin, Javier
collection PubMed
description Market making is a high-frequency trading problem for which solutions based on reinforcement learning (RL) are being explored increasingly. This paper presents an approach to market making using deep reinforcement learning, with the novelty that, rather than to set the bid and ask prices directly, the neural network output is used to tweak the risk aversion parameter and the output of the Avellaneda-Stoikov procedure to obtain bid and ask prices that minimise inventory risk. Two further contributions are, first, that the initial parameters for the Avellaneda-Stoikov equations are optimised with a genetic algorithm, which parameters are also used to create a baseline Avellaneda-Stoikov agent (Gen-AS); and second, that state-defining features forming the RL agent’s neural network input are selected based on their relative importance by means of a random forest. Two variants of the deep RL model (Alpha-AS-1 and Alpha-AS-2) were backtested on real data (L2 tick data from 30 days of bitcoin–dollar pair trading) alongside the Gen-AS model and two other baselines. The performance of the five models was recorded through four indicators (the Sharpe, Sortino and P&L-to-MAP ratios, and the maximum drawdown). Gen-AS outperformed the two other baseline models on all indicators, and in turn the two Alpha-AS models substantially outperformed Gen-AS on Sharpe, Sortino and P&L-to-MAP. Localised excessive risk-taking by the Alpha-AS models, as reflected in a few heavy dropdowns, is a source of concern for which possible solutions are discussed.
format Online
Article
Text
id pubmed-9767337
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-97673372022-12-21 A reinforcement learning approach to improve the performance of the Avellaneda-Stoikov market-making algorithm Falces Marin, Javier Díaz Pardo de Vera, David Lopez Gonzalo, Eduardo PLoS One Research Article Market making is a high-frequency trading problem for which solutions based on reinforcement learning (RL) are being explored increasingly. This paper presents an approach to market making using deep reinforcement learning, with the novelty that, rather than to set the bid and ask prices directly, the neural network output is used to tweak the risk aversion parameter and the output of the Avellaneda-Stoikov procedure to obtain bid and ask prices that minimise inventory risk. Two further contributions are, first, that the initial parameters for the Avellaneda-Stoikov equations are optimised with a genetic algorithm, which parameters are also used to create a baseline Avellaneda-Stoikov agent (Gen-AS); and second, that state-defining features forming the RL agent’s neural network input are selected based on their relative importance by means of a random forest. Two variants of the deep RL model (Alpha-AS-1 and Alpha-AS-2) were backtested on real data (L2 tick data from 30 days of bitcoin–dollar pair trading) alongside the Gen-AS model and two other baselines. The performance of the five models was recorded through four indicators (the Sharpe, Sortino and P&L-to-MAP ratios, and the maximum drawdown). Gen-AS outperformed the two other baseline models on all indicators, and in turn the two Alpha-AS models substantially outperformed Gen-AS on Sharpe, Sortino and P&L-to-MAP. Localised excessive risk-taking by the Alpha-AS models, as reflected in a few heavy dropdowns, is a source of concern for which possible solutions are discussed. Public Library of Science 2022-12-20 /pmc/articles/PMC9767337/ /pubmed/36538547 http://dx.doi.org/10.1371/journal.pone.0277042 Text en © 2022 Falces Marin et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Falces Marin, Javier
Díaz Pardo de Vera, David
Lopez Gonzalo, Eduardo
A reinforcement learning approach to improve the performance of the Avellaneda-Stoikov market-making algorithm
title A reinforcement learning approach to improve the performance of the Avellaneda-Stoikov market-making algorithm
title_full A reinforcement learning approach to improve the performance of the Avellaneda-Stoikov market-making algorithm
title_fullStr A reinforcement learning approach to improve the performance of the Avellaneda-Stoikov market-making algorithm
title_full_unstemmed A reinforcement learning approach to improve the performance of the Avellaneda-Stoikov market-making algorithm
title_short A reinforcement learning approach to improve the performance of the Avellaneda-Stoikov market-making algorithm
title_sort reinforcement learning approach to improve the performance of the avellaneda-stoikov market-making algorithm
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9767337/
https://www.ncbi.nlm.nih.gov/pubmed/36538547
http://dx.doi.org/10.1371/journal.pone.0277042
work_keys_str_mv AT falcesmarinjavier areinforcementlearningapproachtoimprovetheperformanceoftheavellanedastoikovmarketmakingalgorithm
AT diazpardodeveradavid areinforcementlearningapproachtoimprovetheperformanceoftheavellanedastoikovmarketmakingalgorithm
AT lopezgonzaloeduardo areinforcementlearningapproachtoimprovetheperformanceoftheavellanedastoikovmarketmakingalgorithm
AT falcesmarinjavier reinforcementlearningapproachtoimprovetheperformanceoftheavellanedastoikovmarketmakingalgorithm
AT diazpardodeveradavid reinforcementlearningapproachtoimprovetheperformanceoftheavellanedastoikovmarketmakingalgorithm
AT lopezgonzaloeduardo reinforcementlearningapproachtoimprovetheperformanceoftheavellanedastoikovmarketmakingalgorithm