TY - GEN
T1 - MA-FDRNN
T2 - 8th International Conference on Soft Computing and Machine Intelligence, ISCMI 2021
AU - Skeepers, Tarrin
AU - Van Zyl, Terence L.
AU - Paskaramoorthy, Andrew
N1 - Publisher Copyright:
© 2021 IEEE.
PY - 2021
Y1 - 2021
N2 - Reinforcement learning (RL) in the context of portfolio optimisation (PO) aims to generate profits beyond the abilities of human traders. However, there is no definitive framework sufficiently able to generate consistent profits in line with expectations on real-world stock exchanges. Previous research has demonstrated the ability of Deep Deterministic Policy Gradient (DDPG), Proximal Policy Optimisation (PPO), Ensemble of Identical Independent Evaluators (EIIE) and Fuzzy Deep Recurrent Neural Network (FDRNN) methodologies to generate positive results within controlled testing environments. This paper consolidates recent progress by evaluating these RL methodologies applied to five different stock exchanges. The selected methodologies are tested on real-world data within a simulated trading framework to assess their performance under different market conditions. To this end, we extend the FDRNN method to allow it to interact with a multi-asset market. As a result, in contrast to previous methods, our new Multi-Asset FDRNN (MA-FDRNN) can take short positions, which we hypothesised would give it an advantage over DDPG, PPO and EIIE. To assess the performance of the RL methods, we compare their results to a selection of benchmark PO algorithms. The results show the superiority of the MA-FDRNN method within the sideways and bear market, where it can exploit short positions and earn higher returns at lower risk than the other methods. In the bull markets, the benchmark algorithms outperform the RL agents, and in the crashed market, the RL methods and benchmark algorithms demonstrate similar performance. The results indicate that while RL methods do show some promise in achieving the goals of portfolio management, further research is needed to create an RL framework capable of succeeding in real-world market conditions.
AB - Reinforcement learning (RL) in the context of portfolio optimisation (PO) aims to generate profits beyond the abilities of human traders. However, there is no definitive framework sufficiently able to generate consistent profits in line with expectations on real-world stock exchanges. Previous research has demonstrated the ability of Deep Deterministic Policy Gradient (DDPG), Proximal Policy Optimisation (PPO), Ensemble of Identical Independent Evaluators (EIIE) and Fuzzy Deep Recurrent Neural Network (FDRNN) methodologies to generate positive results within controlled testing environments. This paper consolidates recent progress by evaluating these RL methodologies applied to five different stock exchanges. The selected methodologies are tested on real-world data within a simulated trading framework to assess their performance under different market conditions. To this end, we extend the FDRNN method to allow it to interact with a multi-asset market. As a result, in contrast to previous methods, our new Multi-Asset FDRNN (MA-FDRNN) can take short positions, which we hypothesised would give it an advantage over DDPG, PPO and EIIE. To assess the performance of the RL methods, we compare their results to a selection of benchmark PO algorithms. The results show the superiority of the MA-FDRNN method within the sideways and bear market, where it can exploit short positions and earn higher returns at lower risk than the other methods. In the bull markets, the benchmark algorithms outperform the RL agents, and in the crashed market, the RL methods and benchmark algorithms demonstrate similar performance. The results indicate that while RL methods do show some promise in achieving the goals of portfolio management, further research is needed to create an RL framework capable of succeeding in real-world market conditions.
KW - Algorithmic Finance
KW - Algorithmic Trading
KW - Automated Trading
KW - Portfolio Management
KW - Portfolio Optimisation
KW - Reinforcement Learning
UR - http://www.scopus.com/inward/record.url?scp=85124414895&partnerID=8YFLogxK
U2 - 10.1109/ISCMI53840.2021.9654987
DO - 10.1109/ISCMI53840.2021.9654987
M3 - Conference contribution
AN - SCOPUS:85124414895
T3 - 2021 8th International Conference on Soft Computing and Machine Intelligence, ISCMI 2021
SP - 32
EP - 37
BT - 2021 8th International Conference on Soft Computing and Machine Intelligence, ISCMI 2021
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 26 November 2021 through 27 November 2021
ER -