Cardinality and Bounding Constrained Portfolio Optimization Using Safe Reinforcement Learning

Yiran Li, Nanjiang Du, Xingke Song, Xiaoying Yang, Tianxiang Cui*, Ning Xue, Amin Farjudian, Jianfeng Ren, Wooi Ping Cheah

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contribution

13 Downloads (Pure)

Abstract

Portfolio optimization is a strategic approach aiming at achieving an optimal balance between risk and returns through the judicious allocation of limited capital across various assets. In recent years, there has been a growing interest in leveraging Deep Reinforcement Learning (DRL) to tackle the complexities of portfolio optimization. Despite its potential, a notable limitation of DRL algorithms is their inherent difficulty in integrating conflicted objectives with the reward functions throughout the learning process. Typically, DRL's reward function prioritizes the maximization of returns or other performance indicators, often overlooking the integration of risk aspects. Furthermore, the standard DRL framework struggles to incorporate practical constraints, such as cardinality and bounding, into the decision process. Without these constraints, the investment strategies developed might be unrealistic and unmanageable. To this end, in this paper, we propose an adaptive and safe DRL framework, which can dynamically optimize the portfolio weights while strictly respecting practical constraints. In our method, any infeasible action (i.e., one that violates the constraints) decided by the RL agent will be mapped to a feasible region using a safety layer. The extended Markowitz Mean-Variance (M-V) model is explicitly encoded in the safety layer to ensure the feasibility of the actions from the alternative views. In addition, we utilize Projection-based Interior-point Policy Optimization (IPO) to resolve multiple objectives and constraints in the examined problem. Extensive results on real-world datasets show that our method is effective in strictly respecting constraints under dynamic market environments, in contrast to prevailing data- driven trading strategies and conventional model-based static solutions.
Original languageEnglish
Title of host publication2024 International Joint Conference on Neural Networks (IJCNN)
PublisherIEEE
Number of pages8
ISBN (Electronic)9798350359312
ISBN (Print)9798350359329 (PoD)
DOIs
Publication statusPublished - 9 Sept 2024
Event2024 International Joint Conference on Neural Networks (IJCNN) - Yokohama, Japan
Duration: 30 Jun 20245 Jul 2024

Publication series

Name Proceedings of International Joint Conference on Neural Network
PublisherIEEE
Volume2024
ISSN (Print)2161-4393
ISSN (Electronic)2161-4407

Conference

Conference2024 International Joint Conference on Neural Networks (IJCNN)
Country/TerritoryJapan
CityYokohama
Period30/06/245/07/24

Keywords

  • Portfolio Optimization
  • Safe Reinforcement Learning
  • Constraint handling

Fingerprint

Dive into the research topics of 'Cardinality and Bounding Constrained Portfolio Optimization Using Safe Reinforcement Learning'. Together they form a unique fingerprint.

Cite this