TY - GEN
T1 - Cardinality and Bounding Constrained Portfolio Optimization Using Safe Reinforcement Learning
AU - Li, Yiran
AU - Du, Nanjiang
AU - Song, Xingke
AU - Yang, Xiaoying
AU - Cui, Tianxiang
AU - Xue, Ning
AU - Farjudian, Amin
AU - Ren, Jianfeng
AU - Cheah, Wooi Ping
PY - 2024/9/9
Y1 - 2024/9/9
N2 - Portfolio optimization is a strategic approach aiming at achieving an optimal balance between risk and returns through the judicious allocation of limited capital across various assets. In recent years, there has been a growing interest in leveraging Deep Reinforcement Learning (DRL) to tackle the complexities of portfolio optimization. Despite its potential, a notable limitation of DRL algorithms is their inherent difficulty in integrating conflicted objectives with the reward functions throughout the learning process. Typically, DRL's reward function prioritizes the maximization of returns or other performance indicators, often overlooking the integration of risk aspects. Furthermore, the standard DRL framework struggles to incorporate practical constraints, such as cardinality and bounding, into the decision process. Without these constraints, the investment strategies developed might be unrealistic and unmanageable. To this end, in this paper, we propose an adaptive and safe DRL framework, which can dynamically optimize the portfolio weights while strictly respecting practical constraints. In our method, any infeasible action (i.e., one that violates the constraints) decided by the RL agent will be mapped to a feasible region using a safety layer. The extended Markowitz Mean-Variance (M-V) model is explicitly encoded in the safety layer to ensure the feasibility of the actions from the alternative views. In addition, we utilize Projection-based Interior-point Policy Optimization (IPO) to resolve multiple objectives and constraints in the examined problem. Extensive results on real-world datasets show that our method is effective in strictly respecting constraints under dynamic market environments, in contrast to prevailing data- driven trading strategies and conventional model-based static solutions.
AB - Portfolio optimization is a strategic approach aiming at achieving an optimal balance between risk and returns through the judicious allocation of limited capital across various assets. In recent years, there has been a growing interest in leveraging Deep Reinforcement Learning (DRL) to tackle the complexities of portfolio optimization. Despite its potential, a notable limitation of DRL algorithms is their inherent difficulty in integrating conflicted objectives with the reward functions throughout the learning process. Typically, DRL's reward function prioritizes the maximization of returns or other performance indicators, often overlooking the integration of risk aspects. Furthermore, the standard DRL framework struggles to incorporate practical constraints, such as cardinality and bounding, into the decision process. Without these constraints, the investment strategies developed might be unrealistic and unmanageable. To this end, in this paper, we propose an adaptive and safe DRL framework, which can dynamically optimize the portfolio weights while strictly respecting practical constraints. In our method, any infeasible action (i.e., one that violates the constraints) decided by the RL agent will be mapped to a feasible region using a safety layer. The extended Markowitz Mean-Variance (M-V) model is explicitly encoded in the safety layer to ensure the feasibility of the actions from the alternative views. In addition, we utilize Projection-based Interior-point Policy Optimization (IPO) to resolve multiple objectives and constraints in the examined problem. Extensive results on real-world datasets show that our method is effective in strictly respecting constraints under dynamic market environments, in contrast to prevailing data- driven trading strategies and conventional model-based static solutions.
KW - Portfolio Optimization
KW - Safe Reinforcement Learning
KW - Constraint handling
U2 - 10.1109/IJCNN60899.2024.10651491
DO - 10.1109/IJCNN60899.2024.10651491
M3 - Conference contribution
SN - 9798350359329 (PoD)
T3 - Proceedings of International Joint Conference on Neural Network
BT - 2024 International Joint Conference on Neural Networks (IJCNN)
PB - IEEE
T2 - 2024 International Joint Conference on Neural Networks (IJCNN)
Y2 - 30 June 2024 through 5 July 2024
ER -