TY - GEN
T1 - Dual Core Portfolio Strategy
T2 - 2024 International Conference on Mathematics and Machine Learning, ICMML 2024
AU - Cui, Xiangyu
AU - Sun, Ruoyu
AU - Zhou, Mian
AU - Su, Jionglong
AU - Wang, Chengyu
AU - Jiang, Zhengyong
N1 - Publisher Copyright:
© 2024 Copyright held by the owner/author(s). Publication rights licensed to ACM.
PY - 2025/1/13
Y1 - 2025/1/13
N2 - Reinforcement learning gains increasing popularity in portfolio management. However, in a complex stock trading circumstance, agent-based algorithms often face challenges such as slow convergence rates and inadequate cooperation between agents. These lead to learning inefficiencies, increased risk, and higher transaction costs. Finally, the generalizability of the trading strategy is reduced. To address these, we propose a novel multi-agent algorithm called the Dual Core Portfolio Strategy which integrates both deterministic and stochastic policies to capitalize on their complementary strengths. In this strategy, the Deep Deterministic Policy Gradient agent is proficient in deterministic policy learning, while the Soft Actor-Critic agent enhances exploration and generalization through stochastic policy. Multiple agents collaborate by making decisions and interacting with the environment, sharing a centralized critic network and their interaction trajectories. This approach strengthens the robustness and adaptability of the portfolio strategy, improving its generalizability. Experiments demonstrate that the Dual Core Portfolio Strategy model consistently outperforms traditional deep reinforcement learning models. The effectiveness is evaluated using data from 2018 to 2020 and from 2020 to 2022 for all constituent stocks in the DJIA. The DC-PS model achieves state-of-the-art results, with a minimum increase of 15.7% (from 0.213 to 0.247) in accumulated returns in 2021 and 2023, underlining its generalizability in the out-of-sample environment.
AB - Reinforcement learning gains increasing popularity in portfolio management. However, in a complex stock trading circumstance, agent-based algorithms often face challenges such as slow convergence rates and inadequate cooperation between agents. These lead to learning inefficiencies, increased risk, and higher transaction costs. Finally, the generalizability of the trading strategy is reduced. To address these, we propose a novel multi-agent algorithm called the Dual Core Portfolio Strategy which integrates both deterministic and stochastic policies to capitalize on their complementary strengths. In this strategy, the Deep Deterministic Policy Gradient agent is proficient in deterministic policy learning, while the Soft Actor-Critic agent enhances exploration and generalization through stochastic policy. Multiple agents collaborate by making decisions and interacting with the environment, sharing a centralized critic network and their interaction trajectories. This approach strengthens the robustness and adaptability of the portfolio strategy, improving its generalizability. Experiments demonstrate that the Dual Core Portfolio Strategy model consistently outperforms traditional deep reinforcement learning models. The effectiveness is evaluated using data from 2018 to 2020 and from 2020 to 2022 for all constituent stocks in the DJIA. The DC-PS model achieves state-of-the-art results, with a minimum increase of 15.7% (from 0.213 to 0.247) in accumulated returns in 2021 and 2023, underlining its generalizability in the out-of-sample environment.
KW - Centralized Critic Network
KW - Decision Support
KW - Deep Reinforcement Learning
KW - Economics
KW - Multi-Agent Algorithm
UR - http://www.scopus.com/inward/record.url?scp=105005725931&partnerID=8YFLogxK
U2 - 10.1145/3708360.3708384
DO - 10.1145/3708360.3708384
M3 - Conference Proceeding
AN - SCOPUS:105005725931
T3 - Proceedings of 2024 International Conference on Mathematics and Machine Learning, ICMML 2024
SP - 147
EP - 153
BT - Proceedings of 2024 International Conference on Mathematics and Machine Learning, ICMML 2024
PB - Association for Computing Machinery, Inc
Y2 - 8 November 2024 through 10 November 2024
ER -