Correlation-assisted spatio-temporal reinforcement learning for stock revenue maximization

  • Jaehyun Chung
  • , Minjoo Kim
  • , Seokhyeon Min
  • , Hyunseok Choi
  • , Soohyun Park*
  • , Joongheon Kim
  • *Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Investors struggle with the unpredictable, nonlinear nature of stock price volatility. Econometric models based on machine learning algorithms have improved prediction accuracy but remain limited in dynamic and highly correlated markets. This paper builds upon the proximal policy optimization (PPO) algorithm, the well-established deep reinforcement learning (DRL) method, and proposes an enhanced variant called correlation graph-based PPO (CGPPO), which incorporates spatio-temporal stock correlations for more realistic and robust predictions. The reward function, designed based on trading frequency and portfolio value, enhances experimental sophistication by reflecting practical investment objectives. The experiment is conducted in the simulated market environment using four major Korean stocks while explicitly considering the correlations among them. Experimental results show that the proposed CGPPO algorithm outperforms baseline methods, achieving 64.60% reward convergence value during training and 69.04% prediction value during inference.

Original languageEnglish
Article number128361
JournalExpert Systems With Applications
Volume289
DOIs
Publication statusPublished - 2025 Sept 15

Bibliographical note

Publisher Copyright:
© 2025 Elsevier Ltd

Keywords

  • Deep reinforcement learning
  • Proximal policy optimization
  • Stock correlation
  • Stock prediction

ASJC Scopus subject areas

  • General Engineering
  • Computer Science Applications
  • Artificial Intelligence

Fingerprint

Dive into the research topics of 'Correlation-assisted spatio-temporal reinforcement learning for stock revenue maximization'. Together they form a unique fingerprint.

Cite this