File Download

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Portfolio Selection via Deep Reinforcement Learning: Comparative Analysis with Classical Strategies

Author(s)
Park, Jae Yeon
Advisor
Choi, Jin Hyuk
Issued Date
2024-08
URI
https://scholarworks.unist.ac.kr/handle/201301/84130 http://unist.dcollection.net/common/orgView/200000813736
Abstract
Portfolio management involves the strategic allocation of assets to achieve investment objectives while minimizing risk. This study explores portfolio management by comparing traditional Mean-Variance Optimization (MVO), the Black-Litterman Model, and Deep Reinforcement Learning (DRL) for optimizing portfolio allocations. MVO utilizes the Ledoit-Wolf shrinkage method to estimate covariance matrices and applies Efficient Frontier techniques for optimization. The Black-Litterman Model extends MVO by incorporating investor views into the equilibrium market returns derived from the Capital Asset Pricing Model (CAPM), offering a more balanced approach. In contrast, Proximal Policy Optimization (PPO) is used as the DRL method to dynamically adjust portfolio weights. Empirical analysis, based on backtesting with historical market data, shows that the DRL approach significantly outperforms both the MVO and the Black-Litterman Model across various performance metrics, including cumulative return, annual return, volatility, and Sharpe ratio. These results highlight the potential of DRL, particularly PPO strategies, as robust tools for dynamic and adaptive portfolio management, capable of achieving superior returns and effectively managing risks effectively in modern financial markets.
Publisher
Ulsan National Institute of Science and Technology
Degree
Master
Major
Department of Mathematical Sciences

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.