Evaluating Deep Reinforcement Learning Models in Automated Trading Systems

Authors

  • Rama Rao Gose Assistant Professor, Department of CSE, Christu Jyothi Institute of Technology and Science, Jangaon Author
  • A Sandeep Assistant Professor, Department of CSE, Sri Indu College of Engineering and Technology-Hyderabad Author
  • A Sandeep Assistant Professor, Department of CSE, Sri Indu College of Engineering and Technology-Hyderabad Author
  • Shaik Munnisa Begum Assistant Professor, Department of CSE, Sri Indu College of Engineering and Technology-Hyderabad Author

DOI:

https://doi.org/10.33425/3066-1226.1113

Abstract

This research presents a comprehensive evaluation of Deep Reinforcement Learning (DRL) models— specifically, Deep Q-Network (DQN), Deep Deterministic Policy Gradient (DDPG), and Proximal Policy Optimization (PPO)—in the context of automated trading systems. The study compares these models across critical performance metrics, including cumulative returns, Sharpe ratio, maximum drawdown, and the number of profitable trades, to assess their effectiveness in dynamic and complex financial markets. Our findings indicate that PPO outperforms DQN and DDPG in terms of both profitability and risk management, achieving the highest cumulative return and the best risk-adjusted performance. DDPG also demonstrates strong potential, particularly in handling continuous action spaces, while DQN shows effectiveness in simpler, discrete decision-making environments. These results underscore the capability of DRL models to enhance automated trading strategies by adapting to evolving market conditions and optimizing long-term returns.

Published

2025-07-28

Issue

Section

Articles