SCALABILITY OF DEEP REINFORCEMENT LEARNING IN HIGH-DIMENSIONAL STATE SPACES
The swift development of Deep Reinforcement Learning (DRL) has facilitated notable advancements in addressing intricate sequential decision-making challenges. Nonetheless, scalability continues to pose a significant challenge when DRL is utilized in environments with high-dimensional state spaces, like autonomous driving, robotics, and financial modeling. When the dimensionality of state representations rises, conventional reinforcement learning methods encounter the curse of dimensionality, resulting in poor exploration, elevated computational expenses, and unstable convergence (Sutton & Barto, 2018). This study explores the scalability challenges of DRL in these environments and examines cutting-edge methods aimed at tackling these issues. Methods such as function approximation with deep neural networks, dimensionality reduction, representation learning, and hierarchical reinforcement learning are analyzed for their efficacy in enhancing scalability (LeCun et al., 2015). Moreover, recent developments including attention mechanisms, distributed training frameworks, and model-based reinforcement learning are examined to emphasize their contribution to improving performance in high-dimensional environments (Mnih et al., 2015; Silver et al., 2016). The research addresses the trade-offs between computational efficiency and learning precision, offering insights into enhancing DRL frameworks for practical applications. Through the integration of existing studies, this paper seeks to highlight significant limitations and suggest future research paths to facilitate scalable and effective DRL systems
Thakur, A. S. (2026). Scalability of Deep Reinforcement Learning in High-Dimensional State Spaces. International Journal of Science, Strategic Management and Technology, 02(04). https://doi.org/10.55041/ijsmt.v2i4.293
Thakur, Ajay. "Scalability of Deep Reinforcement Learning in High-Dimensional State Spaces." International Journal of Science, Strategic Management and Technology, vol. 02, no. 04, 2026, pp. . doi:https://doi.org/10.55041/ijsmt.v2i4.293.
Thakur, Ajay. "Scalability of Deep Reinforcement Learning in High-Dimensional State Spaces." International Journal of Science, Strategic Management and Technology 02, no. 04 (2026). https://doi.org/https://doi.org/10.55041/ijsmt.v2i4.293.
2.Mnih, V., Kavukcuoglu, K., Silver, D., et al. (2015). Human-level control through deep reinforcement learning. Nature, 518(7540), 529–533.
3.Silver, D., Huang, A., Maddison, C. J., et al. (2016). Mastering the game of Go with deep neural networks and tree search. Nature, 529(7587), 484–489.
4.Sutton, R. S., & Barto, A. G. (2018). Reinforcement Learning: An Introduction (2nd ed.). MIT Press.
5.Richard S. Sutton, & Andrew G. Barto. (2018). Reinforcement learning: An introduction (2nd ed.). MIT Press.
6.Yann LeCun, Yoshua Bengio, & Geoffrey Hinton. (2015). Deep learning. Nature, 521(7553), 436–444.
7.Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Andrei A. Rusu, Joel Veness, Marc G. Bellemare, et al. (2015). Human-level control through deep reinforcement learning. Nature, 518(7540), 529–533.
8.Richard S. Sutton, David A. McAllester, Satinder Singh, & Yishay Mansour. (2000). Policy gradient methods for reinforcement learning with function approximation. Advances in Neural Information Processing Systems, 12, 1057–1063.
9.Timothy P. Lillicrap, Jonathan J. Hunt, Alexander Pritzel, Nicolas Heess, Tom Erez, Yuval Tassa, et al. (2016). Continuous control with deep reinforcement learning. ICLR.
10.Tuomas Haarnoja, Aurick Zhou, Pieter Abbeel, & Sergey Levine. (2018). Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. ICML.