IJSMT Journal

International Journal of Science, Strategic Management and Technology

An International, Peer-Reviewed, Open Access Scholarly Journal Indexed in recognized academic databases · DOI via Crossref The journal adheres to established scholarly publishing, peer-review, and research ethics guidelines set by the UGC

ISSN: 3108-1762 (Online)
webp (1)

Plagiarism Passed
Peer reviewed
Open Access

SCALABILITY OF DEEP REINFORCEMENT LEARNING IN HIGH-DIMENSIONAL STATE SPACES

AUTHORS:
Dr. Ajay Singh Thakur
Mentor
Affiliation
CC BY 4.0 License:
This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Abstract

The swift development of Deep Reinforcement Learning (DRL) has facilitated notable advancements in addressing intricate sequential decision-making challenges. Nonetheless, scalability continues to pose a significant challenge when DRL is utilized in environments with high-dimensional state spaces, like autonomous driving, robotics, and financial modeling. When the dimensionality of state representations rises, conventional reinforcement learning methods encounter the curse of dimensionality, resulting in poor exploration, elevated computational expenses, and unstable convergence (Sutton & Barto, 2018). This study explores the scalability challenges of DRL in these environments and examines cutting-edge methods aimed at tackling these issues. Methods such as function approximation with deep neural networks, dimensionality reduction, representation learning, and hierarchical reinforcement learning are analyzed for their efficacy in enhancing scalability (LeCun et al., 2015). Moreover, recent developments including attention mechanisms, distributed training frameworks, and model-based reinforcement learning are examined to emphasize their contribution to improving performance in high-dimensional environments (Mnih et al., 2015; Silver et al., 2016). The research addresses the trade-offs between computational efficiency and learning precision, offering insights into enhancing DRL frameworks for practical applications. Through the integration of existing studies, this paper seeks to highlight significant limitations and suggest future research paths to facilitate scalable and effective DRL systems

Keywords
Article Metrics
Article Views
16
PDF Downloads
0
HOW TO CITE
APA

MLA

Chicago

Copy

Thakur, A. S. (2026). Scalability of Deep Reinforcement Learning in High-Dimensional State Spaces. International Journal of Science, Strategic Management and Technology, 02(04). https://doi.org/10.55041/ijsmt.v2i4.293

Thakur, Ajay. "Scalability of Deep Reinforcement Learning in High-Dimensional State Spaces." International Journal of Science, Strategic Management and Technology, vol. 02, no. 04, 2026, pp. . doi:https://doi.org/10.55041/ijsmt.v2i4.293.

Thakur, Ajay. "Scalability of Deep Reinforcement Learning in High-Dimensional State Spaces." International Journal of Science, Strategic Management and Technology 02, no. 04 (2026). https://doi.org/https://doi.org/10.55041/ijsmt.v2i4.293.

References
1.LeCun, Y., Bengio, Y., & Hinton, G. (2015). Deep learning. Nature, 521(7553), 436–444.

2.Mnih, V., Kavukcuoglu, K., Silver, D., et al. (2015). Human-level control through deep reinforcement learning. Nature, 518(7540), 529–533.

3.Silver, D., Huang, A., Maddison, C. J., et al. (2016). Mastering the game of Go with deep neural networks and tree search. Nature, 529(7587), 484–489.

4.Sutton, R. S., & Barto, A. G. (2018). Reinforcement Learning: An Introduction (2nd ed.). MIT Press.

5.Richard S. Sutton, & Andrew G. Barto. (2018). Reinforcement learning: An introduction (2nd ed.). MIT Press.

6.Yann LeCun, Yoshua Bengio, & Geoffrey Hinton. (2015). Deep learning. Nature, 521(7553), 436–444.

7.Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Andrei A. Rusu, Joel Veness, Marc G. Bellemare, et al. (2015). Human-level control through deep reinforcement learning. Nature, 518(7540), 529–533.

8.Richard S. Sutton, David A. McAllester, Satinder Singh, & Yishay Mansour. (2000). Policy gradient methods for reinforcement learning with function approximation. Advances in Neural Information Processing Systems, 12, 1057–1063.

9.Timothy P. Lillicrap, Jonathan J. Hunt, Alexander Pritzel, Nicolas Heess, Tom Erez, Yuval Tassa, et al. (2016). Continuous control with deep reinforcement learning. ICLR.

10.Tuomas Haarnoja, Aurick Zhou, Pieter Abbeel, & Sergey Levine. (2018). Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. ICML.
Ethics and Compliance
✓ All ethical standards met
This article has undergone plagiarism screening and double-blind peer review. Editorial policies have been followed. Authors retain copyright under CC BY-NC 4.0 license. The research complies with ethical standards and institutional guidelines.
Indexed In
Similar Articles
Beyond Saturated Fat: A Comprehensive Review of Oleogels as Sustainable Alternatives in Food Systems
string(16) "Anit Chakraborty" Chakraborty, A.
(2026)
DOI: 10.55041/ijsmt.v2i3.387
Block-chain for Fraud detection and Risk Management in Finance
string(15) "K S Purushotham" Purushotham, K. S.et al.
(2026)
DOI: 10.55041/ijsmt.v2i4.172
Smart Wireless Solar Charging Station with Smart Switching Between Grid, Battery and Solar Power
string(21) "Priyanka Ramesh Patil" Patil, P. R.et al.
(2026)
DOI: 10.55041/ijsmt.v2i3.401
Scroll to Top