IJSMT Journal

International Journal of Science, Strategic Management and Technology

An International, Peer-Reviewed, Open Access Scholarly Journal Indexed in recognized academic databases · DOI via Crossref The journal adheres to established scholarly publishing, peer-review, and research ethics guidelines set by the UGC

ISSN: 3108-1762 (Online)
webp (1)

Plagiarism Passed
Peer reviewed
Open Access

SCALABILITY OF DEEP REINFORCEMENT LEARNING IN HIGH-DIMENSIONAL STATE SPACES

AUTHORS:
Dr. Ajay Singh Thakur
Mentor
Affiliation
CC BY 4.0 License:
This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Abstract

The swift development of Deep Reinforcement Learning (DRL) has facilitated notable advancements in addressing intricate sequential decision-making challenges. Nonetheless, scalability continues to pose a significant challenge when DRL is utilized in environments with high-dimensional state spaces, like autonomous driving, robotics, and financial modeling. When the dimensionality of state representations rises, conventional reinforcement learning methods encounter the curse of dimensionality, resulting in poor exploration, elevated computational expenses, and unstable convergence (Sutton & Barto, 2018). This study explores the scalability challenges of DRL in these environments and examines cutting-edge methods aimed at tackling these issues. Methods such as function approximation with deep neural networks, dimensionality reduction, representation learning, and hierarchical reinforcement learning are analyzed for their efficacy in enhancing scalability (LeCun et al., 2015). Moreover, recent developments including attention mechanisms, distributed training frameworks, and model-based reinforcement learning are examined to emphasize their contribution to improving performance in high-dimensional environments (Mnih et al., 2015; Silver et al., 2016). The research addresses the trade-offs between computational efficiency and learning precision, offering insights into enhancing DRL frameworks for practical applications. Through the integration of existing studies, this paper seeks to highlight significant limitations and suggest future research paths to facilitate scalable and effective DRL systems

Keywords
Article Metrics
Article Views
15
PDF Downloads
0
HOW TO CITE
APA

MLA

Chicago

Copy

Thakur, A. S. (2026). Scalability of Deep Reinforcement Learning in High-Dimensional State Spaces. International Journal of Science, Strategic Management and Technology, 02(04). https://doi.org/10.55041/ijsmt.v2i4.293

Thakur, Ajay. "Scalability of Deep Reinforcement Learning in High-Dimensional State Spaces." International Journal of Science, Strategic Management and Technology, vol. 02, no. 04, 2026, pp. . doi:https://doi.org/10.55041/ijsmt.v2i4.293.

Thakur, Ajay. "Scalability of Deep Reinforcement Learning in High-Dimensional State Spaces." International Journal of Science, Strategic Management and Technology 02, no. 04 (2026). https://doi.org/https://doi.org/10.55041/ijsmt.v2i4.293.

References
1.LeCun, Y., Bengio, Y., & Hinton, G. (2015). Deep learning. Nature, 521(7553), 436–444.

2.Mnih, V., Kavukcuoglu, K., Silver, D., et al. (2015). Human-level control through deep reinforcement learning. Nature, 518(7540), 529–533.

3.Silver, D., Huang, A., Maddison, C. J., et al. (2016). Mastering the game of Go with deep neural networks and tree search. Nature, 529(7587), 484–489.

4.Sutton, R. S., & Barto, A. G. (2018). Reinforcement Learning: An Introduction (2nd ed.). MIT Press.

5.Richard S. Sutton, & Andrew G. Barto. (2018). Reinforcement learning: An introduction (2nd ed.). MIT Press.

6.Yann LeCun, Yoshua Bengio, & Geoffrey Hinton. (2015). Deep learning. Nature, 521(7553), 436–444.

7.Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Andrei A. Rusu, Joel Veness, Marc G. Bellemare, et al. (2015). Human-level control through deep reinforcement learning. Nature, 518(7540), 529–533.

8.Richard S. Sutton, David A. McAllester, Satinder Singh, & Yishay Mansour. (2000). Policy gradient methods for reinforcement learning with function approximation. Advances in Neural Information Processing Systems, 12, 1057–1063.

9.Timothy P. Lillicrap, Jonathan J. Hunt, Alexander Pritzel, Nicolas Heess, Tom Erez, Yuval Tassa, et al. (2016). Continuous control with deep reinforcement learning. ICLR.

10.Tuomas Haarnoja, Aurick Zhou, Pieter Abbeel, & Sergey Levine. (2018). Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. ICML.
Ethics and Compliance
✓ All ethical standards met
This article has undergone plagiarism screening and double-blind peer review. Editorial policies have been followed. Authors retain copyright under CC BY-NC 4.0 license. The research complies with ethical standards and institutional guidelines.
Indexed In
Similar Articles
Hardware-Aware Bug Prediction:Leveraging Performance Counters for Early Software Defect Detection
string(9) "TC. Vivek" Vivek, T.
(2026)
DOI: 10.55041/ijsmt.v2i3.302
Relationship Between Emotional Intelligence and Leadership Effectiveness in Higher Education Institutions
string(18) "Neha P. Chatterjee" Chatterjee, N. P.et al.
(2026)
DOI: 10.55041/ijsmt.v1i3.002
Tigers, Trauma, and Livelihood: Understanding the Mental Health of Sundarbans’ Women Warriors
string(12) "Paulami Bose" Bose, P.
(2026)
DOI: 10.55041/ijsmt.v2i3.369
Scroll to Top