| Issue |
ITM Web Conf.
Volume 78, 2025
International Conference on Computer Science and Electronic Information Technology (CSEIT 2025)
|
|
|---|---|---|
| Article Number | 01021 | |
| Number of page(s) | 10 | |
| Section | Deep Learning and Reinforcement Learning – Theories and Applications | |
| DOI | https://doi.org/10.1051/itmconf/20257801021 | |
| Published online | 08 September 2025 | |
Deep Reinforcement Learning in Continuous Control: Advances and Challenges
School of Computer Science, Hangzhou Dianzi University, Hangzhou, China
Multiple media and web development, Fuzhou University, Fuzhou, China
International College, Chongqing University of Posts and Telecommunications, Chongqing, China
With the widespread application of Deep Reinforcement Learning (DRL) in continuous control, its learning and decision-making capabilities in high-dimensional state spaces have brought new opportunities for intelligent control systems. This paper provides a systematic review of the research progress of DRL in continuous control tasks, covering representative algorithms from classic Deep Deterministic Policy Gradient (DDPG) and Twin Delayed DDPG (TD3) to the soft policy method Soft Actor-Critic (SAC), and further focuses on key breakthroughs in algorithm structure and sample efficiency in recent years. In the latest research, one class of methods effectively improves the accuracy and stability of policy generation by aligning the behavior of diffusion models with the Q function; another class of methods employs Euclidean data augmentation techniques to enhance data diversity and generalization performance during the training process; additional studies introduce successor features and concurrent policy combination mechanisms, significantly improving transfer efficiency and adaptability in multi-task learning. These innovative algorithms show great potential in alleviating training instability, enhancing sample utilization, and improving policy generalization capabilities. Looking ahead, DRL research in continuous control can further explore efficient exploration mechanisms, model prediction integration, multi-agent collaboration, and real-world deployment, laying the foundation for building reliable and adaptive intelligent control systems.
© The Authors, published by EDP Sciences, 2025
This is an Open Access article distributed under the terms of the Creative Commons Attribution License 4.0, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.
Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.
Initial download of the metrics may take a while.

