Deep Reinforcement Learning Based Trajectory Planning Under Uncertain Constraints
- PMID: 35586262
- PMCID: PMC9108367
- DOI: 10.3389/fnbot.2022.883562
Deep Reinforcement Learning Based Trajectory Planning Under Uncertain Constraints
Abstract
With the advance in algorithms, deep reinforcement learning (DRL) offers solutions to trajectory planning under uncertain environments. Different from traditional trajectory planning which requires lots of effort to tackle complicated high-dimensional problems, the recently proposed DRL enables the robot manipulator to autonomously learn and discover optimal trajectory planning by interacting with the environment. In this article, we present state-of-the-art DRL-based collision-avoidance trajectory planning for uncertain environments such as a safe human coexistent environment. Since the robot manipulator operates in high dimensional continuous state-action spaces, model-free, policy gradient-based soft actor-critic (SAC), and deep deterministic policy gradient (DDPG) framework are adapted to our scenario for comparison. In order to assess our proposal, we simulate a 7-DOF Panda (Franka Emika) robot manipulator in the PyBullet physics engine and then evaluate its trajectory planning with reward, loss, safe rate, and accuracy. Finally, our final report shows the effectiveness of state-of-the-art DRL algorithms for trajectory planning under uncertain environments with zero collision after 5,000 episodes of training.
Keywords: collision avoidance; neural networks; reinforcement learning; robotics; trajectory planning; uncertain environment.
Copyright © 2022 Chen, Jiang, Cheng, Knoll and Zhou.
Conflict of interest statement
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Figures
References
-
- Adiyatov O., Varol H. A. (2017). A novel RRT*-based algorithm for motion planning in dynamic environments, in 2017 IEEE International Conference on Mechatronics and Automation (ICMA) (Takamatsu: ), 1416–1421. 10.1109/ICMA.2017.8016024 - DOI
-
- Amarjyoti S. (2017). Deep reinforcement learning for robotic manipulation-the state of the art. arXiv preprint arXiv:1701.08878. 10.48550/arXiv.1701.08878 - DOI
-
- Flacco F., Kröger T., De Luca A., Khatib O. (2012). A depth space approach to human-robot collision avoidance, in 2012 IEEE International Conference on Robotics and Automation (St Paul, MN: ), 338–345. 10.1109/ICRA.2012.6225245 - DOI
-
- Gu S., Holly E., Lillicrap T., Levine S. (2017). Deep reinforcement learning for robotic manipulation with asynchronous off-policy updates, in 2017 IEEE International Conference on Robotics and Automation (ICRA) (Singapore: ), 3389–3396. 10.1109/ICRA.2017.7989385 - DOI
-
- Gu S., Lillicrap T., Ghahramani Z., Turner R. E., Levine S. (2016). Q-prop: Sample-efficient policy gradient with an off-policy critic. arXiv preprint arXiv:1611.02247. 10.48550/arXiv.1611.02247 - DOI
LinkOut - more resources
Full Text Sources
