Active Collision Avoidance for Robotic Arm Based on Artificial Potential Field and Deep Reinforcement Learning
<p>The structure and coordinate system of the robot.</p> "> Figure 2
<p>(<b>a</b>) Rock drill arm mod: 1—arm swing joint; 2—arm pitch joint; 3—arm telescoping joint; 4—beam slewing joint; 5—beam pitch joint; 6—beam swing joint; 7—beam telescoping joint. (<b>b</b>) Arm simplified.</p> "> Figure 3
<p>The figure on the left shows the original experience and the figure on the right shows the transformed experience.</p> "> Figure 4
<p>Schematic diagram of the control points of the rock drilling robotic arm.</p> "> Figure 5
<p>Schematic diagram of repulsive forces generated at control points.</p> "> Figure 6
<p>Comparison plot of training success rates for different collision-free experience generation methods. Blue represents obstacle reset red represents obstacle removal, and the shaded area represent one standard deviation.</p> "> Figure 7
<p>Comparison plot of training success rates for different numbers of obstacles, where blue and green correspond to HER-CA and HER, respectively.</p> "> Figure 8
<p>The comparison between HER and HER-CA in three-dimensional pathfinding tasks.</p> "> Figure 9
<p>The process of real-time collision avoidance movement using only the APF method. From 1 to 4 represents the process of the robotic arm from the beginning to the end of the simulation.</p> "> Figure 10
<p>Robotic arm motion process using our guidance APF method. The green arrow represents the direction of the temporary target pose given by the action model. From 1 to 4 represents the process of the robotic arm from the beginning to the end of the simulation.</p> "> Figure 11
<p>The closest distance between the robot and the obstacles. The red dotted line is the distance limit.</p> "> Figure 12
<p>Robotic arm motion process using our guidance APF method. The green arrow represents the direction of the temporary target pose given by the action model. From 1 to 6 represents the process of the robotic arm from the beginning to the end of the simulation.</p> "> Figure 13
<p>The change in the value of each joint.</p> "> Figure 14
<p>The closest distance between the robot and the obstacles. The red dotted line is the distance limit.</p> ">
Abstract
:1. Introduction
- The artificial potential field has the problem of falling into local minima [4].
- DRL training is inefficient because of the need to avoid multiple obstacles.
- Propose a DRL-guided method for collision avoidance in a simplified robotic arm model. The local minima problem of APF is solved by DRL guidance, while the use of a simplified robotic arm reduces the difficulty of DRL training.
- Proposed Hindsight Experience Replay for Collision Avoidance (HER-CA) algorithm to improve the training effect. The algorithm allows the agent to learn about collision avoidance from collision experience.
- Full-body collision avoidance model for rock drilling robotic arm based on artificial potential field.
2. Deep Reinforcement Learning Method
2.1. Simplified Robotic Arm
2.2. Describe Environment
2.3. Hindsight Experience Replay for Collision Avoidance
Algorithm 1 Hindsight Experience Replay for Collision Avoidance |
|
3. Guided Artificial Potential Field
3.1. Attractive Force
3.2. Repulsive Force
3.2.1. Control Point Repulsion
3.2.2. Joint Space Repulsion
3.3. Guidance
Algorithm 2 Guided APF |
|
4. Experiments
4.1. Comparison of HER and HER-CA
4.2. Static Obstacle Avoidance
4.3. Dynamic Obstacle Avoidance
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Li, L.; Zou, H.; Liu, H.; Tu, W.; Chen, Y. Research Status and Development Trends on Intelligent Construction of Drill-and-BlastTunnels. China J. Highw. Transp. 2024, 1–27. Available online: https://kns.cnki.net/kcms/detail/61.1313.U.20231207.1324.002.html (accessed on 3 June 2024).
- Ananthanarayanan, H.S. Complete Path Planning of Higher DOF Manipulators in Human Like Environments. Ph.D. Thesis, University of Dayton, Dayton, OH, USA, 2015. [Google Scholar]
- Zhang, H.; Jia, Y.; Guo, Y.; Qian, K.; Song, A.; Xi, N. Online Sensor Information and Redundancy Resolution Based Obstacle Avoidance for High DOF Mobile Manipulator Teleoperation Regular Paper. Int. J. Adv. Robot. Syst. 2013, 10, 244. [Google Scholar] [CrossRef]
- Song, D.; Qu, J. Particle Swarm Optimization Algorithm Based on Artificial Potential Field. In Proceedings of the 2nd International Conference on Mechanical, Electronic, Control and Automation Engineering (MECAE 2018), Qingdao, China, 30–31 March 2018; pp. 591–594. [Google Scholar]
- Yu, H.; Ning, L. Coordinated Obstacle Avoidance of Multi-AUV Based on Improved Artificial Potential Field Method and Consistency Protocol. J. Mar. Sci. Eng. 2023, 11, 1157. [Google Scholar] [CrossRef]
- Wang, R.; Wang, J.; Wang, N. Research on path planning of mobile robot based on improved artificial potential field. In Proceedings of the 2015 Joint International Mechanical, Electronic and Information Technology Conference, Chongqing China, 18–20 December 2015; pp. 660–663. [Google Scholar] [CrossRef]
- Song, J.; Zhao, M.; Liu, Y.; Liu, H.; Guo, X. Multi-Rotor UAVs Path Planning Method based on Improved Artificial Potential Field Method. In Proceedings of the 2019 Chinese Control Conference (CCC), Guangzhou, China, 27–30 July 2019; pp. 8242–8247. [Google Scholar] [CrossRef]
- Liu, Z.; Wang, X.; Li, K. Research on path planning of multi-rotor UAV based on improved artificial potential field method. MATEC Web Conf. 2021, 336, 07006. [Google Scholar] [CrossRef]
- Chen, Q.; Liu, Y.; Wang, P. An Autonomous Obstacle Avoidance Method for Dual-Arm Surgical Robot Based on the Improved Artificial Potential Field Method. In Proceedings of the 2022 International Conference on Intelligent Robotics and Applications, Harbin, China, 1–3 August 2022; pp. 496–508. [Google Scholar]
- Miao, T.T.; Fang, H.Z.; An, K.; Sun, Y.B.; Fang, Z.H.; Guan, X.Q. Obstacle avoidance planning algorithm of robot arm based on attraction factor of artificial potential field. Comput. Eng. Des. 2024, 45, 578–586. [Google Scholar] [CrossRef]
- Liu, H.; Wu, J.; Fan, S.; Jin, M.; Fan, C. Integrated virtual impedance control based pose correction for a simultaneous three-fingered end-effector. Ind. Robot 2018, 45, 255–266. [Google Scholar] [CrossRef]
- Li, H.; Wang, X.; Huang, X.; Ma, Y.; Jiang, Z. Multi-Joint Active Collision Avoidance for Robot Based on Depth Visual Perception. IEEE/CAA J. Autom. Sin. 2022, 9, 2186–2189. [Google Scholar] [CrossRef]
- Chen, Y.; Chen, L.; Ding, J.; Liu, Y. Research on Real-Time Obstacle Avoidance Motion Planning of Industrial Robotic Arm Based on Artificial Potential Field Method in Joint Space. Appl. Sci. 2023, 13, 6973. [Google Scholar] [CrossRef]
- Park, S.O.; Lee, M.C.; Kim, J. Trajectory Planning with Collision Avoidance for Redundant Robots Using Jacobian and Artificial Potential Field-based Real-time Inverse Kinematics. Int. J. Control. Autom. Syst. 2020, 18, 2095–2107. [Google Scholar] [CrossRef]
- Cao, Y.; Guo, Y.; Li, L.; Zhu, B.; Zhao, Z. Reinforcement Learning-based Trajectory Planning for Manipulator Obstacle Avoidance. J. Mech. Transm. 2023, 47, 40–46+96. [Google Scholar] [CrossRef]
- Li, Y.; Zhang, C.; Chai, L. Collaborative obstacle avoidance trajectory planning for mobile robotic arms based on artificial potential field DDPG algorithm. Comput. Integr. Manuf. Syst. 2024, 1–15. Available online: https://link.cnki.net/urlid/11.5946.TP.20230920.0949.008 (accessed on 3 June 2024).
- Andrychowicz, M.; Wolski, F.; Ray, A.; Schneider, J.; Fong, R.; Welinder, P.; Mcgrew, B.; Tobin, J.; Abbeel, P.; Zaremba, W. Hindsight Experience Replay. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
- Tymoteusz, L.; Andrzej, M. Reinforcement Learning-Based Algorithm to Avoid Obstacles by the Anthropomorphic Robotic Arm. Appl. Sci. 2022, 12, 6629. [Google Scholar] [CrossRef]
- Evan, P.; MyeongSeop, K.; JaeHan, P.; JiHun, B.; JungSu, K. Path Planning for Multi-Arm Manipulators Using Deep Reinforcement Learning: Soft Actor-Critic with Hindsight Experience Replay. Sensors 2020, 20, 5911. [Google Scholar] [CrossRef] [PubMed]
- Bai, C.; Wang, L.; Wang, Y.; Wang, Z.; Zhao, R.; Bai, C.; Liu, P. Addressing Hindsight Bias in Multigoal Reinforcement Learning. IEEE Trans. Cybern. 2023, 53, 392–405. [Google Scholar] [CrossRef] [PubMed]
- Luu, T.M.; Yoo, C.D. Hindsight Goal Ranking on Replay Buffer for Sparse Reward Environment. IEEE Access 2021, 9, 51996–52007. [Google Scholar] [CrossRef]
- Fujimoto, S.; Van Hoof, H.; Meger, D. Addressing Function Approximation Error in Actor-Critic Methods. arXiv 2018, arXiv:1802.09477. [Google Scholar]
- Dankwa, S.; Zheng, W. Twin-Delayed DDPG: A Deep Reinforcement Learning Technique to Model a Continuous Movement of an Intelligent Robot Agent. In Proceedings of the ICVISP 2019: 3rd International Conference on Vision, Image and Signal Processing, Vancouver, BC, Canada, 26–28 August 2019. [Google Scholar]
Joint | (mm) | (mm) | ||
---|---|---|---|---|
1 | 0 | 200 | ||
2 | 0 | −35 | ||
3 | 4297 | 0 | 0 | |
4 | 0 | 0 | ||
5 | 650 | 80 | ||
6 | 679.5 | 241 | ||
7 | 0 | 3058 | 0 | 0 |
Symbol | Representation |
---|---|
x | Current position x-coordinate |
y | Current position y-coordinate |
Target position x-coordinate | |
Target position y-coordinate | |
Obstacle 1 location x-coordinate | |
Obstacle 1 location y-coordinate | |
Obstacle 2 location x-coordinate | |
Obstacle 2 location x-coordinate |
Spherical Obstacle and Robot | Information |
---|---|
Start joint value of the robot arm | [, , 0 mm, , , , 0 mm] |
Goal position and attitude of the end-effector | [8680 mm, 4506 mm, 1700 mm, , , ,] |
Goal joint value of the robot arm | [, −, 1515 mm, , , −, 1300 mm] |
Coordinates of obstacle 1 | [6002 mm, 3106 mm, 1089 mm] |
Coordinates of obstacle 2 | [6292 mm, 2556 mm, 2389 mm] |
Obstacle radius | 300 mm |
Spherical Obstacle | Information |
---|---|
Coordinates of obstacle 1 | [6002 mm, 3106 mm, 1089 mm] |
Coordinates of obstacle 2 | [6292 mm, 2556 mm, 2389 mm] |
Velocity of obstacles 1 | [0, −10, −30] (mm/iteration) |
Velocity of obstacles 2 | [0, −40, −10] (mm/iteration) |
Obstacle radius | 300 mm |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Xu, Q.; Zhang, T.; Zhou, K.; Lin, Y.; Ju, W. Active Collision Avoidance for Robotic Arm Based on Artificial Potential Field and Deep Reinforcement Learning. Appl. Sci. 2024, 14, 4936. https://doi.org/10.3390/app14114936
Xu Q, Zhang T, Zhou K, Lin Y, Ju W. Active Collision Avoidance for Robotic Arm Based on Artificial Potential Field and Deep Reinforcement Learning. Applied Sciences. 2024; 14(11):4936. https://doi.org/10.3390/app14114936
Chicago/Turabian StyleXu, Qiaoyu, Tianle Zhang, Kunpeng Zhou, Yansong Lin, and Wenhao Ju. 2024. "Active Collision Avoidance for Robotic Arm Based on Artificial Potential Field and Deep Reinforcement Learning" Applied Sciences 14, no. 11: 4936. https://doi.org/10.3390/app14114936
APA StyleXu, Q., Zhang, T., Zhou, K., Lin, Y., & Ju, W. (2024). Active Collision Avoidance for Robotic Arm Based on Artificial Potential Field and Deep Reinforcement Learning. Applied Sciences, 14(11), 4936. https://doi.org/10.3390/app14114936