TY - GEN
T1 - Deep Reinforcement Learning Based Mobile Robot Navigation in Crowd Environments
AU - Yang, Guang
AU - Guo, Yi
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Robots are becoming popular in assisting humans. The mobile robot navigation in human crowd environments has become more important. We propose a deep reinforcement learning-based mobile robot navigation method that takes the observation from the robot's onboard Lidar sensor as input and outputs the velocity control to the robot. A customized deep deterministic policy gradient (DDPG) method is developed that incorporates guiding points to guide the robot toward the global goal. We built a 3D simulation environment using an open dataset of real-world pedestrian trajectories that were collected in a large business center. The neural network models are trained and tested in such environments. We compare the performance of our proposed method with existing algorithms that include a classic motion planner, an existing DDPG method, and a generative adversarial imitation learning (GAIL) method. Using the measurement metrics of success rate, the number of times freezing, and normalized path length, extensive simulation results show that our method outperforms other state-of-the-art approaches in both trained and untrained environments. Our method has also better generalizability compared with the GAIL method.
AB - Robots are becoming popular in assisting humans. The mobile robot navigation in human crowd environments has become more important. We propose a deep reinforcement learning-based mobile robot navigation method that takes the observation from the robot's onboard Lidar sensor as input and outputs the velocity control to the robot. A customized deep deterministic policy gradient (DDPG) method is developed that incorporates guiding points to guide the robot toward the global goal. We built a 3D simulation environment using an open dataset of real-world pedestrian trajectories that were collected in a large business center. The neural network models are trained and tested in such environments. We compare the performance of our proposed method with existing algorithms that include a classic motion planner, an existing DDPG method, and a generative adversarial imitation learning (GAIL) method. Using the measurement metrics of success rate, the number of times freezing, and normalized path length, extensive simulation results show that our method outperforms other state-of-the-art approaches in both trained and untrained environments. Our method has also better generalizability compared with the GAIL method.
UR - http://www.scopus.com/inward/record.url?scp=85200707096&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85200707096&partnerID=8YFLogxK
U2 - 10.1109/UR61395.2024.10597481
DO - 10.1109/UR61395.2024.10597481
M3 - Conference contribution
AN - SCOPUS:85200707096
T3 - 2024 21st International Conference on Ubiquitous Robots, UR 2024
SP - 513
EP - 519
BT - 2024 21st International Conference on Ubiquitous Robots, UR 2024
T2 - 21st International Conference on Ubiquitous Robots, UR 2024
Y2 - 24 June 2024 through 27 June 2024
ER -