Abstract:
The collision avoidance in real-time is crucial for unmanned surface vehicles (USVs) in a complex environment. Traditional methods make it hard to ensure the balance of c...Show MoreMetadata
Abstract:
The collision avoidance in real-time is crucial for unmanned surface vehicles (USVs) in a complex environment. Traditional methods make it hard to ensure the balance of control decisions. To balance safety and practicality, a collision avoidance algorithm based on deep reinforcement learning (DRL) and a two-level incentive reward based on the principle of complementarity is proposed. To address the vital sparse reward problem of Deep Deterministic Policy Gradient (DDPG), the trajectory evaluation function of the dynamic window algorithm (DWA) is referred to construct the primary reward strategy, and a secondary incentive reward is constructed based on velocity obstacle (VO) to eliminate potential collision risks. To improve the efficiency of training, the electronic chart (EC) and Unity3D are used to build an immersive simulation platform. Based on it, simulations are made to verify the performance. In addition, field experiments are first conducted in various encounter scenarios to verify the effectiveness. The results show that it can take safe collision avoidance actions and get practical paths in various situations.
Published in: IEEE Transactions on Intelligent Transportation Systems ( Volume: 25, Issue: 12, December 2024)
Funding Agency:
Keywords assist with retrieval of results and provide a means to discovering other relevant content. Learn more.
- IEEE Keywords
- Index Terms
- Complex Environment ,
- Deep Reinforcement Learning ,
- Deep Reinforcement Learning Approach ,
- Balancing Algorithm ,
- Autonomous Surface Vehicles ,
- Collision Avoidance Algorithm ,
- Field Experiments ,
- Evaluation Of Function ,
- Collision Risk ,
- Electronic Charts ,
- Potential Collision ,
- Complementarity Principle ,
- Reward Incentive ,
- State Space ,
- Field Test ,
- Maximum Speed ,
- Simulation Environment ,
- Path Planning ,
- Typical Scenario ,
- Obstacle Avoidance ,
- Dynamic Obstacles ,
- Artificial Potential Field ,
- Cumulative Changes ,
- Deep Q-network ,
- Proximal Policy Optimization ,
- Cumulative Difference ,
- Offshore Environment ,
- Closest Distance ,
- Negative Reward ,
- Lower Left Corner
- Author Keywords
Keywords assist with retrieval of results and provide a means to discovering other relevant content. Learn more.
- IEEE Keywords
- Index Terms
- Complex Environment ,
- Deep Reinforcement Learning ,
- Deep Reinforcement Learning Approach ,
- Balancing Algorithm ,
- Autonomous Surface Vehicles ,
- Collision Avoidance Algorithm ,
- Field Experiments ,
- Evaluation Of Function ,
- Collision Risk ,
- Electronic Charts ,
- Potential Collision ,
- Complementarity Principle ,
- Reward Incentive ,
- State Space ,
- Field Test ,
- Maximum Speed ,
- Simulation Environment ,
- Path Planning ,
- Typical Scenario ,
- Obstacle Avoidance ,
- Dynamic Obstacles ,
- Artificial Potential Field ,
- Cumulative Changes ,
- Deep Q-network ,
- Proximal Policy Optimization ,
- Cumulative Difference ,
- Offshore Environment ,
- Closest Distance ,
- Negative Reward ,
- Lower Left Corner
- Author Keywords