[1]LI Kangbin,ZHU Qidan,MU Jinyou,et al.Automatic ship berthing path-planning method based on improved DDQN[J].CAAI Transactions on Intelligent Systems,2025,20(1):73-80.[doi:10.11992/tis.202401005]
Copy
CAAI Transactions on Intelligent Systems[ISSN 1673-4785/CN 23-1538/TP] Volume:
20
Number of periods:
2025 1
Page number:
73-80
Column:
学术论文—机器学习
Public date:
2025-01-05
- Title:
-
Automatic ship berthing path-planning method based on improved DDQN
- Author(s):
-
LI Kangbin; ZHU Qidan; MU Jinyou; JIAN Ziting
-
School of Intelligent Science and Engineering, Harbin Engineering University, Harbin 150001, China
-
- Keywords:
-
automatic berthing; path planning; deep reinforcement learning; double deep Q network; reward function; current velocity; state exploration; thrust; time; independent repeated experiments
- CLC:
-
TP273
- DOI:
-
10.11992/tis.202401005
- Abstract:
-
During the automatic docking process, ships are influenced by factors such as wind, waves, currents, and quay wall effects, necessitating precise path-planning methods to prevent docking failures. For fully actuated ships, a ship’s automatic docking path-planning method is designed based on the double deep Q network (DDQN) algorithm. Firstly, a three-degree-of-freedom model for the ship is established, and then the reward function is improved by incorporating distance, heading, thrust, time, and collisions as rewards or penalties. DDQN is then introduced to learn the action-reward model and use the learning results to manipulate ship movements. By pursuing higher reward values, the ship can autonomously find the optimal docking path. Experimental results show that under different water flow velocities, ships can reduce both time and thrust while completing docking. Moreover, at the same water flow velocity, compared with Q-learning, SARSA, and deep Q Network (DQN), the DDQN algorithm reduces thrust by 241.940N, 234.614N, and 80.202N respectively during the docking process, with the time being only 252.485 seconds.