Chinese Journal of Ship Research, Volume. 16, Issue 1, 105(2021)

Tracking control of intelligent ship based on deep reinforcement learning

Kang ZHU1, Zhen HUANG1, and Xuming WANG2
Author Affiliations
  • 1School of Automation, Wuhan University of Technology, Wuhan 430070, China
  • 2Intelligent Transport System Research Center, Wuhan University of Technology, Wuhan 430063, China
  • show less

    Objectives

    The tracking control of intelligent ships often faces the problem of low controller stability in complex control environments and manual algorithmic computing. In order to achieve precise tracking control, this paper proposes a controller based on deep reinforcement learning (DRL).

    Methods

    Guided by the line-of-sight (LOS) algorithm and based on the maneuvering characteristics and control requirements of ships, this paper formulates a path of Markov decision processes by following the control problem, designing its state space, action space and reward by applying a deep deterministic policy gradient (DDPG) algorithm to implement the controller. An off-line learning method was used to train the controller. After the training, a comparison was made with BP-PID control to analyze the control effects.

    Results

    Simulation results show that the deep reinforcement learning (DRL) controller can rapidly converge from the training process to meet the control requirements, with the advantages of small yaw error, and a visible reduction in the frequency of changes of the rudder angle.

    Conclusions

    The study results can provide a reference for the tracking control of intelligent ships.

    Tools

    Get Citation

    Copy Citation Text

    Kang ZHU, Zhen HUANG, Xuming WANG. Tracking control of intelligent ship based on deep reinforcement learning[J]. Chinese Journal of Ship Research, 2021, 16(1): 105

    Download Citation

    EndNote(RIS)BibTexPlain Text
    Save article for my favorites
    Paper Information

    Category:

    Received: Apr. 29, 2020

    Accepted: --

    Published Online: Mar. 27, 2025

    The Author Email:

    DOI:10.19693/j.issn.1673-3185.01940

    Topics