Akademska digitalna zbirka SLovenije - logo
E-resources
Full text
Peer reviewed
  • Deterministic Promotion Rei...
    Zhang, Yuxiang; Guo, Lulu; Gao, Bingzhao; Qu, Ting; Chen, Hong

    IEEE transactions on vehicular technology, 2020-Jan., 2020-1-00, 20200101, Volume: 69, Issue: 1
    Journal Article

    Reinforcement learning is regarded as a potential method to be applied in automated vehicles, but the stability and efficiency of algorithms are concerns. To improve them, the deterministic promotion reinforcement learning method is put forward, which can promote the policy determinately. Correspondingly, the policy evaluation in critic and the exploration in actor are improved, which combines a normalization-based evaluation and a model-free search guide. The aim is finding the right action exploration direction by critic, then the direction is used to update and guide action exploration in actor. The modified method decreases the dependencies of exploring a good action for promotional updating and only makes deterministic promotion in policy. Consequently, the efficiency of the algorithm is improved without loss in stability. More notably, it can relieve the cold-start and circumvent the limitations in learning with constrained physical systems. To verify the proposed method, the longitudinal velocity control problem for automated vehicles is considered, which contains car-following and non-car-following conditions in a unitized form. The learning system is established in Carsim. Furthermore, some different reinforcement learning technologies are used to accelerate learning. Real-vehicle experiments for validation are also given. The results indicate that the proposed method can achieve permissible learning performance in the longitudinal velocity continuous control problem.