Computer and Modernization ›› 2023, Vol. 0 ›› Issue (06): 33-38.doi: 10.3969/j.issn.1006-2475.2023.06.006

• DESIGN AND ANALYSIS OF ALGORITHM • Previous Articles     Next Articles

An Experience Replay Strategy Based on Mixed Samples

LAI Jian-bin, FENG Gang   

  1. School of Computer Science, South China Normal University, Guangzhou 510635, China
  • Received:2022-07-11 Revised:2022-08-24 Online:2023-06-28 Published:2023-06-28

Abstract: Experience replay strategy has become an important part of deep reinforcement learning algorithm. It can not only accelerate the convergence of deep reinforcement learning algorithm, but also enhance the performance of agents. Mainstream experience replay strategies use uniform sampling, priority experience replay, expert experience replay and other methods to accelerate learning. In order to further improve the utilization of experience samples in deep reinforcement learning, this paper proposes an experience replay strategy based on mixed samples (ER-MS). This strategy mainly uses two methods: immediate learning of the latest experience and review of successful experience. It immediately learns the latest samples generated by the interaction between the agent and the environment, and uses an additional experience buffer pool to save the samples of successful rounds for experience replay. Experiments show that the experience replay strategy based on mixed samples combined with DDPG algorithm can achieve better results in Open AI mujoco task.

Key words: experience replay, deep reinforcement learning, expert experience

CLC Number: