Actor-Critic with Synthesis Loss for Solving Approximation Biases

Bo Wen Guo, Fei Chao*, Xiang Chang, Changjing Shang, Qiang Shen

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

1 Citation (Scopus)

Abstract

Approximation biases of value functions are considered a key problem in reinforcement learning (RL). In particular, existing RL algorithms are hindered by overestimation and underestimation biases, i.e., value mismatching between RL’s actual returns and action–value approximations limits the performance of RL algorithms. In this article, we first develop a new synthesis loss function for RL’s action–value estimation integrating a regularization term and a modified “clipped double Q -learning” structure for solving overestimation and underestimation biases. To minimize the differences between action–value estimations and actual returns in RL, we develop a new discrepancy function to determine the type and magnitude of approximation biases. Then, two coefficients embedded in the synthesis loss are automatically tuned by minimizing the discrepancy function during training to minimize approximation biases. We further design a new actor–critic (AC) algorithm, named AC with synthesis loss (ACSL), by integrating the synthesis loss function and an error-controlled mechanism. Experimental results on continuous control tasks illustrate that the proposed ACSL algorithm outperforms other cutting-edge RL methods in many tasks and that the proposed synthesis loss function is easily implemented into other algorithms and significantly reduces approximation biases while improving performance. The proposed method can successfully handle many complex continuous control tasks and can greatly outperform other state-of-the-art algorithms on most tasks.
Original languageEnglish
Pages (from-to)5323-5336
Number of pages14
JournalIEEE Transactions on Cybernetics
Volume54
Issue number9
Early online date03 May 2024
Publication statusPublished - 01 Sept 2024

Keywords

  • Actor critic (AC)
  • approximation biases
  • error-controlled mechanism
  • reinforcement learning (RL)
  • synthesis loss function

Fingerprint

Dive into the research topics of 'Actor-Critic with Synthesis Loss for Solving Approximation Biases'. Together they form a unique fingerprint.

Cite this