模式识别与人工智能
Sunday, Apr. 6, 2025 Home      About Journal      Editorial Board      Instructions      Ethics Statement      Contact Us                   中文
  2014, Vol. 27 Issue (2): 103-110    DOI:
Papers and Reports Current Issue| Next Issue| Archive| Adv Search |
Kernel-Based Continuous-Action Actor-Critic Learning
CHEN Xing-Guo, GAO Yang, FAN Shun-Guo, YU Ya-Jun
State Key Laboratory for Novel Software Technology, Nanjing University, Nanjing 210093
Department of Computer Science and Technology, Nanjing University, Nanjing 210093

Download: PDF (488 KB)   HTML (1 KB) 
Export: BibTeX | EndNote (RIS)      
Abstract  In reinforcement learning, the learning algorithms frequently have to deal with both continuous state and continuous action spaces to control accurately. In this paper, the great capacity of kernel method for handling continuous state space problems and the advantage of actor-critic method in dealing with continuous action space problems are combined. Kernel-based continuous-action actor-critic learning(KCACL) is proposed grounded on the combination. In KCACL, the actor updates each action probability based on reward-inaction, and the critic updates the state value function according to online selective kernel-based temporal difference(OSKTD) learning. The experimental results demonstrate the effectiveness of the proposed algorithm.
Received: 13 May 2013     
ZTFLH: TP 181  
Service
E-mail this article
Add to my bookshelf
Add to citation manager
E-mail Alert
RSS
Articles by authors
Cite this article:   
URL:  
http://manu46.magtech.com.cn/Jweb_prai/EN/      OR     http://manu46.magtech.com.cn/Jweb_prai/EN/Y2014/V27/I2/103
Copyright © 2010 Editorial Office of Pattern Recognition and Artificial Intelligence
Address: No.350 Shushanhu Road, Hefei, Anhui Province, P.R. China Tel: 0551-65591176 Fax:0551-65591176 Email: bjb@iim.ac.cn
Supported by Beijing Magtech  Email:support@magtech.com.cn