首页 | 本学科首页   官方微博 | 高级检索  
     检索      

基于多个并行CMAC神经网络的强化学习方法
引用本文:李明爱,焦利芳,郝冬梅,乔俊飞.基于多个并行CMAC神经网络的强化学习方法[J].系统仿真学报,2008,20(24).
作者姓名:李明爱  焦利芳  郝冬梅  乔俊飞
作者单位:北京工业大学电子信息与控制工程学院
基金项目:国家自然科学基金 , 科博启动基金  
摘    要:为解决标准Q学习算法收敛速度慢的问题,提出一种基于多个并行小脑模型(Cerebellar Model Articulation Controller:CMAC)神经网络的强化学习方法。该方法通过对输入状态变量进行分割,在不改变状态分辨率的前提下,降低每个状态变量的量化级数,有效减少CMAC的存储空间,将之与Q学习方法相结合,其输出用于逼近状态变量的Q值,从而提高了Q学习方法的学习速度和控制精度,并实现了连续状态的泛化。将该方法用于直线倒立摆的平衡控制中,仿真结果表明了其正确性和有效性。

关 键 词:强化学习  小脑模型  神经网络  收敛性  倒立摆

Reinforcement Learning Based on Many Parallel CMAC Neural Networks
LI Ming-ai,JIAO Li-fang,HAO Dong-mei,QIAO Jun-fei.Reinforcement Learning Based on Many Parallel CMAC Neural Networks[J].Journal of System Simulation,2008,20(24).
Authors:LI Ming-ai  JIAO Li-fang  HAO Dong-mei  QIAO Jun-fei
Abstract:To solve the problem of the slow convergent rate of standard Q-learning, a reinforcement learning algorithm based on many parallel Cerebellar Model Articulation Controller (CMAC) neural networks was proposed. The input state variables were divided to decrease the grades of quantization without changing the resolution. Therefore, the storage spaces of CMAC were reduced effectively, and the outputs of CMAC with lower storage spaces were used to approximate the Q-functions of the corresponding input state variables by integrating CMAC with Q-learning method. So, the learning rate and control precision of Q-algorithm were improved simultaneity, and the generalization of continuous state variables was realized. The method was applied to control the balance of inverted pendulum, and the simulation results show its correctness and efficiency.
Keywords:reinforcement learning  CMAC  neural network  convergence  inverted pendulum
本文献已被 CNKI 万方数据 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号