Back to Search
Start Over
TD algorithm based on double-layer fuzzy partitioning
- Source :
- Tongxin xuebao, Vol 34, Pp 92-99 (2013)
- Publication Year :
- 2013
- Publisher :
- Editorial Department of Journal on Communications, 2013.
-
Abstract
- When dealing with the continuous space problems,the traditional Q-iteration algorithms based on lookup-table or function approximation converge slowly and are diff lt to get a continuous policy.To overcome the above weak-nesses,an on-policy TD algorithm named DFP-OPTD was proposed based on double-layer fuzzy partitioning and its convergence was proved.The first layer of fuzzy partitioning was applied for state space,the second layer of fuzzy parti-tioning was applied for action space,and Q-value functions were computed by the combination of the two layer fuzzy partitioning.Based on the Q-value function,the consequent parameters of fuzzy rules were updated by gradient descent method.Applying DFP-OPTD on two classical reinforcement learning problems,experimental results show that the algo-rithm not only can be used to get a continuous action policy,but also has a better convergence performance.
Details
- Language :
- Chinese
- ISSN :
- 1000436X
- Volume :
- 34
- Database :
- Directory of Open Access Journals
- Journal :
- Tongxin xuebao
- Publication Type :
- Academic Journal
- Accession number :
- edsdoj.49e18507a544d198c1dfea28a07910
- Document Type :
- article
- Full Text :
- https://doi.org/10.3969/j.issn.1000-436x.2013.10.011