计算机工程与应用 ›› 2014, Vol. 50 ›› Issue (11): 115-119.

• 数据库、数据挖掘、机器学习 • 上一篇    下一篇

一种结合演示数据和演化优化的强化学习方法

宋  拴,俞  扬   

  1. 南京大学 计算机软件新技术国家重点实验室,南京 210023
  • 出版日期:2014-06-01 发布日期:2015-04-08

Reinforcement learning method via combining demonstration data and evolutionary optimization

SONG Shuan, YU Yang   

  1. National Key Laboratory for Novel Software Technology, Nanjing University, Nanjing 210023, China
  • Online:2014-06-01 Published:2015-04-08

摘要: 强化学习研究智能体如何从与环境的交互中学习最优的策略,以最大化长期奖赏。由于环境反馈的滞后性,强化学习问题面临巨大的决策空间,进行有效的搜索是获得成功学习的关键。以往的研究从多个角度对策略的搜索进行了探索,在搜索算法方面,研究结果表明基于演化优化的直接策略搜索方法能够获得优于传统方法的性能;在引入外部信息方面,通过加入用户提供的演示,可以有效帮助强化学习提高性能。然而,这两种有效方法的结合却鲜有研究。对用户演示与演化优化的结合进行研究,提出iNEAT+Q算法,尝试将演示数据通过预训练神经网络和引导演化优化的适应值函数的方式与演化强化学习方法结合。初步实验表明,iNEAT+Q较不使用演示数据的演化强化学习方法NEAT+Q有明显的性能改善。

关键词: 强化学习, 演化算法, 从演示中学习, 神经网络

Abstract: Reinforcement learning aims at learning an optimal policy that maximizes the long term rewards, from interactions with the environment. Since the environment feedbacks commonly delay after a sequences of actions, reinforcement learning has to tackle the problem of searching in a huge policy space, and thus an effective search is the key to a success approach. Previous studies explore various ways to achieve effective search methods, one effective way is employing the evolutionary algorithm as the search method, and another direction is introducing user demonstration data to guide the search. In this work, it investigates the combination of the two directions, and proposes the iNEAT+Q approach, which trains a neural network using the demonstration data as well as integrating the demonstration data into the fitness function for the evolutionary algorithm. Preliminary empirical study shows that iNEAT+Q is superior to NEAT+Q, which is an classical evolutionary reinforcement learning approach.

Key words: reinforcement learning, evolutionary algorithm, learning from demonstrations, neural network