Internal reinforcement in a connectionist genetic programming approach

作者:

摘要

Genetic programming (GP) can learn complex concepts by searching for the target concept through evolution of a population of candidate hypothesis programs. However, unlike some learning techniques, such as Artificial Neural Networks (ANNs), GP does not have a principled procedure for changing parts of a learned structure based on that structure's performance on the training data. GP is missing a clear, locally optimal update procedure, the equivalent of gradient-descent backpropagation for ANNs. This article introduces a new algorithm, “internal reinforcement”, for defining and using performance feedback on program evolution. This internal reinforcement principled mechanism is developed within a new connectionist representation for evolving parameterized programs, namely “neural programming”. We present the algorithms for the generation of credit and blame assignment in the process of learning programs using neural programming and internal reinforcement. The article includes a comprehensive overview of genetic programming and empirical experiments that demonstrate the increased learning rate obtained by using our principled program evolution approach.

论文关键词:Machine learning,Evolutionary computation,Genetic programming,Signal understanding,Internal reinforcement,Neural programming,Bucket brigade

论文评审过程:Received 14 April 1998, Revised 24 January 2000, Available online 27 July 2000.

论文官网地址:https://doi.org/10.1016/S0004-3702(00)00023-0