Created by W.Langdon from gp-bibliography.bib Revision:1.8098
All functions are polynomials (quadratics in the examples), terminals are either inputs or memories. Each memory terminals hold the value of a function node on the previous time step.
The coeffients of the polynomials are learnt by trying to match the training data using a 'Generalised Error Proporgation Algorithm'. This is determinstic. Seems like STROGANOFF's (but different?), time sequence based, based on back-propagation. The coefficients are recalculated each generation (assuming tree has changed).
Fitness function used 'minimum description length' (MDL).
Quadratic coefficients mya be limited to 0<=x<=1 to avoid divergence.
Examples: 2 step 0-1 oscilator, 4 Tomita languages (on binary alphabet).
Tree could be converted to finite state automata, which was more general than tree, ie works in all cases including those not in the training set.
On the tomita languages problems 'R-STROGANOFF works almost as well as (the best) best recurrent networks'
",
Genetic Programming entries for Hitoshi Iba Hugo de Garis Taisuke Sato