Abstract
OrdinalGP (2006) [4] embraced a fail-fast philosophy to efficiently model very large data sets. Recently, we realized that it was also effective against small data sets to reward model generalization. ESSENCE (2009) [6] extended the OrdinalGP concept to handle imbalanced data by using the SMITS algorithm to rank data records according to their information content to avoid locking into the behavior of heavily sampled data regions but had the disadvantage of computationally-intensive data conditioning with a corresponding fixed data ranking. With BalancedGP (2019) we shifted to a stochastic sampling to achieve a similar benefit. Trustable models (2007) [3] exploited the diversity of model forms developed during symbolic regression to define ensembles that feature both accurate prediction as well as detection of extrapolation into new regions of parameter space as well as changes in the underlying system. Although the deployed implementation has been effective, the diversity metric used was data-centric so alternatives have been explored to improve the robustness of ensemble definition. This chapter documents our latest thinking, realizations, and benefits of revisiting OrdinalGP and trustable models.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Cava, W. L., Spector, L., Danai, K.: Epsilon-lexicase selection for regression. In: Proceedings of the Genetic and Evolutionary Computation Conference (2016)
Keijzer, M., Foster, J.: Crossover bias in genetic programming. In: Genetic Programming, pp. 33–44. 10th European Conference, EuroGP (2007)
Kotanchek, M., Smits, G., Vladislavleva, E.: Exploiting trustable models via pareto GP For targeted data collection. In: Genetic Programming Theory and Practice VI, pp. 145–162. Springer, New York (2009)
Kotanchek, M., Smits, G., Vladislavleva, E.: Pursuing the pareto paradigm: tournaments, algorithm variations, and ordinal optimization. In: Genetic Programming Theory and Practice IV, pp. 167–185. Springer, New York (2007)
Kotanchek, M., Vladislavleva, E., Smits, G.: Symbolic regression is not enough: it takes a village to raise a model. In: Genetic Programming Theory and Practice X, pp. 187–203. Springer, New York (2013)
Kotanchek, M., Vladislavleva, E., Smits, G.: Symbolic regression via genetic programming as a discovery engine: insights on outliers and prototypes. In: Genetic Programming Theory and Practice VII, pp. 55–72. Springer, New York (2010)
Smits, G., Kotanchek, M.: Pareto-front exploitation in symbolic regression. In: Genetic Programming Theory and Practice II, pp. 283–299. Springer, New York (2005)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this chapter
Cite this chapter
Kotanchek, M., Haut, N. (2022). Back to the Future—Revisiting OrdinalGP and Trustable Models After a Decade. In: Banzhaf, W., Trujillo, L., Winkler, S., Worzel, B. (eds) Genetic Programming Theory and Practice XVIII. Genetic and Evolutionary Computation. Springer, Singapore. https://doi.org/10.1007/978-981-16-8113-4_7
Download citation
DOI: https://doi.org/10.1007/978-981-16-8113-4_7
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-16-8112-7
Online ISBN: 978-981-16-8113-4
eBook Packages: Computer ScienceComputer Science (R0)