skip to main content
10.1145/3583133.3590748acmconferencesArticle/Chapter ViewAbstractPublication PagesgeccoConference Proceedingsconference-collections
poster

Performance Analysis of Self-Supervised Strategies for Standard Genetic Programming

Published:24 July 2023Publication History

ABSTRACT

Self-supervised learning (SSL) methods have been widely used to train deep learning models for computer vision and natural language processing domains. They leverage large amounts of unlabeled data to help pretrain models by learning patterns implicit in the data. Recently, new SSL techniques for tabular data have been developed, using new pretext tasks that typically aim to reconstruct a corrupted input sample and yielding models which are, ideally, robust feature transforms. In this paper, we pose the research question of whether genetic programming is capable of leveraging data processed using SSL methods to improve its performance. We test this hypothesis by assuming different amounts of labeled data on seven different datasets (five OpenML benchmarking datasets and two real-world datasets). The obtained results show that in almost all problems, standard genetic programming is not able to capitalize on the learned representations, producing results equal to or worse than using the labeled partitions.

References

  1. L. Breiman. 2004. Random Forests. Machine Learning 45 (2004), 5--32.Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Carlota Cardoso, Rita T Sousa, Sebastian Köhler, and Catia Pesquita. 2020. A Collection of Benchmark Data Sets for Knowledge Graph-based Similarity in the Biomedical Domain. Database 2020 (11 2020). Google ScholarGoogle ScholarCross RefCross Ref
  3. Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey E. Hinton. 2020. A Simple Framework for Contrastive Learning of Visual Representations. ArXiv abs/2002.05709 (2020).Google ScholarGoogle Scholar
  4. Sajad Darabi, Shayan Fazeli, Ali Pazokitoroudi, Sriram Sankararaman, and Majid Sarrafzadeh. 2021. Contrastive Mixup: Self- and Semi-Supervised learning for Tabular Domain. ArXiv abs/2108.12296 (2021).Google ScholarGoogle Scholar
  5. Carl Doersch, Abhinav Kumar Gupta, and Alexei A. Efros. 2015. Unsupervised Visual Representation Learning by Context Prediction. 2015 IEEE International Conference on Computer Vision (ICCV) (2015), 1422--1430.Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Spyros Gidaris, Praveer Singh, and Nikos Komodakis. 2018. Unsupervised Representation Learning by Predicting Image Rotations. ArXiv abs/1803.07728 (2018).Google ScholarGoogle Scholar
  7. Alexander Kolesnikov, Xiaohua Zhai, and Lucas Beyer. 2019. Revisiting Self-Supervised Visual Representation Learning. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2019), 1920--1929.Google ScholarGoogle Scholar
  8. Mehdi Noroozi and Paolo Favaro. 2016. Unsupervised Learning of Visual Representations by Solving Jigsaw Puzzles. In European Conference on Computer Vision.Google ScholarGoogle ScholarCross RefCross Ref
  9. F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, and E. Duchesnay. 2011. Scikit-learn: Machine Learning in Python. Journal of Machine Learning Research 12 (2011), 2825--2830.Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Riccardo Poli, William B. Langdon, and Nicholas Freitag McPhee. 2008. A Field Guide to Genetic Programming. lulu.com.Google ScholarGoogle Scholar
  11. Alec Radford and Karthik Narasimhan. 2018. Improving Language Understanding by Generative Pre-Training.Google ScholarGoogle Scholar
  12. Rita T. Sousa, Sara Silva, and Catia Pesquita. 2021. Supervised Semantic Similarity. bioRxiv (2021). Google ScholarGoogle ScholarCross RefCross Ref
  13. Joaquin Vanschoren, Jan N. van Rijn, Bernd Bischl, and Luis Torgo. 2013. OpenML: networked science in machine learning. SIGKDD Explorations 15, 2 (2013), 49--60. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Jinsung Yoon, Yao Zhang, James Jordon, and Mihaela van der Schaar. 2020. VIME: Extending the Success of Self- and Semi-supervised Learning to Tabular Domain. In Advances in Neural Information Processing Systems, H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin (Eds.), Vol. 33. Curran Associates, Inc., 11033--11043.Google ScholarGoogle Scholar

Index Terms

  1. Performance Analysis of Self-Supervised Strategies for Standard Genetic Programming

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      GECCO '23 Companion: Proceedings of the Companion Conference on Genetic and Evolutionary Computation
      July 2023
      2519 pages
      ISBN:9798400701207
      DOI:10.1145/3583133

      Copyright © 2023 Owner/Author(s)

      Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the owner/author(s).

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 24 July 2023

      Check for updates

      Qualifiers

      • poster

      Acceptance Rates

      Overall Acceptance Rate1,669of4,410submissions,38%

      Upcoming Conference

      GECCO '24
      Genetic and Evolutionary Computation Conference
      July 14 - 18, 2024
      Melbourne , VIC , Australia
    • Article Metrics

      • Downloads (Last 12 months)33
      • Downloads (Last 6 weeks)4

      Other Metrics

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader