Mini-Batching, Gradient-Clipping, First-versus Second-Order: What Works in Gradient-Based Coefficient Optimisation for Symbolic Regression'

Joe Harrison, Marco Virgolin, Tanja Alderliesten, Peter Bosman

Research output: Chapter in Book/Conference proceedings/Edited volumeConference contributionScientificpeer-review

1 Citation (Scopus)
40 Downloads (Pure)

Abstract

The aim of Symbolic Regression (SR) is to discover interpretable expressions that accurately describe data. The accuracy of an expression depends on both its structure and coefficients. To keep the structure simple enough to be interpretable, effective coefficient optimisation becomes key. Gradient-based optimisation is clearly effective at training neural networks in Deep Learning (DL), which can essentially be viewed as large, over-parameterised expressions: in this paper, we study how gradient-based optimisation techniques as often used in DL transfer to SR. In particular, we first assess what techniques work well across random SR expressions, independent of any specific SR algorithm. We find that mini-batching and gradient-clipping can be helpful (similar to DL), while second-order optimisers outperform first-order ones (different from DL). Next, we consider whether including gradient-based optimisation in Genetic Programming (GP), a classic SR algorithm, is beneficial. On five real-world datasets, in a generation-based comparison, we find that second-order optimisation outperforms coefficient mutation (or no optimisation). However, in time-based comparisons, performance gaps shrink substantially because the computational expensiveness of second-order optimisation causes GP to perform fewer generations. The interplay of computational costs between the optimisation of structure and coefficients is thus a critical aspect to consider.

Original languageEnglish
Title of host publicationGECCO 2023 - Proceedings of the 2023 Genetic and Evolutionary Computation Conference
PublisherAssociation for Computing Machinery (ACM)
Pages1127-1136
Number of pages10
ISBN (Electronic)9798400701191
DOIs
Publication statusPublished - 2023
Event2023 Genetic and Evolutionary Computation Conference, GECCO 2023 - Lisbon, Portugal
Duration: 15 Jul 202319 Jul 2023

Publication series

NameGECCO 2023 - Proceedings of the 2023 Genetic and Evolutionary Computation Conference

Conference

Conference2023 Genetic and Evolutionary Computation Conference, GECCO 2023
Country/TerritoryPortugal
CityLisbon
Period15/07/2319/07/23

Keywords

  • coefficient optimisation
  • explainable AI
  • genetic programming
  • gradient descent
  • symbolic regression

Fingerprint

Dive into the research topics of 'Mini-Batching, Gradient-Clipping, First-versus Second-Order: What Works in Gradient-Based Coefficient Optimisation for Symbolic Regression''. Together they form a unique fingerprint.

Cite this