This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Genetic Programming for
Explainable Manifold Learning

Ben Cravens, Andrew Lensen,  Paula Maddigan, Bing Xue, 
This work was supported by the University Research Fund at Te Herenga Waka–Victoria University of Wellington under grant number 410128/4223.The authors are with the Centre for Data Science and Artificial Intelligence, and School of Engineering and Computer Science; Victoria University of Wellington; Wellington 6140; New Zealand (e-mail: ben.conor.reese.cravens@gmail.com; andrew.lensen@ecs.vuw.ac.nz; paula.maddigan@ecs.vuw.ac.nz; bing.xue@ecs.vuw.ac.nz).
Abstract

Manifold learning techniques play a pivotal role in machine learning by revealing lower-dimensional embeddings within high-dimensional data, thus enhancing both the efficiency and interpretability of data analysis by transforming the data into a lower-dimensional representation. However, a notable challenge with current manifold learning methods is their lack of explicit functional mappings, crucial for explainability in many real-world applications. Genetic programming, known for its interpretable functional tree-based models, has emerged as a promising approach to address this challenge. Previous research leveraged multi-objective GP to balance manifold quality against embedding dimensionality, producing functional mappings across a range of embedding sizes. Yet, these mapping trees often became complex, hindering explainability. In response, in this paper, we introduce Genetic Programming for Explainable Manifold Learning (GP-EMaL), a novel approach that directly penalises tree complexity. Our new method is able to maintain high manifold quality while significantly enhancing explainability and also allows customisation of complexity measures, such as symmetry balancing, scaling, and node complexity, catering to diverse application needs. Our experimental analysis demonstrates that GP-EMaL is able to match the performance of the existing approach in most cases, while using simpler, smaller, and more interpretable tree structures. This advancement marks a significant step towards achieving interpretable manifold learning.

Index Terms:
Manifold Learning, Genetic Programming, Dimensionality Reduction, Explainable Artificial Intelligence

I Introduction

Many real-world high-dimensional datasets have highly complex topological structure [1, 2, 3]. Manifold learning (MaL) methods, a category of Nonlinear Dimensionality Reduction (NLDR) techniques, are crucial for transforming these datasets into reduced embedding spaces, thereby aiding in the comprehension of the data’s intrinsic structure. Techniques like functional mapping and graph-based methods facilitate the visualization of these embeddings but often result in significant data loss [4]. In contrast, non-mapping and deep learning methods, while mitigating data loss, compromise interpretability due to their increased complexity. The application of complex, unexplainable NLDR methods, especially in regulated sectors, can have substantial ethical, legal, scientific, and commercial implications, as outlined in legislation such as the EU’s General Data Protection Regulation (GDPR) [5].

In high-risk domains such as healthcare, it is essential for practitioners to understand the features driving model outcomes [6, 7, 8, 9]. This necessity underscores the growing importance of interpretable NLDR techniques. Genetic Programming (GP), an evolutionary computation (EC) method, evolves symbolic functional mappings represented by syntax trees. Prior research [10, 11, 12] has shown that GP-based methods are effective in creating interpretable manifold learning models, particularly when the symbolic trees maintain low complexity. A recent study [13] further extends the explainability of GP trees from manifold learning models by integrating large language models such as ChatGPT to provide conversational explanations. GP, as a mapping method, provides an explicit function enabling the reproduction of embeddings and the implementation of sensitivity analysis. Techniques like automatic program simplification can further simplify these functions. Hence, GP is highly effective in scenarios where understanding the relative importance of features is crucial for ethical and legal reasons.

Interpretability in machine learning, however, remains a subjective concept [14, 15]. Defining exactly what makes a model interpretable is challenging [16], yet there are clear indications of a model being uninterpretable, such as having very large numbers of parameters or complex operations. If we characterise an interpretable model as having a low level of complexity in its tree structure, we can then define complexity by a metric, replacing subjectiveness with a measurable quantity. This, however, creates an inherent trade-off: a less complex model will generally produce lower-quality embeddings. Existing GP-NLDR work [11] used a multi-objective approach to balance embedding quality and the number of embedding dimensions. In this work, we propose an explainability-focused approach, where the complexity of the GP model is explicitly represented as an objective alongside embedding quality.

Major Contributions

  • We introduce GP-EMaL, a novel GP approach aiming at improving the explainability of GP trees/models for multi-objective manifold learning.

  • Through comprehensive experiments, we demonstrate that GP-EMaL retains performance comparable to existing methods while significantly enhancing interpretability.

  • To support further research and practical applications, we provide open-source code and a web-based application for further experimentation and exploration in this field.

II Background and Related Work

II-A Dimensionality Reduction

Dimensionality reduction, a key technique in machine learning, focuses on reducing the number of features in a dataset while preserving as much original information as possible. This process is essential for two main reasons: firstly, it significantly reduces the computational burden associated with large datasets; secondly, it helps overcome the “curse of dimensionality” [17], a phenomenon where the feature space becomes so sparse in high dimensions that the model’s performance degrades, leading to overfitting.

Dimensionality reduction can be classified into two broad approaches: feature selection [18] and feature extraction/construction [19]. Feature selection involves choosing a subset of relevant features from the dataset. This method selects features without altering their original form, thus maintaining the interpretability of the data.

Feature extraction or construction, on the other hand, transforms the original features into a new set of features. This transformation can be further divided into mapping and non-mapping methods. Mapping methods, such as Principal Component Analysis (PCA), create a new, reduced feature space through an explicit, interpretable functional mapping. This mapping is reversible (and reusable), allowing for a reconstruction of the original features from the reduced set. Non-mapping methods such as t-SNE [20] and UMAP [21], in contrast, focus on reducing the dimensionality without providing a functional mapping for reversing the process. These methods typically prioritize data compression and efficient representation over interpretability.

II-B Nonlinear Dimensionality Reduction

Nonlinear dimensionality reduction (NLDR) is employed when relationships in a dataset are too complex to be captured by linear methods alone [20]. Notable among NLDR techniques are deep neural networks, such as autoencoders, which consist of an encoder compressing data into a lower-dimensional space and a decoder reconstructing it back. This process is assessed using a loss function that measures the discrepancy between the input and output, enabling the autoencoder to efficiently map the data while preserving structural integrity [22].

Another significant category of NLDR methods is graph-based approaches, such as t-distributed stochastic neighbour embedding (t-SNE) [23]. These methods create a nearest neighbour graph to capture the inherent structure of the data. The graph is then embedded into a lower-dimensional space, preserving its structural characteristics. t-SNE specifically constructs a graph by considering pairwise similarities between high-dimensional data points and then subsequently optimising a 2-D or 3-D embedding to align with this graph. This method is particularly effective in preserving local data structures and for visualising high-dimensional data [24].

The integration of GP in NLDR was pioneered by the GP for Manifold Learning (GP-MaL) method [10]. Subsequent studies such as GP-MaL-MO [11] have continued to build on the concept, further advancing the research in this field. GP-Mal-MO applies a multi-objective approach to GP-based NLDR, balancing between preserving the nearest neighbour structure in the embedding and minimising the embedding dimensionality. This paper builds upon these developments to enhance the explainability of GP-based NLDR models.

II-C GP and Evolutionary Multi-Objective Optimisation

GP is an evolutionary computation technique that evolves programs, typically represented as tree structures, to solve problems or model data [25]. GP is particularly notable for its ability to evolve interpretable models, making it a valuable tool in domains where understanding the model’s decision-making process is crucial [26, 27]. In manifold learning, GP has been effectively used to evolve mappings that reduce data dimensionality while preserving its intrinsic structure [10].

Evolutionary Multi-Objective Optimisation (EMO) extends the EC paradigm to handle multiple, often conflicting, objectives. EMO algorithms evolve a population of solutions, aiming to find a set of Pareto-optimal solutions that represent the best possible trade-offs between the objectives [28]. This approach is particularly beneficial when dealing with complex problems where optimising a single objective could lead to sub-optimal or undesirable solutions.

A critical development in the EMO field was the MOEA/D (Multi-Objective Evolutionary Algorithm based on Decomposition) method [29]. MOEA/D decomposes a multi-objective optimisation problem into a number of scalar optimisation sub-problems and optimises them simultaneously. Each sub-problem focuses on a specific region of the Pareto front, enabling MOEA/D to effectively explore and exploit the search space. This method has gained popularity due to its efficiency and effectiveness, especially in problems with a complex Pareto front landscape.

In the context of GP for manifold learning, MOEA/D offers a robust framework for balancing the trade-offs between manifold quality, embedding complexity, and interpretability. By applying MOEA/D, GP is expected to evolve a diverse set of Pareto-optimal solutions, each representing a different balance between these objectives. This flexibility allows users to select a solution that best fits their specific needs, whether prioritising interpretability for a domain expert’s analysis or optimising performance for automated tasks.

II-D Tree Complexity

In GP, the complexity of the syntax trees is a crucial factor affecting both the performance and interpretability of the generated models. One method for calculating tree complexity starts at the root node and progresses recursively [30]. This method, outlined in Table I, aggregates the complexity of each node based on specific rules, facilitating efficient (𝒪(t)\mathcal{O}(t) for a tree containing tt nodes) and consistent calculations (e.g. as opposed to time-based complexity measures [31]). While this approach generates simpler trees, it has its limitations: it does not constrain the shape of the embedding tree, potentially resulting in asymmetrical and visually confusing/uninterpretable structures. Asymmetrical trees are harder to understand than the equivalent symmetrical tree with the same number of nodes, as they contain more layers of nested functions. Additionally, while this existing approach penalises complex operations (function nodes), it does not explicitly minimise the tree size, often leading to large trees using many simpler operators. The fixed function set also does not consider the varying interpretability of functions across different contexts.

TABLE I: Complexity Values [30]
Complexity(nn) Symbol of node(n)(n)
1 constant
2 variable
ccnComplexity(c)\sum_{c\in c_{n}}\text{Complexity}(c) +,+,-
ccnComplexity(c)\prod_{c\in c_{n}}\text{Complexity}(c) + 1 ,/*,/
Complexity(n1)2(n_{1})^{2} square
Complexity(n1)3(n_{1})^{3} square root
2Complexity(n1)2^{\text{Complexity}(n_{1})} sin, cos, tan, exp, log

where cc is a child of node nn, and n1n_{1} is the first child of nn.

II-E Tree Complexity Metrics

Various tree complexity metrics have been proposed in the literature, falling into two main categories: structural complexity and functional (semantic) complexity metrics [32, 33, 34, 35]. Structural complexity assesses the tree at a node level, while functional complexity is based on the complexity of the tree’s overall behaviour.

Research on structural complexity has explored factors like penalizing large trees through parsimony pressure [36] and reducing nested functions [37]. However, the integration of other metrics, such as assessing the asymmetry of trees, has been less explored. Our work aims to fill this gap.

In terms of functional complexity, the prevalence of many nonlinear operators, especially when nested, has often been identified as a factor of complexity [30]. The use of Tikhonov regularization to apply a global smoothing function to the tree has also been studied [38]. This method penalizes the function based on the norm of its higher-order derivatives, resulting in a smoother function.

The distinction between structural and functional complexity is not absolute, as these aspects can overlap. For example, a tree represented by a nested function like sin(cos(exp(x)))sin(cos(exp(x))) may be considered both functionally complex (due to its nonlinear nature) and structurally complex (due to its nested structure). Thus, these complexity measures should be viewed as complementary heuristics rather than distinct categories. For measuring explainability, structural complexity approaches are generally more appropriate as they directly consider the complexity of the tree structure, which is what the user will be trying to interpret.

III Proposed Method: GP-EMaL

Refer to caption
Figure 1: GP-EMaL Architecture

In our proposed method, GP-EMaL, we build upon the framework established by GP-MaL-MO [11], targeting a key limitation: the tendency towards complex tree structures which can impede interpretability. Our GP-EMaL approach introduces a novel complexity metric for expression trees, replacing the manifold dimensionality objective in GP-MaL-MO that focuses on minimizing the complexity while retaining the essential manifold quality cost metric based on neighbourhood ordering [11, 10]. This change aims to enhance the tree interpretability without compromising the effectiveness of the manifold learning process. The core algorithm for multi-objective optimisation is similar to GP-MaL-MO, as illustrated in Fig. 1, which outlines the architecture of GP-EMaL.

III-A GP Design

The evolutionary process begins by randomly initialising a population of GP individuals. These individuals are evaluated according to two objectives: one measuring manifold embedding quality and the other measuring tree complexity. The MOEA/D algorithm is used to find the population Front, which is used to select individuals for breeding in the next generation. This process repeats for a number of generations, with the Front produced by the final generation representing the different trade-offs between manifold quality and GP individual complexity.

GP-EMaL, like GP-MaL-MO, uses a multi-tree structure where each individual contains several trees. Each of these trees produces a single dimension of the lower-dimensional embedding. An individual is randomly initialised with a number of trees in the range [2,m÷2][2,m\div 2] for mm features. By utilising a specialised mutation operator (“Add/Remove Tree”), individuals are able to add and remove trees during evolution. This should allow the evolutionary search to balance manifold embedding quality with the number of trees and tree complexity. We also retain a “‘Standard” mutation operator, which picks a random subtree from any tree in the GP individual to perform mutation on. Likewise, crossover picks a random subtree from each of the two individuals being mated.

III-B Complexity Metric

The complexity metric in GP-EMaL has been enhanced to include parameterisation of the function set, allowing users to tweak the complexity cost of operators according to their specific needs/domain. For instance, linear operations like addition and subtraction could be assigned lower costs due to their straightforward interpretation in financial contexts, such as calculating balances or changes in value. Conversely, in theoretical physics applications, complex mathematical functions like trigonometric, exponential, and logarithmic operators are frequently used to model and understand phenomena such as wave mechanics, quantum physics, or signal processing. These complex operators, though less intuitive for a general audience, are essential and interpretable within the context of these specialized fields, as they accurately represent the underlying principles and behaviours being studied. We also provide a set of default parameters

To further improve interpretability, our new metric also focuses on achieving balanced tree structures through an asymmetry penalty term (AA) and encourages the development of smaller trees via a step-wise scaling term (SS). These improvements in GP-EMaL capture several crucial dimensions of tree complexity:

  1. 1.

    Structure: The AiA_{i} term encourages tree structures to be balanced, improving interpretability.

  2. 2.

    Size: Smaller trees, being easier to interpret, are favoured. The SiS_{i} scaling term proportionally increases the cost for larger trees.

  3. 3.

    Content: Our metric allows for a strategic selection of operators within the tree, prioritising those that enhance simplicity and context-specific understandability.

  4. 4.

    Semantics: The placement of complex operators within the tree directly impacts its interpretability. Placing complex operators closer to the leaves of the tree improves interpretability by avoiding nested complex functions. By using a hierarchical cost assignment to operators, our method aims to avoid complexity near the leaves and simplify the overall structure.

This refined metric in GP-EMaL enhances the interpretability of the evolved trees, addressing critical aspects of tree complexity while preserving the efficacy of the manifold learning process. We discuss the design of each part of this metric in the following subsections.

III-B1 Symmetry Balancing Term

In GP-EMaL, the symmetry balancing term AiA_{i} is applied to each node, calculated based on the size disparity between the left and right subtrees, denoted as SizeLeft\text{Size}_{\text{Left}} and SizeRight\text{Size}_{\text{Right}} respectively. This term is crucial for promoting symmetrical tree structures, as it imposes a penalty proportional to the asymmetry between subtrees. So that symmetric trees are not penalised, a deduction of 1 is applied to the term. The formula for this penalty is:

Ai=2|Δi|1A_{i}=2^{|\Delta_{i}|}-1 (1)

where Δ=SizeLeftSizeRight\Delta=\text{Size}_{\text{Left}}-\text{Size}_{\text{Right}}. This method contrasts with GP-MaL-MO, where trees often become unbalanced and visually complex. The application of AiA_{i} at every node ensures that balance is considered at all levels of the tree, leading to structures that are easier to interpret and analyse.

III-B2 Scaling term

Recognising that larger trees pose a challenge to interpretability, GP-EMaL incorporates a scaling term STS_{T} to penalise excessive tree sizes. The term is governed by the parameter α\alpha, representing tree T’s size (tt) relative to the pre-defined maximum tree size (i.e. SizeMax\text{Size}_{\text{Max}}). The punishment for larger trees begins when α\alpha exceeds a predefined threshold μ\mu, set by default to 0.75. An increase in μ\mu results in a stronger penalty for larger trees. The scaling term is defined as:

ST={1,if αi<μ,2αi,if αi>μαi=tSizeMax\quad{S_{T}}=\begin{cases}1,&\text{if }\alpha_{i}<\mu,\\ 2\alpha_{i},&\text{if }\alpha_{i}>\mu\end{cases}\quad\alpha_{i}=\frac{t}{\text{Size}_{\text{Max}}}\quad (2)
Refer to caption
Figure 2: Scaling term for S([0,1]) with μ=0.5\mu=0.5

III-B3 Tree Complexity Term

The GP-EMaL approach introduces enhanced flexibility in measuring tree complexity by introducing a parameterisable function set. This set allows users to define both the tree operators and their respective contributions to overall tree complexity. The cost assigned to each operator is proportional to the height of the respective subtree, with the scaling of these costs categorised as linear, quadratic, or exponential.

For example, consider a function set comprising three basic arithmetic operators and a nonlinear operator: [+,,×,sigmoid][+,-,\times,sigmoid], paired with a function cost set [sum,sum,prod,exp][sum,sum,prod,exp]. This configuration results in a higher prevalence of the simpler arithmetic operators [+,][+,-] within the evolved trees due to their lower cost. The multiplication operator ×\times would appear less frequently, typically higher in the tree, while the nonlinear operator sigmoidsigmoid would be least common, generally positioned near the top of the tree, directly affecting the features.

This methodological approach in GP-EMaL allows for nuanced control over the complexity of the generated trees. It enables the specification of not only which operators are present but also their preferred locations within the tree structure, thereby influencing both the functional complexity and the interpretability of the trees.

TABLE II: Summary of Function Cost Sets
Cost Operator Complexity Cost Scaling
sum ni(+,)n_{i}\in(+,-) Li+RiL_{i}+R_{i} 𝒪(t)\mathcal{O}(t)
prod ni(×,÷)n_{i}\in(\times,\div) max(Li×Ri,Li,RiL_{i}\times R_{i},L_{i},R_{i}) 𝒪(t2)\mathcal{O}(t^{2})
exp ni(f1,..,fn)n_{i}\in(f_{1},..,f_{n}) 2(Li+Ri)2^{(L_{i}+R_{i})} 𝒪(2t)\mathcal{O}(2^{t})

Table II summarises the settings for these function cost sets. Here, the complexity of the left and right subtrees is denoted as LiL_{i} and RiR_{i}, respectively, with f1,..,fnf_{1},..,f_{n} representing the set of non-arithmetic or “special” operations (e.g. trigonometric functions). The notation 𝒪\mathcal{O} indicates the asymptotic scaling complexity in relation to the average size (number of nodes, tt) of the left and right child subtrees based on a binary tree structure.

The tree complexity function, denoted as F(T)F(T), is defined as:

F(T)=ST×[ni(+,)(Li+Ri+Ai)+nj(×,÷)(max(Lj×Rj,Lj,Rj)+Aj)+nk(f1,..,fn)(2(Lk+Rk)+Ak)]\begin{split}F(T)&=S_{T}\times[\sum_{n_{i}\in(+,-)}(L_{i}+R_{i}+A_{i})\\ &+\sum_{n_{j}\in(\times,\div)}(\text{max}(L_{j}\times R_{j},L_{j},R_{j})+A_{j})\\ &+\sum_{n_{k}\in(f_{1},..,f_{n})}(2^{(L_{k}+R_{k})}+A_{k})]\end{split} (3)

where AiA_{i} refers to the asymmetry penalty at each node as defined in Eq. 1, and StS_{t} is the scaling term for tree size as per Eq. 2.

III-C Function and terminal sets

Our default function set contains a range of basic and more complex operators. It includes the arithmetic functions (++,-,×\times, protected ÷\div), conditional functions (max and min), and nonlinear operators (absolute value, ReLU and the Sigmoid function). The terminal set (leaf nodes) contains each of the features of the dataset.

III-D Objective functions

As a multi-objective GP-MaL approach, our method has two objectives: one that minimises the difference between the input feature space and the embedding space; and a second that minimises the complexity of the trees used to transform the input space into the embedding space.

As our focus in this work is on reducing the complexity of the learned trees, we utilise the same first objective (Cost) as in GP-MaL-MO, which has shown to be an effective measure:

Cost(I,X)=1|X|xX(1corr(N,N))2Cost(I,X)=\frac{1}{|X|}\sum_{x\in X}\frac{(1-\text{corr}(N,N^{\prime}))}{2} (4)

for GP individual II and dataset XX with data point xXx\in X. corr represents the Spearman’s rank correlation coefficient, |X||X| is the number of data points, and the 22 in the denominator is a scaling factor. This equation penalises GP individuals where the embedding has a different neighbourhood structure from the input dataset.

Whereas GP-MaL-MO simply uses embedding dimensionality (number of trees) as the complexity measure, our proposed GP-EMaL uses the new complexity metric introduced in Section III-B. Given that a GP-EMaL individual contains multiple trees (embedding dimensions), we sum across the tree complexity metric (Eq. 3) to get the overall complexity of a given GP individual (II):

Complexity(I)=TIF(T)Complexity(I)=\sum_{T\in I}F(T) (5)

where TT is a tree in II.

These two objective functions are generally conflicting, as less complex trees represent functions with fewer degrees of freedom and, therefore, cannot retain the neighbourhood structure of the input space as accurately as more complex trees, leading to a higher cost. By using the MOEA/D algorithm to evolve GP individuals according to these two objectives, we are able to produce a population of GP individuals to obtain an approximate Pareto front, where individuals represent different (non-dominated) tradeoffs between embedding quality (Cost) and tree complexity.

III-E Implementation

The GP-EMaL algorithm is built using Python with the code publicly available111https://github.com/cravies/GP-EMaL. A Streamlit web-based application is also accessible for ease of use222https://gp-emal.streamlit.app/ depicted in Fig. 3.

Refer to caption
Figure 3: Streamlit Application Interface.

IV Experiment Design

We evaluated our new GP-EMaL approach using the following methodology, split into three different analyses.

IV-A Predictive Performance

In our study, we adopt a widely used approach for comparing manifold learning (MaL) algorithms, focusing on the classification accuracy attainable in the embedded spaces they produce. This method avoids biases associated with unsupervised comparisons based on specific cost or quality formulations. While other supervised learning tasks like regression could also serve as indicators of MaL quality, classification is particularly effective in evaluating the preservation of data structures in embeddings.

Our evaluation involves training KNN and RF classifiers on embeddings from both the proposed GP-EMaL and baseline GP-MaL-MO methods, using ten-fold cross-validation with random forest (RF) and k-nearest neighbour (KNN) classifier. The random forest, comprising an ensemble of 100 trees, is chosen for its reliable and stable performance, making it a suitable tool for our analysis. KNN was selected as a simple method that heavily relies on high-quality embeddings for achieving high classification accuracy, due to its distance-based approach.

This evaluation utilized seven datasets, listed in Table III, primarily from the UCI repository [39], and included the COIL20 dataset [40]. These datasets were chosen as a broadly representative set containing varying numbers of features, instances, and classes. Classification accuracy was recorded over the 30 runs for each dataset with parameter settings as outlined in Table IV.

IV-B Complexity Summary Statistics

We also gathered summary statistics across all the datasets studied. For both GP-EMaL and GP-MaL-MO, we measured the average number of nodes, the number of unique features used, and the complexity of the nodes chosen. This provides a high-level comparison of the complexity/explainability of the two methods.

IV-C Visual Comparisons

Finally, to more directly contrast the explainability of the two approaches, we compare a typical example individual generated by each of GP-MaL-MO and GP-EMaL on the COIL-20 dataset. This dataset is the largest one (by number of features) in our tests and so represents a very difficult challenge for producing an explainable NLDR model.

TABLE III: Classification Datasets for Evaluation
Dataset Features Instances Classes
Wine 13 178 3
Dermatology 34 358 6
Image Segmentation 19 2310 7
MFEAT 649 2000 10
MNIST 784 2000 2
COIL20 1024 1440 20
Isolet 617 1560 26
TABLE IV: GP-EMaL Parameter Settings
Parameter Setting
Generations 1000
Population Size 100
“Standard” Mutation 15%
Add/Remove Tree Mutation 15%
Crossover 70%
Max. No. Trees max(2, [m÷2][m\div 2])
Min. Tree Depth 2
Max. Tree Depth 14
Decomposition Tchebycheff
Pop. Initialisation Ramped Half-and-half
Function Set +×÷+-\times\div ReLu sigmoid max min abs
Cost Set sum sum prod prod exp exp exp exp exp

V Results

V-A Predictive Performance

We begin by exploring the trade-off between complexity and classification accuracy in the individuals produced by GP-EMaL in Section V-A1. We do not directly compare with GP-MaL-MO on our proposed complexity metric, as it would represent an unfair comparison — we would expect GP-MaL-MO to perform substantially worse given that, unlike GP-EMaL, it was not optimised according to this metric.

Instead, we compare and contrast the performance of GP-EMaL and GP-MaL-MO according to simple measures of tree complexity in Section V-A2. These include the number of nodes, the number of complex operators (those with exponential cost scaling), and the number of unique features used by the GP individuals. While none of these provides a wholly unbiased measure of complexity, analysing these different measures provides a clearer view of the differences between our proposed GP-EMaL method and the GP-MaL-MO baseline.

V-A1 GP-EMaL Complexity vs. Classification Accuracy

Figures 4 to 10 show the results for GP-EMaL using the seven datasets, in order of roughly increasing dataset complexity. For each dataset, we show two visualisations. The first depicts the approximated Pareto front. The cost (reduction in manifold quality) is measured on the y-axis, with the tree complexity metric (in log scale) shown on the x-axis. As both cost and complexity should be minimised, points closer to the bottom-left are of higher quality. The second plot shows the Classif. Performances of both KNN (lighter blue) and RF (darker blue) predictors using the new embeddings. A smoothing spline is run on the results to average and smooth the curves over the 30 runs. As classification accuracy should be maximised, points closer to the top-left of the second set of plots are better-performing. We also include a dotted orange horizontal line, which is the performance of the RF classifier when using all features. This line represents a benchmark of what a good classifier could achieve on the source high-dimensional dataset.

Refer to caption
Figure 4: Front and Classif. Performance on Wine
Refer to caption
Figure 5: Front and Classif. Performance on Dermatology
Refer to caption
Figure 6: Front and Classif. Performance on Image Segmentation
Refer to caption
Figure 7: Front and Classif. Performance on the MFEAT dataset
Refer to caption
Figure 8: Front and Classif. Performance on MNIST
Refer to caption
Figure 9: Front and Classif. Performance on COIL20
Refer to caption
Figure 10: Front and Classif. Performance on Isolet

The Wine dataset (Fig. 4) is the least complex of our tested datasets, with only 13 features and 178 instances. The approximated Pareto front shows that GP individuals with very low complexity ({\approx}30 as measured by Eq. 4 ) are able to achieve a very low cost, strongly retaining the structure of the original high-dimensional space by using only simple operators. A similar result is seen in the classification plots, where both the KNN and Random Forest classifiers perform well with above 90% accuracy using the embeddings produced by a simple GP individual.

The Dermatology dataset (Fig. 5) exhibits similar behaviour to that of Wine but requires a slightly more complex solution ({\approx}75) to generate an embedding with a cost near zero. Both RF and KNN achieve in excess of 90% classification accuracy, matching the baseline performance of all features.

The Image Segmentation results (Fig. 6) resemble those of Wine, given the similar number of features in both datasets. This is the first set of results where there is a substantial performance difference between RF and KNN. The RF classifer achieves over 95% accuracy with a GP complexity of only {\approx}30, whereas the KNN classifier is slightly worse across all complexities.

The MFEAT and MNIST datasets show comparable patterns in their results (Figs. 7 to 8). They show convergence to low-cost solutions at a slightly higher complexity (\approx100) than the simpler datasets, achieving over 90% classification accuracy on RF and KNN. These are the first datasets where there is a (very small) gap in performance between RF using all features and RF using the embedding produced by even the most complex GP trees. This is an inherent trade-off: on sufficiently complex problems, a more complex (less explainable) model is needed to achieve the best performance. However, as we show in the rest of this paper, a much simpler model often can still achieve very high performance — a trade-off that is easily preferable in many applications.

The most complex datasets, COIL20 and Isolet (Figs. 9 to 10) also show an asymptotic pattern, where cost approaches zero at a complexity of 75{\approx}75. While the RF classification performance on COIL20 is near the baseline, on Isolet, there is a gap of around 10%~{}10\%, even at the highest complexities. If very high performance is crucial on Isolet, the original GP-MaL-MO approach may be required (as discussed later) instead.

V-A2 Comparing GP-EMaL to GP-MaL-MO

To compare GP-EMaL to GP-MaL-MO as fairly as possible, we evaluate the performance of each method along three different signals of complexity/interpretability. These are:

  1. 1.

    The total number of nodes used in the GP individual, which gives a “raw” measure of tree complexity;

  2. 2.

    The total number of exp cost operators used in the GP individual, which measures how many of the most difficult-to-understand functions are used; and

  3. 3.

    The number of unique features used, i.e. the number of dimensions in the high-dimensional input data that were used to construct the embedding. Using fewer unique features will make a GP individual easier to understand, as it requires less interpretation of different aspects of the data domain.

Refer to caption
Figure 11: Performance Comparisons using the Wine dataset
Refer to caption
Figure 12: Performance Comparisons using the Dermatology dataset
Refer to caption
Figure 13: Performance Comparisons using the Image Segmentation dataset
Refer to caption
Figure 14: Performance Comparisons using the MFEAT dataset
Refer to caption
Figure 15: Performance Comparisons using the MNIST dataset
Refer to caption
Figure 16: Performance Comparisons using the COIL20 dataset
Refer to caption
Figure 17: Performance Comparisons using the Isolet dataset

We plot each of these measures against the classification accuracy for all individuals in the fronts evolved by both GP-EMaL and GP-MaL-MO across all our runs. We use only RF for computing classification accuracy, as it always outperforms KNN and hence provides a more accurate measure of embedding quality. The plots for each of the datasets are shown in Figs. 11 to 17, where each orange ×\bm{\times} symbol shows one GP individual evolved by GP-EMaL and each \bm{\circ} symbol shows one GP individual evolved by GP-MaL-MO. We also include an overlay within each plot to show a closer perspective of the boundary between the GP-EMaL and GP-MaL-MO regions. For all of these plots, points closer to the top-left of the plot are of higher quality (having higher classification accuracy at a lower complexity). As with our earlier results, we include a dotted horizontal line to represent the “baseline” performance, which uses the same RF classifier with all the original features of the dataset.

Across nearly all of the datasets, the proposed GP-EMaL method is able to achieve similar performance to the existing GP-MaL-MO method despite having smaller trees with fewer complex operators and using fewer unique features. This is a clear testament to the complexity function proposed in this work — by using a more sophisticated measure of complexity (compared to just the number of trees in GP-MaL-MO) as our second objective, we can produce explainable trees that effectively preserve data structure in the embedded space.

This improvement in GP-EMaL is especially clear-cut in the first three datasets (Figs. 11 to 13). On these, GP-EMaL is able to retain sufficient manifold structure to reach the baseline accuracy while clearly using much less complex trees than GP-MaL-MO — in other words, GP-EMaL effectively dominates GP-MaL-MO across the three different complexity measures. GP-EMaL is consistently less complex at high levels of accuracy; on the Image Segmentation dataset, for example, GP-MaL-MO has accuracies as low as 70%, despite using all features in the dataset and using trees with many more nodes. Finally, GP-EMaL provides a much better approximation of the low-complexity part of the Pareto front: it produces extremely simple individuals (e.g. using only a few nodes and/or one or two features) that often can achieve over 80% accuracy — a trade-off that may be desirable on some problem domains.

On the MFEAT dataset (Fig. 14), we observe that GP-MaL-MO has a small but consistent decrease in RF accuracy for less-complex GP individuals. At similar levels of complexity, GP-EMaL outperforms GP-MaL-MO. For example, GP-MaL-MO achieves around 93% accuracy for a GP individual with 100 nodes, whereas GP-EMaL achieves close to the baseline accuracy (around 97%) with as few as 40 nodes. Similar patterns are seen for the number of exponential operators and unique features. Indeed, GP-EMaL can consistently reach the baseline accuracy when using only 25 unique features, whereas GP-MaL-MO has significant variance in accuracy.

As the complexity of our datasets further increases, a small gap in classification accuracy between the two methods begins to emerge. On MNIST (Fig. 15), GP-EMaL is very slightly (1%\leq 1\%) below the baseline accuracy, whereas GP-MaL-MO is able to reach baseline performance. There is, however, a marked complexity gap between the two methods: GP-EMaL is nearly an order of magnitude less complex than GP-MaL-MO (e.g. 100100 vs 500500 nodes) despite the only 1% difference in performance. A very similar result is observed on the COIL20 dataset (Fig. 16), albeit with a smaller gap in complexity and performance between the most-complex GP-EMaL individual and the least-complex GP-MaL-MO individual.

On the most complex dataset, Isolet (Fig. 17), GP-MaL-MO is clearly able to attain higher classification accuracy than GP-EMaL — but not without using more complex GP trees. GP-MaL-MO requires very complex trees (1000{\approx}1000 nodes, 100{\approx}100 complex operators, 300{\approx}300 unique features) to achieve a baseline level of accuracy. It is interesting to note that together, GP-EMaL and GP-MaL-MO almost form a combined front, with neither method having individuals that Pareto-dominate the other. On large, complex datasets such as Isolet, it may be desirable to run both GP-EMaL and GP-MaL-MO and produce a combined front to provide the user with a comprehensive set of trade-offs from very simple to very complex models.

In conclusion, the results demonstrate a clear trade-off between complexity and predictive accuracy across various datasets. This trade-off is more evident in datasets with higher complexity. For most datasets, GP-EMaL offers an improvement over GP-MaL-MO by producing substantially more explainable GP individuals without compromising on predictive accuracy. For the more complex datasets, GP-MaL-MO may have an edge in accuracy, but GP-EMaL stands out as a robust alternative, especially when the focus is on interpretability, with only a minimal performance decrease.

TABLE V: Summary Statistics
Statistic Description
# 𝒪(2n)\mathcal{O}(2^{n}) nodes total number of 𝒪(2n)\mathcal{O}(2^{n}) cost operators
# 𝒪(n2)\mathcal{O}(n^{2}) nodes total number of 𝒪(n2)\mathcal{O}(n^{2}) cost operators
# 𝒪(n)\mathcal{O}(n) nodes total number of 𝒪(n)\mathcal{O}(n) cost operators
# 𝒪(1)\mathcal{O}(1) nodes total number of 𝒪(1)\mathcal{O}(1) cost operators (feature nodes)
# nodes total number of nodes
# unique feat number of unique feature nodes
Refer to caption
Refer to caption
Figure 18: Summary statistics for the two GP methods across all datasets. The red line represents the median result.

V-B Summary Statistics

To understand the overall complexity of the two methods across all the datasets, we calculated several summary statistics for both GP-MaL-MO and GP-EMaL. These statistics are described in Table V.

Fig. 18 shows the summary statistics in the form of histograms for each of the two GP methods. By looking at the x-axis values, we can see that GP-MaL-MO produces exceptionally more complex trees than GP-EMaL. Indeed, the maximum value for GP-EMaL on each of the graphs is actually below the median value on the corresponding graph for GP-MaL-MO. The median and maximum number of nodes used by GP-EMaL are around 100×100\times smaller than that of GP-MaL-MO.

V-C Visual Comparisons

Refer to caption
Figure 19: A typical GP-EMaL individual containing 20 trees on COIL20, with 94.2% test accuracy and a complexity of 78.
Refer to caption
Figure 20: A typical GP-MaL-MO individual containing 29 trees on COIL20, with 99.1% test accuracy.

Finally, we present a typical individual evolved by GP-EMaL (Fig. 20) and GP-MaL-MO (Fig. 20) on the complex COIL20 dataset333Further sample tree outputs are available at github.com/cravies/GP-EMaL. We chose the individual with median classification accuracy across all runs to represent the “elbow” of the approximated Pareto Front. While GP-EMaL performs slightly worse than GP-MaL-MO on classification accuracy (mirroring our earlier results), the individual itself is markedly less complex: the biggest tree produced by GP-EMaL has seven nodes, compared to 31 in GP-MaL-MO. The deepest tree in GP-EMaL is only a depth of four; GP-MaL-MO produces a tree with a depth of ten. While GP-MaL-MO may be technically interpretable to a domain expert, it is clearly very difficult to explain compared to the one produced by GP-EMaL. We can also see that the tree structure in GP-EMaL is much more interpretable: the most complex sigmoid and ReLU operators are always at the bottom of the tree, with more human-understandable subtraction and multiplication operators used nearer the root.

VI Conclusions

This study’s exploration into GP-EMaL represents a significant stride in the realm of interpretable manifold learning using genetic programming. GP-EMaL distinguishes itself by its emphasis on interpretability, a critical aspect often overlooked in traditional approaches. Notably, while GP-EMaL demonstrates a slight decrease in performance on higher-complexity datasets, this reduction is generally modest, typically under 5%. This is a small price to pay considering the complexity of previous methods, which were so intricate that they were impractical for tasks where interpretability is paramount.

Furthermore, our findings underline the delicate balance between interpretability and performance in machine learning models. The open-source availability of GP-EMaL is intended to foster more research and practical applications in this field. Looking ahead, future work could involve user testing to validate interpretability and investigate ways to enhance embedding quality, potentially through the development of more efficient and specific evolutionary search operators. Furthermore, expanding GP-EMaL to optimise three objectives — neighbourhood structure preservation, embedding complexity, and dimensionality — simultaneously would offer a more nuanced approach, allowing for solutions tailored to specific requirements and contexts.

References

  • [1] J. Birjandtalab, M. B. Pouyan, and M. Nourani, “Nonlinear dimension reduction for eeg-based epileptic seizure detection,” in 2016 IEEE-EMBS International Conference on Biomedical and Health Informatics (BHI), 2016, pp. 595–598.
  • [2] D. Singh, H. Climente-Gonzalez, M. Petrovich, E. Kawakami, and M. Yamada, “Fsnet: Feature selection network on high-dimensional biological data,” in 2023 International Joint Conference on Neural Networks (IJCNN), 2023, pp. 1–9.
  • [3] S. Galelli and A. Castelletti, “Tree-based iterative input variable selection for hydrological modeling,” Water Resources Research, vol. 49, no. 7, pp. 4295–4310, 2013.
  • [4] A. Gracia, S. González, V. Robles, and E. Menasalvas, “A methodology to compare dimensionality reduction algorithms in terms of loss of quality,” Information Sciences, vol. 270, pp. 1–27, 2014.
  • [5] European Parliament and Council of the European Union. Regulation (EU) 2016/679 of the European Parliament and of the Council. [Online]. Available: https://data.europa.eu/eli/reg/2016/679/oj
  • [6] L. Longo, M. Brcic, F. Cabitza, J. Choi, R. Confalonieri, J. D. Ser, R. Guidotti, Y. Hayashi, F. Herrera, A. Holzinger, R. Jiang, H. Khosravi, F. Lecue, G. Malgieri, A. Páez, W. Samek, J. Schneider, T. Speith, and S. Stumpf, “Explainable artificial intelligence (XAI) 2.0: A manifesto of open challenges and interdisciplinary research directions,” Information Fusion, vol. 106, p. 102301, 2024.
  • [7] S. Ali, T. Abuhmed, S. El-Sappagh, K. Muhammad, J. M. Alonso-Moral, R. Confalonieri, R. Guidotti, J. Del Ser, N. Díaz-Rodríguez, and F. Herrera, “Explainable artificial intelligence (XAI): What we know and what is left to attain trustworthy artificial intelligence,” Information Fusion, vol. 99, p. 101805, 2023.
  • [8] M. A. Ahmad, C. Eckert, and A. Teredesai, “Interpretable machine learning in healthcare,” in Proceedings of the 2018 ACM international conference on bioinformatics, computational biology, and health informatics, 2018, pp. 559–560.
  • [9] P. Maddigan and T. Susnjak, “Forecasting patient demand at urgent care clinics using machine learning,” arXiv preprint arXiv:2205.13067, 2022.
  • [10] A. Lensen, B. Xue, and M. Zhang, “Can genetic programming do manifold learning too?” in Proceedings of the European Conference on Genetic Programming (EuroGP), Lecture Notes in Computer Science.   Springer International Publishing, 2019, vol. 11451, pp. 114–130.
  • [11] A. Lensen, M. Zhang, and B. Xue, “Multi-objective genetic programming for manifold learning: balancing quality and dimensionality,” Genetic Programming and Evolvable Machines, vol. 21, pp. 399–431, 2020.
  • [12] Y. Mei, Q. Chen, A. Lensen, B. Xue, and M. Zhang, “Explainable artificial intelligence by genetic programming: A survey,” IEEE Transactions on Evolutionary Computation, vol. 27, no. 3, pp. 621–641, 2023.
  • [13] P. Maddigan, A. Lensen, and B. Xue, “Explaining genetic programming trees using large language models,” arXiv preprint arXiv:2403.03397, 2024.
  • [14] D. V. Carvalho, E. M. Pereira, and J. S. Cardoso, “Machine learning interpretability: A survey on methods and metrics,” Electronics, vol. 8, no. 8, 2019.
  • [15] C. Singh, J. P. Inala, M. Galley, R. Caruana, and J. Gao, “Rethinking interpretability in the era of large language models,” arXiv preprint arXiv:2402.01761, 2024.
  • [16] C. Molnar, G. Casalicchio, and B. Bischl, “Interpretable machine learning – a brief history, state-of-the-art and challenges,” in ECML PKDD 2020 Workshops.   Cham: Springer International Publishing, 2020, pp. 417–431.
  • [17] P. Domingos, “A few useful things to know about machine learning,” Communications of the ACM, vol. 55, no. 10, pp. 78–87, 2012.
  • [18] J. Huang, W. Qian, C.-M. Vong, W. Ding, W. Shu, and Q. Huang, “Multi-label feature selection via label enhancement and analytic hierarchy process,” IEEE Transactions on Emerging Topics in Computational Intelligence, vol. 7, no. 5, pp. 1377–1393, 2023.
  • [19] R. Zebari, A. Abdulazeez, D. Zeebaree, D. Zebari, and J. Saeed, “A comprehensive review of dimensionality reduction techniques for feature selection and feature extraction,” Journal of Applied Science and Technology Trends, vol. 1, no. 2, pp. 56 – 70, May 2020.
  • [20] L. Van Der Maaten, E. O. Postma, H. J. van den Herik et al., “Dimensionality reduction: A comparative review,” Journal of Machine Learning Research, vol. 10, no. 66-71, p. 13, 2009.
  • [21] L. McInnes, J. Healy, and J. Melville, “Umap: Uniform manifold approximation and projection for dimension reduction,” 2020.
  • [22] G. E. Hinton and R. Salakhutdinov, “Reducing the dimensionality of data with neural networks,” Science, vol. 313, pp. 504 – 507, 2006.
  • [23] L. van der Maaten and G. E. Hinton, “Visualizing data using t-sne,” Journal of Machine Learning Research, vol. 9, pp. 2579–2605, 2008.
  • [24] M. Wattenberg, F. Viégas, and I. Johnson, “How to use t-sne effectively,” Distill, 2016.
  • [25] J. Dong, J. Zhong, W.-N. Chen, and J. Zhang, “An efficient federated genetic programming framework for symbolic regression,” IEEE Transactions on Emerging Topics in Computational Intelligence, vol. 7, no. 3, pp. 858–871, 2023.
  • [26] R. Poli, W. B. Langdon, and N. F. McPhee, A Field Guide to Genetic Programming.   lulu.com, 2008.
  • [27] B. Al-Helali, Q. Chen, B. Xue, and M. Zhang, “Genetic programming for feature selection based on feature removal impact in high-dimensional symbolic regression,” IEEE Transactions on Emerging Topics in Computational Intelligence, pp. 1–14, 2024, early Access.
  • [28] C. A. C. Coello, “Evolutionary multi-objective optimization: a historical view of the field,” IEEE Comput. Intell. Mag., vol. 1, no. 1, pp. 28–36, 2006.
  • [29] Q. Zhang and H. Li, “MOEA/D: A multiobjective evolutionary algorithm based on decomposition,” IEEE Trans. Evol. Comput., vol. 11, no. 6, pp. 712–731, 2007.
  • [30] M. Kommenda, G. Kronberger, M. Affenzeller, S. M. Winkler, and B. Burlacu, Evolving Simple Symbolic Regression Models by Multi-Objective Genetic Programming.   Springer International Publishing, 2016, ch. Genetic Programming Theory and Practice XIII, pp. 1–19.
  • [31] A. S. Sambo, R. M. A. Azad, Y. Kovalchuk, V. P. Indramohan, and H. Shah, “Time control or size control? reducing complexity and improving the accuracy of genetic programming models,” 2020.
  • [32] N. Le, H. N. Xuan, A. Brabazon, and T. P. Thi, “Complexity measures in genetic programming learning: A brief review,” in 2016 IEEE Congress on Evolutionary Computation (CEC), 2016, pp. 2409–2416.
  • [33] J. Chadalawada, V. Havlicek, and V. Babovic, “A genetic programming approach to system identification of rainfall-runoff models,” Water Resour Manage, vol. 31, p. 3975–3992, 2017.
  • [34] G. Campobello, D. Dell’Aquila, M. Russo, and A. Segreto, “Neuro-genetic programming for multigenre classification of music content,” Applied Soft Computing, vol. 94, p. 106488, 2020.
  • [35] Y. Tang, H. Jia, and N. Verma, “Reducing energy of approximate feature extraction in heterogeneous architectures for sensor inference via energy-aware genetic programming,” IEEE Transactions on Circuits and Systems I: Regular Papers, vol. 67, no. 5, pp. 1576–1587, 2020.
  • [36] T. Soule and J. A. Foster, “Effects of code growth and parsimony pressure on populations in genetic programming,” Evolutionary Computation, vol. 6, pp. 293–309, 1998.
  • [37] G. F. Smits and M. Kotanchek, Pareto-Front Exploitation in Symbolic Regression.   Springer US, 2005, pp. 283–299.
  • [38] J.Ni and P.Rockett, “Tikhonov regularization as a complexity measure in multiobjective genetic programming,” IEEE Transactions on Evolutionary Computation, vol. 19, no. 2, pp. 157–166, 2015.
  • [39] D. Dheeru and E. Karra Taniskidou, “UCI machine learning repository,” 2017. [Online]. Available: http://archive.ics.uci.edu/ml
  • [40] S. A. Nene, S. K. Nayar, and H. Murase, “Columbia object image library (coil-20),” Columbia University, Tech. Rep., 1996.