This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

A Distribution Evolutionary Algorithm for the Graph Coloring Problem

Yongjian Xu 3414727121@qq.com Huabin Cheng 2020111132@wsyu.edu.cn Ning Xu xuning@whut.edu.cn Yu Chen ychen@whut.edu.cn Chengwang Xie chengwangxie@m.scnu.edu.cn School of Science, Wuhan University of Technology, Wuhan, 430070, China Department of Basic Science, Wuchang Shouyi University, Wuhan, 430064, China School of Information Engineering, Wuhan University of Technology, Wuhan, 430070, China School of Data Science and Engineering, South China Normal University, Guangdong, 516600, China
Abstract
\color

blueGraph coloring is a challenging combinatorial optimization problem with a wide range of applications. In this paper, a distribution evolutionary algorithm based on a population of probability model (DEA-PPM) is developed to address it efficiently. Unlike existing estimation of distribution algorithms where a probability model is updated by generated solutions, DEA-PPM employs a distribution population based on a novel probability model, and an orthogonal exploration strategy is introduced to search the distribution space with the assistance of an refinement strategy. By sampling the distribution population, efficient search in the solution space is realized based on a tabu search process. Meanwhile, DEA-PPM introduces an iterative vertex removal strategy to improve the efficiency of kk-coloring, and an inherited initialization strategy is implemented to address the chromatic problem well. The cooperative evolution of the distribution population and the solution population leads to a good balance between exploration and exploitation. Numerical results demonstrate that the DEA-PPM of small population size is competitive to the state-of-the-art metaheuristics.

keywords:
distribution evolutionary algorithm, orthogonal exploration, inherited initialization, graph coloring, estimation of distribution algorithm
journal: Swarm and Evolutionary Computation
\UseRawInputEncoding
mytitlenotemytitlenotefootnotetext: Y. Xu and H. Cheng contributed equally to this research.

1 Introduction

Given an undirected graph G=(V,E)G=(V,E) with a vertex set VV and a edge set EE, the (vertex) graph coloring problem (GCP) assigns colors to vertexes such that no adjacent vertexes share the same color. If GG can be colored by kk different colors without color conflicts, it is kk-colorable. The smallest value of color number kk such that GG is kk-colorable is its chromatic number, denoted by χ(G)\chi(G). There are two instances of the GCP, the kk-coloring problem attempting to color a graph with kk colors and the chromatic number problem trying to get the chromatic number of GG, both of which are extensively applied in scientific and engineering fields. \colorredDue to the NP-completeness of GCPs, some relaxation methods were proposed to transform the combinatorial GCPs to continuous optimization problems [1, 2, 3]. However, the transformation will lead to continuous problems with distinct landscapes, and global optimal solutions of the original GCPs could be quite different from those of the relaxed problems.

Accordingly, a variety of metaheuristics have been developed to address the original GCPs efficiently [4]. Individual-based metaheuristics search the solution space by single-point iteration schemes, contributing to their fast convergence and low complexity [5]. However, their performance relies heavily on the initial solution and the definition of neighborhood, which makes it challenging to balance the exploration and the exploitation. Population-based metaheuristics perform cooperative multi-point search in the solution space, but a comparatively large population is usually necessary for the efficient search in the solution space, which makes it inapplicable to large-scale GCPs [4].

Recently, metaheuristics based on probability models have been widely employed to solve complicated optimization problems [6, 7, 8]. As two popular instances, the ant colony optimization (ACO) [6] and the estimation of distribution algorithm (EDA) [7] employ a single probability model that is gradually updated during the iteration process, which makes it difficult to balance the global exploration and the local exploitation in the distribution space. The quantum-inspired evolutionary algorithm (QEA) performs an active update of probability model by the Q-gate rotation, whereas it is a kind of local exploitation that cannot explore the distribution space efficiently [8]. To remedy the aformentioned issues, we propose a distribution evolutionary algorithm based on a population of probability model (DEA-PPM), where a balance between the convergence performance and the computational complexity could be kept by evolution of small populations. \colorblue Contributions of this work are as follows.

  • 1.

    We propose a novel distribution model that incorporates the advantages of EDAs and QEAs.

  • 2.

    Based on the proposed distribution model, an orthogonal exploration strategy is introduced to search the probability space with the assistance of a tailored refinement strategy.

  • 3.

    For the chromatic problem, an inherited initialization is presented to accelerate the convergence process.

\color

red Rest of this paper is organized as follows. Section 2 presents a brief review on related works. The proposed distribution model is presented in Section 3, and Section 4 elaborates details of DEA-PPM. Section 5 investigates the influence of parameter and the distribution evolution strategies, and the competitiveness of DEA-PPM is verified by numerical experiments. Finally, we summarize the work in Section 6.

2 Literature Review

2.1 Individual-based metaheuristics for GCPs

Besides the simulated annealing [9, 10] and the variable neighborhood search [11], the tabu search (TS) is one of the most popular individual-based metaheuristics applied to solve the GCPs [12]. Porumbel et al. [13] improved the performance of TS by evaluation functions that incorporates the structural or dynamic information in addition to the number of conflicting edges. Blöchliger and Zufferey [14] proposed a TS-based constructive strategy, which constructs feasible but partial solutions and gradually increases its size to get the optimal color assignment of a GCP. Hypothesizing that high quality solutions of GCPs could be grouped in clusters within spheres of a specific diameter, Porumbel et al. [15] proposed two improved TS variants using a learning process and a tree-like structure of the connected spheres. Assuming that each vertex only interacts with a limited number of components, Galán [16] developed a decentralized coloring algorithm, where colors of vertexes are modified according to those of the adjacent vertexes to iteratively reduce the number of edge conflicts. Sun et al. [17] established a solution-driven multilevel optimization framework for GCP, where an innovative coarsening strategy that merges vertexes based on the solution provided by the TS, and the uncoarsening phase is performed on obtained coarsened results to get the coloring results of the original graph. To color vertexes with a given color number kk, Peng et al. [18] partitioned a graph into a set of connected components and a vertex cut component, and combined the separately local colors by an optimized maximum matching based method.

Since a probability model can provide a bird’s-eye view for the landscape of optimization problem, Zhou et al. [19, 20] proposed to enhance the global exploration of individual-based metaheuristics by the introduction of probability models. They deployed a probabilistic model for the colors of vertexes, which is updated with the assistance of a reinforcement learning technology based on discovered local optimal solutions [19]. Moreover, they improved the learning strategy of probability model to develop a three-phase local search, that is, a starting coloring generation phase based on a probability matrix, a heuristic coloring improvement phase and a learning based probability updating phase [20].

2.2 Population-based metaheuristics for GCPs

Population-based iteration mechanisms are incorporated to the improve exploration abilities of metaheuristics as well. Hsu et al. [21] proposed a modified turbulent particle swarm optimization algorithm for the planar graph coloring problem, where a three-stage turbulent model is employed to strike a balance between exploration and exploitation. Hernández and Blum [22] dealt with the problem of finding valid graphs colorings in a distributed way, and the assignment of different colors to neighboring nodes is asynchronously implemented by simulating the calling behavior of Japanese tree frogs. Rebollo-Ruiz and M. Graña [23] addressed the GCP by a gravitational swarm intelligence algorithm, where nodes of a graph are mapped to agents, and its connectivity is mapped into a repulsive force between the agents corresponding to adjacent nodes. Based on the conflict matrices of candidate solutions, Zhao et al. [24] developed a dimension-by-dimension update method, by which a discrete selfish herd optimizer was proposed to address GCPs efficiently. Aiming to develop an efficient parameter-free algorithm, Chalupa and Nielsen [25] proposed to improve the global exploration by a multiple cooperative searching strategy. For the four-colormap problem, Zhong et al. [26] proposed an enhanced discrete dragonfly algorithm that performs a global search and a local search alternately to color maps efficiently.

The incorporation of probability models are likewise employed to improve the performance of population-based metaheuristics. Bui et al. [27] developed a constructive strategy of coloring scheme based on an ant colony, where an ant colors just a portion of the graph unsing only local information. Djelloul et al. [28] took a collection of quantum matrices as the population of the cuckoo search algorithm, and an adapted hybrid quantum mutation operation was introduced to get enhanced performance of the cuckoo search algorithm.

2.3 Hybrid metaheuristics for GCPs

The population-based metaheuristics can be further improved by hybrid search strategies. Paying particular attention to ensuring the population diversity, Lü and Hao [29] proposed an adaptive multi-parent crossover operator and a diversity-preserving strategy to improve the searching efficiency of evolutionary algorithms, and proposed a memetic algorithm that takes the TS as a local search engine. Porumbel et al. [30] developed a population management strategy that decides whether an offspring should be accepted in the population, which individual needs to be replaced and when mutation is applied. Mahmoudi and Lotfi [31] proposed a discrete cuckoo optimization algorithm for the GCP, where a neighborhood search in radius of the lay egg causes the algorithm hardly trapped in local minimum and producing new eggs. Accordingly, it provides a good balance between diversification and centralizing.

Wu and Hao [32] proposed a preprocessing method that extracts large independent sets by the TS, and the memetic algorithm proposed by Lü and Hao [29] was employed to color the residual graph. For the chromatic problem, Douiri and Elbernoussi [33] initialized the color number by the coloring result of a heuristic algorithm and generated the initial population of genetic algorithm (GA) by finding a maximal independent set approximation of the investigated graph.

Bessedik et al. [34] addressed the GCPs within the framework of the honey bees optimization, where a local search, a tabu search and an ant colony system are implemented as workers and queens are randomly generated. Mirsaleh and Meybodi [35] proposed a Michigan memetic algorithm for GCPs, where each chromosome is associated to a vertex of the input graph. Accordingly, each chromosome is a part of the solution and represents a color for its corresponding vertex, and each chromosome locally evolves by evolutionary operators and improves by a learning automata based local search. Moalic and Gondran [36] integrated a TS procedure with an evolutionary algorithm equipped with the greedy partition crossover, by which the hybrid algorithm can performs well with a population consisting of two individuals. Silva et al. [37] developed a hybrid algorithm iColourAnt, which addresses the GCP using an ant colony optimization procedure with assistance of a local search performed by the reactive TS.

2.4 Related work on the estimation of distribution algorithm

A large number of works have been reported to improve the general performance of EDAs. To improve the general precision of a distribution model, Shim et al. [38] modelled the restricted Boltzmann machine as a novel EDA, where the probabilistic model is constructed using its energy function, and the kk-means clustering was employed to group the population into small clusters. Approximating the Boltzmann distribution by a Gaussian model, Valdez et al. [39] proposed a Boltzmann univariate marginal distribution algorithm, where the Gaussian distribution obtains a better bias to sample intensively the most promising regions. Considering the multivariate dependencies between continuous random variables, PourMohammadBagher et al. [40] proposed a parallel model of some subgraphs with a smaller number of variables to avoid complex approximations of learning a probabilistic graphical model. Dong et al. [41] proposed a latent space-based EDA, which transforms the multivariate probabilistic model of Gaussian-based EDA into its principal component latent subspace of lower dimensionality to improve its performance on large-scale optimization problems.

To enhance the local exploitation of an EDA, Zhou et al. [42] suggested to combine an estimation of distribution algorithm with cheap and expensive local search methods for making use of both global statistical information and individual location information. Considering that the random sampling of Gaussian EDA usually suffers from the poor diversity and the premature convergence, Dang et al. [43] developed an efficient mixture sampling model to achieve a good tradeoff between the diversity and the convergence, by which it can explore more promising regions and utilize the unsuccessful mutation vectors.

The performance of EDA can also improved by designing tailored update strategies of the probability model. To address the multiple global optima of multimodal problem optimizations, Pẽna et al. [44] introduced the unsupervised learning of Bayesian metwokrs in EDA, which makes it able to model simultaneously the different basins represented by the selected individuals, whereas preventing genetic drift as much as possible. Peng et al. [45] developed an explicit detection mechanism of the promising areas, by which function evaluations for exploration can be significantly reduced. To prevent the Gaussian EDAs from premature convergence, Ren et al. [46] proposed to tune the main search direction by the anisotropic adaptive variance that is scaled along different eigendirections based on the landscape characteristics captured by a simple topology-based detection method. Liang et al. [47] proposed to archive a certain number of high-quality solutions generated in the previous generations, by which fewer individuals are needed in the current population for model estimation. In order to address the mixed-variable newsvendor problem, Wang et al. [48] developed a histogram model-based estimation of distribution algorithm, where an adaptive-width histogram model is used to deal with the continuous variables and a learning-based histogram model is applied to deal with the discrete variables. Liu et al. [49] embedded within the search procedure a learning mechanism based on an incremental Gaussian mixture model, by which all new solutions generated during the evolution are fed incrementally into the learning model to adaptively discover the structure of the Pareto set of an MOP.

3 The Distribution Model for the Graph Coloring Problem

3.1 The graph coloring problem

Let n=|V|n=|V| be the vertex number of a graph G=(V,E)G=(V,E). An assignment of vertexes with kk colors can be represented by an integer vector 𝐱=(x1,,xn)\mathbf{x}=(x_{1},\dots,x_{n}), where xjx_{j} denotes the assigned color of vertex vjv_{j}. Then, the kk-coloring problem can be modelled as a minimization problem

minfk(𝐱)=j1=1nj2=1nδ(j1,j2)s.t.{δ(j1,j2)={1,if (vj1,vj2)Exj1=xj2,0,otherwise ,𝐱=(x1,,xn),xj{1,,k},j=1,,n,ji{1,2,,n},i=1,2.\begin{array}[]{l}\min\quad f_{k}(\mathbf{x})=\sum\limits_{j_{1}=1}^{n}\sum\limits_{j_{2}=1}^{n}\delta(j_{1},j_{2})\\ s.t.\left\{\begin{aligned} &\delta(j_{1},j_{2})=\left\{\begin{aligned} &1,&&\mbox{if }(v_{j_{1}},v_{j_{2}})\in E\wedge x_{j_{1}}=x_{j_{2}},\\ &0,&&\mbox{otherwise },\end{aligned}\right.\\ &\mathbf{x}=(x_{1},\dots,x_{n}),x_{j}\in\{1,\dots,k\},j=1,\dots,n,\\ &j_{i}\in\{1,2,\dots,n\},i=1,2.\end{aligned}\right.\end{array} (1)

While δ(j1,j2)=1\delta(j_{1},j_{2})=1, the adjacent vertexes vj1v_{j_{1}} and vj2v_{j_{2}} are assigned with the same color, and (vj1,vj2)(v_{j_{1}},v_{j_{2}}) is called a conflicting edge. Accordingly, the objective values fk(𝐱)f_{k}(\mathbf{x}) represents the total conflicting number of the color assignment 𝐱\mathbf{x}. While GG is kk-colorable, there exists an optimal partition 𝐱\mathbf{x}^{*} such that fk(𝐱)=0f_{k}(\mathbf{x}^{*})=0, and 𝐱\mathbf{x}^{*} is called a legal kk-color assignment of graph GG. Thus, the chromatic problem is modelled as

minks.t.fk(𝐱)=0,\begin{array}[]{l}\min\quad k\\ s.t.\quad f_{k}(\mathbf{x}^{*})=0,\end{array} (2)

where 𝐱\mathbf{x}^{*} represents a legal kk-color assignment that is an optimal color assignment of problem (1).

\color

red

3.2 The Q-bit model and the Q-gate transformation

Different from the ACO and the EDA, the QEA employs a quantum matrix for probabilistic modelling of the solution space, modelling the probability distribution of a binary variable bxbx by a Q-bit (α,β)T(\alpha,\beta)^{T} satisfying |α|2+|β|2=1|\alpha|^{2}+|\beta|^{2}=1 [50]. That is, |α|2|\alpha|^{2} and |β|2|\beta|^{2} give the probabilities of bx=1bx=1 and bx=0bx=0, respectively. Accordingly, the probability distribution of an nn-dimensional binary vector 𝐛𝐱=(bx1,,bxn)\mathbf{bx}=(bx_{1},\dots,bx_{n}) is represented by

𝐪=(q1,q2,,qn)=[α1α2αnβ1β2βn],\mathbf{q}=(\vec{q}_{1},\vec{q}_{2},\dots,\vec{q}_{n})=\left[\begin{array}[]{cccc}\alpha_{1}&\alpha_{2}&\cdots&\alpha_{n}\\ \beta_{1}&\beta_{2}&\cdots&\beta_{n}\end{array}\right],

where |α|j2+|β|j2=1|\alpha|_{j}^{2}+|\beta|_{j}^{2}=1, j=1,,nj=1,\dots,n. Then, the value of bxjbx_{j} can be obtained by sampling the probability distribution (|α|j2,|β|j2)(|\alpha|_{j}^{2},|\beta|_{j}^{2}).

The probability distribution of binary variable is modified in the QEA by the Q-gate, a 2×22\times 2 orthogonal matrix

U(Δθ)=[cos(Δθ)sin(Δθ)sin(Δθ)cos(Δθ)].U(\Delta\theta)=\left[\begin{array}[]{lr}\cos(\Delta\theta)&-\sin(\Delta\theta)\\ \sin(\Delta\theta)&\cos(\Delta\theta)\end{array}\right].

Premultiplying qj\vec{q}_{j} by U(Δθ)U(\Delta\theta), the probability distribution of bxjbx_{j} is modified as

qj=U(Δθ)qj,i=j,,n.\vec{q}^{\prime}_{j}=U(\Delta\theta)\cdot\vec{q}_{j},\quad i=j,\dots,n.

To coloring a graph of nn vertexes by kk colors, Djelloul et al. [28] modelled the probability distribution of color assignment by a quantum matrix

𝐪=(q1,q2,,qn)=[q1,1q1,2q1,nq2,1q2,2q2,nq2k1,1q2k1,2q2k1,nq2k,1q2k,2q2k,n],\mathbf{q}=(\vec{q}_{1},\vec{q}_{2},\dots,\vec{q}_{n})=\left[\begin{array}[]{cccc}q_{1,1}&q_{1,2}&\cdots&q_{1,n}\\ q_{2,1}&q_{2,2}&\cdots&q_{2,n}\\ \cdots&\cdots&\cdots&\\ q_{2k-1,1}&q_{2k-1,2}&\cdots&q_{2k-1,n}\\ q_{2k,1}&q_{2k,2}&\cdots&q_{2k,n}\\ \end{array}\right],

where

(q2i1,j)2+(q2i,j)2=1,i{1,,k},j{1,,n}.(q_{2i-1,j})^{2}+(q_{2i,j})^{2}=1,\quad\forall\,i\in\{1,\dots,k\},j\in\{1,\dots,n\}.

In this way, the Q-gate can be deployed for each unit vector (q2i1,j,q2i,j)T(q_{2i-1,j},q_{2i,j})^{T} to regulate the distribution of color assignment.

3.3 The proposed distribution model of color assignment

Since the unit vector (q2i1,j,q2i,j)T(q_{2i-1,j},q_{2i,j})^{T} models each candidate color ii of vertex jj independently, the sampling process would lead to multiple assignments for vertex jj, and an additional repair strategy is needed to get a feasible kk-coloring assignment [28]. To address this defect, we propose to model the color distribution of vertex jj by a kk-dimensional unit vector qj\vec{q}_{j}, and present the color distribution of nn vertexes as

𝐪=(q1,,qn)=[q11q12q1nq21q22q2nqk1qk2qkn],\mathbf{q}=(\vec{q}_{1},\dots,\vec{q}_{n})=\begin{bmatrix}q_{11}&q_{12}&...&q_{1n}\\ q_{21}&q_{22}&...&q_{2n}\\ \vdots&\vdots&\vdots&\vdots\\ q_{k1}&q_{k2}&...&q_{kn}\end{bmatrix}, (3)

where qj\vec{q}_{j} satisfies

qj22=i=1kqij2=1,j=1,,n.\|\vec{q}_{j}\|_{2}^{2}={\sum_{i=1}^{k}q_{ij}^{2}}=1,\quad\forall\,j=1,\dots,n. (4)

The distribution model confirmed by (3) and (4) incorporates the advantages of models in EDAs and QEAs.

  • 1.

    An feasible kk-coloring of nn vertexes can be achieved by successively sampling nn columns of 𝐪\mathbf{q}.

  • 2.

    The update of probability distribution can be implemented by both orthogonal transformations performed on column vectors of 𝐪\mathbf{q} and direct manipulations of components that do not change the norms of columns vectors 111Details for the update process are presented in Section 4.4..

4 A Distribution Evolutionary Algorithm Based on a Population of Probability Model

\color

blue

The proposed DEA-PPM solves the GCP based on a distribution population and a solution population. The distribution population consists of individuals representing distribution models of graph coloring, which are updated by an orthogonal exploration strategy and an composite exploitation strategy. Meanwhile, an associated solution population is deployed to exploit the solution space by the TS-based local search. Moreover, an iterative vertex removal strategy and a tailored inherited initialization strategy are introduced to accelerate the procedure of kk-coloring, which in turn contributes to its high efficiency of addressing the chromatic problem. Thanks to the cooperative interplay between the distribution population and the solution population, the DEA-PPM with small populations is expected to achieve competitive results of GCPs.

4.1 The framework of DEA-PPM

Input: an undirected graph G=(V,E)G=(V,E);
Output: the color number kk, the obtained color assignment 𝐱G\mathbf{x}^{*}_{G};
1
2gen0gen\leftarrow 0;
3 initialize the color number kk;
4 while termination-condition 1 is not satisfied do
5       reduce G=(V,E)G=(V,E) to G=(V,E)G^{\prime}=(V^{\prime},E^{\prime}) by the IVR strategy ;
6       if gen=0gen=0 then
7            initialize 𝐐(0)\mathbf{Q}(0) by (5);
8             sample 𝐐(0)\mathbf{Q}(0) to generate 𝐏(0)\mathbf{P}(0); //* uinform initialization /*/
9      else
10             (𝐐(0),𝐏(0))=InherInit(𝐐,𝐏,k)(\mathbf{Q}(0),\mathbf{P}(0))=InherInit(\mathbf{Q},\mathbf{P},k); //* inherited initialization /*/ k=k1k=k-1;
11       end if
12      
13      set 𝐱G\mathbf{x}^{*}_{G^{\prime}} as the best solution in 𝐏(0)\mathbf{P}(0);
14       𝐩1=𝐱G\mathbf{p}_{1}=\mathbf{x}^{*}_{G^{\prime}}, 𝐩2=𝐱G\mathbf{p}_{2}=\mathbf{x}^{*}_{G^{\prime}};
15      
16      t1t\leftarrow 1;
17      
18      while termination-condition 2 is not satisfied do
19             𝐐(t)=OrthExpQ(𝐐(t1),𝐏(t1))\mathbf{Q}^{\prime}(t)=OrthExpQ(\mathbf{Q}(t-1),\mathbf{P}(t-1)); //* orthogonal exploration /*/
20             𝐏(t)=SampleP(𝐐(t),𝐏(t1)\mathbf{P}^{\prime}(t)=SampleP(\mathbf{Q}^{\prime}(t),\mathbf{P}(t-1); //* sampling with inheritance /*/
21             (𝐏(t),𝐩1,𝐩2,𝐱G)=RefineP(𝐏(t),𝐩1,𝐩2,𝐱G)(\mathbf{P}(t),\mathbf{p}_{1},\mathbf{p}_{2},\mathbf{x}^{*}_{G^{\prime}})=RefineP(\mathbf{P}^{\prime}(t),\mathbf{p}_{1},\mathbf{p}_{2},\mathbf{x}^{*}_{G^{\prime}}) //* refinement of the solution population /*/
22            
23            𝐐(t)=RefineQ(𝐏(t),𝐏(t),𝐐(t))\mathbf{Q}(t)=RefineQ(\mathbf{P}^{\prime}(t),\mathbf{P}(t),\mathbf{Q}^{\prime}(t)); //* refinement of the distribution population /*/
24             tt+1t\leftarrow t+1;
25            
26       end while
27      recover 𝐱G\mathbf{x}^{*}_{G^{\prime}} by the IR strategy to get 𝐱G\mathbf{x}^{*}_{G};
28       𝐐=𝐐(t)\mathbf{Q}=\mathbf{Q}(t), 𝐏=𝐏(t)\mathbf{P}=\mathbf{P}(t);
29       gengen+1gen\leftarrow gen+1;
30      
31 end while
Algorithm 1 The framework of DEA-PPM
\color

red As presented in Algorithm 1, DEA-PPM is implemented as two nested loops: the inner loop addressing the kk-coloring problem and the outer loop decreasing kk to get the chromatic number χ(G)\chi(G). Based on a distribution population 𝐐(t)=(𝐪[1](t),,𝐪[np](t))\mathbf{Q}(t)=(\mathbf{q}^{[1]}(t),\dots,\mathbf{q}^{[np]}(t)) and the corresponding solution population 𝐏(t)=(𝐱[1](t),,𝐱[np](t))\mathbf{P}(t)=(\mathbf{x}^{[1]}(t),\dots,\mathbf{x}^{[np]}(t)), it starts with the initialization of the color number kk, which is then minimized by the outer loop to get the chromatic number χ(G)\chi(G).

Each iteration of the outer loop begins with the iterative vertex removal (IVR) strategy [51], by which the investigated graph GG could be transformed into a reduced graph G=(V,E)G^{\prime}=(V^{\prime},E^{\prime}), and the complexity of the coloring process could be reduced as well. Then, Lines 5-11 of Algorithm 1 initialize a distribution population 𝐐(0)\mathbf{Q}(0) and the corresponding solution population 𝐏(0)\mathbf{P}(0) for GG^{\prime}. Once GG^{\prime} is colored by Lines 12-20 of Algorithm 1, DEA-PPM recovers the obtained color assignment 𝐱G\mathbf{x}^{*}_{G^{\prime}} to get an color assignment 𝐱G\mathbf{x}^{*}_{G} of GG, and 𝐐(t)\mathbf{Q}(t) as well as 𝐏(t)\mathbf{P}(t) is archived for the inherited initialization performed at the next generation. Repeating the aforementioned process until the termination-condition 1 is satisfied, DEA-PPM returns a color number kk and the corresponding color assignment 𝐱G\mathbf{x}^{*}_{G}.

After the initialization of 𝐱G\mathbf{x}^{*}_{G^{\prime}}, 𝐩1\mathbf{p}_{1}, 𝐩2\mathbf{p}_{2} and tt, the inner loop tries to get a legal kk-coloring assignment for the reduced graph GG^{\prime} by evolving both the distribution population 𝐐(t)\mathbf{Q}(t) and the solution population 𝐏(t)\mathbf{P}(t). It first performs the orthogonal exploration on 𝐐(t)\mathbf{Q}(t) to generate 𝐐(t)\mathbf{Q}^{\prime}(t), and then, generates an intermediate solution population 𝐏(t)\mathbf{P}^{\prime}(t), which is further refined to get 𝐏(t+1)\mathbf{P}(t+1). Meanwhile, 𝐐(t+1)\mathbf{Q}(t+1) is generated by refining 𝐐(t)\mathbf{Q}^{\prime}(t). The inner loop repeats until the termination-condition 2 is satisfied.

The outer loop of DEA-PPM is implemented only once for the kk-coloring problem. To address the chromatic problem, the termination condition 1 is satisfied if the chromatic number has been identified or the inner loop fails to get a legal kk-coloring assignment for a given iteration budget. The termination condition 2 is met while a legal kk-coloring assignment is obtained or the maximum iteration number is reached.

For the kk-coloring problem, DEA-PPM initializes the color number kk by a given positive integer. While it is employed to address the chromatic number problem, we set k=ΔG+1k=\Delta G+1222Here ΔG\Delta G is the maximum vertex degree of graph GG. because an undirected graph GG is sure to be (ΔG+1)(\Delta G+1)-colorable [52].

4.2 The iterative vertex removal strategy and the inverse recovery strategy

To reduce the time complexity of the kk-coloring algorithm, Yu et al. [51] proposed an iterative vertex removal (IVR) strategy to reduce the size of the investigated graph. By successively removing vertexes with degrees less than kk, IVR generates a reduced graph G=(V,E)G^{\prime}=(V^{\prime},E^{\prime}), and put the removed vertexes into a stack SS. In this way, one could get a graph GG^{\prime} where degrees of vertexes are greater than or equal to kk, and its size could be significantly smaller than that of GG.

While a kk-coloring assignment 𝐱G\mathbf{x}^{*}_{G^{\prime}} is obtained for the reduced graph GG^{\prime}, the inverse recovery (IR) operation is implemented by recovering vertexes in the stack SS. The IR process is initialized by assigning any legal color to the vertex at the top of SS. Because the IVR process removes vertexes with degree less than kk, the IR process can get all recovered vertexes colored without conflicting. An illustration for the implement of the IVR and the IR is presented in Fig. 1.

Refer to caption
Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
(d)
Refer to caption
(e)
Figure 1: The iterative vertex removal and the inverse recovery process for the 3-coloring of G=(V,E)G=(V,E). (a) Graph G=(V,E)G=(V,E) consisting of 8 vertexes and 14 edges. (b)(c) Vertexes BB and AA are successively removed from GG to get the reduced graph G=(V,E)G^{\prime}=(V^{\prime},E^{\prime}). (d) A legal 3-color assignment of GG^{\prime}. (e)(f) Recover vertexes AA and BB to get a legal 3-color assignment of GG.

4.3 Population initialization

Depending on the iteration stage of DEA-PPM, the initialization of populations is implemented by the uniform initialization or the inherited initialization.

At the beginning, the uniform initialization generates npnp individuals of 𝐐(0)\mathbf{Q}(0) as

𝐪(0)=[1k1k1k1k1k1k1k1k1k].\mathbf{q}^{(0)}=\begin{bmatrix}\frac{1}{\sqrt{k}}&\frac{1}{\sqrt{k}}&...&\frac{1}{\sqrt{k}}\\ \frac{1}{\sqrt{k}}&\frac{1}{\sqrt{k}}&...&\frac{1}{\sqrt{k}}\\ \vdots&\vdots&\vdots&\vdots\\ \frac{1}{\sqrt{k}}&\frac{1}{\sqrt{k}}&...&\frac{1}{\sqrt{k}}\end{bmatrix}. (5)

𝐏(0)\mathbf{P}(0) are generated by sampling model (5) npnp times.

While gen>0gen>0, the inherited initialization gets 𝐐(0)\mathbf{Q}(0) and 𝐏(0)\mathbf{P}(0) with the assistance of the distribution populations 𝐐\mathbf{Q} and the solution population 𝐏\mathbf{P} archived at the last generation. The graph GG has been colored with kk colors, and it is anticipated to get a legal k1k-1-color assignment. To get an initial color assignment of k1k-1 colors, a color index lml_{m} that corresponds to the minimum vertex independent set is identified for 𝐱[i]=(x1[i],,xn[i])𝐏\mathbf{x}^{[i]}=({x}^{[i]}_{1},\dots,{x}^{[i]}_{n})\in\mathbf{P}. Then, we get the initial color assignment 𝐲[i]=(y1[i],,yn[i])\mathbf{y}^{[i]}=({y}^{[i]}_{1},\dots,{y}^{[i]}_{n}) by

y[i]={x[i]1,if x[i]lm,x[i],otherwise.y^{[i]}=\begin{cases}x^{[i]}-1,&\mbox{if }x^{[i]}\geq l_{m},\\ x^{[i]},&\mbox{otherwise.}\end{cases}

Meanwhile, delete the lml_{m}-th row of 𝐪[i]\mathbf{q}^{[i]}, and normalize its columns to get an initial distribution 𝐫[i]\mathbf{r}^{[i]} corresponding to k1k-1 colors. Details of the inherited initialization are presented in Algorithm 2.

Input: a distribution population 𝐐\mathbf{Q}, a solution population 𝐏\mathbf{P}, a color number kk;
Output: the initialized distribution population 𝐐\mathbf{Q}^{\prime}, the initialized solution population 𝐏\mathbf{P}^{\prime};
1 for i=1,,npi=1,\dots,np do
2      𝐱[i]𝐏\mathbf{x}^{[i]}\in\mathbf{P},𝐪[i]𝐐\mathbf{q}^{[i]}\in\mathbf{Q};
3       transform 𝐱[i]=(x1[i],,xn[i])\mathbf{x}^{[i]}=(x^{[i]}_{1},\dots,x^{[i]}_{n}) to a vertex partition s={V1,,Vk}s=\{V_{1},\dots,V_{k}\};
4       lmargmin{|Vj|}l_{m}\leftarrow\arg\min\{|V_{j}|\};
5       for j=1,,nj=1,\dots,n do
6             if xj[i]lmx^{[i]}_{j}\geq l_{m} then
7                  xj[i]=xj[i]1x^{[i]}_{j}=x^{[i]}_{j}-1;
8             end if
9            
10       end for
11      𝐲[i]=𝐱[i]\mathbf{y}^{[i]}=\mathbf{x}^{[i]};
12       delete the lml_{m}-th row of 𝐪[i]\mathbf{q}^{[i]} and normalize 𝐪[i]\mathbf{q}^{[i]} to get 𝐫[i]\mathbf{r}^{[i]};
13      
14 end for
15𝐐=i=1np𝐫[i]\mathbf{Q}^{\prime}=\bigcup_{i=1}^{np}\mathbf{r}^{[i]}, 𝐏=i=1np𝐲[i]\mathbf{P}^{\prime}=\bigcup_{i=1}^{np}\mathbf{y}^{[i]};
Algorithm 2 (𝐐,𝐏)=InherInit(𝐐,𝐏,k)(\mathbf{Q}^{\prime},\mathbf{P}^{\prime})=InherInit(\mathbf{Q},\mathbf{P},k)

4.4 Evolution of the distribution population

Based on the distribution model defined by (3) and (4), DEA-PPM performs the orthogonal exploration on individuals of 𝐐(t)\mathbf{Q}(t) to explore the probability space. Moreover, distribution individuals of 𝐐(t)\mathbf{Q}(t) are refined by an exploitation strategy or a disturbance strategy.

4.4.1 Orthogonal transformation

An orthogonal transformation on a column vector is performed by premultiplying an orthogonal matrix M{M}, a square matrix satisfying

MTM=MMT=I,M^{T}\cdot M=M\cdot M^{T}=I,

where II is the identity matrix. Because an orthogonal transformation preserves the 2-norm [53], we know

Mv2=v2,vn.\|M\vec{v}\|_{2}=\|{\vec{v}}\|_{2},\quad\forall\vec{v}\,\in\mathbb{R}^{n}. (6)

Then, by performing orthogonal transformations on columns of the distribution individuals, DEA-PPM can explore the distribution space flexibly.

Since columns of an orthogonal matrix are orthonormal [53], one can get an orthogonal matrix by performing the QR decomposition on an invertible matrix [54].

4.4.2 Orthogonal exploration in the distribution space

To perform the orthogonal exploration in the distribution space, DEA-PPM generates an orthogonal matrix by performing the QR decomposition on an invertible matrix that is generated randomly. As presented in Algorithm 3, mm worst individuals of 𝐐\mathbf{Q} are modified by random orthogonal transformations performed on cc randomly selected columns. As an initial study, mm is set as a random integer in [1,np/2][1,np/2], and cc is an integer randomly sampled in [1,n/10][1,n/10].

Input: a distribution population 𝐐\mathbf{Q}, a solution population 𝐏\mathbf{P};
Output: the updated distribution population 𝐐\mathbf{Q}^{\prime};
1 sorting 𝐐\mathbf{Q} by fitness values of corresponding individuals in 𝐏\mathbf{P};
2 take 𝐐w\mathbf{Q}_{w} as the collection of mm worst individuals of 𝐐\mathbf{Q};
3 𝐐=𝐐𝐐w\mathbf{Q}^{\prime}=\mathbf{Q}\setminus\mathbf{Q}_{w};
4 for 𝐪𝐐w\mathbf{q}\in\mathbf{Q}_{w} do
5       𝐪𝐪\mathbf{q}^{\prime}\leftarrow\mathbf{q};
6       randomly select cc columns qjl(l=1,,c)\vec{q}^{\prime}_{j_{l}}(l=1,...,c) from 𝐪\mathbf{q}^{\prime};
7      
8      for l=1,,cl=1,...,c do
9             generate a random orthogonal matrix MlM_{l} ;
10             qjl=Mlqjl\vec{q_{j_{l}}}^{\prime}=M_{l}\vec{q_{j_{l}}}^{\prime};
11            
12       end for
13      𝐐=𝐐𝐪\mathbf{Q}^{\prime}=\mathbf{Q}^{\prime}\cup\mathbf{q}^{\prime};
14 end for
15
Algorithm 3 𝐐=OrthExpQ(𝐐,𝐏)\mathbf{Q^{\prime}}=OrthExpQ(\mathbf{Q},\mathbf{P})

4.4.3 Refinement of the distribution population

Input: two solution populations 𝐏\mathbf{P} and 𝐏\mathbf{P}^{\prime}, a distribution population 𝐐\mathbf{Q};
Output: the refined distribution population 𝐐\mathbf{Q}^{\prime};
1 for i=1,,npi=1,\dots,np do
2       𝐪[i]𝐐\mathbf{q}^{[i]}\in\mathbf{Q}^{\prime}, 𝐱[i]𝐏\mathbf{x}^{[i]}\in\mathbf{P}^{\prime},𝐲[i]𝐏\mathbf{y}^{[i]}\in\mathbf{P};
3       for j=1,,nj=1,\dots,n do
4             set rndjU(0,1)rnd_{j}\sim U(0,1);
5             if rndjp0rnd_{j}\leq p_{0} then
6                  rj[i]\vec{r}^{[i]}_{j} is generated by the exploitation strategy confirmed by Eqs. (7) and (8);
7            else
8                  rj[i]\vec{r}^{[i]}_{j} is generated by the exploitation strategy defined by Eq. (9);
9             end if
10            
11       end for
12      𝐫[i]=(r1[i],,rn[i])\mathbf{r}^{[i]}=(\vec{r_{1}}^{[i]},\dots,\vec{r_{n}}^{[i]});
13      
14 end for
𝐐=i=1np𝐫[i]\mathbf{Q}=\bigcup_{i=1}^{np}\mathbf{r}^{[i]}.
Algorithm 4 𝐐=RefineQ(𝐏,𝐏,𝐐)\mathbf{Q}=RefineQ(\mathbf{P}^{\prime},\mathbf{P},\mathbf{Q}^{\prime})

As presented in Algorithm 4, the distribution population 𝐐\mathbf{Q}^{\prime} is refined to generate 𝐐\mathbf{Q}. 𝐪[i]=(qi,j[i])k×n𝐐\forall\,\mathbf{q}^{[i]}=({q}^{[i]}_{i,j})_{k\times n}\in\mathbf{Q}^{\prime}, DEA-PPM refines its jj-th column qj[i]\vec{q}^{[i]}_{j} with the assistance of the jj-th components of 𝐱[i]=(x1[i],,xn[i])𝐏\mathbf{x}^{[i]}=(x^{[i]}_{1},\dots,x^{[i]}_{n})\in\mathbf{P}^{\prime} and 𝐲[i]=(y1[i],,yn[i])𝐏\mathbf{y}^{[i]}=(y^{[i]}_{1},\dots,y^{[i]}_{n})\in\mathbf{P}. With probability p0p_{0}, qj[i]\vec{q}^{[i]}_{j} is refined by an exploitation strategy; otherwise, its refinement is implemented by a disturbance strategy.

The exploitation strategy

Similar to the probability learning procedure proposed in [20], the first phase of the exploitation strategy is implemented by

rl,j[i]={α+(1α)(ql,j[i])2 if l=yj[i],(1α)(ql,j[i])2 if lyj[i],l=1,,k,r^{[i]}_{l,j}=\begin{cases}\sqrt{\alpha+(1-\alpha)(q^{[i]}_{l,j})^{2}}&\text{ if }l=y^{[i]}_{j},\\ \sqrt{(1-\alpha)(q^{[i]}_{l,j})^{2}}&\text{ if }l\neq y^{[i]}_{j},\end{cases}\quad l=1,\dots,k, (7)

where yj[i]y^{[i]}_{j} is the jj-th component of 𝐲[i]\mathbf{y}^{[i]}. Then, an local orthogonal transformation is performed as

[rl1,j[i]rl2,j[i]]=U(Δθj)×[rl1,j[i]rl2,j[i]],\begin{bmatrix}r^{[i]}_{l_{1},j}\\ r^{[i]}_{l_{2},j}\end{bmatrix}=U(\Delta\theta_{j})\times\begin{bmatrix}r^{[i]}_{l_{1},j}\\ r^{[i]}_{l_{2},j}\end{bmatrix}, (8)

where

U(Δθj)=[cos(Δθj)sin(Δθj)sin(Δθj)cos(Δθj)],U(\Delta\theta_{j})=\begin{bmatrix}cos(\Delta\theta_{j})&-sin(\Delta\theta_{j})\\ sin(\Delta\theta_{j})&cos(\Delta\theta_{j})\end{bmatrix},

l1=xj[i]l_{1}=x^{[i]}_{j}, l2=yj[i]l_{2}=y^{[i]}_{j}. Equation (7) conducts an overall regulation controlled by the parameter α\alpha, and equation (8) rotates the subvector (rl1,j[i],rl2,j[i])T(r^{[i]}_{l_{1},j},r^{[i]}_{l_{2},j})^{T} counterclockwise by Δθi\Delta\theta_{i} to regulate it slightly.

The disturbance strategy

j{1,2,,n}\forall\,j\in\{1,2,\dots,n\}, rj[i]=(r1,j[i],,rk,j[i])\vec{r}^{[i]}_{j}=(r^{[i]}_{1,j},\dots,r^{[i]}_{k,j}) is generated by

rl,j[i]={λ(ql0,j[i])21(1λ)(ql0,j[i])2 if l=l0,(ql,j[i])21(1λ)(ql0,j[i])2 if ll0,r^{[i]}_{l,j}=\begin{cases}\sqrt{\frac{\lambda(q^{[i]}_{l_{0},j})^{2}}{1-(1-\lambda)(q^{[i]}_{l_{0},j})^{2}}}&\text{ if }l=l_{0},\\ \sqrt{\frac{(q^{[i]}_{l,j})^{2}}{1-(1-\lambda)(q^{[i]}_{l_{0},j})^{2}}}&\text{ if }l\neq l_{0},\end{cases} (9)

l=1,,kl=1,\dots,k. For 0<λ<10<\lambda<1, the l0l_{0}-th components of rj[i]\vec{r}^{[i]}_{j} is smaller than that of qj[i]\vec{q}^{[i]}_{j}, and others are greater. Thus, we set l0=yj[i]l_{0}=y^{[i]}_{j} to prevent DEA-PPM from premature convergence.

4.5 Efficient search in the solution space

To search the solution space efficiently, DEA-PPM generates a solution population by sampling with inheritance, and then, refines it using a multi-parent crossover operation followed by the TS search proposed in Ref. [55].

4.5.1 The strategy of sampling with inheritance

Inspired by the group selection strategy [19], components of new solution 𝐲[i]=(y1[i],,yn[i])\mathbf{y}^{[i]}=({y}_{1}^{[i]},\dots,{y}_{n}^{[i]}) are either generated by sampling the distribution 𝐪[i]=(q1[i],,qn[i])\mathbf{q}^{[i]}=(\vec{q}^{[i]}_{1},\dots,\vec{q}^{[i]}_{n}) or inheriting from the corresponding solution 𝐱[i]=(x1[i],,xn[i])\mathbf{x}^{[i]}=({x}_{1}^{[i]},\dots,{x}_{n}^{[i]}). The strategy of sampling with inheritance is presented in Algorithm 5, where rr is the probability of generating yj[i]y^{[i]}_{j} by sampling qj[i]\vec{q}_{j}^{[i]}.

Input: a distribution population 𝐐\mathbf{Q}, a solution population 𝐏\mathbf{P};
Output: the generated solution population 𝐏\mathbf{P}^{\prime};
1 for i=1,,npi=1,\dots,np do
2       𝐪[i]𝐐\mathbf{q}^{[i]}\in\mathbf{Q}, 𝐱[i]𝐏\mathbf{x}^{[i]}\in\mathbf{P};
3       for j=1,,nj=1,\dots,n do
4             set rndjU(0,1)rnd_{j}\sim U(0,1);
5             if rndj<rrnd_{j}<r then
6                  sampling qj[i]\vec{q}^{[i]}_{j} to get yj[i]{y^{[i]}_{j}};
7            else
8                  yj[i]=xj[i]{y}^{[i]}_{j}={x}^{[i]}_{j};
9             end if
10            
11       end for
12      𝐲[i]=(y1[i],,yn[i])\mathbf{y}^{[i]}=({y}^{[i]}_{1},\dots,{y}^{[i]}_{n});
13      
14 end for
𝐏=i=1np𝐲[i]\mathbf{P}^{\prime}=\bigcup_{i=1}^{np}\mathbf{y}^{[i]}.
Algorithm 5 𝐏=SampleP(𝐐,𝐏)\mathbf{P}^{\prime}=SampleP(\mathbf{Q},\mathbf{P})

4.5.2 Refinement of the solution population

The quality of generated solutions is further improved by a refinement strategy presented in Algorithm 6, which is an iterative process consisting of a multi-parent greedy partition crossover guided by two promising solutions 𝐩1\mathbf{p}_{1} and 𝐩2\mathbf{p}_{2} as well as the TS process presented in Ref. [55]. Meanwhile, two promising solutions 𝐩1\mathbf{p}_{1} and 𝐩2\mathbf{p}_{2} are updated. The refinement process ceases once it stagnates for 20 consecutive iterations.

1
Input: a solution population PP^{\prime}, two reference solutions 𝐩1\mathbf{p}_{1} and 𝐩2\mathbf{p}_{2}, the best color assignment 𝐱G\mathbf{x}^{*}_{G^{\prime}};
Output: the updated solution population PP, two updated reference solutions 𝐩1\mathbf{p}_{1} and 𝐩2\mathbf{p}_{2}, the updated best color assignment 𝐱G\mathbf{x}^{*}_{G^{\prime}};
2 iter0iter\leftarrow 0, iter_stag0iter\_stag\leftarrow 0;
3 𝐜1=𝐱G\mathbf{c}_{1}=\mathbf{x}^{*}_{G^{\prime}};
4
5while iter_stag<20iter\_stag<20 do
6       𝐏=MGPX(𝐏,𝐩1,𝐩2)\mathbf{P}=MGPX(\mathbf{P}^{\prime},\mathbf{p}_{1},\mathbf{p}_{2});
7       𝐏=Tabu(𝐏)\mathbf{P}=Tabu(\mathbf{P});
8       record the best solution in 𝐏\mathbf{P} as 𝐛\mathbf{b};
9       if f(𝐛)<f(𝐩1)f(\mathbf{b})<f(\mathbf{p}_{1}) then
10             iter_stag=0iter\_stag=0;
11             𝐜1=𝐩1,𝐩1=𝐛\mathbf{c}_{1}=\mathbf{p}_{1},\mathbf{p}_{1}=\mathbf{b};
12      else
13            iter_stag=iter_stag+1iter\_stag=iter\_stag+1;
14       end if
15      
16      if f(𝐛)<f(𝐱G)f(\mathbf{b})<f(\mathbf{x}^{*}_{G^{\prime}}) then
17            𝐱G=𝐛\mathbf{x}^{*}_{G^{\prime}}=\mathbf{b};
18            
19       end if
20      
21      if mod(iter,10)=0mod(iter,10)=0 then
22            𝐩2=𝐜1\mathbf{p}_{2}=\mathbf{c}_{1};
23       end if
24      𝐏=𝐏\mathbf{P}^{\prime}=\mathbf{P};
25       iter=iter+1iter=iter+1;
26      
27 end while
Algorithm 6 (𝐏,𝐩1,𝐩2,𝐱G)=RefineP(𝐏,𝐩1,𝐩2,𝐱G)(\mathbf{P},\mathbf{p}_{1},\mathbf{p}_{2},\mathbf{x}^{*}_{G^{\prime}})=RefineP(\mathbf{P}^{\prime},\mathbf{p}_{1},\mathbf{p}_{2},\mathbf{x}^{*}_{G^{\prime}})
Multi-parent greedy partition crossover (MGPX)

Inspired by the motivation of greedy partition crossover (GPX) for graph coloring [55], we propose the multi-parent greedy partition crossover (MGPX) presented in Algorithm 7. For 𝐱1𝐏\mathbf{x}_{1}\in\mathbf{P}, two mutually different solutions 𝐱2\mathbf{x}_{2} and 𝐱3\mathbf{x}_{3} are selected from 𝐏{𝐱1}{𝐩1,𝐩2}\mathbf{P}\setminus\{\mathbf{x}_{1}\}\cup\{\mathbf{p}_{1},\mathbf{p}_{2}\}. Then, the MGPX is performed on 𝐱1\mathbf{x}_{1}, 𝐱2\mathbf{x}_{2} and 𝐱3\mathbf{x}_{3} to generate a new solution 𝐲\mathbf{y}. After the traversal of the solution population 𝐏\mathbf{P}, all generated solutions construct the intermediate solution population 𝐏\mathbf{P}^{\prime}.

Update of the promising solutions

The promising solution 𝐩1\mathbf{p}_{1} is updated if a better solution 𝐛\mathbf{b} is obtained. Then, 𝐩2\mathbf{p}_{2} is set as the original values of 𝐩1\mathbf{p}_{1}. To fully exploits the promising information incorporated by 𝐩2\mathbf{p}_{2}, it is updated once every 10 iterations.

Input: a solution population 𝐏\mathbf{P}, two reference solutions 𝐩1\mathbf{p}_{1} and 𝐩2\mathbf{p}_{2};
Output: the updated solution population 𝐏\mathbf{P}^{\prime};
1 𝐏=\mathbf{P}^{\prime}=\emptyset;
2 for 𝐱𝐏\mathbf{x}\in\mathbf{P} do
3       let 𝐱1=𝐱\mathbf{x}_{1}=\mathbf{x}, 𝐱2\mathbf{x}_{2} and 𝐱3\mathbf{x}_{3} be two different solutions selected from 𝐏{𝐱}{𝐩1,𝐩𝟐}\mathbf{P}\setminus\{\mathbf{x}\}\cup\{\mathbf{p}_{1},\mathbf{p_{2}}\};
4       transform 𝐱i\mathbf{x}_{i} to the corresponding vertex partition si={V1i,,Vki}s_{i}=\{V_{1}^{i},\dots,V_{k}^{i}\}, i=1,2,3i=1,2,3;
5       for l=1,,kl=1,\dots,k do
6             randomly select an index i0{1,2,3}i_{0}\in\{1,2,3\} according to the probability distribution {Pc,(1Pc)/2,(1Pc)/2}\{P_{c},(1-P_{c})/2,(1-P_{c})/2\};
7             choose l0l_{0} such that Vl0i0V_{l_{0}}^{i_{0}} has a maximum cardinality;
8             Vl:=Vl0i0V_{l}:=V_{l_{0}}^{i_{0}};
9             remove the vertices of VlV_{l} from s1s_{1}, s2s_{2} and s3s_{3};
10            
11       end for
12      assign the vertices of Vl=1kVlV^{\prime}\setminus\bigcup_{l=1}^{k}V_{l} by s1s_{1};
13       transform s=(V1,,Vk)s=(V_{1},\dots,V_{k}) to a solution 𝐲\mathbf{y};
14       𝐏=𝐏{𝐲}\mathbf{P}^{\prime}=\mathbf{P}^{\prime}\cup\{\mathbf{y}\};
15      
16 end for
Algorithm 7 𝐏=MGPX(𝐏,𝐩1,𝐩2)\mathbf{P}^{\prime}=MGPX(\mathbf{P},\mathbf{p}_{1},\mathbf{p}_{2})

5 Numerical Experiments

\color

red The investigated algorithms are evaluated on the benchmark instances from the second DIMACS competition333Publicly available at ftp://dimacs.rutgers.edu/pub/challenge/graph/benchmarks/color/. that were used to test graph coloring algorithms in recent studies [16, 17, 20, 29, 36] . All tested algorithms are developed in C++ programming language, and run in Microsoft Windows 7 on a laptop equipped with the Intel(R) Core(TM) i7 CPU 860 @ 2.80GHz and 8GB system memory. We first perform a parameter study to get appropriate parameter settings of DEA-PPM, and then, the proposed evolution strategies of distribution population are investigated to demonstrate their impacts on its efficiency. Finally, numerical comparisons for both the chromatic problem and the kk-coloring problem are performed with the state-of-the-art algorithms. For numerical experiments, time budgets of all algorithms are consistently set as 3600 seconds (one hour). \colorred Because performance of the investigated algorithms varies for the selected benchmark problems, we perform the numerical comparison in two different ways. If two compared algorithms achieve inconsistent coloring results for the chromatic problem, numerical comparison is performed by the obtained color numbers; otherwise, we take the running time as the evaluation metric while they get the same coloring results.

5.1 Parameter study

By setting kk as the best known color numbers of the benchmark problems, preliminary experiments for the kk-coloring problem show that the performance of DEA-PPM is significantly influenced by the population size npnp, the regulation parameter α\alpha and the maximum iteration budget ItermaxIter_{max} of the TS. Then, we first demonstrate the univariate influence of parameters by the one-way analysis of variance (ANOVA), and then, perform a descriptive comparison to get a set of parameter for further numerical investigations. \colorblue The benchmark instances selected for the parameter study are the kk-coloring problems of DSJC500.5, flat300_28_0, flat1000_50_0, flat1000_76_0, le450_15c, le450_15d.

\color

red

5.1.1 Analysis of variance on the impacts of parameters

Our preliminary experiments show that DEA-PPM achieves promising results with np=8np=8, α=0.2\alpha=0.2 and itermax=5000iter_{max}=5000, which is taken as the baseline parameter setting of the one-side ANOVA test of running time. With the significance level of 0.05, the significant influences are highlighted in Tab.1 by bold P-values.

Table 1: Results of the one-way ANOVA test.
Tested Parameter P Values
Parameter Settings flat300_28_0 le450_15c le450_15d DSJC500.5 flat1000_50_0 flat1000_76_0
npnp {4,6,8,10,12}\{4,6,8,10,12\} 0.162 0.548 0.404 0.004 0.001 0.001
α\alpha {0.1,0.15,0.2,0.25,0.3}\{0.1,0.15,0.2,0.25,0.3\} 0.622 0.643 0.357 0.003 0.268 0.000
itermaxiter_{max} {0.5,1,1.5,2.0,2.5}×104\{0.5,1,1.5,2.0,2.5\}\times 10^{4} 0.025 0.146 0.000 0.040 0.000 0.000

Generally, the univariate changes of npnp, α\alpha and itermaxiter_{max} do not have significant influence on performance of DEA-PPM for instances flat300_28_0, le450_15c and le450_15d, except that values of itermaxiter_{max} has great impact on the results of le450_15d. But for instances DSJC500.5, flat1000_50_0 and flat1000_76_0, the influence is significant, except that α\alpha does not significantly influence the performance of DEA-PPM on flat1000_50_0. To illustrate the results, we included the curves of expected running time in Fig. 2. The univariate analysis shows that the best results could be achieved by setting np=8np=8, α=0.2\alpha=0.2 and itermax=5000iter_{max}=5000.

Refer to caption
(a) npnp
Refer to caption
(b) α\alpha
Refer to caption
(c) itermaxiter_{max}
Figure 2: Influence of parameters on expected running time of DEA-PPM.

5.1.2 Descriptive statistics on the composite impacts of parameters

Besides the one-way ANOVA test, we also present a descriptive comparison for the composite impact of sevaral parameter settings. With the parameter combinations presented in Tab. 2, statistical results for running time of 3030 independent runs are included in Fig. 3.

Table 2: Candidate parameter settings of DEA-PPM.
Parameter Setting
S1S_{1} S2S_{2} S3S_{3} S4S_{4} S5S_{5} S6S_{6} S7S_{7} S8S_{8} S9S_{9} S10S_{10} S11S_{11} S12S_{12}
npnp 4 4 4 4 4 4 8 8 8 8 8 8
α\alpha 0.1 0.1 0.1 0.2 0.2 0.2 0.1 0.1 0.1 0.2 0.2 0.2
ItermaxIter_{max} 5000 10000 20000 5000 10000 20000 5000 10000 20000 5000 10000 20000
Refer to caption
Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
(d)
Refer to caption
(e)
Figure 3: Statistical results for running time of the DEA-PPM that addresses the kk-coloring problem of selected benchmark problems. (a)DSJC500.5(k=48k=48). (b)flat300_28_0(k=31k=31). (c)flat1000_50_0(k=50k=50). (d)flat1000_76_0(k=86k=86). (e)le450_15c(k=15k=15). (f)le450_15d(k=15k=15).

It indicates that the parameter setting S10S_{10} leads to the most promising results of DEA-PPM. Combining it with the setting of other parameters, we get the parameter setting of DEA-PPM presented in Tab. 3, which is adopted in the following experiments.

Table 3: Parameter setting of the DEA-PPM for numerical experiments.
Parameter Setting Description
npnp 8 Population size of 𝐐(t)\mathbf{Q}(t) and 𝐏(t)\mathbf{P}(t);
α\alpha 0.20.2 Regulation parameter in equation (7);
itermaxiter_{max} 5×1035\times 10^{3} Iteration budget for the TS;
p0p_{0} 0.980.98 Parameter for update of 𝐐(t)\mathbf{Q}(t);
Δθi\Delta\theta_{i} 0.05π0.05\pi Parameter in equation (8);
λ\lambda 0.50.5 Parameter in equation (9);
rr randomly selected from {0.2,0.8}\{0.2,0.8\} Parameter in Algorithm 5;
PcP_{c} 0.40.4 Parameter in Algorithm 7;

5.2 Experiments on the evolution strategies of probability distribution

In DEA-PPM, the evolution of distribution population is implemented by the orthogonal exploration strategy and the exploitation strategy. We try to validate the positive effects of these strategies in this section.

Refer to caption
(a) Easy benchmark problems
Refer to caption
(b) Hard benchmark problems
Figure 4: Comparison of running time between the DEA-PPM-O and the DEA-PPM-N by selected benchmark problems.

To validate the efficiency of the orthogonal exploration strategy, we compare two variants, the DEA-PPM with orthogonal exploration (DEA-PPM-O) and the DEA-PPM without orthogonal exploration (DEA-PPM-N), and show in Fig. 4 the statistical results of running time for kk-coloring of the easy benchmark problems (DSJC500.1 (k=12k=12), le450_15c (k=15k=15), led450_15d (k=15k=15)) and the hard benchmark problems(DSJC500.5 (k=48k=48), DSJC1000.1 (k=20k=20), DSJC1000.9 (k=226k=226)). The box plots imply that with the employment of the orthogonal exploration strategy, DEA-PPM-O performs generally better than DEA-PPM-N, resulting in smaller values of the median value, the quantiles and the standard deviations of running time.

\color

red The positive impact of exploitation strategy is verified by comparing the DEA-PPM with exploitation (DEA-PPM-E) with the variant without exploitation (DEA-PPM-W), and the box plots of running time are included in Fig. 5. It is demonstrated that DEA-PPM-E generally outperforms DEA-PPM-W in terms of the median value, the quantiles and the standard deviation of running time.

Refer to caption
(a) Easy benchmark problems
Refer to caption
(b) Hard benchmark problems
Figure 5: Comparison of running time between the DEA-PPM-E and the DEA-PPM-W by selected benchmark problems.
\color

blue Besides the statistical comparison regarding the exact values of running time, we perform a further comparison by the Wilcoxon rank sum test with a significance level of 0.05, where the statistical test is based on the sorted rank of running time instead of its exact values. The results are included in Tab. 4, where “P” is the p-value of hypothesis test. For the test conclusion “R”, “+”, “-” and “\sim” indicate that the performance of DEA-PPM is better than, worse than and incomparable to that of the compared variant, respectively. The results demonstrate that DEA-PPM outperforms DEA-PPM-N and DEA-PPM-W on two instances, and is not inferior to them for all benchmark problems. It further validates the conclusion that both the exploration strategy and the exploitation strategy significantly improve the performance of DEA-PPM.

Table 4: Wilcoxon rank-sum test for the evolution strategies of distribution population.
Instance DEA-PPM-N DEA-PPM-W
P R P R
DSJC500.1 1.08E-01 \sim 9.25E-01 \sim
le450_15c 6.36E-01 \sim 4.99E-02 +
le450_15d 2.50E-01 \sim 7.76E-01 \sim
DSJC500.5 4.97E-02 + 5.29E-02 \sim
DSJC1000.1 6.56E-03 + 4.57E-01 -
DSJC1000.9 7.15E-01 \sim 4.68E-05 +
+/\sim/- 2/4/0 2/4/0

5.3 Numerical comparison with the state-of-the-art algorithms

To demonstrate the competitiveness of DEA-PPM, we perform numerical comparison for the chromatic problem and the kk-coloring problem with SDGC [16], MACOL [29], SDMA [17], PLSCOL [20], and HEAD [36], the parameter settings of which are presented in Tab. 5. If an algorithm cannot address the chromatic problem or the kk-coloring problem in 3600 seconds, a failed run is recorded by the running time of 3600 seconds.

Table 5: Parameter settings of the compared algorithms.
Algorithms Parameters Description Values
SDGC itit The number of iterations; 1×1051\times 10^{5}
MACOL pp Size of population; 2020
α\alpha Depth of TS; 1×1051\times 10^{5}
mm Number of parents for crossover; A random number in {2,,6}\{2,\dots,6\}
prp_{r} Probability for accepting worse offspring; 0.20.2
λ\lambda Parameter for goodness score function; 0.080.08
SDMA β\beta Search depth of weight tabu coloring; 1×1061\times 10^{6}
ttwtt_{w} Tabu tenure of weight tabu coloring; rand(10)+frand(10)+f^{{}^{\prime}}
tttt Tabu tenure of perturbation; rand(1000)+frand(1000)+f^{{}^{\prime}}
LL Level limit of coarsening phase; 55
λ\lambda Unimproved consecutive rounds for best solution; 1010
PLSCOL ω\omega Noise probability; 0.20.2
α\alpha Reward factor for correct group; 0.10.1
β\beta Penalization factor for incorrect group; [0.05,0.45]\left[0.05,0.45\right]
γ\gamma Compensation factor for expected group; 0.30.3
ρ\rho Smoothing coefficient; 0.50.5
p0p_{0} Smoothing threshold; 0.9950.995
HEAD IterTCIter_{TC} Depth of TS; 1×1051\times 10^{5}
ItercycleIter_{cycle} The number of generations into one cycle; 1010

5.3.1 Comparison for the chromatic number problem

In order to verify the competitiveness of DEA-PPM on the chromatic number problem, we compare it with SDGC, MACOL, SDMA, PLSCOL and HEAD by 8 selected benchmark problems, and the statistical results of 30 independent runs are collected in Tab. 6, where kavek_{ave}, kmink_{min}, kmaxk_{max} and kstdk_{std} represent the average color number, the maximum color number, the minimum color number and the standard deviation of color numbers, respectively. The best results are highlighted by bold texts.

Table 6: Comparison of DEA-PPM with SDGC, MACOL, SDMA, PLSCOL and HEAD for the chromatic number problem.
Instance χ(G)\chi(G) Algorithm kavek_{ave} kmink_{min} kmaxk_{max} kstdk_{std} Instance χ(G)\chi(G) Algorithm kavek_{ave} kmink_{min} kmaxk_{max} kstdk_{std}
fpsol2_i_2 30 SDGC 85.4 79 91 3.71 le450_15c 15 SDGC 29.07 28 32 1.46
MACOL 88.3 88 89 0.46 MACOL 19.4 18 21 1.2
SDMA 59.41 53 73 6.22 SDMA 30.93 28 38 3.12
PLSCOL 73 71 77 1.75 PLSCOL 16.1 15 17 0.4
HEAD 74.7 71 78 1.97 HEAD 15.87 15 16 0.34
DEA-PPM 30 30 30 0 DEA-PPM 15 15 15 0
fpsol2_i_3 30 SDGC 85.5 79 95 4.35 le450_15d 15 SDGC 30.27 27 34 2.24
MACOL 88 87 89 0.45 MACOL 18.7 17 21 1.35
SDMA 57.86 51 65 5.28 SDMA 32.72 31 38 2.17
PLSCOL 71.67 66 77 3.47 PLSCOL 16.33 16 17 0.47
HEAD 75.57 73 78 1.54 HEAD 15.93 15 16 0.25
DEA-PPM 30 30 30 0 DEA-PPM 15 15 15 0
flat300_26_0 26 SDGC 40.5 38 44 1.5 DSJC500_1 12 SDGC 16.67 16 18 0.79
MACOL 31.8 31 32 0.4 MACOL 13 13 13 0
SDMA 44.83 43 51 3.12 SDMA 20.76 17 29 3.39
PLSCOL 26 26 26 0 PLSCOL 12.77 12 13 0.42
HEAD 26 26 26 0 HEAD 13 13 13 0
DEA-PPM 26 26 26 0 DEA-PPM 12 12 12 0
flat300_28_0 28 SDGC 40.83 40 42 0.58 DSJC1000_1 20 SDGC 31.63 31 32 0.48
MACOL 32 32 32 0 MACOL 80.37 76 82 2.79
SDMA 45.72 43 54 3.64 SDMA 86.21 73 94 5.55
PLSCOL 31 30 32 0.73 PLSCOL 21 21 21 0
HEAD 31 31 31 0 HEAD 21 21 21 0
DEA-PPM 31 31 31 0 DEA-PPM 21 21 21 0
\color

redIt is shown that DEA-PPM generally outperforms the other five state-of-the-art algorithms on kavek_{ave}, kmink_{min}, kmaxk_{max} and kstdk_{std} of 30 independent runs. Attributed to the population-based distribution evolution strategy, the global exploration ability of DEA-PPM is enhanced significantly. Moreover, the inherited initialization strategy improve the searching efficiency of the inner loop for search of kk-coloring assignment. As a result, it can address these problems efficiently and obtain χ(G)\chi(G) with a 100% success rate for all of eight selected problems.

It is noteworthy that the competitiveness is partially attributed to the IVR strategy introduced by DEA-PPM, especially for the sparse benchmark graphs fpsol2.i.2 and fpsol2.i.3. Numerical implementation shows that when k=30k=30, introduction of the IVR strategy reduces the vertex number of fpsol2.i.2 and fpsol2.i.3 from 451 and 425 to 90 and 88, respectively. Thus, the scale of the reduced graph GG^{\prime} is significantly cut down for fpsol2.i.2 and fpsol2.i.3, which greatly improves the efficiency of the kk-coloring process validated by the inner loop of DEA-PPM.

\color

red However, it demonstrates that DEA-PPM, PLSCOL and HEAD get consistent results on the instances flag300_26_0 and DSJC1000_1, and the best results of DEA-PPM and HEAD is a bit worse than that of PLSOCL. Accordingly, we further compare their performance by the Wilcoxon rand sum test. If the compared algorithms obtain different results of color number, the sorted rank is calculated according to the color number; while they get consistent results of color number, the rank sum test is performed according to the running time of 30 independent runs.

Table 7: Wilcoxon rank sum test for the comparison of performance on the chromatic number problem.
Instance SDGC MACOL SDMA PLSCOL HEAD
P R P R P R P R P R
fpsol2_i_2 1.09E-12 + 2.90E-13 + 1.13E-12 + 9.96E-13 + 1.11E-12 +
fpsol2_i_3 1.17E-12 + 1.59E-13 + 1.13E-12 + 1.12E-12 + 1.03E-12 +
flat300_26_0 7.98E-13 + 1.55E-13 + 1.11E-12 + 4.81E-11 - 6.73E-01 \sim
flat300_28_0 4.27E-13 + 1.69E-14 + 1.05E-12 + 6.31E-01 \sim 0.028 +
le450_15c 8.93E-13 + 7.31E-13 + 1.13E-12 + 2.05E-10 + 1.97E-11 +
le450_15d 1.05E-12 + 5.16E-13 + 9.63E-13 + 3.37E-13 + 7.15E-13 +
DSJC500_1 6.21E-13 + 1.69E-14 + 9.31E-13 + 1.47E-09 + 1.69E-14 +
DSJC1000_1 3.80E-13 + 1.09E-12 + 1.15E-12 + 2.74E-11 - 3.73E-09 -
+/\sim/- 8/0/0 8/0/0 8/0/0 5/1/2 6/1/1

The test results demonstrate that DEA-PPM does outperform SDGC, MACOL and SDMA on the selected benchmark problems. For the instance flat300_26_0, it is shown in Tab. 6 that DEA-PPM, PLSCOL and HEAD can address the chromatic number in 3600s. While the running time is compared by the rank sum test, we get the conclusion that PLSCOL runs fast than DEA-PPM. Considering the instance DSJC1000_1, DEA-PPM, PLSCOL and HEAD stagnate at the assignment of 21 colors. However, the rank sum test shows that DEA-PPM is inferior to PLSCOL and HEAD in term of the running time.

5.3.2 Comparison for the kk-coloring problem

Numerical results on the chromatic problem imply that DEA-PPM, PLSCOL and HEAD outperform SDGC, MACOL and SDMA, but the superiority of DEA-PPM over PLSCOL and HEAD is dependent on the benchmark instances. To further compare DEA-PPM with PLSCOL and HEAD, we investigate their performance for the kk-coloring problem, where kk is set as the chromatic number of the investigate instance. For 18 selected benchmark problems collected in Tab. 8, we present the success rate (SR) and average runtime (T) of 30 independent runs, and the best results are highlighted by bold texts.

Table 8: Numerical results of DEA-PPM, HEAD and PLSCOL for the kk-coloring problem
Instance kk PLSCOL HEAD DEA-PPM
SR T(s) SR T(s) SR T(s)
DSJC125.5 17 30/30 0.31 30/30 0.55 30/30 0.65
DSJC125.9 44 30/30 0.08 30/30 0.12 30/30 0.39
DSJC250.5 28 30/30 17.60 30/30 37.03 30/30 23.93
DSJC250.9 72 30/30 6.22 30/30 7.20 30/30 42.97
DSJC500.1 12 9/30 3140.21 29/30 1655.11 30/30 226.14
DSJC500.5 48 0/30 3600 30/30 1176.30 30/30 771.39
DSJC500.9 126 0/30 3600 1/30 3504.49 3/30 3346.64
DSJC1000.1 20 0/30 3600 0/30 3600 30/30 902.53
DSJC1000.5 85 0/30 3600 30/30 2575.73 23/30 2271.70
DSJC1000.9 225 0/30 3600 19/30 2784.67 12/30 3240.40
le450_15c 15 0/30 3600 30/30 400.85 30/30 9.34
le450_15d 15 0/30 3600 27/30 1121.86 30/30 24.60
flat300_20_0 20 30/30 0.11 30/30 0.20 30/30 0.81
flat300_26_0 26 30/30 3.47 30/30 8.81 30/30 15.46
flat300_28_0 30 5/30 3196.66 0/30 3600 0/30 3600
flat1000_50_0 50 30/30 159.32 30/30 433.27 30/30 636.96
flat1000_60_0 60 30/30 347.74 30/30 580.71 30/30 843.81
flat1000_76_0 84 0/30 3600 23/30 2834.23 30/30 2139.33
Average Rank 1.94 2 1.38 2.05 1.16 1.83
\color

blue Thanks to the incorporation of the population-based distribution evolution strategy, the global exploration of DEA-PPM has been significantly improved, resulting in better success rate for most of the selected problems except for DSJC1000.9 and flag_300_28_0. Accordingly, the average rank of DEA-PPM is 1.16, better than 1.94 of PLSCOL and 1.38 of HEAD. The global exploration improved by the population-based distribution strategy and the IVR contributes to faster convergence of DEA-PPM for the complicated benchmark problems, however, increases the generational complexity of DEA-PPM, which leads to its slightly increased running time in some small-scale problems. Consequently, DEA-PPM gets the first place with the average running-time rank 1.83.

\color

redFurther investigation of the performance is conducted by the Wilcoxon rank sum test of running time. With a significance level of 0.05, the results are presented in Tab. 9, where “P” is the p-value of hypothesis test. While both HEAD and DEA-PPM cannot get legal color assignments for flat300_28_0, the Wilcoxon rank sum test is conducted by the numbers of conflicts of 30 independent runs.

It is shown that DEA-PPM performs better than PLSCOL for 2 of 9 selected instances with vertex number less than 500, and better than HEAD for 3 of 9 problems, but performs a bit worse than PLSCOL and HEAD for most of small-scale instances. However, It outperforms PLSCOL and HEAD on the vast majority of instances with vertex number greater than or equal to 500. Therefore, we can conclude that DEA-PPM is competitive to PLSCOL and HEAD on large-scale GCPs, which is attributed to the composite function of the population-based distribution evolution mechanism and the IVR strategy.

Table 9: Results of Wilcoxon rank-sum test for performance comparison.
Instance (n<500n<500) PLSCOL HEAD Instance (n500n\geq 500) PLSCOL HEAD
P R P R P R P R
DSJC125.5 4.00E-03 - 0.46 \sim DSJC500.1 1.96E-10 + 1.10E-11 +
DSJC125.9 3.01E-11 - 3.00E-03 - DSJC500.5 1.21E-12 + 2.00E-03 +
DSJC250.5 0.22 \sim 0.38 \sim DSJC500.9 0.08 \sim 0.34 \sim
DSJC250.9 6.01E-08 - 6.53E-08 - DSJC1000.1 1.21E-12 + 1.21E-12 +
le450_15c 5.05E-13 + 1.40E-11 + DSJC1000.5 5.85E-09 + 0.06 \sim
le450_15d 5.05E-13 + 1.40E-11 + DSJC1000.9 1.53E-04 + 0.03 +
flat300_20_0 3.01E-11 - 3.01E-11 - flat1000_50_0 3.02E-11 - 1.75E-05 -
flat300_26_0 8.15E-11 - 1.39E-06 - flat1000_60_0 8.99E-11 - 6.36E-05 -
flat300_28_0 0.02 - 4.62E-05 + flat1000_76_0 3.45E-07 + 0.02 +
+/\sim/- 2/1/6 3/2/4 +/\sim/- 6/1/2 5/2/2

6 Conclusion and Future Work

\color

red To address the graph coloring problems efficiently, this paper develops a distribution evolution algorithm based on a population of probability model (DEA-PPM). Incorporating the merits of the respective probability models in EDAs and QEAs, we introduce a novel distribution model, for which an orthogonal exploration strategy is proposed to explore the probability space efficiently. Meanwhile, an inherited initialization is employed to accelerate the process of color assignment.

  • 1.

    Assisted by an iterative vertex removing strategy and a TS-based local search process, DEA-PPM can achieve excellent performance with small populations, which contributes to its competitiveness on the chromatic problem.

  • 2.

    Since the population-based evolution leads to slightly increased generational time complexity of DEA-PPM, its running time for the small-scale kk-coloring problems is a bit higher than that of the individual-based PLSCOL and HEAD.

  • 3.

    DEA-PPM achieves overall outperformance on benchmark problems with vertex numbers greater than 500, because its enhanced global exploration improves the ability of escaping from the local optimal solutions.

  • 4.

    The iterative vertex removal strategy reduces sizes of the graphs to be colored, which likewise improves the coloring performance of DEA-PPM.

The proposed DEA-PPM could be extended to other complex problems. To further improve the efficiency of DEA-PPM, our future work will focus on the adaptive regulation of population size, and the local exploitation is anticipated to be enhanced by utilizing the mathematical characteristics of graph instance. Moreover, we will try to develop a general framework of DEA-PPM to address a variety of combinatorial optimization problems.

Acknowledgement

This research was supported in part by the National Key R& D Program of China [grant number 2021ZD0114600], in part by the Fundamental Research Funds for the Central Universities [grant number WUT:2020IB006], and in part by the National Nature Science Foundation of China [grant number 61763010] as well as the Natural Science Foundation of Guangxi [grant number 2021GXNSFAA075011].

References

  • [1] G. W. Greenwood, Using differential evolution for a subclass of graph theory problems, IEEE Transactions on Evolutionary Computation 13 (2009) 1190–1192.
  • [2] F. J. A. Artacho, R. Campoy, V. Elser, An enhanced formulation for solving graph coloring problems with the douglas–rachford algorithm, Journal of Global Optimization 77 (2020) 783–403.
  • [3] O. Goudet, B. Duval, J.-K. Hao, Population-based gradient descent weight learning for graph coloring problems, Knowledge-Based Systems 212 (2021) 106581.
  • [4] T. Mostafaie, F. Modarres Khiyabani, N. J. Navimipour, A systematic study on meta-heuristic approaches for solving the graph coloring problem, Computers & Operations Research 120 (2020) 104850.
  • [5] P. Galinier, A. Hertz, A survey of local search methods for graph coloring, Computers & Operations Research 33 (9) (2006) 2547–2562.
  • [6] M. Dorigo, M. Birattari, T. Stuetzle, Ant colony optimization - artificial ants as a computational intelligence technique, IEEE Computational Intelligence Magazine 1 (4) (2006) 28–39.
  • [7] M. Hauschild, M. Pelikan, An introduction and survey of estimation of distribution algorithms, Swarm and Evolutionary Computation 1 (3) (2011) 111–128.
  • [8] H. Xiong, Z. Wu, H. Fan, G. Li, G. Jiang, Quantum rotation gate in quantum-inspired evolutionary algorithm: A review, analysis and comparison study, Swarm and Evolutionary Computation 42 (2018) 43–57.
  • [9] O. Titiloye, A. Crispin, Quantum annealing of the graph coloring problem, Discrete Optimization 8 (2) (2011) 376–384.
  • [10] A. J. Pal, B. Ray, N. Zakaria, S. S. Sarma, Comparative performance of modified simulated annealing with simple simulated annealing for graph coloring problem, Procedia Computer Science 9 (2012) 321–327.
  • [11] C. Avanthay, A. Hertz, N. Zufferey, A variable neighborhood search for graph coloring, European Journal of Operational Research 151 (2) (2003) 379–388.
  • [12] A. Hertz, D. de Werra, Using tabu search techniques for graph coloring, Computing 39 (1987) 345–351.
  • [13] D. C. Porumbel, J.-K. Hao, P. Kuntz, Informed reactive tabu search for graph coloring, Asia-Pacific Journal of Operational Research 30 (04) (2013) 1350010.
  • [14] I. Blöchliger, N. Zufferey, A graph coloring heuristic using partial solutions and a reactive tabu scheme, Computers & Operations Research 35 (3) (2008) 960–975.
  • [15] D. C. Porumbel, J.-K. Hao, P. Kuntz, A search space “cartography” for guiding graph coloring heuristics, Computers & Operations Research 37 (4) (2010) 769–778.
  • [16] S. F. Galán, Simple decentralized graph coloring, Computational Optimization and Applications 66 (2017) 163–185.
  • [17] W. Sun, J.-K. Hao, Y. Zang, X. Lai, A solution-driven multilevel approach for graph coloring, Applied Soft Computing 104 (2021) 107174.
  • [18] Y. Peng, X. Lin, B. Choi, B. He, Vcolor*: a practical approach for coloring large graphs, Frontiers of Computer Science 15 (4) (2021) 1–17.
  • [19] Y. Zhou, J.-K. Hao, B. Duval, Reinforcement learning based local search for grouping problems: A case study on graph coloring, Expert Systems with Applications 64 (2016) 412–422.
  • [20] Y. Zhou, B. Duval, J.-K. Hao, Improving probability learning based local search for graph coloring, Applied Soft Computing 65 (2018) 542–553.
  • [21] L.-Y. Hsu, S.-J. Horng, P. Fan, M. K. Khan, Y.-R. Wang, R.-S. Run, J.-L. Lai, R.-J. Chen, Mtpso algorithm for solving planar graph coloring problem, Expert Systems with Applications 38 (5) (2011) 5525–5531.
  • [22] H. Hernández, C. Blum, Distributed graph coloring: an approach based on the calling behavior of japanese tree frogs, Swarm Intelligence 6 (2) (2012) 117–150.
  • [23] I. Rebollo-Ruiz, M. Graña, An empirical evaluation of gravitational swarm intelligence for graph coloring algorithm, Neurocomputing 132 (2014) 79–84.
  • [24] R. Zhao, Y. Wang, C. Liu, P. Hu, H. Jelodar, M. Rabbani, H. Li, Discrete selfish herd optimizer for solving graph coloring problem, Applied Intelligence 50 (2020) 1633–1656.
  • [25] D. Chalupa, P. Nielsen, Parameter-free and cooperative local search algorithms for graph colouring, Soft Computing 25 (24) (2021) 15035–15050.
  • [26] L. Zhong, Y. Zhou, G. Zhou, Q. Luo, Enhanced discrete dragonfly algorithm for solving four-color map problems, Applied Intelligence 53 (2023) 6372–6400.
  • [27] T. N. Bui, T. Nguyen, C. M. Patel, K.-A. T. Phan, An ant-based algorithm for coloring graphs, Discret Applied Mathematics 156 (2008) 190–200.
  • [28] H. Djelloul, A. Layeb, S. Chikhi, Quantum inspired cuckoo search algorithm for graph colouring problem, International Journal of Bio-Inspired Computation 7 (2015) 183–194.
  • [29] Z. Lü, J.-K. Hao, A memetic algorithm for graph coloring, European Journal of Operational Research 203 (1) (2010) 241–250.
  • [30] D. C. Porumbel, J.-K. Hao, P. Kuntz, An evolutionary approach with diversity guarantee and well-informed grouping recombination for graph coloring, Computers & Operations Research 37 (10) (2010) 1822–1832.
  • [31] S. Mahmoudi, S. Lotfi, Modified cuckoo optimization algorithm (mcoa) to solve graph coloring problem, Applied Soft Computing 33 (2015) 48–64.
  • [32] Q. Wu, J.-K. Hao, Coloring large graphs based on independent set extraction, Computers & Operations Research 39 (2) (2012) 283–290.
  • [33] S. M. Douiri, S. Elbernoussi, Solving the graph coloring problem via hybrid genetic algorithms, Journal of King Saud University: Engineering Sciences 27 (2015) 114–118.
  • [34] M. Bessedik, B. Toufik, H. Drias, How can bees colour graphs, International Journal of Bio-Inspired Computation 3 (2011) 67–76.
  • [35] M. R. Mirsaleh, M. R. Meybodi, A michigan memetic algorithm for solving the vertex coloring problem, Journal of Computational Science 24 (2018) 389–401.
  • [36] L. Moalic, A. Gondran, Variations on memetic algorithms for graph coloring problems, Journal of Heuristics 24 (1) (2018) 1–24.
  • [37] A. F. d. Silva, L. G. A. Rodriguez, J. F. Filho, The improved colourant algorithm: a hybrid algorithm for solving the graph colouring problem, International Journal of Bio-Inspired Computation 16 (1) (2020) 1–12.
  • [38] V. A. Shim, K. C. Tan, C. Y. Cheong, J. Y. Chia, Enhancing the scalability of multi-objective optimization via restricted boltzmann machine-based estimation of distribution algorithm, Information Sciences 248 (2013) 191–213.
  • [39] S. Ivvan Valdez, A. Hernandez, S. Botello, A boltzmann based estimation of distribution algorithm, Information Sciences 236 (2013) 126–137.
  • [40] L. PourMohammadBagher, M. M. Ebadzadeh, R. Safabakhsh, Graphical model based continuous estimation of distribution algorithm, Applied Soft Computing 58 (2017) 388–400.
  • [41] W. Dong, Y. Wang, M. Zhou, A latent space-based estimation of distribution algorithm for large-scale global optimization, Soft Computing 23 (13) (2019) 4593–4615.
  • [42] A. Zhou, J. Sun, Q. Zhang, An estimation of distribution algorithm with cheap and expensive local search methods, IEEE Transactions on Evolutionary Computation 19 (6) (2015) 807–822.
  • [43] Q. Dang, W. Gao, M. Gong, An efficient mixture sampling model for gaussian estimation of distribution algorithm, Information Sciences 608 (2022) 1157–1182.
  • [44] J. Pẽna, J. Lozano, P. Larrañaga, Globally multimodal problem optimization via an estimation of distribution algorithm based on unsupervised learning of bayesian networks, Evolutionary Computation 13 (1) (2005) 43–66.
  • [45] P. Yang, K. Tang, X. Lu, Improving estimation of distribution algorithm on multimodal problems by detecting promising areas, IEEE Transactions on Cybernetics 45 (8) (2015) 1438–1449.
  • [46] Z. Ren, Y. Liang, L. Wang, A. Zhang, B. Pang, B. Li, Anisotropic adaptive variance scaling for gaussian estimation of distribution algorithm, Knowledge-based Systems 146 (2018) 142–151.
  • [47] Y. Liang, Z. Ren, X. Yao, Z. Feng, A. Chen, W. Guo, Enhancing gaussian estimation of distribution algorithm by exploiting evolution direction with archive, IEEE Transactions on Cybernetics 50 (1) (2020) 140–152.
  • [48] F. Wang, Y. Li, A. Zhou, K. Tang, An estimation of distribution algorithm for mixed-variable newsvendor problems, IEEE Transactions on Evolutionary Computation 24 (3) (2020) 479–493.
  • [49] T. Liu, X. Li, L. Tan, S. Song, An incremental-learning model-based multiobjective estimation of distribution algorithm, Information Sciences 569 (2021) 430–449.
  • [50] K.-H. Han, J.-H. Kim, Quantum-inspired evolutionary algorithm for a class of combinatorial optimization, IEEE Transactions on Evolutionary Computation 6 (2002) 580–593.
  • [51] B. Yu, K. Yuan, B. Zhang, D. Ding, D. Z. Pan, Layout decomposition for triple patterning lithography, 2011 IEEE/ACM International Conference on Computer-Aided Design (ICCAD) (2011) 1–8.
  • [52] S. T. Hedetniemi, D. P. Jacobs, P. K. Srimani, Linear time self-stabilizing colorings, Information Processing Letters 87 (2003) 251–255.
  • [53] W. Greub, Linear Algebra, Springer, 1975.
  • [54] R. Kress, Numerical Analysis, Springer, 1998.
  • [55] P. Galinier, J.-K. Hao, Hybrid evolutionary algorithms for graph coloring, Journal of Combinatorial Optimization 3 (1999) 379–397.