Content uploaded by Alexandre Gondran

Author content

All content in this area was uploaded by Alexandre Gondran on Apr 15, 2014

Content may be subject to copyright.

Variations on Memetic Algorithms

for Graph Coloring Problems

Laurent Moalic∗1and Alexandre Gondran†2

1IRTES-SET, UTBM, University of Technology of Belfort-Montbliard, Belfort, France

2MAIAA, ENAC, French Civil Aviation University, Toulouse, France

Abstract

Graph vertices coloring with a given number of colors is a famous and much-

studied NP-complete problem. The best methods to solve this problem are hybrid

algorithms such as memetic algorithms [15, 26, 38] or quantum annealing [34, 35,

36]. Those hybrid algorithms use a powerful local search inside a population-based

algorithm. The balance between intensiﬁcation and diversiﬁcation is essential for

those metaheuristics but diﬃcult to archieve. Customizing metaheuristics takes

long time and is one of the main weak points of these approaches. This paper

studies the impact of the increase and the decrease of diversiﬁcation in one of

the most eﬀective algorithms known: the Hybrid Evolutionary Algorithm (HEA)

from Galinier and Hao [15]. We then propose a modiﬁcation of this memetic

algorithm in order to work with a population of only two individuals. This new

algorithm more eﬀectively manages the correct ‘dose’ of diversiﬁcation to add

into the included local search - TabuCol [20] in the case of the HEA. It has

produced several good results for the well-known DIMACS benchmark graphs,

such as 47-colorings for DSJC500.5, 82-colorings for DSJC1000.5, 222-colorings

for DSJC1000.9 and 81-colorings for ﬂat1000 76 0, which have so far only been

produced by quantum annealing [36] in 2012 with massive multi-CPUs.

1 Introduction

Given a undirected graph G= (V, E) with Va set vertices and Ea set of edges, the

graph vertex coloring involves assigning each vertex with a color so that two adjacent

vertices (linked by an edge) feature diﬀerent colors. The Graph Vertex Coloring Prob-

lem (GVCP) involves ﬁnding the minimum number of colors required to color a given

graph in respect of these binary constraints. The GVCP is a famous and much-studied

problem because this simple formalization can be applied to various issues such as

∗laurent.moalic@utbm.fr

†alexandre.gondran@enac.fr

1

arXiv:1401.2184v1 [cs.AI] 8 Jan 2014

frequency assignment problems [1, 10], scheduling problems [25, 39, 37] and ﬂy level

allocation problems [3]. Most problems that involve sharing a rare resource (colors) be-

tween diﬀerent operators (vertices) can be modeled as a GVCP, such as most resource

allocation problems. GVCP is NP-hard [18]. Given ka positive integer corresponding

to the maximum number of colors, a k-coloring of a given graph Gis a function cthat

assigns to each vertex a color (i.e. an integer included between 1 and k) as follows :

c:V→ {1,2..., k}

v7→ c(v)

One recalls some deﬁnitions : a k-coloring is called legal or proper k-coloring if it

respects the following binary constraints : ∀(u, v)∈E, c(u)6=c(v). Otherwise the k-

coloring is called non legal or non proper; and edges (u, v)∈Esuch as c(u) = c(v) are

called conﬂicting edges, and uand vconﬂicting vertices. A given graph Gis k-colorable

if a proper k-coloring exists. The chromatic number χ(G) of a given graph Gis the

smallest integer ksuch as Gis k-colorable. A coloring is called complete coloring if one

color is assigned to all vertices, otherwise it is called partial coloring, whereby some

vertices may remain uncolored. An independent set or stable set is a set of vertices, no

two of which are adjacent. In this case, it is possible to assign the same color to all the

vertices of an independent set without producing any conﬂicting edges. The problem of

ﬁnding a minimal graph partition of independent sets is then equivalent to the GVCP.

The k-coloring problem - ﬁnding a proper k-coloring of a given graph G- is NP-

complete [24] for k > 2. Therefore, the best performing exact algorithms are generally

not able to ﬁnd a proper k-coloring in reasonable time when the number of vertices is

greater than 100 [22, 11]. For large graphs, one uses heuristics that partially explore

the search space to occasionally ﬁnd a proper k-coloring in a reasonable time frame.

However, this partial search does not guarantee that a better solution exists. Very

interesting and comprehensive surveys on the GVCP and the most eﬀective heuristics

to solve it can be found in [16, 14]. These studies classify heuristics by the search space

used. In order to deﬁne the search space (or that which is termed strategy), one has to

answer to three questions :

1. Is the number of available colors ﬁxed or not ?

2. Is non proper colorings included in the search space ?

3. Does the heuristic use complete colorings or partial colorings ?

Among the eight theoretical possibilities, four main strategies are deﬁned [14] :

•Proper strategy if the number of colors is not ﬁxed and only complete and proper

colorings are taken into account. The aim is to ﬁnd a coloring that minimizes

the number of colors used under constraints of legality and completeness of the

coloring.

•k-ﬁxed partial proper strategy if the number of colors is ﬁxed and partial and

proper colorings are taken into account. The aim is to ﬁnd a coloring that mini-

mizes the number of uncolored vertices under constraints of the number of given

colors and of proper coloring.

2

•k-ﬁxed penalty strategy if the number of colors is ﬁxed and complete and no proper

colorings are taken into account. The aim is to ﬁnd a coloring that minimizes the

number of conﬂicting edges under constraints of the number of given colors and

of completed coloring.

•Penalty strategy if the number of colors is not ﬁxed and no proper and completed

colorings are taken into account. The aim is to ﬁnd a coloring that minimizes the

number of conﬂicting edges and the number of colors used under the constraint

of completed coloring.

The Variable Space Search of [21] is interesting and didactic because it works with three

of the four above strategies. Another more classical means of classifying the diﬀerent

methods is to consider how these methods explore the search space.

•Constructive or exhaustive methods (such as greedy methods, branch and bound,

backtracking and constraint programming) build a coloring step-by-step from

empty coloring; those approaches usually have a k-ﬁxed partial proper strategy.

DSATUR [5] and RLF [25] are the most well-known greedy algorithms. Those

algorithms rapidly provide an upper bound on χ(G), but it is quite distance from

the exact value of χ(G). They are used to initialize solutions before a local search

or an evolutionary algorithm is employed. Some improved algorithm such as

XRFL [22] based on RLF provides much better results. Exact algorithms such as

branch and bound (B&B), backtracking, or constraint programming [6] are k-ﬁxed

partial proper strategies. The B&B implementation of [22] takes too much time

after 100 vertices. Column generation approaches are also exact methods. [27]

divided the problem into two parts: the ﬁrst problem involves generating useful

columns in the form of independent sets. The second problem involves selecting

a minimum number of independent sets (created by the ﬁrst problem) in order to

cover the graph. To obtain better results, exact methods can be hybridized with

local searches [32, 9].

•Local (or neighborhood or trajectory) searches (such as the hillclimbing, the simu-

lated annealing [22], the tabu search [20, 8], the variable neighborhood search [2],

the variable space search [21] and the min-conﬂict) start from an initial coloring

and try to improve it by local moves; usually those approaches have a proper

or k-ﬁxed penalty strategies. A detailed overview of those local search methods

for graph coloring is provided in [16]. In our algorithm, we employ an improved

version [15] of a tabu search algorithm called TabuCol [20]; it is one of the ﬁrst

metaheuristics developed for graph coloring and uses a k-ﬁxed penalty strategy.

•Population-based approaches (such as the evolutionary algorithm, the ant colony

optimization [29], the particle swarm algorithm and the quantum annealing [34,

35]) work with several colorings that can interact together with, for example, a

crossover operator, a share memory, a repulsion or attraction operator, or others

criteria such as sharing. Those methods currently obtain the best results when

they are combined with the aforementioned local search algorithms. However,

used alone, those population-based approaches are limited.

3

The ﬁrst objective of this paper is to present a new and simple algorithm that pro-

vides the best results for coloring DIMACS benchmark graphs. The second objective

is to show why this algorithm eﬀectively controls diversiﬁcation. Indeed, in our work,

the population of this memetic algorithm is reduced to only two individuals, called a

couple-population. It provides the opportunity to focus on the correct ‘dose’ of diversity

to add to a local search. This new approach is simpler that a general evolutionary algo-

rithm because it does not use numerous parameters such as selection, size of population,

crossover and mutation rates.

The organization of this paper is as follows. The issue of diversiﬁcation in heuristics

is presented in Section 2. Section 3 describes our improvement of the memetic algorithm

for the GVCP. The experimental results are presented in Section 4 and some of the

impacts of diversiﬁcation are analyzed in Section 5. Finally, we consider the conclusions

of this study and discuss possible future research in Section 6.

2 How to manage diversity ?

Because the search is not exhaustive, heuristics provide only sub-optimal solutions with

no information about the distance between those sub-solutions and the optimal solution.

Heuristics return the best solution found after partial exploration of the search space. To

produce a good solution, heuristics must alternate exploitation phases and exploration

phases. During an exploitation phase (or an intensiﬁcation phase), a solution produced

by the heuristic is improved. It resembles an exhaustive search but occurs in a small

part of the search space. An exploration phase (or diversiﬁcation phase) involves a

search into numerous uncharted regions of the search space. The balance between

those two phases is diﬃcult to achieve. Metaheuristics therefore deﬁne a framework

that manages this balance. In this section, we classify the main components of several

metaheuristics as intensiﬁcation operators or as diversiﬁcation operators. Of course,

this list, presented in table 1, is not exhaustive and we take into account only well-

known metaheuristics : Local Search (LS), Tabu Search (TS), Simulated Annealing

(SA), Variable Neighborhood Search (VNS), Evolutionary Algorithm (EA), and Ant

Colony (AC). Some components can be shared between several metaheuristics. Other

components can be at the same time an intensiﬁcation operator and a diversiﬁcation

operator, such as parent selection or population update. It depends on the context of

the algorithm, as we will now demonstrate.

Local search algorithms start from an initial solution and then try to improve this

iteratively through local transformation. The simplest LS, hill-climbing, accepts a move

only if the objective function is improved: it is inherently an intensiﬁcation operator.

It is possible to introduce some diversity by generating several diﬀerent initial solu-

tions. This process is called multi-starting. The limit of a simple LS is that after a

given number of iterations, the algorithm is blocked within a local optimum. No local

transformation can improve the solution. SA and TS therefore accept some worsening

moves during the search, with a given criteria for SA and with a tabu list for TS. On

the other hand, VNS, to escape local optimum, changes the neighborhood structure.

The change in neighborhood structure is an eﬀective diversiﬁcation operator because it

4

never worsens the current solution. However, this diversiﬁcation operator is too weak:

VNS must add a shaking process, similar to partial restart (as in Iterated Local Search),

in order to achieve a more global optimization.

The population-based algorithms are often classiﬁed as global optimization algo-

rithms because they work with several candidate solutions, although this does not indi-

cate that the algorithm will ﬁnd the global optimum. Moreover, working with several

candidate solutions can be regarded as a diversiﬁcation operator. EAs are population-

based algorithms. A basic EA can be presented in ﬁve steps as follows: 1) Parents

selection: one selects two individuals of the population, which are called parents. 2)

Crossover: according to a given rate, a crossover operator is applied to those two par-

ents, which creates two new individuals, called children, blending of the two parents.

3) Mutation: according to a given rate, a mutation operator is applied to each children,

which modiﬁes them slightly. 4) Population update: under given conditions, the two

children take the place of two individuals of the population. 5) This cycle of four steps

is called a generation; it is repeated until a given stop condition is realized. The best

individual of the population is the output of the EA. We shall classify the diﬀerent

components of this basic EA as intensiﬁcation operators or diversiﬁcation operators.

Mutation and crossover operators are diﬀerent in nature to selection and population

update processes. Indeed, mutation and crossover operators are chance generators,

while selection and population update processes are higher-level systems that control

the chance interest. These two means of functioning are pithily summarized in an

expression attributed to Democritus: “Everything existing in the universe is the fruit

of chance and necessity”, which is also considered in Jacques Monod’s book Chance and

Necessity [28]. The mutation and the crossover processes are therefore diversiﬁcation

operators in essence. The mutation operator slightly changes one of the individuals

of the population. The crossover operator mixes two individuals of the population in

order to create a new one. The mutation is a unary operator while the crossover is a

binary operator. There exists also trinary operators for some other EAs, such as for

Diﬀerential Evolution. The unary or binary changes (applied by mutation or crossover)

are performed randomly and these are therefore exploration phases. Occasionally, a

random modiﬁcation improves the solution, but this function is driven by the higher-

level system. These changes cannot be considered intensiﬁcation operators, except in

some speciﬁc cases where the modiﬁcations (mutation or crossover) are not performed

randomly. For example, in [12], the crossover is a quasi-deterministic process directly

guided by the separability of the objective function in order to improve the latter.

The aim of a diversiﬁcation operator is to provide diversity to current solution(s), but

there is a risk of providing too much diversity, and thus breaking the structure of current

solution(s). The crossover operator generally provides more diversity than the mutation

operator.

The classiﬁcation of operators such as the parents selection and the population

update depends on the choice of the EA parameters. Indeed, if one chooses an elitist

parents selection policy (the choice of best individuals of the population for the crossover

and the mutation), the selection is then an intensiﬁcation operator. Conversely, a policy

of random parents selection (a random selection of individuals from the population

for the crossover and the mutation) indicates a diversiﬁcation operator. This double

5

Intensiﬁcation/Exploitation Diversiﬁcation/Exploration

LS: accept improving moves LS: multi-starts

SA, TS: accept worsening moves

VNS: change of neighborhood structure VNS: change of neighborhood structure

shaking

p-bA: population (several candidate solutions)

EA: parents selection (the best) EA: parents selection (random)

population update (if child is better) population update (systematic)

mutation, crossover

sharing, elitism

AC: collective memory (pheromone) AC: random choice

Table 1: Classiﬁcation of main components of some well-known metaheuristics as inten-

siﬁcation operators or as diversiﬁcation operators. LS: Local Search, TS: Tabu Search,

SA: Simulated Annealing, VNS: Variable Neighborhood Search, EA: Evolutionary Al-

gorithm, AC: Ant Colony, p-bA: population based Algorithm.

role of the parents selection can be very interesting but also very diﬃcult to properly

control. The population update has the same feature. If one chooses to include in

the population the individuals created by the crossover and/or the mutation (children)

only if they are better than the individuals of the population that one removes, then

the population update is an intensiﬁcation operator. Conversely, if one chooses to

systematically replace some individuals of the population by the created children (even

if they are worse), then the population update is a diversiﬁcation operator. The sharing

and the elitism are two others mechanisms of EAs playing diversiﬁcation operators roles.

In AC, the interactions between individuals of the population occur through the

sharing of a collective memory. The phenomenon mechanism (deposit and evaporation)

plays this role, while the random choice of the instantiation of variables plays the role

of diversiﬁcation.

Table 1 summarizes the classiﬁcation of main components of several metaheuristics

as diversiﬁcation operators or as intensiﬁcation operators. The separation is not always

very clear, such as in the case of EAs, in which parents selection or update population

processes play both roles. Controlling the correct balance between intensiﬁcation and

diversiﬁcation is a diﬃcult step; Much ﬁne tuning is required to obtain good results

from heuristics.

An interesting feature of a diversiﬁcation operator is its ability to explore new areas

of the search space without breaking the structure of current solution(s). We recall a

quotation of the physiologist Claude Bernard about hormones [4] and that corresponds

well with the role of a diversiﬁcation operator :

“Everything is poisonous, nothing is poisonous, it is all a matter of dose.”

Claude Bernard - 1872

The level or dose of diversity provided by a diversiﬁcation operator is diﬃcult to eval-

uate. However, it is possible to compare the diversiﬁcation operators with their dose of

diversity. For example, the mutation operator generally provides less diversity than the

crossover operator. For each operator, there are many parameters that aﬀect the dose

6

Figure 1: Fitness of the child solution in function of the hamming distance separating

the two parents having the same ﬁtness. The ﬁtness is the number of conﬂicting edges

in a completed 48-coloring. The ﬁtness of the parents is equal to f= 40. The hamming

distance separating two k-colorings is the minimum number of moves (changing of one

vertex of color class) to pass from one k-coloring to the other (it is independent of the

colors symmetry).

of diversity; One of the easily identiﬁable parameters for the crossover operator is the

Hamming distance between the two parents. Indeed, if we highlight two good solutions

that have very good objective functions but that are very diﬀerent (in terms of Ham-

ming distance), then they would have a high probability of producing children with bad

objective functions following crossover. The danger of the crossover is of completely

destroying the solution structure. On the other hand, two solutions very closed (in

terms of Hamming distance) produce a child with quasi-similar ﬁtness. Chart 1 shows

the correlation between the Hamming distance separating the two parents of the same

ﬁtness (abscissa axis) and the ﬁtness of the child (ordinate axis). This chart is obtained

for k-coloring problem where the objective is to minimize the number of conﬂicting

edges (the ﬁtness) in a complete but non legal k-coloring with k= 48. We consider the

DSJC500.5 graph from the DIMACS benchmark [23] and the GPX crossover of [15].

The parents used for the crossover have the same ﬁtness value, equal to 40 conﬂicting

edges. There is a quasi-linear correlation between these two parameters.

The key question is : how does one manage the correct ‘dose’ of diversiﬁcation in a

heuristic? In the next section, we present a memetic algorithm with only 2 individuals,

which simpliﬁes the management of the diversity dose.

7

3 Memetic Algorithms with only two individuals

Memetic Algorithms [19] (MA) are hybrid metaheuristics using a local search algorithm

inside a population-based algorithm. They can also be viewed as speciﬁc EAs where

all individuals of the population are local minimum (of a speciﬁc neighborhood). This

hybrid metaheuristic is a low-level relay hybrid (LRH) in the Talbi taxonomy [33].

A low-level hybridization means that a section of one metaheuristic is replaced by a

second metaheuristic. In MA, the mutation of the EA is replaced by a local search

algorithm. Conversely, a high-level hybridization refers to a combination of several

metaheuristics, none of which are modiﬁed. A relay (respectively teamwork ) hybrid

means that metaheuristics are used successively (respectively in parallel).

In graph coloring, the Hybrid Evolutionary Algorithm (HEA) of Galinier and Hao [15]

is a MA; the mutation of the EA is replaced by a tabu search. HEA is one of the best

algorithms for solving the GVCP; From 1999 until 2012, it provided the majority of

the best results for DIMACS benchmark graphs [23], particularly for diﬃcult graphs

such as DSJC500.5 and DSJC1000.5 (see table 2). These results were obtained with a

population of 8 individuals.

The tabu search used is an improvement of the TabuCol of [20]. This version of

TabuCol is a very powerful tabu search, which even obtained very good results for the

GVCP when it is used alone (see table 2). Another beneﬁt of this version of TabuCol is

that it has only two parameters to adjust in order to control the tabu tenure. Moreover,

[15] have demonstrated on a very large number of instances that with the same setting,

TabuCol obtained very good k-colorings. Indeed, one of the main disadvantages of

heuristics is that the number of parameters to set is high and diﬃcult to adjust. This

version of TabuCol is very robust. Thus we retain the setting of [15] in all our tests

and consider TabuCol as an atomic algorithm.

The mutation, which is a diversiﬁcation operator, is replaced by a local search,

an intensiﬁcation operator. The balance between intensiﬁcation and diversiﬁcation is

restored because in the case of the graph coloring problem, all crossovers are too strong

and destroy too much of a solution’s structure.

The crossover used in HEA is called the Greedy Partition Crossover (GPX); it is

based on color classes. Its aim is to add slightly more diversiﬁcation to the TabuCol

algorithm.

These hybridizations combine the beneﬁts of population-based methods, which are

better at diversiﬁcation by means of a crossover operator, and local search methods,

which are better at intensiﬁcation.

The intensiﬁcation/diversiﬁcation balance is diﬃcult to achieve. In order to simplify

the numerous parameters involved in EAs, we have chosen to consider a population with

only two individuals. This implies 1) no selection process, and 2) a simpler management

of the diversiﬁcation because there is only one diversiﬁcation operator: the crossover.

8

3.1 General Pattern - H2O: Hybrid approach with 2 trajectories-

based Optimization

The basic blocks of HEA are the TabuCol algorithm, which is a very powerful local

search for intensiﬁcation, and the Greedy Partion Crossover (GPX), which adds a little

more diversiﬁcation.

The idea is to combine as simply as possible these two blocks. The local search is a

unary operator while the crossover is a binary operator. We present with algorithm 1

the pseudo code of a ﬁrst version of this simple algorithm, which can be seen as a

2 trajectories-based algorithm. We then call the algorithm ‘H2O’, which signiﬁes a

Hybrid approach with 2 trajectories-based Optimization.

Algorithm 1 First version of H2O: Hybrid approach with 2 trajectories-based Opti-

mization

1: Input: an asymmetric crossover, a local search.

2: Output: the best solution found best

3: p1, p2, best ←init() {initialize with random solutions}

4: generation ←0

5: repeat

6: c0

1←crossover(p1,p2)

7: c0

2←crossover(p2,p1)

8: c1←localSearch(c0

1)

9: c2←localSearch(c0

2)

10: best ←saveBest(c1, c2, best)

11: p1, p2←replace(c1, c2)

12: generation + +

13: until stop condition()

14: return best

The algorithm 1 needs an asymmetric crossover, which means: crossover(p1,p2)6=

crossover(p2,p1). After randomly initializing the two solutions, the algorithm repeats

an instructions loop until a stop criteria occurres. First, we introduce some diversity

with the crossover operator, then the two oﬀspring c0

1and c0

2are improved by means of

the local search. Next, we register the best solution and we systematically replace the

parents by the two children. An iteration of this algorithm is called a generation.

In order to add more diversity on the algorithm 1, we present a second version of

H2Owith two extra elite solutions.

9

Algorithm 2 Second version of H2O: Hybrid approach with 2 trajectories-based

Optimization

1: Input: an asymmetric crossover, a local search, Itercycle .

2: Output: the best solution found

3: p1, p2, elite1, elite2, best ←init() {initialize with random solutions}

4: generation ←0

5: repeat

6: c0

1←crossover(p1,p2)

7: c0

2←crossover(p2,p1)

8: c1←localSearch(c0

1)

9: c2←localSearch(c0

2)

10: elite1←saveBest(c1, c2, elite1)

11: p1, p2←replace(c1, c2)

12: best ←saveBest(elite1, best)

13: if generation%Itercy cle = 0 then

14: p1←elite2

15: elite2←elite1

16: elite1←init()

17: end if

18: generation + +

19: until stop condition()

20: return best

We add two other candidate solutions (similar to elite solutions), elite1and elite2,

in order to reintroduce some diversity in the couple-population. Indeed, after a given

number of generations, the two individuals of the population become increasingly similar

within the search space. To maintain a given diversity in the couple-population, the

elite2elite solution replaces one of the population individual after I tercycle generations

i.e. one cycle. elite1is the best solution found during the current cycle and elite2the

best solution found during the previous cycle. The ﬁgure 2 represents the graphic view

of the algorithm 2.

3.2 Application to graph coloring : H2col

H2col is the application of H2Oto the k-coloring problem with the GPX as crossover

and the TabuCol as local search. It uses only one parameter: IterT C , the number of

iterations performed by the TabuCol algorithm.

10

Crossover'

Local'search'

P1# P2#

cycle'itera/ons'

best1#

?#

?#

best2#

Figure 2: Diagram of H2O.

Algorithm 3 H2col : Hybrid approach with 2 trajectories-based Optimization for

graph coloring with GPX and TabuCol

1: Input: the asymmetric crossover: GPX, the local search: TabuCol (with IterT C

iterations), Itercycle = 10.

2: Output: the best conﬁguration found

3: p1, p2, elite1, elite2←init() {initialize with random colorings}

4: generation ←0

5: repeat

6: c0

1←GPX(p1,p2)

7: c0

2←GPX(p2,p1)

8: c1←TabuCol(c0

1)

9: c2←TabuCol(c0

2)

10: elite1←saveBest(c1,c2,elite1)

11: p1←c1

12: p2←c2

13: best ←saveBest(elite1, best)

14: if generation%Itercy cle = 0 then

15: p1←elite2

16: elite2←elite1

17: elite1←init()

18: end if

19: generation + +

20: until nbConf licts > 0

21: return best

11

This algorithm is an improvement of the TabuCol algorithm; these are two parallel

TabuCol algorithms that interact periodically by crossover. We will now brieﬂy recall

the principles of the TabuCol algorithm and the GPX crossover.

3.2.1 TabuCol

In 1987, [20] presented the TabuCol algorithm, one year after Fred Glover introduced

the tabu search. This algorithm, which solves k-coloring problems, was enhanced in

1999 by [15]. The three basic features of this trajectory-based algorithm are as follows:

•Search Space and Objective Function: the algorithm is a k-ﬁxed penalty strategy.

The objective function minimizes the number of conﬂicting edges.

•Neighborhood: ak-coloring solution is a neighbor of an other k-coloring solution

if the color of only one conﬂicting vertex is diﬀerent. This move is called a critic

1-move. Therefore the neighborhood size depends on the number of conﬂicting

vertices.

•Move Strategy: the move strategy is the standard tabu search strategy. Even if

the objective function is worse, at each iteration, one decides to move to the best

neighbors, which are not inside the tabu list. Note that all the neighborhood is

explored. If there are several best moves, one chooses one of them at random.

This is the only random aspect of this metaheuristic. The tabu list is not the

list of each already-visited solution because this is computationally expensive. It

is more eﬃcient to put only the reverse moves inside the tabu list. Indeed, the

aim is to forbid returning to previous solutions, and it is possible to reach this

goal by forbidding the reverse moves during a given number of iterations (i.e.

the tabu tenure). The tabu tenure is dynamic: it depends on the neighborhood

size. A basic aspiration criteria is also implemented: it accepts a tabu move

to a k-coloring, which has a better objective function than the best k-coloring

encountered so far.

Data structures have a major impact on algorithm eﬃciency, constituting one of the

main diﬀerences between the Hertz and de Werra version of TabuCol [20] and the

Galinier and Hao version [15]. Checking that a 1-move is tabu or not and updating the

tabu list are operations in constant time (ﬁgure 3b). TabuCol also uses an incremental

evaluation [13]: the objective function of the neighbors is not computed from scratch,

but only the diﬀerence between the two solutions is computed. This is a very important

feature for local search eﬃciency. Finding the best 1-move corresponds to ﬁnding the

maximum value of a matrix (cf. ﬁgure 3a).

3.2.2 Greedy Partition Crossover (GPX)

The second block of H2col is the Greedy Partition Crossover (GPX) from the Hybrid

Evolutionary Algorithm HEA [15]. The two main principles of GPX are: 1) a coloring

is a partition of vertices and not an assignment of colors to vertices, and 2) large color

classes should be transmitted to the child. The ﬁgure 4 gives an example of GPX for

12

(a) Matrix of incremental evaluation. It gives

the beneﬁt of the all 1-moves; for example,

changing the color of the C vertex in green adds

2 conﬂicts; the cross, for example in (A,blue),

means that the A vertex is colored in blue in

the current solution; therefore it is not a real

1-move.

(b) Matrix representing the tabu list and

the tabu tenure. If the current iteration

is the 21st iteration, therefore all the 1-

moves of a value greater than 21 means

that this 1-move is tabu; for example,

(B,blue)-move is tabu because 34 >21.

Figure 3: Data structures for TabuCol algorithm.

a problem with three colors (red, blue and green) and 10 vertices (A, B, C, D, E, F,

G, H, I and J). The ﬁrst step is to transmit to the child the largest color class of the

ﬁrst parent. After having withdrawn those vertices in the second parent, one proceeds

to step 2 where one transmits to the child the largest color class of the second parent.

This process is repeated until all the colors are used. There are most probably still

some uncolored vertices in the child solution. The ﬁnal step is to randomly add those

vertices to the color classes.

4 Experimental Results

In this section we present the results obtained with the two versions of the proposed

memetic algorithm; the ﬁrst version of H2col without elites solutions is indicated as

H0

2col, and the second version with the two extra elites solutions is simply indicated

H2col. Test instances are selected among the most studied graphs since the 1990s, which

are known to be very diﬃcult (DIMACS [23]). To validate the proposed approach, the

results of H2col are compared with the results obtained by some of the best methods

currently known.

4.1 Instances and Benchmarks

We study some graphs from the second DIMACS challenge of 1992-1993 [23]. This is

to date the most widely-used benchmark for solving the graph coloring problem. These

instances are available at the following address: ftp://dimacs.rutgers.edu

We focus on two main types of graphs from the DIMACS benchmark: DSJC and

FLAT, which are randomly or quasi-randomly generated. DSJCn.d graphs are graphs

with nvertices, wich each vertex connected to an average of n×dvertices; dis the

13

Figure 4: An example of GPX crossover for a graph of 10 vertices (A, B, C, D, E, F,

G, H, I and J) and three colors (red, blue and green). This example comes from [15].

graph density. The chromatic number of these graphs is unknown. FLAT graphs have

another structure: they are built for a known chromatic number. The ﬂatn χ graph

has nvertices and χis the chromatic number.

4.2 Computational Results

H2col and H0

2col were programmed in C++ standard. The results presented in this

section were obtained on a computer with an Intel Xeon 3.10GHz processor - 4 cores

and 8GB of RAM. Note that the RAM size has no impact on the calculations: even

for large graphs such as DSJC1000.9 (with 1000 vertices and high density of 0.9), the

memory used does not exceed 125 MB. The main characteristic is the processor speed.

As shown in Section 3, the proposed algorithms have two successive calls to local

search (lines 8 and 9 of the algorithms 1, 2 and 3), one for each child of the current gen-

eration. Almost all of the time is spent performing the local search. It is possible and

moreover easy to parallelize both local searches when using multi-core processor archi-

tecture. This is what we have done using the OpenMP API (Open Multi-Processing),

which has the advantage of being cross-platform (Linux, Windows, MacOS, etc.) and

simple to use. The execution times provided in the following table are in CPU time.

Thus, when we give an execution time of 30 minutes, the required time is actually close

to 15 minutes using two processing cores.

Table 2 presents results of the principal methods known to date. For each graph, it

indicates the lowest number of colors found by each algorithm. For TabuCol [20] the

reported results are from [21] which are better than those of 1987. The most recent al-

gorithm, QA-col (Quantum Annealing for graph coloring [36]), provides the best results

14

Graphs H2COL

LS Hybrid algorithm

2013 1987/2008 1999 2008 2010 2011 nov. 2012

H2col TabuCol [20, 21] HEA [15] AmaCol [17] MACOL[26] EXTRACOL [38] QA-col [36]

DSJC500.1 12 13 - 12 12 - -

DSJC500.5 47 50 48 48 48 - 47

DSJC500.9 126 127 - 126 126 -126

DSJC1000.1 20 -20 20 20 20 20

DSJC1000.5 82 89 83 84 83 83 82

DSJC1000.9 222 227 224 224 222 222 222

ﬂat1000 50 0 50 50 -50 50 50 -

ﬂat1000 60 0 60 60 -60 60 60 -

ﬂat1000 76 0 81 88 83 84 82 82 81

Table 2: Best coloring found

but is based on a cluster of PC using 10 processing cores simultaneously. Note that

HEA [15], AmaCol [17], MACOL [26] and EXTRACOL [38] are also population-based

algorithms also using TabuCol and GPX crossover or an improvement of GPX (GPX

with n>2 parents for MACOL and EXTRACOL and the GPX process is replaced in

AmaCol by a selection of kcolor classes among a very large pool of color classes). Only

QA-col has another approach based on several parallel simulated annealing algorithms

interacting together with sharing criteria.

In a standard Simulating Annealing algorithm (SA), the probability of accepting

a candidate solution is managed through a temperature criteria. The value of the

temperature decreases during the SA iterations. A Quantum Annealing (QA) is a

memetic algorithm without crossover and in which the local search is a SA. The only

interaction between the individuals of the population occurs through a speciﬁc sharing

process. A standard sharing process involves penalizing solutions that are too ‘close’

within the search space (the simplest sharing is to forbid having two similar solutions

in the population). However, comparing a solution with all solutions of the population

can be computationally expensive, especially for a large population. This is why QA-col

compares a solution with only two others of the population: this comparison provides

the solution-population distance. The value of this measure is integrated into the

temperature value of each SA. If the solution-population distance is greater, then the

temperature will be higher, and there will be a higher probability that the solution will

be accepted. [30, 31] have also developed a speciﬁc population spacing management

that resembles this one. However, there are diﬀerent ways to calculate the distance

between two solutions. For a set of solutions, [7] deﬁne frozen same pairs (respectively

frozen diﬀerent pairs) as pairs of vertices that are in the same color class (respectively

in the diﬀerent color class) for all solutions. In QA-col, authors use these deﬁnitions

with a set of two solutions in order to deﬁne a distance between these two solutions

as the diﬀerence between the number of frozen same pairs and the number of frozen

diﬀerent pairs.

Table 3 presents the results obtained with H0

2col, the ﬁrst version of H2col (without

elites). This simplest version ﬁnds very good solutions (+1 color compared to the

best known results) for diﬃcult graphs within the literature. Only one method, QA-

col, occasionally ﬁnds a solution with less color. The column IterTC indicates the

number of iterations of the TabuCol algorithm (this is the stop criteria of TabuCol).

The column Success evaluates the robustness of this method, providing the success

rate: success runs/total runs. A success run is that which ﬁnds a legal k-coloring. The

15

Instances k IterTC Success Iter Gene Time

DSJC500.1 12 8000 15/20 2.5×106158 0.2 min

DSJC500.5 48 8000 9/20 5 ×106356 0.5 min

DSJC500.9 126 25000 10/20 15 ×106317 2 min

DSJC1000.1 20 7000 6/20 6 ×106472 0.9 min

DSJC1000.5 83 40000 16/20 137 ×1061723 36 min

DSJC1000.9 223 30000 4/20 56 ×106939 12 min

222 60000 1/20 516 ×1064304 131 min

ﬂat1000 50 0 50 130000 20/20 1.1×1065 0.5 min

ﬂat1000 60 0 60 130000 20/20 2.4×1069 1 min

ﬂat1000 76 0 82 40000 19/20 152 ×1061905 38 min

81 30000 1/20 380 ×1066333 118 min

Table 3: Results of H0

2col, the ﬁrst version of H2col algorithm (without elites)

average number of generations or crossovers performed during one success run is given

by Gene value. The total average number of iterations of TabuCol preformed during

H0

2Ois Iter =IterTC ×Gene ×2. The column Time indicates the average CPU time

in minutes of success runs.

H0

2col does not ﬁnd the solutions each time for these graphs, but when it does, it is

generally very rapid. For example, for the graph coloring DSJC1000.1 with 20 colors,

recent results reported in the literature are:

•MACOL in 108 minutes (CPU 3.4GHz) [26]

•EXTRACOL in 93 minutes (CPU 2.8GHz) [38]

Our algorithm, H0

2col, achieves solutions in less than one minute (CPU 3.1GHz).

The main drawback of H0

2col is that it converges sometimes too quickly. In such

instances it cannot ﬁnd a solution before the two individuals in a generation become

identical. The second version, H2Col, adds more diversity while performing an intensi-

fying role.

Table 4 shows the results obtained with H2Col. Of primary important is that

H2Col ﬁnds solutions with fewer colors than all the best-known methods. Only the

Quantum Annealing algorithm, using ten CPU cores simultaneously, achieves this level

of performance. In particular, DSJC500.5 is solved with only 47 colors and ﬂat1000 76 0

with 81 colors. We noted 1* the number of success runs when the solution is occasionally

found, but on average this occurs in less than one in 20 runs. This is the case for graphs

DSJC500.5 and ﬂat1000 76 0 with 47 colors and 81 respectively.

The computation time of H2Col is generally close to that of H0

2Col but the former

algorithm is more robust with quasi-100% of success. In particular, the two graphs

DSJC500.5 and DSJC1000.1 with 48 and 20 colors respectively are resolved each time,

and in less than one CPU minute on average. Using a multicore CPU, these instances

are solved in less than 30 seconds on average, often in less than 10 seconds. As a

comparison, the shortest time reported in the literature for DSJC1000.1 is 93 minutes

16

Instances k IterTC Success Iter Gene Time

DSJC500.1 12 4000 20/20 3.8×106483 0.3 min

DSJC500.5 48 8000 20/20 7 ×106494 0.9 min

47 12000 1* 20 ×106850 3 min

DSJC500.9 126 15000 13/20 29 ×106970 4 min

DSJC1000.1 20 3000 20/20 4 ×106346 0.7 min

DSJC1000.5 83 40000 20/20 96 ×1061200 16 min

82 40000 2/20 548 ×1066854 96 min

DSJC1000.9 223 30000 19/20 126 ×1062107 32 min

222 50000 1/20 1.4×10914208 354 min

ﬂat1000 50 0 50 130000 20/20 1.1×1065 0.5 min

ﬂat1000 60 0 60 130000 20/20 2.2×1069 1 min

ﬂat1000 76 0 82 40000 20/20 84 ×1061052 16 min

81 40000 2/20 716 ×1068961 116 min

Table 4: Results of the second version of H2col algorithm (with elites) with the indica-

tion of CPU time

for EXTRACOL with a 2.8GHz processor (and 108 minutes for MACOL with a 3.4GHz

processor).

5 Analysis of diversiﬁcation

In this section we perform several tests in order to analyze the role of diversiﬁcation

in the H2col algorithm. We increase or decrease the dose of diversiﬁcation within the

H2col algorithm. There are only two operators that lead to diversiﬁcation: the GPX

crossover and the population update process. In a ﬁrst set of tests, we slightly modify

the dose of diversiﬁcation in the GPX crossover and analyze the results. In a second

set of tests, we focus on the population update process: in H2col, the two produced

children systematically replace both parents, even if they have worse ﬁtness values than

their parents. If we lighten this rule, the diversiﬁcation decreases.

5.1 Dose of diversiﬁcation in the GPX crossover

Some modiﬁcations are performed on the GPX crossover in order to increase (as for

the ﬁrst test) or decrease (as for the second test) the dose of diversiﬁcation within this

operator.

5.1.1 Test on GPX with increased chance: random draw of a color classes

number

In order to increase the levels of chance within the GPX crossover, one randomizes

the GPX. One recalls (cf. section 3.2.2) that at each step of the GPX, the selected

parent transmits the largest color class to the child. In this test, we begin by randomly

17

transmitting xcolor classes chosen from the parents to the child; after those xsteps,

one starts again by alternately transmitting the largest color class from each parent. x

is the random level. If x= 0, then the crossover is the same as the initial GPX. If x

increases, then the chance and the diversity also increase. To evaluate this modiﬁcation

of the crossover, we count the cumulative iterations number of TabuCol that one H2col

run requires in order to ﬁnd a legal k-coloring. For each xvalue, the algorithm runs

ten times in order to produce more robust results. For the test, we consider the 48-

coloring problem for graph DSJC500.5 of the DIMACS benchmark. The ﬁgure 5 shows

in abscissa the random level xand in ordinate the average necessary iterations number

required to ﬁnd a legal 48-coloring.

Figure 5: Average necessary iterations number to ﬁnd a legal 48-coloring for DSJC500.5

graph in function of the randomness level; abscissa: x, the randomness level; ordinate:

average iterations number

First, 0 ≤x≤k, with kthe number of colors, but we stop the computation for

x≥15, because with x= 15, the algorithm does not ﬁnd a 48-coloring within acceptable

computing time limit. This means that when we introduce too much diversiﬁcation,

the algorithm cannot ﬁnd a legal solution. Indeed, for xhigh value, the crossover does

not transmit the good features of the parents, so that the child appears to be a random

initial solution. For 0 ≤x≤8, the algorithm ﬁnds a legal coloring in more or less 10

million iterations. It is not easy to decide which x-value obtains the quickest result.

5.1.2 Test on GPX with decreased chance: imbalanced crossover and im-

portant diﬀerent between the two parents

In the standard GPX, the role of each parent is balanced: they alternatively transmit

their largest color class to the child. Of course, the parent, which ﬁrst transmits its

largest class, has more importance than the other; this is why it is an asymmetric

18

crossover. In this test, we give a higher importance to one of the parents. At each

step, we randomly draw the parent that transmits its largest color class with a diﬀerent

probability for each parent. We introduce x, the probability of selecting the ﬁrst parent;

1−xis the probability of selecting the second parent. For example, if x= 0.75, parent

1 always has a 3 in 4 chance of being selected to transmit its largest color class (parent

2 only has a 1 in 4 chance). If x= 0.5, it means that both parents have an equal

probability (a ﬁfty-ﬁfty chance) to be chosen; this almost corresponds to the standard

GPX. If x= 1, it means that the child is a copy of parent 1; there are no more crossovers

and therefore H2col is a TabuCol with two initial solutions. When xget away from

0.5, the chance and the diversity bring by the crossover decrease. Figure 6 shows in

abscissa the probability xand in ordinate the average number of necessary iterations

required to ﬁnd a legal 48-coloring (as in the previous test).

Figure 6: Average number of necessary iterations required to ﬁnd a legal 48-coloring

for DSJC500.5 graph in function of the imbalanced crossover; abscissa: x, probability

to select the ﬁrst parent; ordinate: average iterations number

First, it is evident that the results are symmetric according to x. The best results

are obtained with approximately x= 0.5 (0.4≤x≤0.6). The impact of this parameter

is weaker than that of the previous one: the control of the reduction of diversiﬁcation

is ﬁner.

5.2 Test on parents’ replacement: systematic or not

In H2col, the two produced children systematically replace both parents, even if they

have worse ﬁtness values than their parents. We modify this replacement rule in this

test. If the ﬁtness value of the child is better than that of its parents, the child automati-

cally replaces one of the parents. Otherwise, we introduce a probability xcorresponding

to the probability of the parents’ replacement, even if the child is worse than his par-

ents. If x= 1, the replacement is systematic as in standard H2col and if x= 0, the

19

replacement is performed only if the children are better. When the x-value decreases,

the diversity also decreases. Figure 7 shows in abscissa the parents’ replacement prob-

ability xand in ordinate the average number of necessary iterations required to ﬁnd a

legal 48-coloring (as in the previous test).

Figure 7: Average number of necessary iterations required to ﬁnd a legal 48-coloring

for DSJC500.5 graph in function of the parents’ replacement policy; abscissa: parents’

replacement probability; ordinate: average number of iterations

If the parents’ replacement probability x= 0 or a very low 0.1, then more time is

required to produce the results. The absence or the lack of diversiﬁcation is shown

to penalize the search. However, for a large range of values: 0.3≤x≤1, it is not

possible to deﬁne the best policy for xcriteria. The dramatic change in behavior of

H2col happens very quickly around 0.2.

These studies enable us to better understand the role of the diversiﬁcation operators

(crossover and parent updating). Some interesting criteria are identiﬁed as the random

level of the crossover or the imbalanced level of the crossover. We will integrate these

criteria that we have studied in this paper into future algorithms in order to dynamically

manage the diversity.

6 Conclusion

We have proposed a new algorithm for the graph coloring problem, called H2col. It is

a variation of a memetic algorithm with only two candidate solutions. This simpliﬁ-

cation has the great advantage of clarifying the role of the diversiﬁcation and intensi-

ﬁcation operators and of more eﬀectively managing the right ‘dose’ of diversiﬁcation.

H2col combines a local search algorithm (TabuCol) as an intensiﬁcation operator with

a crossover operator (GPX) as a diversiﬁcation operator, two main building blocks of

Galinier and Hao’s memetic algorithm [15]. The computational experiments, carried

20

out on a set of challenging DIMACS graphs, show that H2col ﬁnds the best existing

results, such as 47-colorings for DSJC500.5, 82-colorings for DSJC1000.5, 222-colorings

for DSJC1000.9 and 81-colorings for ﬂat1000 76 0, which have so far only been found

by quantum annealing [36] with a massive multi-CPU.

We have performed an in-depth analysis on the crossover operator in order to bet-

ter understand its role in the diversiﬁcation process. Some interesting criteria have

been identiﬁed, such as the crossover’s levels of randomness and imbalance. Those

criteria pave the way for our further research. We have generalized this optimization

methodology, which is a speciﬁc memetic algorithm with only two individuals in its

population. This Hybrid approach with 2 trajectories-based Optimization (H2O) im-

proves the local search algorithm through a crossover operator. The crossover inserts a

dose of diversiﬁcation that is then easy to manage.

References

[1] KarenI. Aardal, StanP.M. Hoesel, ArieM.C.A. Koster, Carlo Mannino, and Anto-

nio Sassano. Models and solution techniques for frequency assignment problems.

Quarterly Journal of the Belgian, French and Italian Operations Research Soci-

eties, 1(4):261–317, 2003.

[2] C. Avanthay, Alain Hertz, and Nicolas Zuﬀerey. A variable neighborhood search

for graph coloring. European Journal of Operational Research, 2003.

[3] Nicolas Barnier and Pascal Brisset. Graph Coloring for Air Traﬃc Flow Manage-

ment. Annals of Operations Research, 130(1-4):163–178, 2004.

[4] Claude Bernard. Le¸cons de pathologie exp´erimentale. J.-B. Bailli`ere et ﬁls, Paris,

1872.

[5] D. Br´elaz. New Methods to Color the Vertices of a Graph. Communications of the

ACM, 22(4):251–256, 1979.

[6] Massimiliano Caramia and Paolo Dell’Olmo. Constraint Propagation in Graph

Coloring. Journal of Heuristics, 8(1):83–107, 2002.

[7] Joseph Culberson and Ian P. Gent. Frozen Development in Graph Coloring. The-

oretical Computer Science, 265(1-2), August 2001.

[8] Isabelle Devarenne, Hakim Mabed, and Alexandre Caminada. Intelligent neigh-

borhood exploration in local search heuristics. In Proceedings of the 18th IEEE

International Conference on Tools with Artiﬁcial Intelligence (ICTAI ’06), pages

144–150. IEEE Computer Society, 2006.

[9] Mohammad Dib. Tabu-NG: hybridization of constraint programming and lo-

cal search for solving CSP. PhD thesis, University of Technology of Belfort-

Montb´eliard, December 2010.

21

[10] Mohammad Dib, Alexandre Caminada, and Hakim Mabed. Frequency manage-

ment in Radio military Networks. In INFORMS Telecom 2010, 10th INFORMS

Telecommunications Conference, Montreal, Canada, May 2010.

[11] N. Dubois and D. de Werra. Epcot: An eﬃcient procedure for coloring optimally

with Tabu Search. Computers & Mathematics with Applications, 25(10–11):35–45,

1993.

[12] Nicolas Durand and Jean-Marc Alliot. Genetic crossover operator for partially sep-

arable functions. In Genetic Programming 1998: Proceedings of the Third Annual

Conference, pages 487–494, University of Wisconsin, Madison, Wisconsin, USA,

22-25 July 1998. Morgan Kaufmann.

[13] C. Fleurent and J. Ferland. Genetic and Hybrid Algorithms for Graph Coloring.

Annals of Operations Research, 63:437–464, 1996.

[14] Philippe Galinier, Jean-Philippe Hamiez, Jin-Kao Hao, and Daniel Cosmin Porum-

bel. Recent Advances in Graph Vertex Coloring. In Ivan Zelinka, V´aclav Sn´asel,

and Ajith Abraham, editors, Handbook of Optimization, volume 38 of Intelligent

Systems Reference Library, pages 505–528. Springer, 2013.

[15] Philippe Galinier and Jin-Kao Hao. Hybrid evolutionary algorithms for graph

coloring. Journal of Combinatorial Optimization, 3(4):379–397, 1999.

[16] Philippe Galinier and Alain Hertz. A survey of local search methods for graph

coloring. Computers & Operations Research, 33:2547–2562, 2006.

[17] Philippe Galinier, Alain Hertz, and Nicolas Zuﬀerey. An adaptive memory algo-

rithm for the k-coloring problem. Discrete Applied Mathematics, 156(2):267–279,

2008.

[18] M. R. Garey and David S. Johnson. Computers and Intractability: A Guide to the

Theory of N P-Completeness. Freeman, San Francisco, CA, USA, 1979.

[19] Jin-Kao Hao. Memetic Algorithms in Discrete Optimization. In Ferrante Neri, Car-

los Cotta, and Pablo Moscato, editors, Handbook of Memetic Algorithms, volume

379 of Studies in Computational Intelligence, pages 73–94. Springer, 2012.

[20] Alain Hertz and Dominique de Werra. Using Tabu Search Techniques for Graph

Coloring. Computing, 39(4):345–351, 1987.

[21] Alain Hertz, M. Plumettaz, and Nicolas Zuﬀerey. Variable Space Search for Graph

Coloring. Discrete Applied Mathematics, 156(13):2551 – 2560, 2008.

[22] David S. Johnson, C. R. Aragon, L. A. McGeoch, and C. Schevon. Optimization

by Simulated Annealing: An Experimental Evaluation; Part II, Graph Coloring

and Number Partitioning. Operations Research, 39(3):378–406, 1991.

22

[23] David S. Johnson and Michael Trick, editors. Cliques, Coloring, and Satisﬁability:

Second DIMACS Implementation Challenge, 1993, volume 26 of DIMACS Series in

Discrete Mathematics and Theoretical Computer Science. American Mathematical

Society, Providence, RI, USA, 1996.

[24] R.M. Karp. Reducibility among combinatorial problems. In R. E. Miller and J. W.

Thatcher, editors, Complexity of Computer Computations, pages 85–103. Plenum

Press, New York, USA, 1972.

[25] F. T. Leighton. A Graph Coloring Algorithm for Large Scheduling Problems.

Journal of Research of the National Bureau of Standards, 84(6):489–506, 1979.

[26] Zhipeng L¨u and Jin-Kao Hao. A memetic algorithm for graph coloring. European

Journal of Operational Research, 203(1):241–250, 2010.

[27] A. Mehrotra and Michael A. Trick. A Column Generation Approach for Graph

Coloring. INFORMS Journal On Computing, 8(4):344–354, 1996.

[28] Jacques Monod. Chance and necessity: an essay on the natural philosophy of

modern biology. Knopf, 1971.

[29] M. Plumettaz, D. Schindl, and Nicolas Zuﬀerey. Ant Local Search and its eﬃcient

adaptation to graph colouring. Journal of Operational Research Society, 61(5):819

– 826, 2010.

[30] Daniel Cosmin Porumbel, Jin-Kao Hao, and Pascale Kuntz. Diversity Control and

Multi-Parent Recombination for Evolutionary Graph Coloring Algorithms. In 9th

European Conference on Evolutionary Computation in Combinatorial Optimisation

(Evocop 2009), T¨ubingen, Germany, 2009.

[31] Daniel Cosmin Porumbel, Jin-Kao Hao, and Pascale Kuntz. An evolutionary

approach with diversity guarantee and well-informed grouping recombination for

graph coloring. Computers & Operations Research, 37:1822–1832, 2010.

[32] Steven David Prestwich. Generalised graph colouring by a hybrid of local search

and constraint programming. Discrete Applied Mathematics, 156(2):148–158, 2008.

[33] El-Ghazali Talbi. A Taxonomy of Hybrid Metaheuristics. Journal of Heuristics,

8(5):541–564, September 2002.

[34] Olawale Titiloye and Alan Crispin. Graph Coloring with a Distributed Hybrid

Quantum Annealing Algorithm. In James O’Shea, Ngoc Nguyen, Keeley Crock-

ett, Robert Howlett, and Lakhmi Jain, editors, Agent and Multi-Agent Systems:

Technologies and Applications, volume 6682 of Lecture Notes in Computer Science,

pages 553–562. Springer Berlin / Heidelberg, 2011.

[35] Olawale Titiloye and Alan Crispin. Quantum annealing of the graph coloring

problem. Discrete Optimization, 8(2):376–384, 2011.

23

[36] Olawale Titiloye and Alan Crispin. Parameter Tuning Patterns for Random Graph

Coloring with Quantum Annealing. PLoS ONE, 7(11):e50060, 11 2012.

[37] D. C. Wood. A Technique for Coloring a Graph Applicable to Large-Scale

Timetabling Problems. Computer Journal, 12:317–322, 1969.

[38] Qinghua Wu and Jin-Kao Hao. Coloring large graphs based on independent set

extraction. Computers & Operations Research, 39(2):283–290, 2012.

[39] Nicolas Zuﬀerey, P. Amstutz, and P. Giaccari. Graph Colouring Approaches for a

Satellite Range Scheduling Problem. Journal of Scheduling, 11(4):263 – 277, 2008.

24

- A preview of this full-text is provided by Springer Nature.
- Learn more

Preview content only

Content available from Journal of Heuristics

This content is subject to copyright. Terms and conditions apply.