Download as pdf or txt
Download as pdf or txt
You are on page 1of 26

Information Sciences 541 (2020) 271–296

Contents lists available at ScienceDirect

Information Sciences
journal homepage: www.elsevier.com/locate/ins

A memetic algorithm based on reformulation local search


for minimum sum-of-squares clustering in networks
Qing Zhou a, Una Benlic b, Qinghua Wu a,⇑
a
School of Management, Huazhong University of Science and Technology, No. 1037, Luoyu Road, Wuhan, China
b
SATALIA, 40 High Street, Islington High Street, London N1 8XB, United Kingdom

a r t i c l e i n f o a b s t r a c t

Article history: The edge minimum sum-of-squares clustering problem (E-MSSC) aims at finding p proto-
Received 4 April 2019 types such that the sum of squares of distances from a set of vertices to their closest pro-
Received in revised form 7 April 2020 totype is minimized, where a prototype is either a vertex or an interior point of an edge.
Accepted 27 June 2020
This paper proposes a highly effective memetic algorithm for E-MSSC that combines a ded-
Available online 7 July 2020
icated crossover operator for solution generation with a reformulation local search for
solution improvement. Reformulation local search is a recent algorithmic framework for
Keywords:
continuous location problems that is based on an alternation between original (continu-
Heuristics
Clustering problems
ous) and discrete formulations of a given problem. Furthermore, the proposed algorithm
Memetic algorithm uses a simple strategy for population updating, and a mutation operator to prevent from
Reformulation local search premature convergence. The computational results obtained on three sets of 132 bench-
mark instances show that the proposed algorithm competes very favorably with the exist-
ing state-of-the-art algorithms in terms of both solution quality and computational
efficiency. We also analyze several essential components of the proposed algorithm to
understand their contribution to the algorithm’s performance.
Ó 2020 Published by Elsevier Inc.

1. Introduction

Cluster analysis, which consists in grouping a collection of patterns into clusters based on similarity, is a powerful tool in
many fields such as pattern recognition [5,28], machine learning [3] and image processing [38]. The simplest form of clus-
tering is partition based clustering that aims at partitioning a given data set into disjoint clusters (subsets), such that some
specific clustering criteria are optimized. Among many criteria encountered in cluster analysis, the most well-known is the
Minimum Sum-of-Squares Clustering (MSSC) [14]. Assuming that the entities are given as points in the N-dimensional space
RN , MSSC consists in finding p points, called prototypes, in such a way that the sum of squares of the Euclidean distances
between the entities and their nearest prototypes is minimized. While MSSC is shown to be an NP-hard problem [1], exact
algorithms can be used to solve it on small to medium-sized problem instances [2]. Due to the complexity of the problem,
heuristics and metaheuristics are necessary [6,15,16,23,24,40] for large problem instances.
Recently, Carrizosa et al. [13,14] introduced two Minimum Sum-of-Squares Clustering (MSSC) problems arising in net-
works, namely the Vertex-MSSC (V-MSSC) and the Edge-MSSC (E-MSSC). Given a connected weighted undirected network
G ¼ ðV; EÞ, where V and E are the set of vertices and edges respectively, V-MSSC seeks to determine p prototypes from V such
that the weighted sum of the squared distances from the vertices to their closest prototype is minimized. According to [14],

⇑ Corresponding author.
E-mail addresses: qingzhouqz03@gmail.com (Q. Zhou), una@satalia.com (U. Benlic), qinghuawu1005@gmail.com (Q. Wu).

https://doi.org/10.1016/j.ins.2020.06.056
0020-0255/Ó 2020 Published by Elsevier Inc.
272 Q. Zhou et al. / Information Sciences 541 (2020) 271–296

V-MSSC is very similar to the classic p-median problem and both exact and heuristic algorithms [4,22] can be successfully used
to solve V-MSSC. E-MSSC is similar to V-MSSC, with the distinction that prototypes are sought along the edges [26]. Meanwhile,
the authors also proposed several efficient variable neighborhood search algorithms for the two problems [14]. While these
heuristics apply the same K-means method for local improvement, they make use of different strategies to generate initial solu-
tions. Extensive experiments show that the VNS-based heuristic (denoted as VNS-1 throughout this paper), which employs a
sophisticated strategy for generating its starting solution, considerably outperforms the VNS variant that starts with a random
initial solution. In [34], Nikolaev et al. proposed two additional VNS-based heuristics for E-MSSC, denoted as VNS-2 and VNS-3
in our work, which are based on two different variants of Variable Neighborhood Descent (VND). The three VNS-based heuris-
tics (VNS-1, VNS-2, VNS-3) are considered as the current state-of-the-art approaches for E-MSSC.
In this work, we are interested in the E-MSSC problem. We propose a highly effective population-based memetic algo-
rithm (MA) to enrich the set of solution methods for solving this computationally challenging problem. The main contribu-
tions of this work can be summarized as follows:

 From the algorithmic perspective, the proposed algorithm integrates several original ingredients, including a dedicated
crossover operator and a reformulation local search procedure for local improvement. Motivated by the observation that
high-quality local optima share common building blocks, the crossover preserves the components (prototypes) that are
present in both parents. The reformulation local search is an iterative algorithm that alternates between a K-means algo-
rithm and tabu search. While K-means performs the search in the continuous space and can easily become trapped in a
local optimum, tabu search is applied to further refine the local optimum by exploring a finite set of points from the con-
tinuous space. Through an iterative application of these two complementary search procedures, the algorithm is able to
explore various zones of the search space without being easily confined to a local optimum.
 From the computational perspective, the proposed algorithm MA achieves highly competitive performances in terms of
both solution quality and computational efficiency on three sets of 132 benchmark instances, including 40 OR-Lib
instances, 80 TSP-Lib instances and 12 network graphs from the University of Florida Sparse Matrix Collection.

The rest of the paper is organized as follows. Section 2 gives a mathematical formulation of E-MSSC, followed by a
description of some essential properties of E-MSSC in Section 3. Section 4 describes in detail the main elements of the pro-
posed memetic algorithm. Computational results and comparisons with several state-of-the-art algorithms are given in Sec-
tion 5. Before concluding, Section 6 studies some key components of the proposed algorithm.

2. Problem statement and mathematical formulation

E-MSSC can be formally described as follows. Let G ¼ ðV; EÞ be a connected undirected network, where
V ¼ fv 1 ; v 2 ; . . . ; v n g is the set of n vertices and E is the edge set with a positive length ljk for each edge ðj; kÞ. We consider
the network G to be a spatial network: for any edge e ¼ ðj; kÞ 2 E and any x 2 ½0; 1, the pair ðe; xÞ is identified by a point
on edge e at a distance x  ljk from j, and at a distance ð1  xÞ  ljk from k. Let PðGÞ denote the set of all the pairs ðe; xÞ,
i.e., the set of points of the network G. Then the E-MSSC problem is to determine p prototypes from the set of points
PðGÞ on edges, such that the weighted sum of squared distances from the vertices to their closest prototype is minimized
[14]:
P
v 2V wv d ðv ; fx1 ; . . . ; xp gÞ
2
min
ð1Þ
s:t: x1 ; . . . ; xp 2 PðGÞ
where positive weights wv are given for each v 2 V. The distance between two points u and v is calculated as the shortest
path dðu; v Þ connecting u and v. The distance dðv ; XÞ between v and a non-empty finite set X of points in the network is
defined as the distance to the closest point in X, i.e., dðv ; fx1 ; . . . ; xp gÞ ¼ min dðv ; xi Þ.
16i6p

The V-MSSC problem is analogous to E-MSSC, where prototypes can only be selected from the vertex set
V ¼ fv 1 ; v 2 ; . . . ; v n g. A solution to E-MSSC can be denoted as fx1 ; . . . ; xp g, where x1 ,. . .,xp are the p prototypes. Given the
set of p prototypes, the p clusters can be easily determined by assigning each vertex v 2 V to its closest prototype.
Fig. 1 provides an example of E-MSSC. Consider a graph G ¼ ðV; EÞ with V ¼ fv 1 ; v 2 ; v 3 ; v 4 ; v 5 g and
E ¼ fðv 1 ; v 2 Þ; ðv 2 ; v 3 Þ; ðv 3 ; v 4 Þ; ðv 4 ; v 5 Þg. The lengths of the edges ðv 1 ; v 2 Þ; ðv 2 ; v 3 Þ; ðv 3 ; v 4 Þ; ðv 4 ; v 5 Þ are respectively equal to
5, 4, 1 and 3. In this example, we assume that the number of clusters p ¼ 2 and that the weights of all the vertices are equal
to 1. The top left side of the figure illustrates a feasible solution s ¼ fx1 ; x2 g with clusters C 1 ¼ fv 1 ; v 2 g and C 2 ¼ fv 3 ; v 4 ; v 5 g,
induced by points x1 ¼ fðv 1 ; v 2 Þ; 0:5g and x2 ¼ fðv 4 ; v 5 Þ; 0:2g. The data for calculating the objective value f ðsÞ is displayed on
the right side of the figure.
E-MSSC can be formulated as a Mixed-Integer Program [34] with the following decision variables:
Variables

– xj : Binary variable, where xj ¼ 1 if and only if vertex j is selected to be a prototype, and xj ¼ 0 otherwise.
– zij : Binary variable, where zij ¼ 1 if and only if vertex i is assigned to the vertex prototype j, and zij ¼ 0 otherwise.
Q. Zhou et al. / Information Sciences 541 (2020) 271–296 273

Fig. 1. An illustrative example of E-MSSC.

– yjk : Binary variable, where yjk ¼ 1 if and only if one of the prototypes lies on edge ðj; kÞ, and yjk ¼ 0 otherwise.
– cjk : Continuous variable which defines the location of the prototype on edge ðj; kÞ with a distance ljk cjk from vertex j and
with a distance ljk ð1  cjk Þ from vertex k.
– mijk1 : Binary variable, where mijk1 ¼ 1 if and only if a cluster with a prototype on edge ðj; kÞ contains vertex i such that the
shortest path (denoted as nijk1 ) between i and this prototype passes through vertex j.
- mijk2 : Binary variable, where mijk2 ¼ 1 if and only if vertex i is assigned to a prototype on edge ðj; kÞ such that the shortest
path (denoted by nijk2 ) between i and this prototype passes through vertex k.

Given the above decision variables, E-MSSC can be formulated as the following Mixed-Integer Program [34]:
Xn Xn 2
Xn X X2
minimizef ðsÞ ¼ i¼1 j¼1
½dði; jÞ zij þ i¼1 ðj;kÞ2E
n2 m
s¼1 ijks ijks
ð2Þ
Xn X X2
subject to j¼1 zij þ ðj;kÞ2E
m ¼ 1;
s¼1 ijks
i 2 f1; . . . ; ng ð3Þ

zij 6 xj ; i; j 2 f1; . . . ; ng ð4Þ


X2
m 6 yjk ; i 2 f1; . . . ; ng; 8ðj; kÞ 2 E
s¼1 ijks
ð5Þ
Xn X
x þ
i¼1 i
y ¼p
ðj;kÞ2E jk
ð6Þ

nijk1 ¼ dði; jÞ þ ljk cjk ; 8ðj; kÞ 2 E ð7Þ


nijk2 ¼ dði; kÞ þ ljk ð1  cjk Þ; 8ðj; kÞ 2 E ð8Þ
zij 2 f0; 1g; i; j 2 f1; . . . ; ng ð9Þ
xj 2 f0; 1g; j 2 f1; . . . ; ng ð10Þ
yjk 2 f0; 1g; 8ðj; kÞ 2 E ð11Þ
mijks 2 f0; 1g; i 2 f1; . . . ; ng; 8ðj; kÞ 2 E; s 2 f1; 2g ð12Þ
cjk 2 ½0; 1; 8ðj; kÞ 2 E ð13Þ

where the objective of E-MSSC is expressed as (2), which is equivalent to the function defined in Eq. (1). Constraint (3)
guarantees that each vertex is located in exactly one cluster, while constraints (4) and (5) ensure that each vertex is only
274 Q. Zhou et al. / Information Sciences 541 (2020) 271–296

allocated to the selected prototype. Constraint (6) guarantees that the number of clusters is exactly equal to p. Finally, con-
straints (7) and (8) define the shortest path between vertex i and the prototype on edge ðj; kÞ that passes through vertex j and
vertex k respectively. The binary values for variables zij ; xj ; yjk and mijks are imposed in constraints (9–12), while constraint
(13) enforces a continuous range ½0; 1 for variable cjk .

3. E-MSSC properties

The E-MSSC problem has two important properties that motivate the design choices of the algorithm described in the
next section. The proofs of these properties were presented in [14], and can be formally described as follows.

Property 1. Let e ¼ ðj; kÞ 2 E, at most one optimal prototype x lies on the interior of e. In that case, both endpoints j and k have x
as its nearest prototype.

Property 2. If an optimal prototype x is located at a vertex v 2 V, then there cannot be an optimal prototype on the interior of
any of the edges adjacent to v.

4. Memetic algorithm (MA)

4.1. Main scheme

Algorithm 1 Memetic algorithm


1: Input: G ¼ ðV; EÞ: E-MSSC instance; jpj: population size
2: Output: best solution S found
3: Initial population Pop ¼ fS1 ; . . . ; Sjpj g /* Section 4.2 */
4: S BestðPopÞ /* S records the best solution found so far */
5: while TimeðÞ 6 t max do
6: randomly select two parents Si , Sj
7: So crossov erðSi ; Sj Þ /* Section 4.3, generate an offspring solution So */
8: So mutationðSo Þ /* Section 4.3, mutate the offspring solution So */
9: So reformulation local searchðSo Þ /* Section 4.5, improve the offspring solution */
10: If f ðSo Þ < f ðS Þ then
11: S So /* Update the best solution found so far */
12: end if
13: Pop population updateðPop; So Þ /* Section 4.4, update the population with a simple updating strategy */
14: end while
15: return S

Memetic algorithm has shown to be an effective framework for tackling combinatorial optimization problems
[25,27,30,31] as it provides a good trade-off between search intensification and diversification by means of a combined
use of a local optimization procedure and a crossover operator. Motivated by an observation that high-quality local
optima share a large number of common prototypes, which can be viewed as building blocks of E-MSSC (refer to Sec-
tion 6.4), we design a dedicated crossover to preserve the components (prototypes) that are present in both parents.
The main scheme of the proposed MA algorithm is given in Algorithm 1. Starting with an initial population (line 3), each
iteration of MA (lines 5 to 14) first proceeds with a random selection of two solutions (i.e., parents, see line 6) that are
used as input to a crossover operator that generates an offspring solution (line 7). Additionally, a mutation operator is
further applied to the offspring solution to prevent from premature convergence (line 8). Following crossover and muta-
tion, the offspring is improved with a reformulation local search method (line 9). The best found solution S is updated
each time an improved solution has been found by the reformulation local search (lines 10 to 12). Finally, a population
updating strategy is applied to update the population (line 13). Before presenting a more detailed description of the main
MA components, we first discuss the search space explored by the MA algorithm. As E-MSSC consists in selecting p pro-
totypes from PðGÞ, the explored search space X contains all the possible subsets of PðGÞ of cardinality p, i.e.,
X ¼ fS  PðGÞ : jSj ¼ pg. Given a selection of p prototypes, the p clusters can easily be obtained by assigning each vertex
v 2 V to its closest prototype.
Q. Zhou et al. / Information Sciences 541 (2020) 271–296 275

4.2. Population initialization

Algorithm 2 Random construction algorithm


1: Input: G ¼ ðV; EÞ: E-MSSC instance; p: number of clusters
2: Output: X 0 : initial solution
3: X 0 £; E0 E /* E0 is the set of feasible edges */
4: P0 ðGÞ PðGÞ /* P0 ðGÞ is the set of feasible points of the network */
5: for i ¼ 1; 2; . . . ; p do
6: randomly choose an edge e from the set E0
7: randomly choose a feasible point y on edge e from the set P0 ðGÞ
8: X 0 X 0 [ fyg
9: If y is an endpoint of the edge e then
10: P0 ðGÞ P0 ðGÞ n y
11: for e1 2 E && e1 is adjacent to endpoint y do
12: delete all interior points on edge e1 from P 0 ðGÞ
13: If all points on e1 have been deleted then
14: E0 E0 n e1
15: end if
16: end for
17: else if y is an interior point on edge e then
18: delete all points on edge e from P 0 ðGÞ
19: E0 E0 n e
20: end if
21: end for
22: return X 0

Each solution from the initial population is obtained with a random construction algorithm (see Algorithm 2) from
[14,34], which ensures satisfaction of Properties 1 and 2. Given the set of previously selected prototypes, a point is feasible
provided that Properties 1 and 2 are still satisfied. We say that an edge is feasible if it contains feasible points.
Algorithm 2 is an iterative procedure, where each iteration consists of the following steps. First, a feasible edge e is chosen
at random from the set of feasible edges, followed by the selection of a feasible point y on e from the set of feasible points.
The selected point acts as a prototype that extends the partial solution X 0 . Once prototype y has been selected, we next dis-
card from future consideration all the points and edges which are in conflict with the selected prototype according to Prop-
erties 1 and 2. More precisely, if y is an interior point of e, we discard from future consideration all the points corresponding
to e that are in conflict with Property 1. If y is an endpoint of an edge, we eliminate from future consideration y and all the
interior points of each edge adjacent to y so as to satisfy Property 2. Finally, we remove from the set of feasible edges all the
edges that contain no feasible point. This process is repeated until p points are obtained.
The solution created by Algorithm 2 is then improved with the local optimization procedure described in Section 4.5. It is
then inserted into the population if and only if it is not identical to any existing solution in the population. This procedure
leads to an initial population of relatively high quality. It stops as soon as the number of solutions in the population reaches
jpj (jpj is the population size).

4.3. Crossover and mutation operator

A crossover operator is the key diversification mechanism of a population-based algorithm. It is well known that an effec-
tive crossover operator should not only generate diversified solutions but also transfer meaningful components from parents
to offspring [25,39]. In case of E-MSSC, an analysis conducted on a sample of locally optimal solutions disclosed that high
quality solutions share a large number of common prototypes, which might constitute building blocks of an optimal solution
(refer to Section 6.4). Given two parent solutions P 1 ¼ fx11 ; x12 ; . . . ; x1p g and P 2 ¼ fx21 ; x22 ; . . . ; x2p g, the proposed crossover for E-
MSSC thus aims to preserve common prototypes shared by both P1 and P2 . More precisely, the first step of the crossover is to
simply copy over to offspring S0 the subset of prototypes shared in both parents, i.e., S0 ¼ P 1 \ P2 . Given the partial offspring
solution S0 , a randomized procedure is then applied to extend S0 to a complete solution containing exactly p prototypes. Each
iteration of the procedure consists in randomly selecting and extending S0 with a point in PðGÞ that satisfies both Properties 1
and 2.
Given a graph consisting of 7 vertices (v 1 ; v 2 ; . . . ; v 7 ) and 6 edges (e1 ; e2 ; . . . ; e6 ), Fig. 2 illustrates the crossover procedure
on two parents P1 ¼ fv 1 ; v 3 ; v 0ðe4 ;0:2Þ g and P 2 ¼ fv 1 ; v 6 ; v 0ðe4 ;0:2Þ g, where fv 0ðe4 ;0:2Þ g represents an interior point on e4 at a distance
0:2  le4 from the first-endpoint of e4 , i.e., vertex v 4 . In this example, we assume that the number of clusters p ¼ 3. In the
276 Q. Zhou et al. / Information Sciences 541 (2020) 271–296

Fig. 2. An example of offspring construction.

first step, the common prototypes shared by parents P 1 and P2 (i.e., fv 1 g and fv 0ðe4 ;0:2Þ g) are transferred to the partial offspring
S0 . In the second step, a feasible point (i.e., fv 0ðe5 ;0:6Þ g) is selected at random to obtain a complete solution S0 .
Once an offspring So is obtained with the crossover procedure, a random mutation operator is applied to So consisting of l
random SwapMove moves, where l is a parameter called the mutation strength. A SwapMove move is performed by ran-
domly selecting a point from PðGÞ and then exchanging this point with a randomly chosen prototype in So , such that Prop-
erties 1 and 2 remain satisfied.

4.4. Population updating strategy

When a new offspring solution is generated with the crossover operator, it is first improved by the reformulation local
search procedure (see Section 4.5) and and then inserted into the population if: (i) it is better than the worst individual
in the population in terms of solution cost; and (ii) there is no individual in the population identical to the offspring solution.
If both criteria are satisfied, offspring replaces the worst individual of the population, and is discarded otherwise. One poten-
tial risk of this classic updating strategy is premature convergence as So is added in the population regardless of its similarity
(see Section 6.4) to other individuals of the population. The proposed algorithm uses a random mutation operator as
described in Section 4.3 to avoid premature convergence by providing a diversified offspring solution. Therefore, the pro-
posed algorithm is not sensitive to the population updating strategy.

4.5. Reformulation local search

Local search is another key component to our memetic algorithm that plays an essential role in exploring the search
space. The fast K-means method has been frequently used as a local search method in many clustering applications. As a
K-means algorithm can easily become trapped in a local optimum just as a descent-based local search [23], it is frequently
embedded into a variable neighborhood search framework [14,34]. In this work, we propose an effective reformulation local
search (RLS) for E-MSSC, which has been successfully applied to a wide variety of continuous optimization problems [10,11].
The main idea behind RLS is to exploit the relation between the continuous model and its discrete counterpart. More pre-
cisely, the K-means method is first carried out in the continuous space until a local optimum is reached, followed by a search
in the discretized space of the continuous model in an attempt to find an improved solution. The alternation between the
two problem formulations continues until no further improvement can be achieved. Furthermore, new potential prototypes
produced in the continuous space are added to the discrete space in order to improve the quality of the discrete approxima-
tion. For search exploration in the discrete space, RLS employs a tabu search approach that has shown to be highly effective
Q. Zhou et al. / Information Sciences 541 (2020) 271–296 277

for difficult discrete optimization problems [21]. Let FS denote a finite set of points and let X be a subset of p points from FS.
The discrete formulation explored by our tabu search is then given by:
minXFS f ðXÞ ð14Þ
Before the beginning of the MA search, the finite set FS is initialized with V ¼ fv 1 ; . . . ; v n g as recommended in [10]. During
the reformulation local search, new prototypes found in the continuous space are being added to FS each time a K-means
method attains a local optimum. If the size of FS reaches a threshold value w (w is a parameter), FS is re-initialized to
V ¼ fv 1 ; . . . ; v n g so as to maintain the discrete model at a manageable size.
The general procedure of RLS is summarized in Algorithm 3. Each iteration alternates between the K-means and the tabu
search approach (lines 5–12), where K-means is applied first to identify a local optimum sc in the continuous space and to
expand FS with the newly found prototypes in sc (line 7). This is followed by the improvement of sc with the tabu search
procedure to obtain a new starting point for the next round of the K-means local search. The search stops as soon as no
improvement is achieved with the tabu search, and returns the best solution found during this process.
Additional experimental results demonstrate that the combined use of the discrete and the continuous problem formu-
lations constitutes an effective hybridization that plays an important role for the RLS performance (Section 6.1). Detailed
descriptions of K-means and the tabu search are given in the following subsections.

Algorithm 3 Reformulation local search


1: Input: Input solution s0 , finite set of points FS
2: Output: Best solution sb found during the local search process
3: sd s0 /* improved solution */
4: sc
£ /* local optimum */
5: while sc – sd
6: sc K  meansðn; p; sd Þ /* see Section 4.5.1 */
7: (augmenting FS): FS FS [ sc
8: s d
tabu searchðFS; s Þ /* see Section 4.5.2 */
c

9: If f ðsd Þ < f ðsb Þ then


10: sb sd
11: end if
12: end while
13: return sb

4.5.1. K-means heuristic


The K-means algorithm and its variations are the most commonly used methods for the clustering problems [5,28,35]. In
our work, we adopt the K-means algorithm introduced in [14] to solve the continuous formulation of E-MSSC. Basically, it
proceeds between location and allocation steps until the two steps reach convergence: in the location step (the allocations
are assumed to be given), the optimal locations of the p prototypes are calculated using the 1-OPT approach [14]; in the allo-
cation step (the locations of the prototypes are assumed to be given), all the vertices are assigned to their closest prototype.
More details about the K-means method for E-MSSC can be found in [14].

4.5.2. Neighborhood and move operators of tabu search


The proposed tabu search (TS) goes through the subset of p prototypes from the finite set FS with the aim to reduce the
objective value. To effectively explore the discrete space induced by FS, TS relies on two compound move operators to gen-
erate its neighborhoods. The first operator can be decomposed into an ADD move followed by another DROP move, while the
second one consists of a DROP move followed by another ADD move. Given the current solution S ¼ fx1 ; x2 ; . . . ; xp g, the basic
ADD and DROP moves can be defined as follows:

– DROPðxi Þ: remove a prototype xi ði ¼ 1; 2; . . . ; pÞ from the current solution S ¼ fx1 ; x2 ; . . . ; xp g.


– ADDðpi Þ: add a point pi R S from FS.

Note that by applying these two compound move operators, the number of prototypes in the current solution always
remains equal to p.
The general procedure of our tabu search is given in Algorithm 4. At each iteration, the choice between the two compound
move operators is made in a random manner. Each basic non-tabu move (ADD and DROP) is selected and performed such that
the corresponding move gain value is minimized (see Section 4.5.3). Each time a point is added to the current solution with
the ADD operator, it is forbidden to remove it with the DROP operator for the next tt iterations to prevent from short-term
cycling, where tt is a parameter called the tabu tenure. TS terminates after maxIter iterations (maxIter is a parameter), and
returns the best solution found during the search.
278 Q. Zhou et al. / Information Sciences 541 (2020) 271–296

Algorithm 4 Tabu search


1: Input: FS - finite set of points; s0 - initial solution
2: Output: best solution sb found during tabu search
3: iteration 0
4: Initialize the tabu list
5: while iteration < maxIter do
6: coin randð0; 1Þ /* flips a coin */
7: if coin ¼ 0 then
8: Choose a best non-tabu DROP move (i.e., dropping the prototype xi from s0 )
9: s0 s0 n xi
10: Choose a best non-tabu ADD move (i.e., adding a point pi R s0 from FS into s0 )
11: s0 s0 [ pi
12: else
13: Choose a best non-tabu ADD move (i.e., adding a point pi R s0 from FS into s0 )
14: s0 s0 [ pi
15: Choose a best non-tabu DROP move (i.e., dropping the prototype xi from s0 )
16: s0 s0 n xi
17: end if
18: Update the tabu list
19: iteration iteration þ 1
20: if f ðs0 Þ < f ðsb Þ
21: sb s0
22: end if
23: end while
24: return sb

4.5.3. Fast neighborhood evaluation technique to accelerate tabu search


The key concept related to the basic ADD and DROP moves is the move gain, which indicates the extent of the solution
improvement in terms of the objective value f (Eq. (1)) when a point is selected to be a prototype or is removed from the
solution. To rapidly determine the move gain of a basic move, we use a fast incremental evaluation technique which greatly
improves on the speed of our tabu search approach. The main idea is to maintain a n  2 matrix DS, where the elements
DS½i½0 and DS½i½1 ði ¼ 1; 2; . . . ; nÞ represent respectively the distance between vertex i and its nearest and second nearest
prototype.
Each time a DROP move is performed removing a prototype xi (i ¼ 1; 2; . . . ; p) from the current solution, all vertices
assigned to xi (denoted as Dðxi Þ) will be reallocated to their second nearest prototypes. The resulting move gain value of a
DROP move can then easily be determined as:
X X
Df ¼ j2Dðxi Þ
ðDS½j½1  DS½j½1Þ  j2Dðxi Þ
ðDS½j½0  DS½j½0Þ ð15Þ

Given a solution S, each time an ADD move is performed by adding a point pi R S from the finite set FS into the current
solution S, the set of vertices (denoted as Aðpi Þ) which are closer to pi than to their nearest prototype will be reassigned to pi .
The resulting move gain value is computed as:
X X
Df ¼ j2Aðpi Þ
ðdðpi ; jÞ  dðpi ; jÞÞ  j2Aðpi Þ
ðDS½j½0  DS½j½0Þ ð16Þ

The matrix DS is initialized before each call of the TS procedure in Oðn  p  2Þ time, and is updated by recalculating DS in
Oðn  p  2Þ after each basic move. For the DROP and the ADD move, the best move can be respectively identified in OðnÞ and
Oðn  jFSjÞ with the help of the matrix DS. Therefore, the time complexity of each tabu search iteration is bounded by
Oðn  p  2 þ n þ n  jFSjÞ.

5. Computational experiments

To evaluate the effectiveness of the proposed algorithm, we carry out extensive experiments on well known benchmarks
and make comparisons with three state-of-the-art E-MSSC algorithms from the literature: VNS-1 [14], VNS-2 [34] and VNS-3
[34]. To the best of our knowledge, these are the current best performing approaches for the E-MSSC problem. In addition, we
also make comparisons between MA and two effective heuristics (GRASP [37] and HYBRID [37]), which were originally
developed for the classic p-median problem.
Q. Zhou et al. / Information Sciences 541 (2020) 271–296 279

5.1. Benchmark instances

We tested the proposed algorithm on the following three sets of instances:

 OR-Lib: This set was introduced by Beasley [8] and includes 40 p-median instances. The size of these instances ranges
from 100 to 900 vertices and from 200 to 16200 edges, while the number of clusters is in the ranges from 5 to 200.
 TSP-Lib: The set was originally designed for the traveling salesman problem [36]. Many instances in TSP-Lib have been
used before for the classic p-median problem [20,22]. Following Fast and Hicks [20], we consider 20 instances from this
set with sizes ranging from 400 to 3038 vertices and from 4922 to 282004 edges. For each instance four different values of
p (number of clusters) are used, i.e., p 2 f10; 50; 100; 200g, thus giving a total of 80 TSP-Lib instances.
 University of Florida Sparse Matrix (UFSM) Collection [17]: This set is a widely used set of sparse matrix benchmark
collected from a variety of real applications. According to our knowledge, we are the first to use it for evaluation of an
E-MSSC approach. Three network graphs (named USAir97, Harvard500 and Email respectively) are considered with the
number of vertices ranging from 332 to 1133. For each graph, four values of p (p 2 f10; 50; 100; 200g) were used, giving
a total of 12 graphs.

Table 1 provides details for the three data sets. Column ‘Instance’ indicates the instance name. Columns ‘n’ and ‘m’ respec-
tively refer to the number of vertices and the number of edges of the graph. All the input files are available in the Mendeley’s
public repository1 to facilitate future research.

5.2. Parameter tuning and experimental protocol

The MA search is controlled by means of 5 parameters: the population size jpj, the search depth of tabu search maxIter, the
mutation strength l, the tabu tenure tt and the size w of the finite set of points considered during tabu search. To tune these
parameters, we use the ‘irace’ software package [29] that implements the iterated racing method [7,9] for finding the best
configuration from a finite set of possible parameter configurations. For the tuning process, we select the 20 largest OR-Lib
instances (i.e., instances pmed21-40), and set the tuning budget to 1000 executions each limited to 60 s of run-time. The
tested and the final parameter values (suggested by irace) are shown in Table 2. For all the experiments conducted in this
paper, we use the same setting of parameters determined by irace. The proposed algorithm was programmed in C++ lan-
guage2, running on a computer with a 2.10 GHz processor (Xeon-E5 2695) and 2 GB RAM, under Windows operating system.

5.3. Comparative results with state-of-the-art algorithms

This section provides an extensive computational comparison of MA with several state-of-the-art algorithms. According
to the previous studies [14,34], three heuristics based on the Variable Neighborhood Search (VNS) framework (denoted as
VNS-1, VNS-2 and VNS-3) are considered as the state-of-the-art algorithms for the E-MSSC problem. In addition, two efficient
heuristics GRASP and HYBRID [37] are also considered for the comparative study. Both GRASP and HYBRID are state-of-the-
art algorithms developed for the classic p-median problem, which is similar to E-MSSC. The main difference between the p-
median problem and E-MSSC is that the p-median problem seeks for the p prototypes from the set of vertices only, while E-
MSSC considers not only the vertices but also the points along the edges. As the main references for comparisons, we thus
use these five algorithms, whose source codes were kindly provided by the authors. All the algorithms in our experimental
studies were programmed in CnC++ and compiled using g++ compiler without any option flag. For the five reference algo-
rithms and MA, we followed the same experimental protocol as described in [34], i.e., each algorithm is executed 10 times
per instance with the time limit t max set to 60 s.
Tables 3–5 summarize the statistical results for all the compared algorithms on the OR-Lib benchmark, the TSP-Lib bench-
mark and the UFSM benchmark respectively. Row ‘#Best/Avg’ shows the number of cases for which an algorithm produces
the best result among all the compared algorithms in terms of the best/average objective value. Row ‘Average g best /g av g (%)’
indicates the average percentage gap of the best/average objective value to the best result obtained within this experiment,
while the average running time in seconds required by each algorithm to reach its best objective value across all the tested
instances is given in row ‘Average time(s)’. To verify whether there is a significant statistical difference between MA and the
reference algorithms in terms of the best and the average objective value, the p-values from the non-parametric Friedman
test [19] are provided in rows ‘p-valuebest’ and ‘p-valueavg’. The best performance is indicated in bold. Detailed results for OR-
Lib, TSP-Lib and UFSM benchmarks are respectively presented in Tables 20–22 of the Appendix.
In terms of the best objective value, we observe from Table 3 that VNS-1, VNS-2, VNS-3, GRASP, HYBRID and MA respec-
tively produce the best result in 14, 21, 25, 7, 7 and 39 cases out of the 40 OR-Lib instances. When considering the average
objective value, MA yields the best result on 39 instances, while VNS-1, VNS-2, VNS-3, GRASP and HYBRID respectively give
the best result on 3, 19, 20, 7 and 7 instances. On this bechmark, MA further reports the smallest average percentage gap of

1
https://data.mendeley.com/datasets/5sk2ps2d23/1.
2
the source code of our algorithm is going to be made available for future research.
280 Q. Zhou et al. / Information Sciences 541 (2020) 271–296

Table 1
List of benchmarks.

OR-Lib benchmark TSP-Lib benchmark UFSM benchmark


Instance n m Instance n m Instance n m Instance n m
pmed01 100 200 pmed21 500 5000 rd400 400 4922 USAir97 332 2126
pmed02 100 200 pmed22 500 5000 pr439 439 5973 Harvard500 500 2042
pmed03 100 200 pmed23 500 5000 pcb442 442 6067 Email 1133 5451
pmed04 100 200 pmed24 500 5000 ali535 535 8850
pmed05 100 200 pmed25 500 5000 rat575 575 10188
pmed06 200 800 pmed26 600 7200 gr666 666 13616
pmed07 200 800 pmed27 600 7200 u724 724 16018
pmed08 200 800 pmed28 600 7200 rat783 783 18745
pmed09 200 800 pmed29 600 7200 dsj1000 1000 30560
pmed10 200 800 pmed30 600 7200 pr1002 1002 30701
pmed11 300 1800 pmed31 700 9800 pcb1173 1173 42059
pmed12 300 1800 pmed32 700 9800 rl1304 1304 51943
pmed13 300 1800 pmed33 700 9800 nrw1379 1379 58167
pmed14 300 1800 pmed34 700 9800 fl1400 1400 59990
pmed15 300 1800 pmed35 800 12800 u1432 1432 62685
pmed16 400 3200 pmed36 800 12800 fl1577 1577 75820
pmed17 400 3200 pmed37 800 12800 d1655 1655 83591
pmed18 400 3200 pmed38 900 16200 vm1748 1748 93143
pmed19 400 3200 pmed39 900 16200 rl1889 1889 108759
pmed20 400 3200 pmed40 900 16200 pcb3038 3038 282004

It acheives the best performance.

Table 2
Setting of parameters.

Parameter Sections Description Considered values Final value


jpj 4.1, 4.2 Population size {3,5,7,10,15} 5
l 4.3 Mutation strength {0.05,0.1,0.15,0.2,0.25}*p 0.1*p
maxIter 4.5.2 Search depth of tabu search {20,50,100,150,200} 100
tt 4.5.2 Tabu tenure {10,15,20,25,30} 20
w 4.5 Size of the finite set of points considered during TS n+{5,10,15,20,30}*p n + 20*p

It acheives the best performance.

Table 3
Statistical results for MA and five state-of-the-art algorithms on the OR-Lib instances with a time limit of 60 s.

Instance set VNS-1 VNS-2 VNS-3 GRASP HYBRID MA


OR-Lib #Best/Avg 14/3 21/19 25/20 7/7 7/7 39/39
p-valuebest 3.41e-7 2.21e-5 1.83e-4 4.07e-8 4.07e-8
p-valueavg 1.18e-9 4.59e-6 1.31e-5 4.07e-8 4.07e-8
Average g best /g av g (%) 2.33/4.21 0.42/1.14 0.32/0.76 3.84/3.96 3.86/3.87 0.00/0.13
Average time(s) 36.45 23.62 24.12 12.70 18.32 20.36

It acheives the best performance.

Table 4
Statistical results for MA and five state-of-the-art algorithms on the TSP-Lib instances with a time limit of 60 s.

Instance set VNS-1 VNS-2 VNS-3 GRASP HYBRID MA


TSP-Lib #Best/Avg 0/0 21/13 22/13 15/8 12/27 53/48
p-valuebest 2.77e-18 4.10e-10 1.49e-8 8.68e-7 8.68e-7
p-valueavg 3.74e-19 2.18e-11 1.54e-12 7.96–4 1.75e-3
Average g best /g av g (%) 31.29/38.82 15.23/22.03 15.33/21.82 2.72/2.83 2.70/2.76 0.15/0.69
Average time(s) 46.01 49.53 49.99 33.12 34.12 38.50

It acheives the best performance.

the best/average objective value to the best solution (0.00%/0.13%). The non-parametric Friedman tests (rows p-valuebest and
p-valueavg) reveal that there exists a statistically significant difference between MA and all the reference algorithms in both
the best and the average performance (p-value 60.05). Table 6 presents the results of the Wilcoxon signed ranks test
Q. Zhou et al. / Information Sciences 541 (2020) 271–296 281

Table 5
Statistical results for MA and five state-of-the-art algorithms on the University of Florida Sparse Matrix instances with a time limit of 60 s.

Instance set VNS-1 VNS-2 VNS-3 GRASP HYBRID MA


UFSM #Best/Avg 4/2 3/3 7/6 4/3 4/5 8/7
p-valuebest 3.39e-2 2.70e-3 0.48 1.96e-2 1.96e-2
p-valueavg 1.96e-2 2.70e-3 0.32 9.56e-2 9.56e-2
Average g best /g av g (%) 3.75/5.23 1.29/2.42 0.86/2.03 4.12/4.33 4.12/4.13 0.06/0.29
Average time(s) 41.92 33.88 33.35 19.07 18.87 28.99

It acheives the best performance.

Table 6
Results of the Wilcoxon signed ranks test for MA and the reference algorithms on the OR-Lib instances with a time limit of 60 s.

Comparison Rþ
best
R
best p-value Rþ
avg
R
avg p-value

MA versus VNS-1 26 0 9.44e-7 37 0 3.25e-9


MA versus VNS-2 18 0 1.96e-4 21 0 5.96e-5
MA versus VNS-3 14 0 9.82e-4 19 0 1.32e-4
MA versus GRASP 33 1 5.23e-7 33 1 5.23e-7
MA versus HYBRID 33 1 5.23e-7 33 1 5.23e-7

It acheives the best performance.

recommended in [12,19] to detect any significant performance difference. Column ‘Rþ þ


best ’ (‘Ravg ’) refers to the sum of ranks for
the instances where MA performs better than the compared algorithm, in terms of the best (average) objective value, while
column ‘R 
best ’ (‘Ravg ’) indicates the sum of ranks for the opposite cases. For the OR-Lib benchmark, we observe that MA indi-
cates a significant improvement over all the compared algorithms with a level of significance a ¼ 0:05. In terms of the aver-
age computing times required to reach the final objective value, MA seems to exhibit a slightly better performance than VNS-
1, VNS-2 and VNS-3 algorithms, but a slightly worse performance than GRASP and HYBRID.
Tables 4 and 5 show that MA records the best objective value for the TSP-Lib instances (UFSM instances) in 53 (8) cases,
while VNS-1, VNS-2, VNS-3, GRASP and HYBRID respectively provide the best result for 0 (4), 21 (3), 22 (7), 15 (4), 12 (4)
instances of the TSP-Lib (UFSM) benchmarks. In terms of the average objective value for the two benchmarks, MA produces
the best result on 48 (7) instances, while VNS-1, VNS-2, VNS-3, GRASP and HYBRID respectively show the best result on 0 (2),
21 (3), 22 (6), 15 (3) and 12 (5) cases. When considering the average percentage gap of the best/average objective value to
the best result obtained across all the compared algorithms, MA reports the smallest gap for both the TSP-Lib instances and
the UFSM instances (0.15%/0.69% and 0.06%/0.29%). The non-parametric Friedman test (rows p-valuebest and p-valueavg)
reveals that there is a significant performance difference between MA and all the compared algorithms on the TSP-Lib
instances, while the Wilcoxon signed ranks test in Table 7 further confirms the superiority of MA. On the UFSM benchmark,
the non-parametric Friedman test discloses a significant difference between MA and all of the compared algorithms, exclud-
ing the VNS-3 algorithm. The same observations are confirmed by the Wilcoxon signed ranks test results presented in
Table 8. These results thus highlight the advantage of MA on the tested instances.
Following the reference work in [34], the time limit for the experiments above was set to 60 s. To further evaluate the
performance of the compared algorithms with extended computing times, Tables 9–14 summarize the comparison results
for a prolonged time limit to 600 s. Detailed results of this study for each benchmark set are shown in Tables 23–25 of
the Appendix. In terms of the best (average) objective value for the OR-lib benchmark, we observe from Table 9 that MA pro-
duces the best result for 39 (36) cases, while VNS-1, VNS-2, VNS-3, GRASP and HYBRID respectively yield the best result for
17 (9), 24 (23), 27 (25), 7 (7), 7 (7) instances. Furthermore, MA achieves the smallest average percentage gap of the best/av-
erage objective value to the best result obtained by all the compared alogirhtms (0.00%/0.11%). Finally, the non-parametric
Friedman test confirms a statistically significant difference in terms of the best and the average objective values (p-value
60.05), while the Wilcoxon signed ranks test results reported in Table 12 confirm that MA outperforms all the reference
algorithms with a 0.05 significance level.
Tables 10,11 and 13,14 present the results of the same comparison on the TSP-Lib and the UFSM benchmark. In terms of
the best objective value, VNS-1, VNS-2, VNS-3, GRASP, HYBRID and MA respectively record the best result on 2 (4), 28 (5), 47
(9), 5 (3), 5 (3), 51 (9) instances of the TSP-Lib (UFSM) benchmark. In terms of the average objective value, VNS-1, VNS-2,
VNS-3, GRASP, HYBRID and MA respectively outperform the competing approaches on 0 (2), 26 (3), 36 (7), 6 (3), 10 (3),
48 (10) instances of TSP-Lib (UFSM). MA reports the smallest average percentage gap of the best/average objective value
to the best solution (0.10%/0.38% and 0.04%/0.15%) for both the TSP-Lib and the UFSM set. The non-parametric Friedman con-
firms this difference with a p-value 60.05, while the Wilcoxon signed ranks test results from Table 13 and Table 14 further
confirm that MA outperforms the reference algorithms VNS-1, VNS-2, GRASP and HYBRID.
282 Q. Zhou et al. / Information Sciences 541 (2020) 271–296

Table 7
Results of the Wilcoxon signed ranks test for MA and the reference algorithms for the TSP-Lib instances with a time limit of 60 s.

Comparison Rþ
best
R
best p-value Rþ
avg
R
avg p-value

MA versus VNS-1 79 1 1.49e-14 80 0 7.85e-15


MA versus VNS-2 57 7 2.86e-10 63 7 3.15e-11
MA versus VNS-3 56 10 3.00e-9 66 8 2.24e-11
MA versus GRASP 62 18 1.04e-5 55 25 6.35e-3
MA versus HYBRID 62 18 9.35e-6 54 26 1.37e-2

It acheives the best performance.

Table 8
Results of the Wilcoxon signed ranks test for MA and the reference algorithms on the University of Florida Sparse Matrix instances with a time limit of 60 s.

Comparison Rþ
best
R
best p-value Rþ
avg
R
avg p-value

MA versus VNS-1 7 1 1.73e-2 8 1 1.09e-2


MA versus VNS-2 9 0 7.69e-3 9 0 7.69e-3
MA versus VNS-3 5 3 0.12 6 3 0.17
MA versus GRASP 8 1 1.52e-2 7 2 2.84e-2
MA versus HYBRID 8 1 1.52e-2 7 2 2.84e-2

It acheives the best performance.

Table 9
Statistical results for MA and five state-of-the-art algorithms on the OR-Lib instances with a time limit of 600 s.

Instance set VNS-1 VNS-2 VNS-3 GRASP HYBRID MA


OR-Lib #Best/Avg 17/9 24/23 27/25 7/7 7/7 39/36
p-valuebest 1.62e-6 1.08e-4 5.32e-4 4.07e-8 4.07e-8
p-valueavg 2.58e-8 5.79e-4 2.70e-3 4.07e-8 4.07e-8
Average g best /g av g (%) 0.85/1.72 0.09/0.24 0.05/0.18 3.87/3.96 3.88/3.89 0.00/0.11
Average time(s) 332.20 164.87 166.78 33.53 42.69 161.23

It acheives the best performance.

Table 10
Statistical results for MA and five state-of-the-art algorithms on the TSP-Lib instances with a time limit of 600 s.

Instance set VNS-1 VNS-2 VNS-3 GRASP HYBRID MA


TSP-Lib #Best/Avg 2/0 28/26 47/36 5/6 5/10 51/48
p-valuebest 7.61e-18 6.40e-4 0.35 8.29e-16 8.29e-16
p-valueavg 2.77e-18 9.60e-4 0.06 1.59e-13 8.34e-13
Average g best /g av g (%) 5.58/7.69 3.50/4.58 3.49/4.51 3.38/3.47 3.38/4.42 0.10/0.38
Average time(s) 481.16 370.36 376.75 93.65 117.95 370.20

It acheives the best performance.

Table 11
Statistical results for MA and five state-of-the-art algorithms on the University of Florida Sparse Matrix instances with a time limit of 600 s.

Instance set VNS-1 VNS-2 VNS-3 GRASP HYBRID MA


OR-Lib #Best/Avg 4/2 5/3 9/7 3/3 3/3 9/10
p-valuebest 8.15e-3 8.15e-3 1.00 2.70e-3 2.70e-3
p-valueavg 1.57e-3 2.70e-3 0.26 2.70e-3 2.70e-3
Average g best /g av g (%) 2.56/3.25 0.48/0.75 0.13/0.37 4.12/4.56 4.41/4.41 0.04/0.15
Average time(s) 315.14 277.10 192.99 194.24 185.24 213.10

It acheives the best performance.

Table 12
Results of the Wilcoxon signed ranks test for MA and the reference algorithms on the OR-Lib instances with a time limit of 600 s.

Comparison Rþ
best
R
best p-value Rþ
avg
R
avg p-value

MA versus VNS-1 23 0 2.70e-5 31 0 1.17e-6


MA versus VNS-2 15 0 6.55e-5 17 2 6.25e-4
MA versus VNS-3 19 0 2.22e-3 14 2 2.28e-3
MA versus GRASP 33 1 5.23e-7 33 1 5.23e-7
MA versus HYBRID 31 0 5.23e-7 33 1 5.23e-7

It acheives the best performance.


Q. Zhou et al. / Information Sciences 541 (2020) 271–296 283

Table 13
Results of the Wilcoxon signed ranks test for MA and the reference algorithms on the TSP-Lib instances with a time limit of 600 s.

Comparison Rþ
best
R
best p-value Rþ
avg
R
avg p-value

MA versus VNS-1 77 1 3.93e-14 79 1 1.10e-14


MA versus VNS-2 43 16 1.87e-4 44 18 1.31e-4
MA versus VNS-3 33 25 0.36 39 24 0.08
MA versus GRASP 76 4 2.98e-13 73 7 3.22e-10
MA versus HYBRID 79 1 3.20e-13 72 8 1.05e-9

It acheives the best performance.

Table 14
Results of the Wilcoxon signed ranks test for MA and the reference algorithms on the University of Florida Sparse Matrix instances with a time limit of 600 s.

Comparison Rþ
best
R
best p-value Rþ
avg
R
avg p-value

MA versus VNS-1 7 0 1.80e-2 10 0 5.06e-3


MA versus VNS-2 7 0 1.80e-2 9 0 7.69e-3
MA versus VNS-3 3 3 0.75 5 2 0.31
MA versus GRASP 9 0 7.69e-3 9 0 7.69e-3
MA versus HYBRID 9 0 7.69e-3 9 0 7.69e-3

It acheives the best performance.

6. Analysis

In this section, we study some key elements of the proposed MA algorithm to understand their contribution to the MA’s
performance, including the benefit of using reformulation local search, the role of the memetic framework and the influence
of the mutation operator.

6.1. The usefulness of the reformulation local search

The main objective of this section is to investigate the effectiveness of the combined use of the K-means method and the
tabu search (TS) within the reformulation local search (RLS) framework. For this purpose, we perform experiments with RLS
and its two underlying components (i.e., the K-means method (Section 4.5.1) and and the tabu search method (Section 4.5.2)
on a selection of 20 representative large OR-Lib instances. For each instance and approach, we perform 10 independent runs
with the time limit set to 60 s. Due to a fast convergence of K-means and for a relatively fair comparison, all the three meth-
ods are performed in a multi-start way, until the end of the run-time limit is reached, with a randomly generated solution for
each restart. Specifically, the TS method is restarted every 100 iterations, K-means is restarted as soon as convergence is
reached, while RLS restarts once its two underlying components arrive to the same solution.
To compare the performances of the three algorithms, we show in Fig. 3 the boxplot graphs which represent, for each
algorithm, the distribution of the obtained results for the 20 selected instances. These results are expressed in the form of
percentage gaps, computed as: y ¼ ðf AVG  f BK Þ=f BK  100, where f BK is the best-known solution reported in the literature

Fig. 3. Comparison of RLS, K-means and TS methods.


284 Q. Zhou et al. / Information Sciences 541 (2020) 271–296

Table 15
Comparisons between multi-start versions of RLS, K-means and TS. The best performance is indicated in bold. The average computing time to reach the final
objective value is in seconds.

Instance p RLS K-means TS

f best f av g t av g f best f av g t av g f best f av g t av g

pmed21 5 202777.00 202777.00 0.0 237038.00 254410.13 28.5 202777.00 202777.00 0.0
pmed22 10 183329.00 183329.00 0.6 238380.36 245311.65 28.1 183329.00 183329.00 1.5
pmed23 50 62794.58 62847.91 35.1 88601.75 91061.68 33.9 63855.04 64202.22 19.5
pmed24 400 27068.65 27153.62 23.8 40575.68 42034.85 25.1 31786.41 32410.12 35.6
pmed25 167 11995.44 12128.71 42.1 18372.05 19069.63 30.1 19777.38 20702.09 38.1
pmed26 5 194461.00 194461.00 0.2 225027.00 242234.73 28.9 194461.00 194461.00 0.2
pmed27 10 144800.55 144800.70 23.9 183159.85 195214.82 34.3 144802.00 144802.00 0.8
pmed28 60 49970.32 50054.23 21.5 71397.27 74072.05 23.5 51389.91 51539.12 18.3
pmed29 120 24689.11 24842.67 24.2 36988.73 37803.62 24.4 30585.08 30994.26 38.6
pmed30 200 11303.19 11380.08 23.7 16886.28 17845.97 32.0 21165.92 21918.85 33.2
pmed31 5 169806.00 169806.00 0.1 205572.00 223944.33 34.5 169806.00 169806.00 0.1
pmed32 10 154078.00 154078.00 1.1 202279.50 209706.33 15.5 154078.00 154078.00 1.0
pmed33 70 45351.54 45456.13 40.8 68836.89 70202.72 22.6 47652.12 47999.98 31.3
pmed34 140 20436.60 20651.06 23.1 31496.93 32080.91 33.1 27741.06 28566.51 34.6
pmed35 5 157540.00 157540.00 0.3 193846.00 206456.34 26.1 157540.00 157540.00 0.3
pmed36 10 150918.00 150918.00 1.3 200516.95 210999.53 40.1 150918.00 150918.00 0.5
pmed37 80 46789.40 46952.93 24.5 71429.60 73384.82 26.3 50455.34 50870.03 17.1
pmed38 5 157956.74 157956.75 40.7 200485.00 214890.93 39.7 157972.00 157972.00 0.6
pmed39 10 123424.94 123429.19 15.5 158556.74 171186.28 23.9 123425.00 123435.00 11.5
pmed40 90 41675.55 41811.86 30.2 65339.54 66248.96 29.6 46253.78 46760.92 25.7
#Best/Avg 20 20 0 0 7 7
p-value 7.74e-6 7.74e-6 3.11e-4 1.00e-3
Average 99058.28 99118.74 18.63 127739.31 134908.01 29.00 101488.50 101753.61 15.43

It acheives the best performance.

Table 16
Wilcoxon signed ranks test results between RLS and its two underlying components, with a level of significance a ¼ 0:05.

Comparison Rþ
best
R
best p-value Rþ
avg
R
avg p-value

RLS versus K-means 20 0 8.86e-5 20 0 8.86e-5


RLS versus TS 13 0 1.47e-3 13 0 2.37e-3

It acheives the best performance.

Table 17
Comparisons of MA with a multi-start version of RLS. The best performance is marked in bold. The average computing time required to find the final objective
value is in seconds.

Instance p MA RLS

f best f av g t av g f best f av g t av g

pmed21 5 202777.00 202777.00 0.1 202777.00 202777.00 0.0


pmed22 10 183329.00 183329.00 0.8 183329.00 183329.00 0.6
pmed23 50 62775.66 62833.77 31.2 62794.58 62847.91 35.1
pmed24 400 26941.11 27007.93 50.0 27068.65 27153.62 23.8
pmed25 167 11703.91 11815.66 50.5 11995.44 12128.71 42.1
pmed26 5 194461.00 194461.00 0.3 194461.00 194461.00 0.2
pmed27 10 144800.55 144800.55 14.8 144800.55 144800.70 23.9
pmed28 60 49849.96 49864.68 46.8 49970.32 50054.23 21.5
pmed29 120 24396.37 24531.07 48.2 24689.11 24842.67 24.2
pmed30 200 11084.11 11179.53 51.6 11303.19 11380.08 23.7
pmed31 5 169806.00 169806.00 0.2 169806.00 169806.00 0.1
pmed32 10 154078.00 154078.00 0.9 154078.00 154078.00 1.1
pmed33 70 45269.56 45333.24 45.1 45351.54 45456.13 40.8
pmed34 140 20357.16 20450.06 44.1 20436.60 20651.06 23.1
pmed35 5 157540.00 157540.00 0.5 157540.00 157540.00 0.3
pmed36 10 150918.00 150918.00 1.1 150918.00 150918.00 1.3
pmed37 80 46719.93 46769.21 50.7 46789.40 46952.93 24.5
pmed38 5 157956.73 157956.73 23.0 157956.73 157956.75 40.7
pmed39 10 123424.94 123424.94 13.5 123424.94 123429.19 15.5
pmed40 90 41583.64 41660.42 46.4 41675.55 41811.86 30.2
#Best/Avg 20 20 10 7
p-value 0.007 0.001
Average 98988.63 99026.84 25.98 99058.28 99118.74 18.63

It acheives the best performance.


Q. Zhou et al. / Information Sciences 541 (2020) 271–296 285

Table 18
Wilcoxon signed ranks test results between MA and RLS with a level of significance a ¼ 0:05.

Comparison Rþ
best
R
best p-value Rþ
avg
R
avg p-value

MA versus RLS 10 0 5.06e-3 13 0 1.47e-3

It acheives the best performance.

and f AVG is the average objective value over 10 independent runs. The performance of an algorithm improves as its corre-
sponding percentage gaps approach zero. Fig. 3 indicates that RLS is greatly superior to both TS and K-means. Furthermore,
the percentage gaps obtained with RLS are very close to 0, implying that RLS is highly competitive compared to the best per-
forming algorithms in the literature.
Detailed comparison results are given in Tables 15 and 16. Column ‘p’ in Table 15 refers to the number of clusters while
columns ‘f best ’ and ‘f av g ’ present the best and average objective value over 10 independent runs for each compared algorithm.
In column ‘tav g ’, we report the average computing time in seconds required to find the final objective value. Row ‘Average’
shows the averaged value across all the runs and tested instances. We observe that RLS, K-means and TS respectively pro-
duce the best results on 20, 0, and 7 instances in terms of both the best and the average objective value. The p-values from
the non-parametric Friedman test reveal a significant difference between RLS and its two underlying approaches both in
terms of the best and the average performance, while the Wilcoxon signed ranks test in Table 16 indicates that RLS outper-
forms the two compared algorithms with a 0.05 significance level.

6.2. The usefulness of the memetic framework

As shown in Section 6.1, our reformulation local search (RLS) method (see Section 4.5) appears to be highly effective for
the E-MSSC problem. We next evaluate if the integration of RLS within a memetic framework can further improve on the

Fig. 4. Percentage gap between the best-known and the average results obtained by the two versions of MA with and without the mutation operator.

Table 19
Analysis of structural similarity between high-quality solutions for 10 large E-MSSC instances.

Instance shq sall slo

pmed23 0.81 0.64 0.52


pmed24 0.84 0.59 0.45
pmed25 0.85 0.61 0.48
pmed28 0.74 0.59 0.35
pmed29 0.82 0.63 0.41
pmed30 0.72 0.57 0.44
pmed33 0.77 0.59 0.43
pmed34 0.83 0.58 0.44
pmed36 0.96 0.75 0.54
pmed37 0.66 0.48 0.39

It acheives the best performance.


286 Q. Zhou et al. / Information Sciences 541 (2020) 271–296

search capacity. For this purpose, we use the same experimental protocol as described in Section 6.1 and report the results of
this study in Tables 17 and 18 for the 20 selected instances. In terms of the best objective value, MA and RLS respectively
yield the best result on 20 and 10 instances. MA reaches a better average objective value in 13 cases, while the average
results of RLS are either equal to or worse than that of MA in all the 20 cases. Moreover, the non-parametric Friedman test
reveals a significant difference between MA and RLS in terms of both the best and the average performance with p-values of
0.007 and 0.001 respectively. The Wilcoxon signed ranks test in Table 18 further discloses the advantage of MA with a 0.05
significance level. This experiment indicates that the memetic element of our algorithm has an important contribution to the
overall algorithm’s performance.

6.3. The influence of the mutation operator

As described in Section 4.3, a random mutation operator is applied to prevent from premature convergence. To evaluate
the importance of the mutation operator, Fig. 4 provides comparative results of the two MA versions with and without the
mutation operator on the selection of 20 large OR-Lib instances (pmed21-40). For each instance and algorithm, the y-axis
shows the percentage gap of the average results to the best-known solutions. These results clearly reveal the importance
of the mutation operator.

6.4. Motivation behind the proposed crossover

To explain the reason behind the efficiency of the proposed crossover, this section provides an analysis of structural sim-
ilarity between local optima of different qualities, in terms of the percentage of commonly shared prototypes. The analysis is
a commonly used and effective tool for justifying the use of the recombination operator [32]. Given two local optima sa and
a b
sb , their similarity is defined as the proportion of the commonly shared prototypes: simðsa ; sb Þ ¼ js js\s j
aj .

This analysis is performed on a random selection of 10 large OR-Lib instances. For each instance, we produce 1000 local
optima of different qualities using MA and its underlying reformulation local search with different time limits. We then take
the top 10% (100) local optima with the smallest objective values to represent the set of ‘high-quality solutions’, and the bot-
tom 10% (100) with the largest objective values to form the set of ‘low-quality solutions’.
Table 19 summarizes the average degree of similarity between the 100 high-quality solutions (column ‘shq ’), the average
degree of similarity between all the 1000 sampled local optima (column ‘sall ’) and the average degree of similarity between
the 100 low-quality solutions (column ‘slo ’). We observe that the average degree of similarity between high-quality solutions
is very large, ranging from 0.66 to 0.96, implying that a significant number of these prototypes might constitute building
blocks of a globally optimal solution.

7. Conclusions

This work proposes the first population-based memetic algorithm (MA) for the Edge Minimum Sum-of-Squares Cluster-
ing (E-MSSC) problem, which integrates a dedicated crossover operator for offspring generation with a reformulation local
search method for local optimization. Computational results on three sets of benchmarks show that the proposed algorithm
is very competitive with the state-of-the-art algorithms from the literature.
An analysis of the key components of the proposed MA reveals the following points. First, the reformulation local search,
that combines the K-means algorithm with tabu search, shows to be a very effective local optimization procedure within our
MA framework. Second, the population based memetic framework is very powerful for tackling E-MSSC. Finally, the muta-
tion operator helps maintain a healthy diversity of the population.
As memetic algorithm shows to be an effective approach for E-MSSC, it is worth considering other population-based evo-
lutionary computing algorithms for the given problem, such as the distributed Evolutionary Algorithms (dEAs) [18,33]. Our
proposed algorithm could also be easily adapted to similar and more complex versions of the network clustering problems
and tested on real-life instances.

CRediT authorship contribution statement

Qing Zhou: Writing - original draft, Software, Visualization. Una Benlic: Writing - review & editing, Supervision. Qinghua
Wu: Conceptualization, Methodology, Project administration, Funding acquisition.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have
appeared to influence the work reported in this paper.
Table 20
Comparisons of MA with five state-of-the-art algorithms on the OR-Lib instances with a time limit of 60 s.

Instance p cost VNS-1 VNS-2 VNS-3 GRASP HYBRID MA


g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g

pmed01 5 434190.94 0.00 0.00 0.20 0.00 0.00 0.00 0.00 0.00 0.00 0.04 0.04 17.91 0.04 0.04 23.21 0.00 0.00 0.09
pmed02 10 252639.75 0.00 0.00 0.90 0.00 0.00 0.10 0.00 0.00 0.00 1.38 1.38 11.27 1.38 1.38 14.56 0.00 0.00 0.99
pmed03 10 259643.17 0.00 0.00 1.20 0.00 0.00 0.10 0.00 0.00 0.00 8.26 8.26 11.99 8.26 8.26 15.29 0.00 0.00 2.21
pmed04 20 145461.30 0.00 0.09 20.90 0.00 0.00 21.70 0.00 0.00 15.40 5.70 5.70 9.22 5.70 5.70 11.89 0.00 0.00 5.11
pmed05 33 39211.95 0.00 0.08 13.70 0.30 0.47 17.80 0.00 0.08 36.50 13.02 13.02 8.90 13.02 13.02 11.43 0.00 0.00 15.66
pmed06 5 383058.00 0.00 0.05 20.10 0.00 0.00 0.00 0.00 0.00 0.00 0.98 0.98 19.01 0.98 0.98 24.60 0.00 0.00 0.04
pmed07 10 212248.00 0.00 0.17 28.00 0.00 0.00 4.20 0.00 0.00 5.10 0.83 0.83 15.22 0.83 0.83 20.06 0.00 0.00 0.25
pmed08 20 149175.44 0.00 0.28 39.90 0.00 0.08 20.50 0.00 0.03 21.10 4.48 4.48 12.70 4.48 4.48 16.54 0.00 0.00 3.85
pmed09 40 60454.76 0.00 1.18 40.20 0.00 0.00 26.20 0.00 0.00 11.70 8.74 8.74 9.93 8.74 8.74 12.43 0.00 0.00 16.83

Q. Zhou et al. / Information Sciences 541 (2020) 271–296


pmed10 67 14478.62 0.01 0.61 39.60 0.01 0.23 35.30 0.00 0.11 25.70 8.66 8.66 9.13 8.66 8.66 12.24 0.00 0.10 29.73
pmed11 5 245749.00 0.00 0.47 22.50 0.00 0.00 0.00 0.00 0.00 0.00 1.21 1.21 19.17 1.21 1.21 25.08 0.00 0.00 0.07
pmed12 10 188983.47 0.00 0.57 43.00 0.00 0.00 0.00 0.00 0.00 0.00 0.67 0.67 17.12 0.67 0.67 22.08 0.00 0.00 0.30
pmed13 30 95693.09 1.48 2.71 49.10 0.00 0.00 16.70 0.00 0.00 15.90 2.36 2.36 11.71 2.36 2.36 15.20 0.00 0.00 15.70
pmed14 60 46978.43 1.39 2.69 50.70 0.13 0.43 36.90 0.26 0.35 32.30 5.52 5.60 9.02 5.45 5.46 26.47 0.00 0.05 37.56
pmed15 100 17078.01 0.65 2.11 53.70 1.07 1.86 46.60 0.28 1.18 41.80 13.78 13.82 8.75 13.78 13.78 25.82 0.00 0.47 42.83
pmed16 5 193375.00 0.00 0.53 28.60 0.00 0.00 0.40 0.00 0.00 0.70 0.11 0.11 19.16 0.11 0.11 24.88 0.00 0.00 0.31
pmed17 10 159496.00 0.00 2.11 39.60 0.00 0.00 0.00 0.00 0.00 0.40 0.00 0.00 15.43 0.00 0.00 20.13 0.00 0.00 0.20
pmed18 40 81817.30 2.66 4.60 51.00 0.03 0.36 30.70 0.00 0.29 46.70 3.93 4.09 9.87 3.90 4.08 28.89 0.00 0.08 40.56
pmed19 80 31732.85 2.44 3.30 50.50 0.30 1.36 49.20 0.32 0.85 45.10 6.97 7.23 8.80 6.82 6.95 25.96 0.00 0.12 42.24
pmed20 133 14316.26 1.67 2.82 53.20 1.19 3.60 48.80 1.02 3.72 53.70 12.19 12.52 8.57 12.19 12.30 25.14 0.00 0.62 40.28
pmed21 5 202777.00 0.00 3.24 40.10 0.00 0.00 1.30 0.00 0.00 1.30 0.00 0.00 19.27 0.00 0.00 24.98 0.00 0.00 0.05
pmed22 10 183329.00 2.73 6.05 37.70 0.00 0.00 3.00 0.00 0.00 5.20 0.43 0.43 15.50 0.43 0.43 20.03 0.00 0.00 0.80
pmed23 50 62775.66 2.89 4.82 49.70 0.08 0.35 36.90 0.04 0.24 48.30 2.41 2.46 9.64 2.46 2.47 12.24 0.00 0.09 31.20
pmed24 400 26941.11 2.95 4.90 52.80 0.26 1.77 48.30 0.19 1.70 47.20 5.99 6.30 8.79 6.00 6.06 11.33 0.00 0.25 50.02
pmed25 167 11703.91 4.04 5.62 52.50 3.61 5.11 57.30 1.76 4.02 57.70 14.57 15.34 8.23 14.57 14.57 10.64 0.00 0.95 50.49
pmed26 5 194461.00 0.39 6.28 35.10 0.00 0.00 1.80 0.00 0.00 1.90 0.00 0.00 16.38 0.00 0.00 21.17 0.00 0.00 0.27
pmed27 10 144800.55 4.67 7.22 25.20 0.00 0.00 24.90 0.00 0.00 16.20 0.06 0.06 15.99 0.06 0.06 20.63 0.00 0.00 14.79
pmed28 60 49849.96 6.49 7.46 46.20 0.19 0.32 38.30 0.15 0.25 48.30 2.48 2.57 8.44 2.49 2.53 10.80 0.00 0.03 46.78
pmed29 120 24396.37 4.29 5.91 53.00 1.61 16.54 55.00 1.20 3.94 59.70 5.25 5.27 7.88 5.06 5.10 10.36 0.00 0.55 48.22
pmed30 200 11084.11 1.94 3.64 52.20 3.26 5.37 59.90 3.43 5.40 59.60 9.86 10.61 7.48 9.64 9.83 9.74 0.00 0.86 51.62
pmed31 5 169806.00 3.22 6.24 27.60 0.00 0.00 3.60 0.00 0.00 3.20 0.00 0.00 17.88 0.00 0.00 23.20 0.00 0.00 0.18
pmed32 10 154078.00 3.54 8.41 35.90 0.00 0.00 7.70 0.00 0.00 6.80 0.70 0.70 15.35 0.70 0.70 19.98 0.00 0.00 0.85
pmed33 70 45269.56 5.81 7.62 52.10 0.03 0.34 48.60 0.05 0.29 46.50 2.16 2.74 7.90 2.69 2.85 10.39 0.00 0.14 45.09
pmed34 140 20357.16 5.05 7.06 50.30 3.48 5.00 58.90 2.82 5.19 58.70 4.24 4.67 7.48 4.40 4.48 9.92 0.00 0.46 44.10
pmed35 5 157540.00 5.46 9.99 21.90 0.00 0.00 3.70 0.00 0.00 4.80 0.00 0.00 17.52 0.00 0.00 22.73 0.00 0.00 0.54
pmed36 10 150918.00 7.77 10.67 29.90 0.00 0.00 6.50 0.00 0.00 6.60 0.00 0.00 16.96 0.00 0.00 22.06 0.00 0.00 1.08
pmed37 80 46719.93 6.65 8.23 44.60 0.18 0.51 55.60 0.60 0.88 57.00 2.87 3.28 8.04 3.07 3.13 10.51 0.00 0.11 50.66
pmed38 5 157685.00 1.91 11.43 25.20 0.17 0.18 11.50 0.17 0.17 14.90 0.00 0.00 19.73 0.00 0.00 25.59 0.17 0.17 23.03
pmed39 10 123424.94 5.72 10.21 32.20 0.00 0.00 7.90 0.00 0.00 10.40 1.10 1.10 18.74 1.10 1.10 24.36 0.00 0.00 13.52
pmed40 90 41583.64 7.41 9.13 47.20 1.01 1.78 59.00 0.66 1.58 58.50 2.67 3.03 8.02 3.14 3.14 10.41 0.00 0.18 46.40

It acheives the best performance.

287
288
Table 21
Comparisons of MA with five state-of-the-art algorithms on the TSP-Lib instances with a time limit of 60 s (part I).

Instance p cost VNS-1 VNS-2 VNS-3 GRASP HYBRID MA


g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g

rd400 10 3.649436e + 07 1.86 4.02 39.30 0.00 0.00 21.70 0.00 0.00 15.70 0.03 0.03 47.97 0.03 0.03 48.14 0.00 0.00 0.75
rd400 50 1.339953e + 07 2.89 4.96 48.90 0.08 0.17 42.70 0.00 0.24 45.00 1.29 1.29 31.56 1.29 1.29 31.82 0.05 0.13 28.56
rd400 100 6.191909e + 06 3.95 5.92 45.20 0.76 1.11 50.60 0.00 0.68 51.60 5.49 5.52 29.36 5.49 5.49 29.75 0.29 1.09 51.29
rd400 200 1.673771e + 06 3.51 3.70 46.70 2.23 3.57 48.60 0.10 2.72 53.60 23.46 23.75 28.19 23.46 23.47 26.43 0.00 2.32 53.16
pr439 10 1.987408e + 09 0.69 4.05 34.20 0.00 0.00 21.80 0.00 0.00 18.70 1.97 1.97 40.73 1.97 1.97 40.49 0.00 0.00 7.75
pr439 50 6.814545e + 08 4.21 6.01 50.90 0.21 0.69 44.20 0.00 0.19 46.40 5.82 5.83 27.83 5.82 5.82 28.16 0.59 0.77 37.73
pr439 100 2.724956e + 08 5.51 7.10 51.60 1.74 3.16 58.50 1.31 4.36 58.70 9.46 9.46 26.44 9.46 9.46 26.18 0.00 1.09 47.87
pr439 200 6.502722e + 07 3.08 5.49 53.30 3.27 9.44 56.00 3.09 10.96 56.20 13.32 13.51 23.24 13.32 13.33 22.71 0.00 2.34 48.34
pcb442 10 3.936676e + 08 1.09 2.86 35.80 0.00 0.00 15.90 0.00 0.00 15.00 0.07 0.07 46.13 0.07 0.07 46.60 0.00 0.00 3.88
pcb442 50 1.306841e + 08 3.55 6.06 48.50 0.01 0.40 41.90 0.13 0.38 41.00 0.82 1.15 27.27 0.82 0.85 28.54 0.00 0.19 48.46
pcb442 100 6.347818e + 07 3.25 5.45 53.60 0.45 1.25 52.40 0.00 0.92 46.50 3.66 3.77 25.03 3.67 3.67 24.98 0.10 0.94 51.03
pcb442 200 1.830759e + 07 0.51 4.56 52.10 0.00 2.34 50.30 1.40 2.17 53.70 18.46 18.98 21.12 18.46 18.46 21.17 1.71 3.77 50.19

Q. Zhou et al. / Information Sciences 541 (2020) 271–296


ali535 10 4.888667e + 05 1.55 7.72 30.70 0.00 0.02 26.10 0.00 0.01 18.30 4.04 4.04 48.24 4.04 4.04 48.40 0.00 0.00 5.28
ali535 50 1.633641e + 05 3.27 7.95 41.90 0.00 0.11 36.00 0.00 0.42 42.80 4.74 4.74 31.09 4.74 4.74 31.78 0.06 0.30 39.71
ali535 100 7.472032e + 04 3.54 8.53 45.30 1.81 2.93 57.40 2.16 3.05 56.10 8.17 8.22 27.52 8.01 8.03 27.88 0.00 1.34 48.21
ali535 200 2.065228e + 04 5.49 9.33 49.80 6.17 15.99 59.60 17.65 23.32 58.20 12.24 12.64 24.18 12.24 12.24 24.23 0.00 2.09 49.45
rat575 10 5.640687e + 06 4.71 6.66 36.20 0.00 0.00 15.70 0.00 0.00 11.40 0.00 0.00 50.80 0.00 0.00 51.30 0.00 0.00 1.00
rat575 50 2.057738e + 06 6.06 7.72 44.40 0.01 0.15 48.10 0.00 0.22 53.70 0.43 0.46 27.87 0.43 0.43 28.72 0.08 0.20 46.57
rat575 100 1.103840e + 06 5.02 6.65 53.40 1.49 2.95 59.50 0.97 3.66 54.40 0.83 0.84 26.07 0.83 0.83 26.85 0.00 0.36 47.85
rat575 200 4.624999e + 05 3.19 4.83 44.20 0.00 2.29 55.30 0.32 1.91 59.30 6.96 6.98 24.72 6.96 6.96 23.54 0.78 1.84 54.04
gr666 10 8.436081e + 05 4.51 9.95 37.30 0.39 0.39 24.60 0.39 0.39 28.50 0.39 0.39 54.37 0.39 0.39 54.43 0.00 0.07 15.38
gr666 50 3.318568e + 05 8.08 11.54 41.00 3.24 8.12 50.30 0.41 5.82 57.80 4.54 4.60 28.55 4.50 4.54 30.82 0.00 0.16 47.51
gr666 100 1.726819e + 05 6.28 9.75 42.10 28.82 37.55 50.70 26.87 38.76 50.30 4.80 5.09 26.43 4.63 4.81 28.19 0.00 0.76 49.37
gr666 200 6.406809e + 04 1.63 4.93 42.30 38.60 49.46 52.50 31.23 54.26 50.30 5.18 5.36 25.12 5.15 5.22 26.21 0.00 1.10 48.60
u724 10 2.076783e + 08 0.86 6.52 31.20 0.00 0.00 26.70 0.00 0.00 19.70 0.19 0.19 54.25 0.19 0.19 55.12 0.00 0.00 2.42
u724 50 8.865043e + 07 6.28 8.91 45.80 0.00 0.32 51.00 0.00 0.17 55.20 0.25 0.35 27.46 0.25 0.27 28.64 0.29 0.55 52.02
u724 100 5.118373e + 07 5.47 9.11 45.20 0.00 0.46 59.00 0.21 0.63 59.60 0.41 0.43 26.59 0.41 0.41 27.84 0.11 0.88 43.63
u724 200 2.302944e + 07 2.67 4.98 56.20 3.72 6.02 58.50 2.03 6.22 56.20 4.98 5.26 24.57 4.95 5.03 25.04 0.00 0.97 41.38
rat783 10 7.799115e + 06 2.45 4.52 32.60 0.00 0.00 24.60 0.00 0.02 26.80 0.10 0.10 34.28 0.10 0.10 34.81 0.00 0.00 6.24
rat783 50 3.371065e + 06 9.82 11.66 49.40 0.59 2.14 50.40 0.72 2.59 54.30 0.00 0.08 28.30 0.00 0.02 30.00 0.27 0.51 43.66
rat783 100 1.991687e + 06 6.08 9.82 40.10 4.36 5.32 51.10 4.26 6.01 55.90 0.08 0.11 26.23 0.08 0.08 27.55 0.00 0.54 56.23
rat783 200 9.332400e + 05 3.88 5.67 37.60 10.96 13.16 51.90 9.49 12.27 51.80 2.24 2.28 24.87 2.20 2.24 25.80 0.00 1.51 37.87
dsj1000 10 2.897240e + 13 1.48 7.75 33.30 0.00 0.00 35.20 0.00 0.01 37.50 0.83 1.25 27.81 0.56 1.08 30.26 0.00 0.00 18.96
dsj1000 50 1.267460e + 13 10.97 13.34 41.60 0.25 0.49 57.50 0.00 0.39 52.70 1.82 3.40 30.91 1.82 3.33 33.88 0.08 0.39 43.94
dsj1000 100 7.819500e + 12 8.28 10.77 39.00 1.70 2.92 50.80 2.00 3.34 56.10 1.47 1.82 38.11 1.56 1.92 41.28 0.00 0.53 28.93
dsj1000 200 4.075460e + 12 5.40 7.04 41.90 9.84 12.48 50.30 10.63 12.76 51.10 2.60 3.73 56.62 2.59 3.76 56.88 0.00 0.79 41.15
pr1002 10 5.540610e + 09 1.14 5.33 31.30 0.00 0.00 18.00 0.00 0.00 23.10 0.05 0.05 55.07 0.05 0.05 55.45 0.00 0.00 26.91
pr1002 50 2.534914e + 09 10.55 13.48 37.80 5.13 6.00 50.99 4.49 6.06 52.80 0.36 0.37 27.28 0.36 0.36 28.81 0.00 0.15 55.44
pr1002 100 1.587449e + 09 8.15 10.21 43.90 7.41 8.23 56.60 7.69 9.01 52.10 0.73 0.75 24.65 0.73 0.75 26.47 0.00 0.57 43.38
pr1002 200 8.426084e + 08 4.57 6.86 47.90 19.82 21.20 58.10 18.52 20.92 50.80 1.43 1.57 23.00 1.42 1.42 24.77 0.00 0.64 41.91
Comparisons of MA with five state-of-the-art algorithms on the TSP-Lib instances with a time limit of 60 s (part II).
Instance p 20ptcost VNS-1 VNS-2 VNS-3 GRASP HYBRID MA
g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g

pcb1173 10 2.898943e + 08 3.03 10.67 49.20 0.00 0.00 37.60 0.00 0.02 44.50 0.29 0.29 50.22 0.29 0.29 50.55 0.00 0.01 28.04
pcb1173 50 1.237713e + 08 12.68 15.29 35.00 9.73 12.50 51.90 8.35 10.83 52.60 0.04 0.04 30.00 0.04 0.04 32.09 0.00 0.26 43.91
Table 21 (continued)

Instance p cost VNS-1 VNS-2 VNS-3 GRASP HYBRID MA


g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g

pcb1173 100 7.949999e + 07 10.16 11.83 49.50 10.85 13.13 51.70 11.67 13.92 53.40 0.02 0.06 25.14 0.00 0.00 26.45 0.44 0.79 29.48
pcb1173 200 4.491868e + 07 6.01 8.38 47.50 16.39 17.02 51.00 13.50 14.74 55.50 0.07 0.23 23.86 0.00 0.01 25.18 0.32 1.42 45.95
rl1304 10 1.014519e + 10 4.72 15.64 50.90 0.00 0.00 40.80 0.00 0.00 53.70 0.11 0.11 56.32 0.11 0.11 56.78 0.00 0.00 24.39
rl1304 50 4.724233e + 09 12.95 15.26 41.70 2.37 3.48 55.50 2.03 2.27 52.80 0.42 0.42 31.88 0.42 0.42 34.00 0.00 0.29 32.82
rl1304 100 3.070192e + 09 11.75 13.98 39.40 3.54 7.89 56.50 8.13 9.08 52.20 0.15 0.16 26.28 0.15 0.15 27.74 0.00 0.34 55.16
rl1304 200 1.768823e + 09 10.50 12.71 28.30 13.91 15.08 56.30 13.52 15.58 54.40 0.82 0.96 24.43 0.82 0.83 24.73 0.00 1.00 50.72
nrw1379 10 2.104002e + 08 4.91 11.06 53.30 0.00 0.00 53.80 0.00 0.00 59.30 0.00 0.00 53.40 0.00 0.00 53.13 0.00 0.00 18.80
nrw1379 50 9.565609e + 07 12.38 16.71 41.40 36.95 37.46 58.60 36.88 37.17 51.30 0.34 0.39 30.53 0.34 0.36 32.72 0.00 0.37 24.54
nrw1379 100 6.196648e + 07 12.56 15.06 36.40 43.69 45.35 58.00 45.50 46.25 51.90 0.00 0.00 27.38 0.00 0.00 28.52 0.12 1.18 45.36
nrw1379 200 3.618693e + 07 8.03 9.32 34.80 33.01 35.71 57.70 31.12 33.97 53.10 0.00 0.20 25.03 0.10 0.11 26.19 0.20 1.02 34.30
fl1400 10 3.427747e + 07 43.93 128.50 59.90 0.00 89.24 52.30 0.00 54.92 54.20 4.13 4.13 39.29 4.13 4.13 39.35 0.00 0.01 27.41
fl1400 50 1.073865e + 07 9.21 19.87 55.60 1.10 2.47 51.60 2.63 5.83 58.30 14.62 14.62 29.97 14.62 14.62 31.32 0.00 0.43 55.04
fl1400 100 5.243616e + 06 11.96 17.12 55.00 10.39 23.07 50.20 18.33 24.94 53.60 18.14 18.14 30.30 18.14 18.14 32.00 0.00 1.40 43.05

Q. Zhou et al. / Information Sciences 541 (2020) 271–296


fl1400 200 1.876087e + 06 8.54 10.12 53.00 94.93 110.88 50.00 90.10 117.61 60.00 10.15 10.26 26.13 9.64 9.86 26.98 0.00 1.32 32.58
u1432 10 1.525047e + 09 3.97 11.87 43.20 0.00 0.33 53.10 0.00 5.06 53.70 0.00 0.00 47.31 0.00 0.00 47.52 0.00 0.00 42.36
u1432 50 7.039039e + 08 13.80 17.68 35.50 12.20 15.55 50.70 11.12 13.08 54.10 0.00 0.08 30.10 0.02 0.02 32.25 0.32 0.63 38.67
u1432 100 4.540743e + 08 12.50 14.69 36.60 15.89 17.56 59.60 15.07 16.98 51.80 0.00 0.16 26.63 0.12 0.14 27.38 0.29 1.01 37.96
u1432 200 2.644735e + 08 9.57 11.47 30.00 19.21 21.74 51.30 15.61 18.63 55.20 0.00 0.17 24.74 0.02 0.04 25.35 1.06 2.07 41.92
fl1577 10 8.823213e + 07 86.46 248.11 54.90 0.00 127.62 55.60 66.74 129.72 58.40 0.68 0.68 51.94 0.68 0.68 52.07 0.00 0.07 23.82
fl1577 50 3.706607e + 07 19.98 25.00 59.30 40.60 56.66 50.10 46.09 54.06 60.00 1.09 1.11 28.81 1.09 1.09 30.39 0.00 0.46 36.06
fl1577 100 2.383941e + 07 12.07 19.70 58.60 32.55 43.11 59.50 43.57 48.10 52.40 3.63 3.72 26.69 3.58 3.63 27.89 0.00 0.24 33.23
fl1577 200 1.259517e + 07 10.87 12.85 57.60 71.85 83.09 59.10 78.64 108.74 59.80 3.75 4.07 25.92 3.67 3.68 26.66 0.00 0.69 56.90
d1655 10 3.062937e + 08 26.31 82.20 59.50 12.50 40.74 57.10 11.84 40.41 59.20 0.12 0.12 56.73 0.12 0.12 56.57 0.00 0.13 23.41
d1655 50 1.443487e + 08 21.84 37.75 58.30 20.52 26.85 60.00 18.64 24.49 60.00 1.05 1.05 30.69 1.05 1.05 31.62 0.00 0.36 32.25
d1655 100 9.659401e + 07 25.26 33.83 55.60 20.74 26.90 60.00 26.06 33.18 56.10 0.03 0.06 25.03 0.00 0.05 25.97 0.33 0.97 39.30
d1655 200 5.757109e + 07 46.00 46.44 57.50 29.00 37.81 55.30 38.51 40.12 55.30 0.00 0.09 23.15 0.06 0.06 24.08 0.63 1.39 50.58
vm1748 10 1.206434e + 10 6.84 37.46 58.80 0.00 25.93 56.20 13.60 28.16 56.30 0.16 0.16 57.95 0.16 0.16 58.10 0.00 0.00 36.14
vm1748 50 4.774759e + 09 103.95 104.67 57.00 68.76 78.50 53.90 71.48 80.80 57.60 0.32 0.59 39.47 0.44 0.51 42.70 0.00 0.18 26.98
vm1748 100 3.192875e + 09 20.85 21.28 54.00 109.51 110.78 58.20 16.48 83.34 52.90 0.91 0.91 26.88 0.90 0.90 28.66 0.00 0.87 28.68
vm1748 200 1.902595e + 09 128.66 129.54 51.40 41.32 93.74 56.50 36.98 87.97 54.20 2.15 2.26 23.59 2.09 2.15 25.10 0.00 0.81 49.68
rl1889 10 1.275225e + 10 72.40 83.00 28.50 36.88 46.28 59.60 28.73 49.12 52.50 0.02 0.02 57.77 0.02 0.02 57.87 0.00 0.00 53.20
rl1889 50 5.942926e + 09 40.40 44.61 54.50 34.75 40.39 58.30 36.17 39.12 55.20 0.00 0.03 32.18 0.00 0.00 34.55 0.24 0.42 43.85
rl1889 100 4.041262e + 09 39.13 39.62 52.70 37.86 38.34 50.50 35.88 36.09 57.90 0.43 0.53 27.02 0.38 0.40 28.29 0.00 0.54 48.25
rl1889 200 2.452274e + 09 37.37 37.99 50.30 35.06 36.17 53.10 34.94 35.71 54.60 0.00 0.11 25.93 0.03 0.04 28.02 0.67 1.23 52.31
pcb3038 10 1.003337e + 09 26.98 27.34 50.70 23.01 24.29 59.20 32.24 32.34 59.50 0.00 0.00 49.46 0.00 0.00 50.52 0.52 1.04 56.03
pcb3038 50 4.364877e + 08 129.19 136.43 53.70 34.85 37.24 50.80 33.39 33.81 54.50 0.00 0.00 29.53 0.00 0.00 33.41 0.57 0.92 52.70
pcb3038 100 2.994432e + 08 41.47 45.48 58.10 46.40 53.27 53.60 46.49 43.40 52.90 0.00 0.04 27.17 0.02 0.04 30.02 1.20 1.41 59.15
pcb3038 200 1.979634e + 08 39.71 43.35 49.90 45.17 45.55 56.80 40.17 41.36 55.40 0.11 0.21 25.06 0.00 0.05 27.16 0.96 1.36 59.31

It acheives the best performance.

289
290
Table 22
Comparisons of MA with five state-of-the-art algorithms on the University of Florida Sparse Matrix instances with a time limit of 60 s.

Q. Zhou et al. / Information Sciences 541 (2020) 271–296


Instance p cost VNS-1 VNS-2 VNS-3 GRASP HYBRID MA
g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g

Harvard500 10 643.00 0.00 2.32 33.70 0.00 0.00 0.10 0.00 0.00 0.30 0.00 0.00 34.75 0.00 0.00 34.01 0.00 0.00 0.07
Harvard500 50 434.00 0.46 1.21 48.70 0.23 0.43 35.20 0.00 0.08 28.80 3.69 3.69 15.88 3.69 3.69 16.25 0.00 0.14 17.53
Harvard500 100 357.50 0.14 0.39 32.90 0.42 1.17 37.70 0.00 0.14 28.90 11.89 11.89 9.34 11.89 11.89 9.92 0.14 0.39 37.02
Harvard500 200 300.00 0.00 0.00 51.26 0.00 0.00 42.39 0.00 0.00 57.64 0.00 0.00 5.50 0.00 0.00 4.91 0.00 0.00 48.32
USAir97 10 472.94 3.88 5.89 43.60 0.65 0.70 9.40 0.63 0.63 1.60 0.86 0.86 36.35 0.86 0.86 34.65 0.00 0.24 7.49
USAir97 50 264.00 0.95 1.74 42.10 1.14 1.53 20.20 0.19 0.38 19.80 6.82 6.82 17.75 6.82 6.82 17.93 0.00 0.25 28.27
USAir97 100 189.00 0.00 0.11 14.70 0.53 1.08 31.30 0.00 0.00 11.00 22.75 22.75 9.63 22.75 22.75 9.03 0.26 0.69 39.22
USAir97 200 132.00 0.00 0.00 47.64 0.00 0.00 45.24 0.00 0.00 54.18 0.00 0.00 6.62 0.00 0.00 6.87 0.00 0.00 35.16
Email 10 4145.09 4.07 7.21 39.80 0.12 0.24 12.80 0.00 0.21 25.50 0.26 0.27 34.90 0.26 0.26 35.13 0.09 0.12 8.24
Email 50 2109.50 8.38 13.39 46.20 0.43 6.65 54.20 0.58 6.12 57.00 0.26 0.30 22.36 0.26 0.26 22.16 0.00 0.34 43.38
Email 100 1456.00 11.23 13.19 50.40 0.71 3.67 58.20 1.61 4.52 57.40 0.00 0.10 19.79 0.00 0.00 21.21 0.19 0.59 34.18
Email 200 935.50 15.93 17.32 52.00 11.20 13.50 59.80 7.27 12.31 58.10 2.94 5.28 16.02 2.94 3.00 14.39 0.00 0.70 48.96

It acheives the best performance.


Table 23
Comparisons of MA with five state-of-the-art algorithms on the OR-Lib instances with a time limit of 600 s.

Instance p cost VNS-1 VNS-2 VNS-3 GRASP HYBRID MA


g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g

pmed01 5 434190.94 0.00 0.00 0.10 0.00 0.00 0.00 0.00 0.00 0.00 0.04 0.04 26.24 0.04 0.04 34.10 0.00 0.00 0.27
pmed02 10 252639.75 0.00 0.00 0.30 0.00 0.00 0.00 0.00 0.00 0.00 1.38 1.38 16.65 1.38 1.38 21.47 0.00 0.00 7.42
pmed03 10 259643.17 0.00 0.00 0.60 0.00 0.00 0.10 0.00 0.00 0.00 8.26 8.26 17.29 8.26 8.26 22.39 0.00 0.00 1.48
pmed04 20 145461.30 0.00 0.02 119.00 0.00 0.00 74.70 0.00 0.00 189.00 5.70 5.70 13.38 5.70 5.70 17.38 0.00 0.00 28.42
pmed05 33 39211.95 0.00 0.00 42.60 0.00 0.00 156.70 0.00 0.00 207.00 13.02 13.02 12.92 13.02 13.02 16.69 0.00 0.00 41.68
pmed06 5 383058.00 0.00 0.00 10.60 0.00 0.00 0.00 0.00 0.00 0.00 0.98 0.98 27.99 0.98 0.98 36.53 0.00 0.00 0.01
pmed07 10 212248.00 0.00 0.00 27.00 0.00 0.00 1.60 0.00 0.00 4.90 0.83 0.83 22.47 0.83 0.83 29.56 0.00 0.00 0.33
pmed08 20 149175.44 0.00 0.00 59.20 0.00 0.00 145.90 0.00 0.00 46.10 4.48 4.48 18.81 4.48 4.48 24.68 0.00 0.00 20.88
pmed09 40 60454.76 0.00 0.09 203.30 0.00 0.00 105.20 0.00 0.00 48.20 8.74 8.74 14.55 8.74 8.74 18.83 0.00 0.00 18.54

Q. Zhou et al. / Information Sciences 541 (2020) 271–296


pmed10 67 14478.62 0.00 0.04 392.40 0.00 0.01 282.30 0.00 0.00 350.10 8.66 8.66 13.52 8.66 8.66 17.46 0.00 0.00 253.51
pmed11 5 245749.00 0.00 0.00 25.40 0.00 0.00 0.00 0.00 0.00 0.00 1.21 1.21 28.21 1.21 1.21 36.62 0.00 0.00 0.02
pmed12 10 188983.47 0.00 0.16 71.10 0.00 0.00 0.00 0.00 0.00 0.00 0.67 0.67 25.17 0.67 0.67 32.81 0.00 0.00 0.46
pmed13 30 95693.09 0.22 0.62 426.10 0.00 0.00 123.40 0.00 0.00 77.50 2.36 2.36 17.28 2.36 2.36 22.43 0.00 0.00 97.36
pmed14 60 46978.43 0.00 0.56 463.50 0.07 0.22 316.10 0.05 0.14 327.80 5.41 5.44 78.84 5.41 5.43 92.46 0.00 0.03 307.29
pmed15 100 17078.01 0.24 0.66 487.40 0.47 0.81 356.40 0.00 0.29 367.30 13.93 13.93 25.71 13.93 13.93 26.78 0.00 0.35 505.02
pmed16 5 193375.00 0.00 0.00 161.60 0.00 0.00 0.10 0.00 0.00 0.20 0.11 0.11 28.18 0.11 0.11 36.65 0.00 0.00 0.17
pmed17 10 159496.00 0.00 0.46 271.70 0.00 0.00 0.20 0.00 0.00 0.40 0.00 0.00 22.66 0.00 0.00 29.56 0.00 0.00 0.12
pmed18 40 81817.30 0.79 1.48 500.80 0.00 0.00 143.50 0.00 0.00 304.30 3.64 3.87 43.09 3.65 3.65 68.78 0.00 0.01 141.86
pmed19 80 31732.85 0.57 1.40 452.20 0.08 0.46 354.60 0.06 0.28 292.60 6.66 6.92 62.13 6.82 6.92 46.79 0.00 0.06 349.92
pmed20 133 14316.26 0.52 0.98 555.90 0.65 1.18 274.20 0.30 0.57 301.00 12.75 13.11 30.05 12.90 12.90 16.25 0.00 0.43 520.26
pmed21 5 202777.00 0.00 0.51 244.40 0.00 0.00 1.00 0.00 0.00 0.80 0.00 0.00 28.38 0.00 0.00 36.86 0.00 0.00 0.04
pmed22 10 183329.00 0.82 1.86 426.30 0.00 0.00 2.90 0.00 0.00 5.00 0.43 0.43 22.71 0.43 0.43 29.46 0.00 0.00 1.89
pmed23 50 62775.66 0.62 2.00 530.90 0.02 0.08 314.70 0.00 0.03 328.60 2.41 2.44 25.14 2.41 2.43 18.20 0.00 0.02 358.63
pmed24 400 26941.11 1.47 2.67 539.40 0.29 0.50 411.60 0.18 0.35 439.80 6.11 6.34 56.20 6.17 6.17 86.74 0.00 0.32 327.61
pmed25 167 11703.91 0.81 1.78 520.10 0.26 0.91 335.80 0.12 0.47 316.70 14.58 15.02 40.59 14.54 14.56 65.32 0.00 0.36 373.30
pmed26 5 194461.00 0.00 2.34 415.80 0.00 0.00 1.80 0.00 0.00 1.90 0.00 0.00 24.14 0.00 0.00 31.34 0.00 0.00 0.45
pmed27 10 144800.55 0.39 2.41 441.00 0.00 0.00 196.60 0.00 0.00 27.80 0.06 0.06 23.57 0.06 0.06 30.55 0.00 0.00 24.01
pmed28 60 49849.96 2.78 3.52 524.80 0.00 0.14 304.80 0.10 0.40 343.30 2.34 2.49 87.48 2.34 2.34 106.11 0.00 0.06 342.19
pmed29 120 24396.37 1.86 2.78 531.50 0.28 1.10 410.70 0.19 1.00 493.70 5.43 5.59 61.88 5.43 5.43 95.29 0.00 0.46 382.98
pmed30 200 11084.11 1.43 2.04 536.10 0.36 1.85 469.90 0.53 1.81 524.50 10.88 11.82 24.43 11.19 11.19 14.26 0.00 1.16 445.69
pmed31 5 169806.00 0.02 1.33 356.10 0.00 0.00 2.20 0.00 0.00 4.10 0.00 0.00 26.35 0.00 0.00 34.24 0.00 0.00 0.59
pmed32 10 154078.00 1.32 3.16 408.90 0.00 0.00 3.00 0.00 0.00 7.70 0.70 0.70 22.64 0.70 0.70 29.32 0.00 0.00 4.69
pmed33 70 45269.56 3.18 4.77 528.40 0.02 0.14 371.30 0.06 0.19 393.60 2.03 2.37 98.93 2.03 2.03 125.16 0.00 0.14 407.31
pmed34 140 20357.16 2.98 4.11 519.40 0.65 1.72 526.90 0.30 0.95 413.80 4.56 4.78 51.84 4.43 4.57 74.39 0.00 0.55 425.70
pmed35 5 157540.00 1.59 3.41 279.90 0.00 0.00 4.70 0.00 0.00 7.10 0.00 0.00 25.97 0.00 0.00 33.53 0.00 0.00 0.22
pmed36 10 150918.00 3.09 4.27 426.40 0.00 0.00 6.80 0.00 0.00 13.20 0.00 0.00 25.07 0.00 0.00 32.45 0.00 0.00 1.69
pmed37 80 46719.93 3.29 4.94 527.60 0.05 0.20 235.80 0.04 0.26 363.00 2.85 3.18 57.38 2.65 2.86 87.41 0.00 0.17 402.94
pmed38 5 157685.00 0.61 4.99 330.30 0.17 0.17 168.60 0.17 0.17 13.30 0.00 0.00 29.16 0.00 0.00 37.80 0.17 0.17 169.13
pmed39 10 123424.94 1.62 3.47 385.50 0.00 0.00 14.70 0.00 0.00 16.30 1.10 1.10 27.85 1.10 1.10 35.94 0.00 0.00 162.78
pmed40 90 41583.64 3.69 5.99 544.40 0.09 0.22 475.90 0.09 0.26 440.70 2.54 2.56 36.38 2.51 2.52 55.35 0.00 0.13 322.52

It acheives the best performance.

291
292
Table 24
Comparisons of MA with five state-of-the-art algorithms on the TSP-Lib instances with a time limit of 600 s (part I).

Instance p cost VNS-1 VNS-2 VNS-3 GRASP HYBRID MA


g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g

rd400 10 3.649436e + 07 0.04 0.95 387.40 0.00 0.00 148.40 0.00 0.00 93.10 0.03 0.03 52.98 0.03 0.03 60.10 0.00 0.00 0.85
rd400 50 1.339953e + 07 0.62 1.48 516.20 0.00 0.01 333.30 0.00 0.01 446.00 1.29 1.29 34.90 1.29 1.29 39.54 0.03 0.05 205.98
rd400 100 6.176798e + 06 0.80 1.84 550.30 0.40 0.73 301.90 0.13 0.32 178.50 5.74 5.74 53.38 5.74 5.74 36.43 0.00 0.40 418.38
rd400 200 1.614626e + 06 2.11 3.08 468.50 0.08 1.53 389.30 0.00 0.96 301.68 27.99 28.17 49.77 27.99 27.99 43.53 0.00 0.80 506.80
pr439 10 1.987408e + 09 0.06 0.47 481.20 0.00 0.00 6.30 0.00 0.00 44.90 1.97 1.97 44.43 1.97 1.97 50.60 0.00 0.00 177.25
pr439 50 6.798665e + 08 2.02 3.21 534.10 0.38 0.51 368.80 0.00 0.13 301.30 6.07 6.07 81.00 6.07 6.07 34.38 0.47 0.70 389.71
pr439 100 2.719918e + 08 1.59 2.65 531.70 0.06 0.23 368.70 0.00 0.16 397.20 9.66 9.66 25.30 9.66 9.66 32.15 0.05 0.41 375.66
pr439 200 6.317237e + 07 2.07 2.84 539.20 0.41 2.08 388.50 0.00 1.35 330.50 16.65 16.71 24.62 16.65 16.65 127.21 0.59 1.78 469.50
pcb442 10 3.936676e + 08 0.00 0.47 414.40 0.00 0.00 39.70 0.00 0.00 60.60 0.07 0.07 51.24 0.07 0.07 58.38 0.00 0.00 0.22
pcb442 50 1.306005e + 08 2.31 3.06 538.00 0.06 0.20 281.80 0.14 0.19 286.50 0.88 1.02 98.88 0.88 0.88 134.36 0.00 0.08 319.60
pcb442 100 6.287998e + 07 1.02 2.32 548.70 0.52 0.92 270.60 0.00 0.38 239.20 4.61 4.70 127.31 4.63 4.63 131.11 0.63 1.09 492.00
pcb442 200 1.773801e + 07 1.05 2.74 541.00 1.92 3.07 263.80 0.00 1.47 307.60 22.26 22.62 83.05 22.26 22.26 26.11 1.79 2.90 510.91

Q. Zhou et al. / Information Sciences 541 (2020) 271–296


ali535 10 4.888667e + 05 0.40 2.18 385.80 0.00 0.00 105.70 0.00 0.00 54.50 4.04 4.04 53.23 4.04 4.04 60.56 0.00 0.00 5.82
ali535 50 1.633641e + 05 2.47 4.14 485.70 0.00 0.02 281.00 0.00 0.00 204.80 4.74 4.74 35.59 4.74 4.74 39.65 0.00 0.09 263.54
ali535 100 7.429903e + 04 2.18 3.23 540.30 0.44 0.60 440.70 0.00 0.21 420.80 8.62 8.73 152.41 8.62 8.67 35.04 0.46 1.49 459.25
ali535 200 2.013859e + 04 4.05 4.81 572.80 1.04 2.87 483.00 0.00 2.05 564.60 15.10 15.34 75.67 15.10 15.10 30.18 0.01 1.69 515.73
rat575 10 5.640687e + 06 1.11 2.24 469.30 0.00 0.00 136.50 0.00 0.00 246.30 0.00 0.00 56.48 0.00 0.00 64.02 0.00 0.00 0.83
rat575 50 2.057818e + 06 1.80 2.79 534.10 0.01 0.04 206.80 0.01 0.03 259.50 0.43 0.44 58.04 0.38 0.40 135.73 0.00 0.04 314.03
rat575 100 1.095323e + 06 2.36 3.29 512.40 0.15 0.41 409.30 0.00 0.18 369.60 1.62 1.62 99.49 1.62 1.62 33.68 0.18 0.37 493.06
rat575 200 4.511390e + 05 2.33 3.91 562.20 1.23 1.77 454.20 0.00 0.72 479.10 9.66 9.66 136.85 9.66 9.66 29.58 0.69 1.27 537.40
gr666 10 8.436081e + 05 3.28 5.00 403.10 0.00 0.35 84.30 0.00 0.31 114.70 0.39 0.39 59.76 0.39 0.39 68.66 0.00 0.00 43.06
gr666 50 3.314771e + 05 5.27 7.29 513.70 0.04 0.04 228.00 0.00 0.02 340.10 4.50 4.56 203.39 4.50 4.54 138.17 0.01 0.09 371.03
gr666 100 1.707351e + 05 4.94 6.11 502.10 0.19 0.65 426.40 0.00 0.65 471.70 5.80 5.96 230.00 5.72 5.74 234.66 0.21 0.48 483.81
gr666 200 6.142418e + 04 1.56 3.28 571.90 0.95 2.51 569.80 0.13 0.94 527.90 9.62 9.74 88.71 9.63 9.64 252.67 0.00 0.63 489.84
u724 10 2.076783e + 08 0.40 1.88 390.80 0.00 0.00 85.80 0.00 0.00 137.40 0.19 0.19 60.63 0.19 0.19 68.79 0.00 0.00 2.38
u724 50 8.860866e + 07 3.44 4.73 530.40 0.00 0.03 199.20 0.00 0.04 235.70 0.30 0.30 131.47 0.30 0.30 155.79 0.00 0.19 386.17
u724 100 5.069868e + 07 4.31 5.79 521.80 0.28 0.50 448.00 0.16 0.55 502.30 1.37 1.37 110.54 1.37 1.37 34.91 0.00 0.19 493.43
u724 200 2.218486e + 07 3.13 5.20 571.10 0.94 1.70 509.40 0.00 0.61 520.50 8.78 9.16 158.04 9.02 9.04 30.99 0.26 0.95 511.55
rat783 10 7.799115e + 06 0.51 1.81 440.40 0.00 0.00 74.40 0.00 0.00 125.50 0.10 0.10 37.95 0.10 0.10 43.36 0.00 0.00 5.00
rat783 50 3.365960e + 06 5.35 7.49 532.60 0.00 0.23 394.90 0.00 0.15 359.30 0.15 0.15 92.62 0.15 0.15 186.38 0.00 0.07 309.04
rat783 100 1.970780e + 06 4.67 6.98 548.40 0.02 0.20 470.80 0.00 0.21 428.40 1.06 1.13 180.01 1.01 1.11 224.32 0.21 0.41 478.36
rat783 200 9.045165e + 05 2.40 4.88 553.10 1.01 1.66 563.40 0.00 0.96 548.30 5.44 5.48 188.99 5.46 5.47 32.61 0.43 0.98 494.25
dsj1000 10 2.897240e + 13 0.05 1.30 401.40 0.00 0.00 143.50 0.00 0.00 225.80 0.55 1.09 33.15 0.41 0.95 248.00 0.00 0.00 226.25
dsj1000 50 1.267510e + 13 6.44 8.42 513.90 0.05 0.16 332.50 0.05 0.25 387.40 1.50 2.83 37.33 1.50 2.64 252.39 0.00 0.05 455.76
dsj1000 100 7.749330e + 12 6.65 8.07 480.60 0.12 0.32 468.70 0.00 0.24 448.50 2.45 2.67 45.33 2.45 2.65 251.65 0.31 0.55 476.62
dsj1000 200 3.965550e + 12 5.18 6.06 555.80 1.21 1.48 578.00 0.00 0.87 505.50 4.07 4.99 62.53 4.07 4.81 271.06 0.22 0.94 546.09
pr1002 10 5.540610e + 09 0.00 1.76 474.50 0.00 0.00 37.20 0.00 0.00 38.60 0.05 0.05 61.07 0.05 0.05 69.43 0.00 0.00 202.22
pr1002 50 2.529714e + 09 5.06 7.77 451.30 0.00 0.08 288.60 0.00 0.10 410.90 0.56 0.56 91.53 0.56 0.56 35.40 0.00 0.10 337.37
pr1002 100 1.567371e + 09 5.40 7.20 529.10 0.22 0.42 418.40 0.00 0.27 476.40 2.01 2.02 167.65 2.00 2.01 232.84 0.44 0.78 525.14
pr1002 200 8.238555e + 08 4.65 6.03 568.70 3.08 5.15 504.90 4.30 4.98 504.80 3.51 3.60 286.74 3.52 3.56 130.72 0.00 0.72 502.83
Comparisons of MA with five state-of-the-art algorithms on the TSP-Lib instances with a time limit of 600 s (part II).
Instance p cost VNS-1 VNS-2 VNS-3 GRASP HYBRID MA
g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g

pcb1173 10 2.898943e + 08 0.94 3.52 414.20 0.00 0.00 97.20 0.00 0.00 79.50 0.29 0.29 55.66 0.29 0.29 63.38 0.00 0.00 78.87
pcb1173 50 1.234250e + 08 7.47 9.20 457.00 0.00 0.01 319.20 0.00 0.00 406.80 0.32 0.32 35.05 0.32 0.32 39.81 0.04 0.13 383.38
Table 24 (continued)

Instance p cost VNS-1 VNS-2 VNS-3 GRASP HYBRID MA


g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g

pcb1173 100 7.888991e + 07 6.11 7.95 543.50 0.29 0.94 596.10 0.07 1.63 505.70 0.72 0.77 188.93 0.72 0.72 282.88 0.00 0.28 461.09
pcb1173 200 4.435643e + 07 5.12 6.73 519.50 3.03 4.13 505.70 3.20 4.31 506.80 1.30 1.35 87.61 1.25 1.28 182.49 0.00 0.54 545.05
rl1304 10 1.014519e + 10 0.26 5.09 273.50 0.00 0.00 160.00 0.00 0.00 123.70 0.11 0.11 62.40 0.11 0.11 70.95 0.00 0.00 144.48
rl1304 50 4.713066e + 09 9.24 10.98 515.50 0.00 0.04 313.10 0.00 0.07 322.20 0.66 0.66 38.12 0.66 0.66 42.70 0.00 0.18 508.52
rl1304 100 3.052560e + 09 8.39 11.13 516.90 0.00 0.21 479.00 0.12 0.23 528.80 0.73 0.73 85.32 0.73 0.73 34.72 0.09 0.22 460.47
rl1304 200 1.745278e + 09 7.39 9.02 501.90 0.78 2.00 506.30 2.12 2.68 505.40 2.01 2.14 158.57 2.07 2.12 293.56 0.00 0.58 463.33
nrw1379 10 2.104002e + 08 2.50 4.11 440.30 0.00 0.00 272.10 0.00 0.00 241.90 0.00 0.00 58.96 0.00 0.00 67.02 0.00 0.00 22.24
nrw1379 50 9.551156e + 07 7.53 10.91 396.60 0.02 0.94 548.00 0.04 1.09 504.80 0.49 0.51 95.45 0.49 0.49 94.45 0.00 0.11 370.68
nrw1379 100 6.161060e + 07 7.85 10.40 494.40 23.51 26.35 580.70 24.33 25.86 502.70 0.58 0.58 31.50 0.58 0.58 35.64 0.00 0.36 485.23
nrw1379 200 3.569954e + 07 6.41 7.66 514.80 20.56 23.78 504.70 20.74 23.38 505.20 1.31 1.43 108.25 1.39 1.40 183.15 0.00 0.50 512.08
fl1400 10 3.427747e + 07 0.05 2.15 405.20 0.00 0.00 129.70 0.00 0.00 126.50 4.13 4.13 43.09 4.13 4.13 49.37 0.00 0.00 112.98
fl1400 50 1.071524e + 07 10.74 12.96 356.50 0.00 0.05 292.20 0.00 0.17 433.60 14.87 14.87 34.63 14.87 14.87 38.89 0.00 0.07 272.14
fl1400 100 5.211615e + 06 8.55 12.36 364.80 0.62 1.34 560.20 0.00 1.81 570.90 18.63 18.82 194.13 18.62 18.68 138.76 0.26 0.98 518.16

Q. Zhou et al. / Information Sciences 541 (2020) 271–296


fl1400 200 1.868231e + 06 3.30 7.33 450.30 19.89 28.78 509.00 25.26 28.66 507.70 10.24 10.45 230.02 10.14 10.14 134.21 0.00 0.66 469.56
u1432 10 1.525047e + 09 2.54 4.73 310.10 0.00 0.00 220.10 0.00 0.00 257.50 0.00 0.00 52.03 0.00 0.00 59.70 0.00 0.00 282.81
u1432 50 7.027056e + 08 11.68 12.58 454.60 0.01 0.09 443.20 0.00 0.19 469.50 0.17 0.21 85.14 0.17 0.17 239.54 0.00 0.08 451.61
u1432 100 4.526464e + 08 8.72 11.05 495.40 5.27 6.31 504.50 5.43 6.49 505.00 0.32 0.43 129.91 0.32 0.32 234.08 0.00 0.27 466.72
u1432 200 2.629027e + 08 8.17 9.34 489.00 11.04 13.03 507.50 10.32 12.16 509.40 0.54 0.73 157.96 0.44 0.51 235.98 0.00 0.65 505.03
fl1577 10 8.823213e + 07 0.21 4.57 243.50 0.00 0.00 199.20 0.00 0.00 282.50 0.68 0.68 57.05 0.68 0.68 64.81 0.00 0.00 8.52
fl1577 50 3.694969e + 07 7.76 10.90 431.70 0.01 0.11 468.10 0.03 0.12 505.50 1.41 1.41 83.19 1.41 1.41 37.26 0.00 0.09 353.82
fl1577 100 2.369881e + 07 9.34 11.99 517.10 14.31 20.00 584.90 15.89 21.24 504.90 4.19 4.24 189.96 4.23 4.24 134.28 0.00 0.15 492.77
fl1577 200 1.247028e + 07 7.24 8.74 414.70 36.02 40.86 501.90 36.56 44.17 501.70 4.78 4.78 29.49 4.43 4.60 230.08 0.00 0.20 495.47
d1655 10 3.062937e + 08 2.38 6.04 329.00 0.00 0.00 184.50 0.00 0.00 199.30 0.12 0.12 62.17 0.12 0.12 70.44 0.00 0.00 68.50
d1655 50 1.440223e + 08 9.27 12.33 488.30 0.00 0.23 496.60 0.00 0.21 495.00 1.28 1.28 34.06 1.28 1.28 38.74 0.02 0.13 321.60
d1655 100 9.623458e + 07 10.72 12.48 500.60 12.87 14.22 505.00 12.68 13.89 506.90 0.37 0.38 128.53 0.37 0.37 99.18 0.00 0.19 541.10
d1655 200 5.727806e + 07 9.09 10.39 495.30 14.16 16.14 505.80 14.69 15.94 508.20 0.50 0.61 126.05 0.48 0.49 263.03 0.00 0.31 537.88
vm1748 10 1.206434e + 10 1.38 2.75 423.00 0.00 0.00 272.50 0.00 0.15 243.60 0.16 0.16 63.73 0.16 0.16 72.37 0.00 0.00 269.40
vm1748 50 4.763207e + 09 8.79 11.33 458.40 0.07 0.18 479.20 0.07 0.24 456.10 0.56 0.68 166.40 0.69 0.69 153.39 0.00 0.11 431.97
vm1748 100 3.175321e + 09 10.10 11.30 484.70 0.44 0.67 507.50 0.58 1.92 510.90 1.38 1.45 231.08 1.38 1.38 253.41 0.00 0.18 507.21
vm1748 200 1.887680e + 09 8.44 9.52 533.10 2.89 5.32 505.60 3.89 4.86 508.30 2.61 2.94 97.29 2.76 2.78 231.69 0.00 0.48 554.10
rl1889 10 1.275225e + 10 3.15 5.93 424.20 0.00 0.00 270.30 0.00 0.00 263.20 0.02 0.02 63.62 0.02 0.02 72.34 0.00 0.00 381.39
rl1889 50 5.935066e + 09 13.15 14.79 465.90 0.00 0.14 453.70 0.00 0.14 449.70 0.13 0.13 127.88 0.13 0.13 43.45 0.12 0.17 489.29
rl1889 100 4.024244e + 09 11.02 13.16 470.80 2.57 3.05 509.40 2.66 3.32 404.80 0.73 0.80 130.66 0.74 0.78 235.93 0.00 0.26 483.01
rl1889 200 2.444401e + 09 10.98 12.53 503.10 5.72 7.02 406.70 5.75 6.74 546.60 0.34 0.35 59.03 0.30 0.30 133.39 0.00 0.42 595.17
pcb3038 10 1.002111e + 09 27.21 27.31 523.30 21.16 21.51 529.30 21.16 21.46 560.60 0.12 0.12 54.97 0.12 0.12 62.20 0.00 0.20 366.36
pcb3038 50 4.364877e + 08 39.13 39.50 584.20 9.24 28.73 549.90 9.21 28.24 552.60 0.00 0.00 36.44 0.00 0.00 41.67 0.12 0.30 483.24
pcb3038 100 2.994432e + 08 16.10 32.21 542.50 36.45 36.60 567.50 24.62 34.42 568.60 0.01 0.02 34.24 0.00 0.01 297.81 0.07 0.44 546.89
pcb3038 200 1.979078e + 08 14.59 29.89 577.40 24.63 33.50 577.80 34.76 35.42 561.30 0.00 0.08 61.38 0.01 0.03 213.95 0.30 1.22 376.84

It acheives the best performance.

293
294
Table 25

Q. Zhou et al. / Information Sciences 541 (2020) 271–296


Comparisons of MA with five state-of-the-art algorithms on the University of Florida Sparse Matrix instances with a time limit of 600 s.

Instance p cost VNS-1 VNS-2 VNS-3 GRASP HYBRID MA


g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g g best g av g t av g

Harvard500 10 643.00 0.00 0.28 216.60 0.00 0.00 0.20 0.00 0.00 0.10 0.00 0.00 352.04 0.00 0.00 326.02 0.00 0.00 0.15
Harvard500 50 433.50 0.12 0.29 391.50 0.00 0.23 197.90 0.00 0.06 157.80 3.81 3.81 160.95 3.81 3.81 158.69 0.00 0.06 246.41
Harvard500 100 356.50 0.14 0.42 232.00 0.42 0.81 361.00 0.00 0.27 93.70 12.20 12.20 99.23 12.20 12.20 96.82 0.14 0.27 334.17
Harvard500 200 300.00 0.00 0.00 67.18 0.00 0.00 59.65 0.00 0.00 84.63 0.00 0.00 68.86 0.00 0.00 53.93 0.00 0.00 49.17
USAir97 10 472.94 0.71 0.98 273.90 0.00 0.42 164.50 0.00 0.44 93.10 0.86 0.86 383.91 0.86 0.86 324.83 0.00 0.00 118.10
USAir97 50 264.00 0.19 0.63 257.00 0.38 0.83 428.70 0.00 0.21 151.10 6.82 6.82 176.70 6.82 6.82 178.00 0.00 0.04 154.87
USAir97 100 188.50 0.00 0.45 268.70 0.53 0.66 358.70 0.27 0.27 8.00 23.08 23.08 98.74 23.08 23.08 96.89 0.00 0.34 307.90
USAir97 200 132.00 0.00 0.00 64.11 0.00 0.00 58.49 0.00 0.00 74.14 0.00 0.00 61.92 0.00 0.00 67.64 0.00 0.00 67.18
Email 10 4145.09 1.74 3.36 397.70 0.12 0.12 214.70 0.00 0.19 234.20 0.26 0.26 354.70 0.26 0.26 346.90 0.05 0.11 33.09
Email 50 2107.29 5.45 7.55 537.30 0.29 0.69 424.90 0.09 0.30 372.60 0.37 0.39 222.73 0.37 0.37 222.89 0.07 0.28 374.69
Email 100 1443.14 8.45 9.97 551.90 0.93 1.32 507.40 0.00 0.22 466.10 0.89 1.02 198.68 0.89 0.89 196.65 0.22 0.44 397.53
Email 200 920.10 13.98 15.06 523.80 3.03 3.89 549.10 1.21 2.43 580.40 4.66 6.33 152.48 4.66 4.66 153.64 0.00 0.27 473.97

It acheives the best performance.


Q. Zhou et al. / Information Sciences 541 (2020) 271–296 295

Acknowledgments

We are grateful to the reviewers for their valuable comments which helped us to improve the paper. We thank Dr. A.
Nikolaev for providing us with the source codes of their VNS-1, VNS-2 and VNS-3 algorithms and Dr. I. Vasil’eV for sending
us the source codes of GRASP and HYBRID heuristics. This work is partially supported by the National Natural Science Foun-
dation of China (Grant Nos. 71850410545, 71821001, 71771099, 71810107003, 71620107002).

Appendix A

The purpose of this appendix is to present computational results and comparisons with the proposed algorithm on OR-
Lib, TSP-Lib and University of Florida Sparse Matrix Collection benchmarks (Tables 20–25). Columns ‘Instance’ and ‘p’ show
the instance name and the number of clusters. Column ‘cost’ indicates the best objective values obtained across all the com-
pared algorithms. Columns ‘g best ’ and ‘g av g ’ respectively refer to the percentage gap of the best objective value and the aver-
age objective value across 10 independent runs to the best solution value from column ‘cost’. The percentage gaps of the best
  
and the average objective value are calculated as ðf  f Þ=f  100, where f is the best result obtained across all the com-
pared algorithms and f is the best or the average objective value achieved by the given algorithm. Column ‘tav g ’ shows
the average run-time in seconds to reach the final objective value. The best result among all the compared algorithms is
marked in bold.

References

[1] D. Aloise, A. Deshpande, P. Hansen, P. Popat, NP-hardness of Euclidean sum-of-squares clustering, Mach. Learn. 75 (2) (2009) 245–248.
[2] D. Aloise, P. Hansen, L. Liberti, An improved column generation algorithm for minimum sum-of-squares clustering, Math. Program. 131 (1–2) (2012)
195–220.
[3] A.Y. Al-Omary, M.S. Jamil, A new approach of clustering based machine-learning algorithm, Knowl.-Based Syst. 19 (4) (2006) 248–258.
[4] P. Avella, A. Sassano, I. Vasil’ev, Computational study of large-scale p-Median problems, Math. Program. 109 (1) (2007) 89–114.
[5] A.M. Bagirov, Modified global k-means algorithm for minimum sum-of-squares clustering problems, Pattern Recogn. 41 (10) (2008) 3192–3199.
[6] A.M. Bagirov, S. Taheri, J. Ugon, Nonsmooth DC programming approach to the minimum sum-of-squares clustering problems, Pattern Recogn. 53
(2016) 12–24.
[7] P. Balaprakash, M. Birattari, T. Stützle, Improvement Strategies for the F-Race Algorithm: Sampling Design and Iterative Refinement, in: T. Bartz-
Beielstein et al. (Eds.), Hybrid Metaheuristics, Lect. Notes Comput. Sci., Vol 4771, Springer, Berlin, 2007, pp. 108–122.
[8] J.E. Beasley, A note on solving large p-median problems, Eur. J. Oper. Res. 21 (2) (1985) 270–273.
[9] M. Birattari, T. Stützle, L. Paquete, K. Varrentrapp, A racing algorithm for configuring metaheuristics, in: Proceedings of the 4th Annual Conference on
Genetic and Evolutionary Computation, Morgan Kaufmann Publishers Inc., 2002, pp. 11–18.
[10] J. Brimberg, Z. Drezner, N. Mladenović, S. Salhi, A new local search for continuous location problems, Eur. J. Oper. Res. 232 (2) (2014) 256–265.
[11] J. Brimberg, Z. Drezner, N. Mladenović, S. Salhi, Using injection points in reformulation local search for solving continuous location problems, Yugoslav
J.Oper. Res. 27 (3) (2016) 291–300.
[12] J. Carrasco, S. García, M.M. Rueda, S. Das, F. Herrera, Recent trends in the use of statistical tests for comparing swarm and evolutionary computing
algorithms: Practical guidelines and a critical review, Swarm Evolut. Comput. (2020), https://doi.org/10.1016/j.swevo.2020.100665, in press.
[13] E.J. Carrizosa Priego, N. Mladenović, R. Todosijević, Sum-of-squares clustering on networks, Yugoslav J. Operat. Res. 21 (2) (2011) 157–161.
[14] E. Carrizosa, N. Mladenović, R. Todosijević, Variable neighborhood search for minimum sum-of-squares clustering on networks, Eur. J. Oper. Res. 230
(2) (2013) 356–363.
[15] L.R. Costa, D. Aloise, N. Mladenović, Less is more: basic variable neighborhood search heuristic for balanced minimum sum-of-squares clustering, Inf.
Sci. 415–416 (2017) 247–253.
[16] T.H. Cuong, J.C. Yao, N.D. Yen, Qualitative Properties of the Minimum Sum-of-Squares Clustering Problem, 2018. arXiv preprint arXiv:1810.02057..
[17] T.A. Davis, Y. Hu, The university of Florida sparse matrix collection, ACM Trans. Math. Software (TOMS) 38 (1) (2011) 1–25.
[18] J. Del Ser, E. Osaba, D. Molina, X.S. Yang, S. Salcedo-Sanz, D. Camacho, S. Das, P.N. Suganthan, C.A.C. Coello, F. Herrera, Bio-inspired computation: Where
we stand and what’s next, Swarm Evolut. Comput. 48 (2019) 220–250.
[19] J. Derrac, S. García, D. Molina, F. Herrera, A practical tutorial on the use of nonparametric statistical tests as a methodology for comparing evolutionary
and swarm intelligence algorithms, Swarm Evolut. Comput. 1 (1) (2011) 3–18.
[20] C.C. Fast, I.V. Hicks, A Branch Decomposition Algorithm for the p-Median Problem, INFORMS J. Comput. 29 (3) (2017) 474–488.
[21] F. Glover, M. Laguna, Tabu Search, in: D.Z. Du, P.M. Pardalos (Eds.), Handbook of Combinatorial Optimization, Springer, Boston, 1998, pp. 2093–2229.
[22] P. Hansen, N. Mladenović, Variable neighborhood search for the p-median, Location Sci. 5 (4) (1997) 207–226.
[23] P. Hansen, N. Mladenović, J-MEANS: a new local search heuristic for minimum sum of squares clustering, Pattern Recogn. 34 (2) (2001) 405–413.
[24] P. Hansen, B.K. Cheung, E. Ngai, N. Mladenović, Analysis of global K-means, an incremental heuristic for minimum sum-of-squares clustering, J. Classif.
22 (2) (2005) 287–310.
[25] J.K. Hao, Memetic Algorithms in Discrete Optimization, in: F. Neri, C. Cotta, P. Moscato (Eds.), Handbook of Memetic Algorithms, Studies in
Computational Intelligence, Vol 379, Springer, Berlin, 2012, pp. 73–94.
[26] J.N. Hooker, R.S. Garfinkel, C.K. Chen, Finite dominating sets for network location problems, Oper. Res. 39 (1) (1991) 100–118.
[27] X. Lai, J.K. Hao, A tabu search based memetic algorithm for the max-mean dispersion problem, Comput. Oper. Res. 72 (2016) 118–127.
[28] A. Likas, N. Vlassis, J.J. Verbeek, The global k-means clustering algorithm, Pattern Recogn. 36 (2) (2003) 451–461.
[29] M. López-Ibáñez, J. Dubois-Lacoste, L.P. Cáceres, M. Birattari, T. Stützle, The irace package: Iterated racing for automatic algorithm configuration, Oper.
Res. Perspect. 3 (2016) 43–58.
[30] Y. Lu, U. Benlic, Q. Wu, A memetic algorithm for the orienteering problem with mandatory visits and exclusionary constraints, Eur. J. Oper. Res. 268 (1)
(2018) 54–69.
[31] L. Lu, J.K. Hao, Q. Wu, Hybrid evolutionary search for the traveling repairman problem with profits, Information Sciences 502 (2019) 91–108.
[32] P. Moscato, An introduction to population approaches for optimization and hierarchical objective functions: a discussion on the role of tabu search,
Ann. Oper. Res. 41 (2) (1993) 85–121.
[33] M.C. Naldi, R.J.G.B. Campello, Comparison of distributed evolutionary k-means clustering algorithms, Neurocomputing 163 (2015) 78–93.
[34] A. Nikolaev, N. Mladenović, R. Todosijević, J-means and I-means for minimum sum-of-squares clustering on networks, Optim. Lett. 11 (2) (2017) 359–
376.
[35] B. Ordin, A.M. Bagirov, A heuristic algorithm for solving the minimum sum-of-squares clustering problems, J. Global Optim. 61 (2) (2015) 341–361.
296 Q. Zhou et al. / Information Sciences 541 (2020) 271–296

[36] G. Reinelt, TSPLIB – a traveling salesman problem library, ORSA J. Comput. 3 (4) (1991) 376–384.
[37] M.G.C. Resende, R.F. Werneck, A hybrid heuristic for the p-median problem, J. Heuristics 10 (1) (2004) 59–88.
[38] J.M. Squyres, A. Lumsdaine, R.L. Stevenson, Cluster-based parallel image processing toolkit, in: Image and Video Processing III, International Society for
Optics and Photonics, vol. 2421, 1995, pp. 228–240..
[39] Q. Wu, J.K. Hao, A hybrid metaheuristic method for the Maximum Diversity Problem, Eur. J. Oper. Res. 231 (2) (2013) 452–464.
[40] Q. Wu, J.K. Hao, A clique-based exact method for optimal winner determination in combinatorial auctions, Inf. Sci. 334–335 (2016) 103–121.

You might also like