Professional Documents
Culture Documents
A Modification Piecewise Convexification Method For Box-Constrained Non-Convex Optimization Programs
A Modification Piecewise Convexification Method For Box-Constrained Non-Convex Optimization Programs
Abstract: This paper presents a piecewise convexification method to approximate the whole approximate
optimal solution set of non-convex optimization problems with box constraints. In the process of box division,
we first classify the sub-boxes and only continue to divide only some sub-boxes in the subsequent division.
At the same time, applying the α-based Branch-and-Bound (αBB) method, we construct a series of piecewise
convex relax sub-problems, which are collectively called the piecewise convexification problem of the original
problem. Then, we define the (approximate) solution set of the piecewise convexification problem based on
the classification result of sub-boxes. Subsequently, we derive that these sets can be used to approximate
the global solution set with a predefined quality. Finally, a piecewise convexification algorithm with a new
selection rule of sub-box for the division and two new termination tests is proposed. Several instances verify
that these techniques are beneficial to improve the performance of the algorithm.
Keywords: Non-convex programming, Global optimization, Optimization solution set, αBB method, piece-
wise convexification, approximation
Mathematics Subject Classification: 90C26, 90C30, 90C90
1 Introduction
Non-convex optimization problems arise frequently in machine learning [21, 7, 17], and in many other appli-
cations. Meanwhile, how to effectively solve non-convex optimization problems has gained much attention.
So far, the majority of global approximation algorithms are designed, see, [9, 10, 14, 20, 11, 16, 18, 19]. In
particular, the αBB method has become an increasingly important global optimization method in the de-
sign of efficient and computationally tractable numerical algorithms for non-convex optimization problems,
see, [3, 12, 6, 8]. It is worth noticing that these algorithms aim in general at determining a single globally
optimal solution, and the majority of application problems may be existed with many or even infinite of
many globally optimal solutions. To the best of our knowledge, the number of algorithms for determining
representations of the set of approximate globally optimal solutions seems to be still quite limited. In [4],
Eichfelder and Gerlach generalize the classical αBB method to find a representation of the whole optimal
solution set with predefined quality for a non-convex optimization problem. As they pointed out, however,
some additional variables and the additional while loop are necessary.
Motivated by [4], we develop a piecewise convexification method to approximate the whole global op-
timal solution set of non-convex optimization problems. Incorporating the αBB method and the interval
subdivision technique, we firstly piecewise relax the original problem and obtain a series of convex relax-
ation sub-problems, collectively called the piecewise convexification problem of the original problem. Then,
we construct the (approximate) solution sets of the piecewise convexification problem by comparing all
the (approximate) optimal solutions of the convex relaxation sub-problems on each sub-box, and we show
that these constructed sets can be used to approximate the global optimal solution set with a predefined
quality. Finally, a new piecewise convexification algorithm is proposed, which incorporates a new selection
rule of sub-box for the division and two new termination rules. Furthermore, several instances verify that
these selection rule and termination tests are conducive to improving the effectiveness and the speed of the
algorithm.
This paper is organized as follows. Section 2 summarizes some basic definitions of the optimization
problem. It also introduces the αBB method and the interval division. In Section 3, we firstly propose a
∗ Corresponding Author
1
piecewise convexification method for the non-convex optimization problem and then analyze the solution
set of this piecewise convexification optimization problem. More importantly, some relationships between
the (approximate) optimal solution set of the convexification problem and the original optimization problem
are also stated in detail. A new algorithm that generates the subset of approximation global solutions is
presented in Section 4. Finally, we report and discuss several numerical experiments in section 5.
2 Preliminaries
Let I denoted the set of all real nonempty closed boxes and In denote the set of all n-dimensional boxes.
n
For a given box X ∈ In , we set X = [a, b] :=
Q
[ai , bi ] where a = (a1 , · · · , an ) and b = (b1 , · · · , bn ). Thus,
i=1
x ∈ X denotes xi ∈ [ai , bi ] for any i ∈ {1, · · · , n}. In this paper, let X = [a, b], and then we consider the
following non-convex optimization problem(NCOP):
(NCOP) min f (x),
x∈X
Definition 2.1. (see [4]) Let ε > 0, X be a nonempty subset of Rn , and f : X → R such that arg min f (x) 6=
x∈X
∅.
(a) A point x̃ ∈ X is an optimal solution of f w.r.t. X, if
f (x̃) ≤ f (y) ∀y ∈ X.
Xop := {x ∈ X : f (x) ≤ f (y), ∀y ∈ X} is the optimal solution set.
(b) A point x̄ ∈ X is an ε-minimal point of f w.r.t. X, if
f (x̄) ≤ f (y) + ε ∀y ∈ X.
ε
Xop := {x ∈ X : f (x) ≤ f (y) + ε, ∀y ∈ X} is the approximate solution set.
Before proceeding further, we hereby give brief descriptions of the αBB method and the interval subdi-
vision, because they will play an important role in designing the piecewise convexification method.
For estimating the value of α several methods already exist, see, [2, 13, 15]. In this article, we directly
adopt the following method from [2] to roughly calculate the value αi (i = 1, · · · , n), as defined by
1 X
αi := max 0, − min ∇2 f (x)ii − (X, i, d) , i = 1, · · · , n (2.2)
2 x∈X
where d := b − a ∈ Rn , Hessian matrix ∇2 f (x) = ∇2 f (x)ij and
X X dj
max min ∇2 f (x)ij , max ∇2 f (x)ij
(X, i, d) := . (2.3)
i6=j
x∈X x∈X di
Noting that αi is a finite value. Let 1 = (1, · · · , 1), a lower bound of the minimum eigenvalue of ∇2 f (x)
w.r.t. X is given in [2], i.e.,
X
λX 2
min ≥ min min ∇ f (x)ii − (X, i, 1) .
i x∈X
2
then f is obviously convex on X, not vice versa. In addition, for any boxes X1 and X2 with X1 ⊆ X2 , it is
clear that λ̃X2 ≤ λ̃X1 .
Moreover, in [13], the maximum separation distance between f and FX α over X is typically of the form
n b − a 2
i i
X
α
D(X) = max kf (x) − FX (x)k = αi , (2.5)
x∈X
i=1
2
where µn denotes the Lebesgue measure on Rn and the subinterval Y kt abbreviate as Y kt = [akt , bkt ] =
n
[aki t , bki t ]. It is worth noting that the construction of the subdivision Y (t+1) of X w.r.t (t + 1) is based
Q
i=1
on Y t , which is to select one or more sub-intervals from Y t to divide. In what follows, we introduce the
interval division method of any subinterval Y kt .
For a given box Y kt = [akt , bkt ] ∈ In , the branching index l is defined by
Clearly, Y kt = Y kt ,1 ∪ Y kt ,2 and Y kt ,1 , Y kt ,2 ∈ Y (t+1) . For simplicity, we may define the splitting operator
Sp(Y kt ) = {Y kt ,1 , Y kt ,2 }. In this paper, we define the length of the subdivision Y t of X by
( n )
X k kt 2
t kt kt 2 t
|T (Y )| := max {ka − b k2 } = max (bi − ai ) .
kt ∈{1,··· ,Mt } kt ∈{1,··· ,Mt }
i=1
3
Let Y t := {Y 1 , Y 2 , · · · , Y Mt } be a subdivision of X. Then we consider the same convex relaxation
subproblem on Y kt = [akt , bkt ] ∈ Y t as [2], that is,
n
kt X
min Fkαt (x) := f (x) + αki t (aki t − xi )(bki t − xi ), (3.1)
x∈Y kt i=1
where if λ̃Y kt ≥ 0 estimated by (2.4), then αki t = 0. Otherwise, αki t is computed by (2.2) for any {1, · · · , n}.
kt
Let αkt := (αk1 t , · · · , αknt ). Apparently, Fkαt is a convex lower estimation function of f (x) on Y kt and
αkt
Fkt (x) = f (x) for any x ∈ Y kt when λ̃Y kt ≥ 0. Let Xap kt
and Xapkt ,ε
denote the set of all optimal solutions
and the set of all approximate solutions of (3.1), respectively, that is,
kt kt
kt
Xap := {x ∈ Y kt : Fkαt (x) ≤ Fkαt (y) for any y ∈ Y kt }, (3.2)
kt kt
kt ,ε kt
Xap := {x ∈ Y : Fkαt (x) ≤ Fkαt (y) + ε for any y ∈ Y kt
}, (3.3)
kt kt ,ε
where ε > 0. Obviously, Xap and Xap are not empty sets.
k
For any kt ∈ {1, · · · , Mt ∈ Y t , Fkαt t is a convex relaxation sub-problem of (NCOP) w.r.t.
} and Y kt
Y kt . Obviously, all convex sub-problems compose the piecewise convexification optimization problem of
(NCOP) w.r.t. X.
where λ̃Y kt is defined by (2.4). Moreover, Mt = M1 (t) ∪ M2 (t). Clearly, f is convex on any subset of Y kt
for any kt ∈ M1 (t). However, for any kt ∈ M2 (t) one cannot assert that f must be non-convex on Y kt
kt
because we only obtain λ̃Y kt < 0 rather than λY min < 0, that is, the convexity of f is uncertain on Y
kt for
any kt ∈ M2 (t). Thus, we need more concerned about the indexes in M2 (t) rather than in M1 (t) for any t.
Our idea of choosing the box to divide is that no subdivision is applied to the box when f is convex on it,
and the box on which f is non-convex is only selected to be divided.
Then, some notations about the union of solution sets corresponding to the above indexes sets, are
represented, which help us to clearly define the solution set of the piecewise convexification problem.
[ [
C kt NC kt
Xap (t) := Xap and Xap (t) := Xap (3.5)
kt ∈M1 (t) kt ∈M2 (t)
kt
where Xap is an optimal solution set of the convex relaxation optimization problem (3.1) w.r.t. Y kt . Final,
in this paper we directly define the set of the piecewise convexification problem Xap (t) of the following form
n o
C∪N C C∪N C
Xap (t) = x ∈ Xap (t) : f (x) ≤ f (y) for any y ∈ Xap (t) (3.6)
4
which indicate that we classify the boxes generated in the division process, and put the index of the new
box that makes f convex into Mtt+1,C , otherwise, put it into Mtt+1,N C . Obviously, these definitions imply
that
M1 (t + 1) = M1 (t) ∪ Mtt+1,C and M2 (t + 1) = Mtt+1,N C . (3.9)
The union of solution sets about (3.7) and (3.8) are similarly represented by
[ [
newC kt newN C kt
Xap (t, t + 1) := Xap and Xap (t, t + 1) := Xap .
t+1,C t+1,N C
kt ∈Mt kt ∈Mt
always based on the result of subdivision t, it follows that this relationship is reasonable.
(iii) From the subdivision t to t + 1, we do not consider the boxes that make f convex in the subdivision
t. Moreover, we directly use Xap C (t) from the result of subdivision t to construct the set X (t + 1), instead
ap
of solving these sub-problems repeatedly. These techniques may reduce the number of sub-problems to be
solved in the piecewise convexification method.
Next, we will discuss the relationship between Xap (t) and Xop when f has the piecewise convex property
on X.
Theorem 3.2. If there exists the subdivision Y t0 of X such that M2 (t0 ) = ∅, then Xop = Xap (t0 ).
N C (t ) = ∅, M
Proof. M2 (t0 ) = ∅ implies that Xap 0 t0 = M1 (t0 ) and
C C
Xap (t0 ) = {x ∈ Xap (t0 ) : f (x) ≤ f (y) for any y ∈ Xap (t0 )}. (3.10)
According to Y t0 k
S
is a subdivision of X, then X = Y . In what follows, we will proof Xap (t0 ) = Xop .
t
kt ∈Mt0
The proof is by contradiction.
If Xop * Xap (t0 ), then there exists x̂ ∈ Xop such that x̂ ∈ / Xap (t0 ). Based on the definition of Xap (t0 )
as shown in (3.10), it easy to verify that x̂ ∈ / Xop .
Next, we assume that Xap (t0 ) * Xop , that is, there exists x̂ ∈ Xap (t0 ) with x̂ ∈ / Xop . Thus, one can
find ŷ ∈ X satisfied f (ŷ) < f (x̂). Since Y t0 is a subdivision, then let ŷ ∈ Y jk0 where kt0 ∈ Mt0 = M1 (t0 ).
kt kt
Based on the non-emptiness of Xap0 , there must exist ẑ such that ẑ ∈ Xap0 . It follows that f (ẑ) ≤ f (ŷ),
ktC (t ). This apparently contradicts the fact that x̂ ∈ X (t ).
which yields to f (ẑ) < f (x̂) with ẑ ∈ Xap0
⊆ Xap 0 ap 0
Thus Xap (t0 ) ⊆ Xop .
Obviously, by the above analysis one can conduct Xap (t0 ) = Xop .
This theorem shows that, for the non-convex problem with the piecewise convex properties, i.e., f is
non-convex on X and f is convex on each sub-box of X for some subdivision t0 , the proposed piecewise
convexification method can explore all global optimal solutions of this non-convex optimization problem.
5
Proof. First, by the Lemma 5 in [4], there exists the subdivision Y t such that |T (Y t )| → 0 as t → ∞.
Since αi is a finite value for any i, then one can yield that there exist t0 and a subdivision Y t0 such that
n
X kt0
akt0 − bkt0 2
i i
max αi ≤ ε. (3.11)
kt0 ∈M2 (t0 )
i=1
2
ε . Suppose that X (t ) * X ε , that is, there exists x̂ ∈ X (t )
Another step in the proof is Xap (t0 ) ⊆ Xop ap 0 op ap 0
such that x̂ ∈ ε C∪N
/ Xop . Obviously, x̂ ∈ Xap C (t0 ) and then one can find ŷ ∈ X satisfied
n
X kt0 kt0 kt0
f (ẑ) + αi (ai − ẑi )(bi − ẑi ) < f (x̂) − ε,
i=1
N C (t ). kt
which implies that f (ẑ) < f (x̂) by (3.11). This contradicts x̂ ∈ Xap (t0 ) because of ẑ ∈ Xap0 ⊂ Xap 0
kt
/ Y kt0 \Xap0 .
Consequently, we infer that ŷ ∈
Apparently, the above proof are contrary to ŷ ∈ Y kt0 . Therefore, the theorem is now evident from what
we have proved.
The above theorems show that the solution set of the piecewise convexification optimization problem is
a lower bound set of Xopε . In order to construct the upper bound set of X ε , the approximation solution set
op
of the piecewise convexification optimization problem is introduced, as defined by
n o
ε C∪N C,ε C∪N C,ε
Xap (t) := x ∈ Xap (t) : f (x) ≤ f (y) + ε for any y ∈ Xap (t) ,
kt ,ε k
Xap denotes the approximate optimal solution set of the convex relaxation sub-problem Fkαt t on Y kt with
ε (t) is a upper bound set of X ε .
the quality ε, as defined by (3.3). In what follows, we present that set Xap op
Theorem 3.4. For any ε > 0, there exists t ∈ N and the subdivision Y t of X satisfied
n
X bkt − akt 2
max αki t i i
≤ε
kt ∈M2 (t)
i=1
2
such that ε
Xop ⊆ ε (t).
Xap
Proof. Similarly, there exist t0 and a subdivision Y t0 of X such that (3.11) holds. It remains to show that
ε ⊆ X ε (t ). Assume that there exists t such that X ε * X ε (t ). Then one can find x̂ ∈ X ε and
Xop ap 0 0 op ap 0 op
x̂ ∈ ε (t ). Now, x̂ ∈
/ Xap ε (t ) can be distinguished two cases, the first of which is x̂ ∈ X C∪N C,ε (t ). It
/ Xap
0 0 ap 0
C∪N C,ε ε ,
implies that there exists ŷ ∈ Xap (t0 ) ⊆ X satisfied f (ŷ) + ε < f (x̂). This contradicts the fact x̂ ∈ Xop
that is, the first case is not true.
C∪N C,ε
The second case is x̂ ∈ / Xap (t0 ). For this subdivision Y t0 , there exists kt0 ∈ Mt0 = M1 (t0 )∪M2 (t0 )
kt ,ε C,ε C∪N C,ε
such that x̂ ∈ Y kt0 . If kt0 ∈ M1 (t0 ), then Xap0 ⊆ Xap (t0 ). Moreover, x̂ ∈
/ Xap (t0 ) indicates that
kt ,ε αk t αk t
x̂ ∈
/ Xap0 . This means that there exists ŷ ∈ Y kt0 satisfied Fkt 0
(ŷ) + ε < Fkt 0
(x̂). Obviously, f is
0 0
k
convex on Y kt0 for kt0 ∈ M1 (t0 ), that is, Fkt α t0 (x) = f (x) for any x ∈ Y kt0 . Then it conducts that
0
6
kt kt
f (ŷ) + ε = Fkαt 0
(ŷ) + ε < Fkαt 0 ε . This yields to k
(x̂) = f (x̂), which contradicts x̂ ∈ Xop t0 ∈
/ M1 (t0 ), that
0 0
kt ,2ε N C,ε kt ,2ε C∪N C,ε
is, kt0 ∈ M2 (t0 ). This indicates that Xap0 ⊆ Xap / Xap0
(t0 ) and x̂ ∈ by x̂ ∈
/ Xap (t0 ). Thus
k k
there exists ŷ ∈ Y kt satisfied Fkt α t0 (ŷ) + 2ε < Fkt α t0 (x̂), which implies that f (ŷ) + ε < f (x̂) from (3.11).
0 0
ε , which means that k
This contradicts x̂ ∈ Xop t0 ∈ M2 (t0 ) is also false. Consequently, the second case would
not hold.
ε ⊆ X ε (t ) holds.
Therefore, the assumption is not true, that is, Xop ap 0
Remark 3.5. Combing Theorems 3.3 and 3.4, the lower and upper bound sets of the approximate solution
set of the original non-convex optimization problem are obtained by constructing solution sets of the piecewise
convexification optimization problem, that is,
ε ε
Xap (t) ⊆ Xop ⊆ Xap (t), ∀ε > 0,
n akt −bkt 2
αki t
P i i
where t satisfies that max 2
≤ ε.
kt ∈M2 (t) i=1
Abbreviation Denotation
Xb The sub-box of X
α
µ̂ The function values of FXb on X
b b
X α
Xap The optimal solution set of FXb on X
b b b
vglob The smallest objective function value found for all current sub-boxes
n 2
bi abi − bi
P
w(X,b α
b) The modified width of X b and w(X,
b αb) := α
b
2
i=1
λ̃Xb A lower bound of λmin (X)
b computed by (2.4)
In what follows, we present the piecewise convexification algorithm to obtain a subset of the approximate
global optimal solution set, which uses selection rule, discarding and termination tests, as shown in Algorithm
1.
This algorithm terminates after finitely many iterations because there exists the number of subdivisions
t such that max e α̃) ≤ ε. Noting that this algorithm applies the same discarding technique to
w(X,
(X,x̃,
e µ̃,α̃)∈LN C
deleting sub-boxes as the modified αBB method [4]. Thus it holds that X∩X op = ∅ when µ̂ > vglob . However,
b
the termination conditions and the selection way of sub-box for the division are different. Compared with
[4], there is only one termination condition, that is, L = ∅. But this piecewise convexification algorithm sets
two termination conditions, that is, max w(X,
e αe) ≤ ε or LN C = ∅. Clearly, if LN C = ∅, then
(X,x̃,
e µ̃,α̃)∈LN C
max e α̃) = 0, and not vice versa, i.e., LN C 6= ∅ may be hold when
w(X, max e α̃) ≤ ε.
w(X,
(X,x̃,
e µ̃,α̃)∈LN C (X,x̃,
e µ̃,α̃)∈LN C
The following numerical experiment results, in section 5, will show that for some complex problems, these two
termination conditions are more conducive to speeding up the algorithm than having only one termination
condition in the modified αBB method [4]. In this algorithm, we put a new selection rule of sub-box for the
division, that is, the box with the maximum modified width in LN C is selected to divide into two sub-boxes
as shown in line 5. This selection approach is different from the one proposed in [4]. Moreover, different
from the modified αBB method [4], this piecewise convexification algorithm only requires a while loop and
does not require additional parameters.
S
Remark 4.1. From the piecewise convexification algorithm framework, the union of sets Xb
(X,b
b x,b
µ,α)∈M
b 1
is a subdivision of X where M1 = M ∪ LN C .
new is a subset of the approximate
The following theorem is illustrated that the output of this algorithm Xap
global solution set.
7
Algorithm 1 The piecewise convexification algorithm for (NCOP).
Input: X 0 = [a, b] ∈ In , f ∈ C2 (Rn , R), ε > 0; Output: Xapnew
;
1: Compute an α0 of f on X 0 according to (2.2)-(2.3);
2: Set X ∗ := X 0 , x∗ := a+b
2
, µ∗ := −∞, α∗ := α0 , xact := x∗ , vglob = vact = +∞;
∗ ∗ ∗ ∗
3: LN C := {(X , x , µ , α )}, MC = MD = ∅ and k := 0.
4: while LN C 6= ∅ and max w(X,
e αe) > ε do
(X,e
e x,e
µ,α)∈L
e NC
5: k := k + 1;
6: Define (X ∗ , x∗ , µ∗ , α∗ ) at the first element of LN C with max w(X,
e αe).
(X,e
e x,e
µ,α)∈L
e NC
7: Delete (X ∗ , x∗ , µ∗ , α∗ ) from LN C .
8: b ∈ Sp(X ∗ ) do
for all X
9: Compute λ̃Xb by (2.4) and α b by (2.2), respectively.
α α
10: Compute x b ∈ arg min FX b (x). Let µ
b
b = min FX b (x).
b
x∈X
b x∈X
b
11: b ≤ vglob then
if µ
12: if λ̃Xb ≥ 0 then
X
Add (X,b xb, µ
b, α
b) as the last element to MC and add xb to Xap .
b
13:
14: else
X
Add (X,b xb, µ
b, α b to Xap
b) as the last element to LN C and add x .
b
15:
16: x) ≤ vact then
if f (b
17: Set xact = x b, vact = f (xact ), vglob = min{vact , vglob }
18: Delete (X, x, µ, α) ∈ LN C with µ > vglob from LN C , and
19: add (X, x, µ, α) ∈ LN C with µ > vglob to MD .
20: else
21: Add (X, b xb, µ
b, α
b) as the last element to MD .
S e
22: M := {X : (X,
e x
e, µ e) ∈ LU C } ∪ MC .
e, α
( )
new S X S X
Xap := x ∈ Xap : f (x) ≤ f (y), ∀ y ∈ Xap .
b b
23:
X∈M
b X∈M
b
Proof. The proof process is similar to Theorem 3.3 and will be omitted.
5 Numerical Experiments
In this section, we demonstrate the efficiency of this piecewise convexification algorithm and compare it with
[4]
the mod αloci,d=u−l BB . All computations have been performed on a computer with Iter(R)Core(TM)i5-
8250U CPU and 8 Gbytes RAM. In numerical results, we use the following notations.
Abbreviation Denotation
iter Number of required iterations
CPU Required CPU time in seconds
Nε Number of ε-optimal solutions from algorithm, let ε = 10−3
f lagter The indicator of termination condition of algorithm, f lagter = 1 or 0
PC-NCOP The piecewise convexification algorithm for (NCOP), i.e. Algorithm 1
- The algorithm does not record a certain value
In this paper, f lagter = 1 indicates that the termination condition of this algorithm is LN C = ∅,
8
otherwise, f lagter = 0 when max w(X̃, α̃) ≤ ε. Moreover, we replace µ̄ ≤ vglob in line 11 by
(X̃,x̃,µ̃,α̃)∈LN C
µ̄ ≤ vglob + 10−6
for all instances.
As stated in [4], authors use INTLAB ToolBox to automaticly compute the elements ∇2 f (X)ij . In fact,
we directly solve the optimization problem min ∇2 f (X̄)ij , to estimate αki n on each sub-box rather than
x∈X̄
using INTLAB. Therefore, we rewrite the modified αBB method [4], without INTLAB.
First, we demonstrate the performance of two approaches on nine test instances with finite optimal
solutions, in which most test instances have multiple optimal solutions and all examples are two-dimensional.
Some information of these test instances about the objective functions f , feasible sets X 0 , number of globally
solutions, and global optimal values are listed in Tables 1.
[4] −5.12 5.12
Rastrigin 20 + x21 + x22 − 10(cos(2πx1 ) + cos(2πx2 )) , 1 0
−5.12 5.12
[5] −1.9 1.9
6-Hump (4 − 2.1x21 + 31 x41 )x21 + x1 x2 − (4 − 4x22 )x22 , 2 ≈ −1.031629
−1.1 1.1
[5]
5.1 2 −5 10
Branin (x2 − x
4π 2 1
+ 5
x
π 1
− 6)2 + 10(1 − 1
8π
) cos(x1 ) + 10 , 3 ≈ 0.397886
0 15
[5] −6 6
Himmelblau (x21 + x2 − 11)2 + (x1 + x22 − 7)2 , 4 0
−6 6
[5] −5.12 5.12
Rastrigin mod 20 + x21 + x22 + 10(cos(2πx1 ) + cos(2πx2 )) , 4 ≈ 0.497480
−5.12 5.12
5 5
[5] P P −10 10
Shubert [i cos((i + 1)x1 + 1)] [j cos((j + 1)x2 + j)] , 18 ≈ −186.730909
i=1 j=1 −10 10
[5] 0 1
Deb 1 − 12 (sin6 (5πx1 ) + sin6 (5πx2 )) , 25 −1
0 1
[5] 0.25 10
Vincent − 21 (sin(10 ln(x1 )) + sin(10 ln(x2 ))) , 36 −1
0.25 10
Numerical results of these two algorithms are presented in Table 2. It is easy to see that for all the test
[4]
examples, the iter and CPU values of PC-NCOP are significantly less than those of the mod αloc
i,d=u−l BB .
As for the |Nε | values, only a slight difference for Branin exists, that is, two algorithms can only find
two globally optimal solutions, not three. These results demonstrate that the proposed algorithm can
find almost all the optimal solutions of the original non-convex problem, and PC-NCOP is better than
[4]
the mod αloc
i,d=u−l BB . In addition, for Branin, Himmelblau and Shubert the termination condition
of PC-NCOP is the same, that is, max w(X̃, α̃) ≤ ε. The termination condition of PC-NCOP for
(X̃,x̃,µ̃,α̃)∈LN C
other remaining instances is LN C = ∅.
9
The division results of X are clearly shown in Fig. 1, where the first and third columns show the
[4] [4]
(a) Rastrigin (b) Rastrigin (c) 6-Hump (d) 6-Hump
[4] [4]
(e) Brain (f) Brain (g) Himmelblau (h) Himmelblau
[4] [4]
(i) Rastrigin mod (j) Rastrigin mod (k) Shubert (l) Shubert
[4] [4]
(m) Deb 1 (n) Deb 1 (o) Vincent (p) Vincent
[4]
results obtained by Algorithm 1, while the results of the mod αloc i,d=u−l BB are shown in the second and
fourth columns. The red star denotes the ε-optimal solutions. Obviously, it can be seen from Fig. 1
that the selection way of the boxes to be divided can effectively reduce the number of iterations. In fact,
[4]
the mod αloci,d=u−l BB has numerous subdivisions of the box near the optimal solution, while our algorithm
has only a few subdivisions, and the following partial graph intuitively reflects this assertion.
In what follows, we consider four numerical tests with infinite number of globally optimal solutions listed
in [4], as defined by Table 3.
Table 4 shows the numerical results of the instance tests of Table 3. The iter and CPU values of PC-
[4]
NCOP are significantly better than the value of the mod αloc i,d=u−l BB . Except for Test03, the number
[4]
of solutions of PC-NCOP is also higher than that of the mod αloc
i,d=u−l BB . In addition, the termination
condition max w(X, b α̂) ≤ ε, in PC-NCOP, is satisfied for these test problems as f lagter = 0. In other
(X,x̂,
b µ̂,α̂)
words, this termination condition is meaningful in Algorithm 1 and could be helpful to reduce the number
of iterations. Moreover, the results of the interval subdivision and solutions in set Xapnew for Algorithm 1
are showed in Fig.3. This figure shows that the distribution of these optimal solutions obtained from the
PC-NCOP can be used to describe the distribution of the optimal solutions of the original problem.
Finally, in order to verify the efficiency of the proposed algorithm for the high-dimensional instances,
Table 5 is shown a high-dimensional test problem, which is selected from the literature [4].
[4]
Table 6 gives results by applying the PC-NCOP and the mod αloc
i,d=u−l BB to TestDimd . From Table
10
(a) Rastrigin mod (b) Rastrigin mod [4]
11
(a) Test01 (b) Test02
12
[4]
6, the experimental results demonstrate that both the proposed algorithm and the mod αloc
i,d=u−l BB lead
[4]
to the same values of Nε . However, compared with Algorithm 1, the mod αloc i,d=u−l BB requires more
iterations and CPU, and the advantage of Algorithm 1 is more prominent as the dimension increases. In
these high-dimensional problems, f lagter = 0 indicates that the stop condition max w(X, b α̂) ≤ ε satisfies
(X,x̂,
b µ̂,α̂)
and LN C 6= ∅ holds. These help to reduce the number of iterations and the CPU. Furthermore, we found
an interesting phenomenon, that is, the number of iterations of Algorithm 1 is one less than the number of
globally optimal solutions.
From the above numerical experiments, for these instances with infinite number of optimal solutions or
with high dimensional, the termination condition max w(X, b α̂) ≤ ε is easier to satisfied than LN C = ∅.
(X,x̂,
b µ̂,α̂)
This means that two termination conditions in the proposed algorithm can be used to reduce the number
of iterations of the algorithm.
6 Conclusions
An αBB convexification method based on box classification strategy is studied for non-convex single-objective
optimization problems. The box classification strategy is proposed based on the convexity of the objective
function on the sub-boxes when dividing the boxes, which helps to reduce the number of box divisions and
improve the computational efficiency. The αBB method is introduced to construct the piecewise convexifi-
cation problem of the non-convex optimization problem, and the solution set of the piecewise convexification
problem is used to approximate the globally optimal solution set of the original problem. Based on the
theoretical results, an αBB convexification algorithm with two termination conditions is proposed, and nu-
merical experiments show that this algorithm can obtain a large number of globally optimal solutions more
quickly than other algorithms.
References
[1] C.S. Adjiman, I.P. Androulakis, C.A. Floudas. A global optimization method, αBB, for general twice-
differentiable constrained NLPs-II. Implementation and computational results. Computers & Chemical
Engineering, 1998, 22: 1159–1179.
[2] C.S. Adjiman, S. Dallwig, C.A. Floudas, A. Neumaier. A global optimization method, αBB, for general
twice-differentiable constrained NLPs-I. Theoretical advances. Computers & Chemical Engineering,
1998, 22: 1137–1158.
[3] I.P. Androulakis, C.D. Maranas, C.A. Floudas. αBB: A global optimization method for general con-
strained nonconvex problems. Journal of Global Optimization, 1995, 7: 337–363.
[4] G. Eichfelder, T. Gerlach, S. Sumi. A modification of the αBB method for box-constrained optimization
and an application to inverse kinematics. EURO Journal on Computational Optimization, 2016, 4: 93–
121.
[5] M.G. Epitropakis, V.P. Plagianakos, M.N. Vrahatis. Finding multiple global optima exploiting differ-
ential evolution’s niching capability. In: Proceedings of IEEE SDE, Paris, France, 2011: pp80–87.
[6] M. Hladı́k. An extension of the αBB-type underestimation to linear parametric Hessian matrices. Jour-
nal of Global Optimization, 2016, 64: 217–231.
[7] P. Jain, , P. Kar , et al.. Non-convex Optimization for Machine Learning. Foundations and Trends in
Machine Learning, 2017, 10: 142–363.
[8] N. Kazazakis, C.S. Adjiman. Arbitrarily tight αBB underestimators of general non-linear functions over
sub-optimal domains. Journal of Global Optimization, 2018, 71: 815–844.
[9] G. Liuzzi, M. Locatelli, V. Piccialli. A new branch-and-bound algorithm for standard quadratic pro-
gramming problems. Optimization Methods & Software, 2019, 34: 79–97.
[10] M. Locatelli, F. Schoen. Global optimization: Historical notes and recent developments. EURO Journal
on Computational Optimization, 2021, 9: 100012.
[11] A. Marmin, M. Castella,J.C. Pesquet. How to globally solve non-convex optimization problems involving
an approximate l0 penalization. In: ICASSP 2019-2019 IEEE International Conference on Acoustics,
Speech and Signal Processing (ICASSP),2019, pp5601–5605.
[12] H. Milan. On the efficient gerschgorin inclusion usage in the global optimization αBB method. Journal
of Global Optimization, 2014, 61: 235–253.
13
[13] D. Nerantzis, C.S. Adjiman. Tighter αBB relaxations through a refinement scheme for the scaled
gerschgorin theorem. Journal of Global Optimization, 2019, 73: 467–483.
[14] F. Schoen, L. Tigli. Efficient large scale global optimization through clustering-based population meth-
ods. Computers and Operations Research, 2021, 127: 105–165.
[15] A. Skjäl, T. Westerlund. New methods for calculating αBB-type underestimators. Journal of Global
Optimization, 2014, 58: 411–427.
[16] X.L. Sun, K.I.M. McKinnon, D. Li. A convexification method for a class of global optimization problems
with applications to reliability optimization. Journal of Global Optimization, 2001, 21: 185–199.
[17] F. Wen, L. Chu, P. Liu, R.C. Qiu. A survey on nonconvex regularization-based sparse and low-rank
recovery in signal processing, statistics, and machine learning. IEEE Access, 2018, 6: 69883–69906.
[18] Z.Y. Wu, F.S. Bai, L.S. Zhang. Convexification and concavification for a general class of global opti-
mization problems. Journal of Global Optimization, 2005, 31: 45–60.
[19] Y. Xia. A survey of hidden convex optimization. Journal of the Operations Research Society of China.
2020, 8: 1–28.
[20] Y. Yang, M. Pesavento, S. Chatzinotas, B. Ottersten. Successive convex approximation algorithms for
sparse signal estimation with nonconvex regularizations. IEEE Journal of Selected Topics in Signal
Processing, 2018, 12: 1286–1302.
[21] P. Zhong. Training robust support vector regression with smooth non-convex loss function. Optimization
Methods & Software, 2012, 27: 1039–1058.
Qiao Zhu
College of Mathematics, Sichuan University, 610065, Chengdu Sichuan, China
E-mail address: math qiaozhu@163.com
Liping Tang
National Center for Applied Mathematics, Chongqing Normal University, 401331 Chongqing, China.
Email address: tanglipings@163.com
Xinmin Yang
National Center for Applied Mathematics, Chongqing Normal University, 401331 Chongqing, China.
Email: xmyang@cqnu.edu.cn
14