Professional Documents
Culture Documents
Application of HNN For Max Cut Problem
Application of HNN For Max Cut Problem
There have been more than ten years since artifi- Sometimes, a variation of the signum function is used
cial neural networks were applied in the field of op- in DHNN, called unipolar binary function:
timization. As we known, the application of neural
1 if y ≥ 0
networks in the field of optimization was initiated sgn(y) =
0 otherwise.
by Hopfield and Tank in 1985. Hopfield and Tank’s
seminal paper [13] demonstrated that the Travelling There are two categories of DHNN identified by
Salesman Problem (TSP) could be solved by using a their operating modes: synchronous mode (parallel
Hopfield neural network. Since then, Hopfield neu- mode) and asynchronous mode (serial mode). The
ral network have always been the majority neural general algorithms of the DHNN in the two modes
network in solving optimization problems. A va- are as follows,
riety of feedback neural networks similar to Hop-
Algorithm 1. (Synchronous DHNN)
field neural network have been proposed to solve lin-
ear programming problems (for example, [21, 22, 25]) 1. given a stop criterion C
and quadratic programming problems (for example,
[3,18,25]) as well as combinatorial optimization prob- 2. given an initial vector x(0), t = 0
lems (for example, [1, 8, 13, 14, 16, 23]) because of 3. xi (t + 1) = sgn(Wi x(t) − ti ) i = 1, · · · , n
the potential of extremely rapid computation power
and speed of neural networks, which can be obtained 4. if x(t + 1) satisfies C then stop
through hardware implementation. Cichocki et al. [6] else t := t + 1 and go to step 3
and Zhang [24] are two comprehensive books about
the neural networks and optimization, and a concise Algorithm 2. (Asynchronous DHNN)
review of neural networks for combinatorial optimiza- 1. given a stop criterion C and a
tion was written by Smith [20]. point-to-point map
The discrete Hopfield-type neural network
(DHNN) denoted by N = (W , T ) can be described M : {1, · · · , n} → {1, · · · , n}
Pn
2. given an initial vector x(0) and set where i=1 ∆wii is a constant which does not con-
i(0) = 1, t = 0 tribute to the solution set. But according to the
above theorem, the diagonal elements of the matrix
3. take i(t) = M (i(t − 1)) ∈ {1, · · · , n} if t > 0 W take very important role in the convergence anal-
ysis. So we can simply assign proper values to the
sgn(Wi x(t) − ti ) i = i(t)
4. xi (t + 1) = diagonal elements of W in order to make the net-
xi (t) i 6= i(t)
work convergent.
5. if x(t + 1) satisfies C then stop On the other hand, the values of the diagonal el-
else t := t + 1 and go to step 3 ements are closely related to the number of stable
states of the network. The following theorem tells
where Wi , i = 1, · · · , n are the rows of W .
us that smaller diagonal elements may result in less
In this paper, we only consider the DHNN running
stable states of a network.
in asynchronous mode.
Theorem 2. ( [15], also [24]) Let N 1 = (W 1 , T ),
N 2 = (W 2 , T ) be two DHNN, where wij 1 2
= wij for
2 Theoretical Results 1 2
i 6= j, and wii ≤ wii for i = 1, · · · , n. Then ΩN 1 ⊆
Definition 1. A DHNN N = (W , T ) is called con- ΩN 2 where ΩN i is the set of all stable state of the
i
vergent from a given initial state x(0) if there exists network N (i = 1, 2) respectively.
an integer k such that x(k) = sgn{W x(k) − T }.
Generally speaking, since there is a finite number of
x∗ = x(k) is referred to as a stable state. A net-
states for a network, the network will always converge
work is called convergent if it converges to a sta-
to a stable state or a stable cycle. However, while
ble state from any given initial state. A DHNN is
a network is used to solve a optimization problem,
called convergent to a stable cycle from a given ini-
we are concerned about whether the energy function
tial state x(0) if there exists integers k and q such
is decreasing as the network is updating rather than
that x(k + q) = sgn{W x(k) − T }.
whether the network converge to a stable state. Thus
The following theorem is given by Hopfield in [12]. the following definition are proposed.
Theorem 1. ( [12]) Let W be a symmetric matrix Definition 2. For a given energy function E(t), a
with nonnegative diagonal. The DHNN N = (W , T ), network N is called E-convergent form a given initial
which runs in the asynchronous mode, will always state x(0) if it converges to a stable state or a stable
converge to a stable state. cycle with ∆E(t) ≤ 0 for t = 0, 1, · · ·. A network
The asynchronous DHNN can be used to solve such is called totally E-convergent if it is convergent from
combinatorial optimization problem any given initial state.
which can be solved by a DHNN N = (Ŵ , 0) running In Table 3, the results are compared with the re-
in asynchronous mode. sults of DSDP method [5], one of the latest algorithms
by solving the Max-Cut SDP relaxation, and CirCut nal weight matrix to solve Max-Cut problem. Ele-
method [4], a new heuristic algorithm for the Max- mentary numerical experiments show that the per-
Cut problem. To our knowledge, DSDP and CirCut formance of such DHNN is exciting. Comparing with
are two of the most efficient algorithms for the Max- other methods solving Max-Cut, for example SDP
Cut problem at present. The results in Table 3 are method and CirCut method which are two of the
the best values obtained in 100 tests with random ini- most efficient solving Max-Cut problem at present,
tial points. The quality of solutions given by DHNN the solution quality of DHNN is competitive or even
is almost better than those of DSDP except on prob- better than those of them. Meanwhile, we note that
lems G48, G49 and G50. On the problem G49, all although we simply change the diagonal elements to
three algorithms obtained the global optimal value. negative, the quality are improved dramatically.
Moreover, on problems G11, G13, G33, G55 and G62, Unlike the complicate energy function used to solve
the solution found by DHNN are best. Note that our TSP (see [1, 13, 16, 23] etc.), the energy function for
code is simple without any further improving tech- the Max-Cut problem is simpler, and suitable for neu-
nique and CirCut is a well developed heuristic code, ral network. It is possibly another reason that DHNN
the quality of the solutions obtained by DHNN is can produce good solutions for the Max-Cut prob-
competitive. lem. There are many other combinatorial optimiza-
tion problems can be formulated to simple integer
Graph DSDP CirCut DHNN programming problems. Another further work is to
G11 542 554 560 extend the idea in this paper to such combinatorial
G12 540 552 548 optimization problems.
G13 564 572 574 However, like traditional Hopfield-type neural net-
G14 2922 3053 3024 work, the neural network proposed in this paper con-
G15 2938 3039 3013 verge to the first stable state they encounter, which
G20 838 939 895 will decrease the solution quality. There are many
G21 841 921 881 techniques to prevent network to be trapped in stable
G22 12960 13331 13167 state corresponding poor quality solution, for exam-
G23 13006 13269 13157 ple “hill-climbing”. If we combine these techniques
G24 12933 13287 13140 with DHNN, we expect that the quality of the solu-
G30 3038 3377 3200 tion given by network is better. We shall study this
G31 2851 3255 3111 problem further.
G32 1338 1380 1378
G33 1330 1352 1354
G34 1334 1358 1356 References
G48 6000 6000 5992
G49 6000 6000 6000 [1] S. V. B. Aiyer, M. Niranjan, and F. Fallside.
G50 5880 5856 5846 A theoretical investigation into the performance
G55 9960 10240 11639 of the Hopfield model. IEEE Trans. on Neural
G56 3634 3943 3700 Networks, Vol. 1, No. 2, pp. 204-215, 1990.
G57 3320 3412 3394
[2] S. Benson, Y. Ye, and X. Zhang. Solving large-
G60 13610 14081 13718
scale sparse semidefinite programs for combina-
G61 5252 5690 5316 torial optimization. SIAM J. Optim., Vol. 10,
G62 4612 4740 4894 No. 2, pp. 443-461, 2000.
G64 7624 8575 8055
[3] A. Bouzerdoum and T. R. Pattison. Neural net-
Table 3: Comparison with DSDP and CirCut work for quadratic optimization with bound con-
straints. IEEE Trans. on Neural Networks, Vol.
4, pp. 293-304, 1993.
[12] J. J. Hopfield. Neural networks and physical [25] X. S. Zhang and H. C. Zhu. A neural network
systems with emergent collective computational model for quadratic programming with simple
abilities. Proc. Natl. Acad. Sci. USA, Vol. 79, pp. upper and lower bounds and its application to
2554-2558, 1982. linear programming. Lecture Notes in Computer
Science, 834, pp. 119-127, Springer-Verlag, 1994.
[13] J. J. Hopfield and D.W. Tank. Neural computa-
tion of decisions in optimization problems. Bio- [26] X. J. Zhuo and X. S. Zhang. Hopfield-type neural
logical Cybernetics, Vol. 52, pp. 141-152, 1985. network for solving four-coloring map problems.
OR Transactions, Vol. 3, No. 3, pp. 35-43, 1999.
[14] A. Jagota. Approximating maximum clique with (in Chinese)
a Hopfield network. IEEE Trans. on Neural Net-
works, Vol. 6, pp. 724-735, 1995.
[15] L. C. Jiao. System theory of neural networks,
Xian Electronic Scientific University Publishing
House, Xian, China, 1990. (in Chinese)
[16] A. Joppe, H. R. A. Cardon, and J. C. Bioch. A
neural network for solving the Traveling Sales-
man Problem on the basis of city adjacency in
the tour. IJCNN, Vol. 3, pp. 961-964, 1990.
[17] R. M. Karp. Reducibility among combinatorial
problems. In R. E. Miller and J. W. Thather,
editors, Complexity of Computer Computation,
pp. 85-103. Plenum Press, New York, 1972.