Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
Abstract. Chaotic neural networks have been proved to be powerful tools for escaping from local minima. In this paper, we first retrospect Chens chaotic neural network and then propose a novel Gauss-Morlet-Sigmoid chaotic neural network model. Second, we make an analysis of the largest Lyapunov exponents of the neural units of Chens and the Gauss-Morlet-Sigmoid model. Third, 10-city traveling salesman problem (TSP) is given to make a comparison between them. Finally we conclude that the novel chaotic neural network model is more effective.
1 Introduction
Many combinatorial optimization problems arising from science and technology are often difficult to solve entirely. Hopfield and Tank first applied the continuous-time, continuous-output Hopfield neural network (HNN) to solve TSP [1], thereby initiating a new approach to optimization problems [2, 3]. The Hopfield neural network, one of the well-known models of this type, converges to a stable equilibrium point due to its gradient decent dynamics; however, it causes sever local-minimum problems whenever it is applied to optimization problems. Chaotic neural networks have been proved to be powerful tools for escaping from local minima. M-SCNN has a more powerful performance than Chens in solving combinatorial optimization problems, especially in searching global minima of continuous nonlinear function and traveling salesman problems [4]. Now we do research on the activation function in detail. In this paper, we mainly make research on the effect of Gauss function on the performance of the network. We first review the Chens chaotic neural network. Second, we propose a novel chaotic neural network model. Third, the time evolution figures of their largest Lyapunov exponents of the neural units are given. At last, we apply them to 10-city traveling salesman problem (TSP) in order to make a comparison. Finally we conclude that the novel chaotic neural network model we proposed is more valid. Because the wavelet function is a kind of basic function, for any function f ( x) L2 ( R) and any wavelet , the known formula can be described as follows:
f ( x) =
j,k =
c j,k j,k ( x)
(1)
D.-S. Huang, K. Li, and G.W. Irwin (Eds.): ICIC 2006, LNCS 4113, pp. 115 125, 2006. Springer-Verlag Berlin Heidelberg 2006
116
Chen and Aiharas transiently chaotic neural network [5] is described as follows:
x i (t ) = f ( y i (t )) =
1 1+ e
yi ( t ) /
y i (t + 1) = ky i (t ) + Wij x j + I i z i (t )( x i (t ) I 0 ) j z i (t + 1) = (1 ) z i (t )
where xi (t ) is output of neuron i ; y i (t ) denotes internal state of neuron i ; Wij describes connection weight from neuron j to neuron i , Wij = W ji ; I i is input bias of neuron i , a a positive scaling parameter for neural inputs, k damping factor of nerve membrane, 0 k 1, z i (t ) self-feedback connection weight (refractory strength) 0, damping factor of z i (t ) , 0< <1, I 0 a positive parameter, steepness parameter of the output function, >0.
2.2 Gauss-Morlet-Sigmoid Chaotic Neural Network Model (G-M-SCNN)
Gauss-Sigmoid chaotic neural network is a novel model proposed by ourselves, described as follows:
y i (t + 1) = ky i (t ) + Wij x j + I i z i (t )( x i (t ) I 0 ) j z i (t + 1) = (1 1 ) z i (t )
x i (t ) = f ( y i (t ))
f ( y i (t )) = BG ((1 + ri (t )) y i (t )) + S ( 0 (1 + i (t )) y i (t )) + M ( 1 (1 + i (t )) y i (t )) (8)
i (t + 1) =
i (t ) ln(e + (1 i (t ))) ri (t + 1) = 2 ri (t )
G (u ) = e S (u ) =
h (u a ) 2
(9) (10)
(11) (12)
1 1 + e u
2
M (u ) = e u
/2
cos(5u )
(13)
117
where xi (t ) , y i (t ) , Wij , , k , I i , z i (t ) , I 0 are the same with the above. And is a positive parameter which controls the speed of this annealing process; ri (0) is an important parameter of activation function which should be varied with kinds of special 1 , 2 A , 0 , 1 are positive parameters of optimization problems, Morlet wavelet function and Sigmoid function. h , a , b , c are important parameters of Gauss function.
The time evolution figure of the largest Lyapunov exponent is shown as Fig.1:
We make an analysis of the time evolution figures of the neural unit G-M-SCNN with the change of 2 . (1) The parameters are set as follows:
2 =0.1, k =0.092, I 0 =0.8, y (0) =0.283, z (0) =0.8,= 0.008, 1 =0.008, 0 =0.8, 1 =20, B =10, r (0) =200.5, (0) =0.8, h ,a ,b ,c .
0.5= 0.5= 1.2-= 2.0=
.1
0 1
118
The time evolution figure of the largest Lyapunov exponent is shown as Fig.2.
2 =0.5, k =0.092, I 0 =0.8, y (0) =0.283, z (0) =0.8,= 0.008, 1 =0.008, 0 =0.8, 1 =20, B =10, r (0) =200.5, (0) =0.8, h ,a ,b ,c .
The time evolution figure of the largest Lyapunov exponent is shown as Fig.3.
2 =0.9, k =0.092, I 0 =0.8, y (0) =0.283, z (0) =0.8,= 0.008, 1 =0.008, 0 =0.8, 1 =20, B =10, r (0) =200.5, (0) =0.8, h ,a ,b ,c .
0.5= 0.5= 1.2-= 2.0=
0.5=
0.5=
1.2-=
2.0=
119
The time evolution figure of the largest Lyapunov exponent is shown as Fig.4.
Seen from the above analysis, we can conclude that the change of 2 does have a profound effect on the time evolution of Lyapunov exponent. So, in this paper we will only make analysis of the performance of the network with different 2 in solving traveling salesman problem (TSP)
where v i 0 = v in and v i , n +1 = v i1 . A and D are the coupling parameters corresponding to the constraint function and the cost function of the tour length, respectively. d xy is the distance between city x and city y . The coordinates of 10-city is as follows: (0.4, 0.4439),(0.2439, 0.1463),(0.1707, 0.2293),(0.2293, 0.716),(0.5171, 0.9414), (0.8732, 0.6536),(0.6878, 0.5219),(0.8488, 0.3609),(0.6683, 0.2536), (0.6195, 0.2634). The shortest distance of the 10-city is 2.6776. Here are the results of the test about Chens and G-M-SCNN.
(14)
120
The coupling parameters corresponding to the constraint function and the cost function of the tour length we adopt are set as follows: A =2.5, D =1. (1) The parameters of Chens are set as follows :
Model
Chens
average
VN 188 185 183 184 181 175 180 189 187 178 183
GN 188 185 183 184 180 175 179 189 186 178 182.7
VP 94% 92.5% 91.5% 92% 90.5% 87.5% 90% 94.5% 93.5% 89% 91.5%
GP 94% 92.5% 91.5% 92% 90% 87.5% 89.5% 94.5% 93% 89% 91.35%
In order to gain insight into the effect of 2 on the proposed model, the tests are shown as follows: (2) The parameters of G-M-SCNN are set as follows:
k =1, I 0 =0.5, z (0) =0.1,= 0.008, 1 =0.008, 0 =0.8, 1 =20, B =10, r (0) =200.5,
Model
Average
VN 187 193 194 188 190 193 190 192 188 189 190.4
0.5=
0.5=
1.2-=
2.0=
(0) =0.8, h
,a
,b
,c
GN 178 188 187 183 176 183 184 181 183 182 182.5
VP 93.5% 96.5% 97% 94% 95% 96.5% 95% 96% 94% 94.5% 95.2%
GP 89% 94% 93.5% 91.5% 88% 91.5% 92% 90.5% 91.5% 91% 91.25%
121
Model
average
VN 187 187 191 187 193 194 188 190 192 190 189.9
GN 180 175 184 180 188 187 183 176 182 184 181.9
VP 93.5% 93.5% 95.5% 93.5% 96.5% 97% 94% 95% 96% 95% 94.95%
GP 90% 87.5% 92% 90% 94% 93.5% 91.5% 88% 91% 92% 90.95%
Model
average
VN 190 192 187 189 189 190 187 187 188 190 188.9
GN 183 181 182 183 175 183 180 178 175 181 180.1
VP 95% 96% 93.5% 94.5% 94.5% 95% 93.5% 93.5% 94% 95% 94.45%
GP 91.5% 90.5% 91% 91.5% 87.5% 91.5% 90% 89% 87.5% 90.5% 90.05%
Model
average
VN 193 188 182 191 191 194 193 194 187 188 190.1
GN 183 179 171 187 180 186 188 187 181 178 182
VP 96.5% 94% 91% 95.5% 95.5% 97% 96.5% 97% 93.5% 94% 95.05%
GP 91.5% 89.5% 85.5% 93.5% 90% 93% 94% 93.5% 90.5% 89% 91%
122
Model
average
VN 193 191 193 193 193 187 189 188 189 188 190.4
GN 183 179 183 183 181 182 183 175 183 181 181.5
VP 96.5% 90.5% 96.5% 96.5% 96.5% 93.5% 94.5% 94% 94.5% 94% 95.2%
GP 91.5% 89.5% 91.5% 91.5% 90.5% 91% 91.5% 87.5% 91.5% 90.5% 90.75%
Model
average
VN 188 185 189 193 181 188 188 190 193 187 188.2
GN 180 171 180 183 171 177 182 183 186 179 178.2
VP 94% 92.5% 94.5% 96.5% 90.5% 94% 94% 95% 96.5% 93.5% 94.1%
GP 90% 85.5% 90% 91.5% 85.5% 88.5% 91% 91.5% 93% 89.5% 89.1%
Model
average
VN 190 181 193 189 190 194 191 191 189 184 189.2
GN 182 173 181 180 185 187 189 183 185 170 181.5
VP 95% 90.5% 96.5% 94.5% 95% 97% 95.5% 95.5% 94.5% 92% 94.6%
GP 91% 86.5% 90.5% 90% 92.5% 93.5% 94.5% 91.5% 92.5% 85% 90.75%
123
Model
average
VN 193 192 192 190 185 193 191 192 193 183 190.4
GN 185 180 185 180 175 183 181 178 180 174 180.1
VP 96.5% 96% 96% 95% 92.5% 96.5% 95.5% 96% 96.5% 91.5% 95.2%
GP 92.5% 90% 92.5% 90% 87.5% 91.5% 90.5% 89% 90% 87% 90.05%
Model
average
VN 188 192 189 194 187 186 192 190 190 190 189.8
GN 180 182 180 183 175 173 177 179 181 182 179.2
VP 94% 96% 94.5% 97% 93.5% 93% 96% 95% 95% 95% 94.9%
GP 90% 91% 90% 91.5% 87.5% 86.5% 88.5% 89.5% 90.5% 91% 89.6%
AVP 91.5% 95.2% 94.95% 94.45% 95.05% 95.2% 94.1% 94.6% 95.2% 94.9%
AGP 91.35% 91.25% 90.95% 90.05% 91% 90.75% 89.1% 90.75% 90.05% 89.6%
124
(AVP=average valid path percent, AGP=average global path percent) Seen from the Table11, the proposed model can solve the TSP with a higher valid minimum-path percent and a little lower valid-path percent under all the values of 2 . This means that the proposed model is more valid in only solving TSP than Chens. The time evolution figures of the energy function of G-M-SCNN and Chens in solving TSP are respectively given in Fig.5 and Fig.6.
By comparison, it is concluded that M-SCNN is superior to Chens model. From the Fig.5, Fig.6, one can see that the velocity of convergence of G-M-SCNN is faster than that of Chens in solving TSP. Because, the state of Fig.5 become stable in 60 iterations while that of Fig.6 become stable in 1500 iterations. The superiority of G-M-SCNN contributes to several factors: First, because of the higher nonlinear nature of Gauss function, the activation function of G-M-SCNN has a further performance in solving combinatorial optimization problems than Chens.
125
Second, it is easier to produce chaotic phenomenon [8] in that the activation function is non-monotonic. Third, the activation function of G-M-SCNN is varied with time. Third, the wavelet function is a kind of basic function.
5 Conclusions
We have introduced two models of chaotic neural networks. To verify the effectiveness of it, we have made comparison with Chens model in optimization problems. By comparison, one can conclude that G-M-SCNN is superior to Chens in searching global minima of continuous nonlinear function. Different from Chens model, the activation function of G-M-SCNN is composed by Gauss, Morlet wavelet and Sigmoid. So, besides it has the quality of sigmoid activation, the activation function of G-M-SCNN has a higher nonlinear nature than Sigmoid, which is easier to produce chaotic phenomenon [7] because of its non-monotonic. Due to these factors, G-M-SCNN is superior to Chens in solving TSP.
References
1. Hopfield, J.J., Tank, D.W.: Neural Computation of Decision in Optimization Problems. Biol. Cybern.52 (1985)141-152 2. Hopfield, J.: Neural Networks and Physical Systems with Emergent Collective Computational Abilities. In: Proc. Natl. Acad. Sci., 79 (1982) 2554-2558 3. Hopfield, J.: Neural Networks and Physical Systems with Emergent Collective Computational Abilities. In: Proc. Natl. Acad. Sci., 81 (1984) 3088-3092 4. Xu, Y.Q., Ming, S., Duan, G.R.: Wavelet Chaotic Neural Networks and Their Application to Optimization Problems. Lecture Notes in Computer Science, 3971 Springer Berlin (2006) 379-384 5. Wang, L.P., Tian, F.Y.: Noisy Chaotic Neural Networks for Solving Combinatorial Optimization Problems. International Joint Conference on Neural Networks. Italy: IEEE, (2000)37-40 6. Sun, S.Y., Zhen, J.L.: A Kind of Improved Algorithm and Theory Testify of Solving TSP in Hopfield Neural Network. Journal of electron, 1 (1995)73-78 7. Potapove, A., Kali, M.: Robust Chaos in Neural Networks. Physics Letters A, 277 (2000) 310-322