Elsevier

Neural Networks

Volume 23, Issue 2, March 2010, Pages 189-200
Neural Networks

Coexistence and local stability of multiple equilibria in neural networks with piecewise linear nondecreasing activation functions

https://doi.org/10.1016/j.neunet.2009.11.010Get rights and content

Abstract

In this paper, we investigate the neural networks with a class of nondecreasing piecewise linear activation functions with 2r corner points. It is proposed that the n-neuron dynamical systems can have and only have (2r+1)n equilibria under some conditions, of which (r+1)n are locally exponentially stable and others are unstable. Furthermore, the attraction basins of these stationary equilibria are estimated. In the case of n=2, the precise attraction basin of each stable equilibrium point can be figured out, and their boundaries are composed of the stable manifolds of unstable equilibrium points. Simulations are also provided to illustrate the effectiveness of our results.

Introduction

In past decades, neural networks have been extensively studied due to their applications in image processing, pattern recognition, associative memories and many other fields. Wilson and Cowan (1972) studied the dynamics of spatially localized neural populations, and introduced two functions E(t) and I(t) to characterize the states of excitatory neurons and inhibitory neurons, respectively. And in Wilson and Cowan (1973), authors derived the following differential equations {μtE(x,t)=E(x,t)+[1reE(x,t)]Se[αμ[ϱeE(x,t)βee(x)ϱiI(x,t)βie(x)±P(x,t)]]μtI(x,t)=I(x,t)+[1riI(x,t)]Si[αμ[ϱeE(x,t)βei(x)ϱiI(x,t)βii(x)±Q(x,t)]] where E(x,t),I(x,t) represent time coarse-grained excitatory and inhibitory activities, respectively; Se[],Si[] are the expected proportions of excitatory neurons and inhibitory neurons receiving at least threshold excitation per unit time; βjj(x) stands for the probability that cells of class j be connected with cells of class j a distance x away; denotes spatial convolution, P(x,t),Q(x,t) are the afferent stimuli to excitatory neurons and inhibitory neurons, respectively. The results obtained are closely related to the biological systems and succeeded in providing qualitative descriptions of several neural processes. Grossberg (1973) introduced another class of recurrent on-center off-surround networks, which were shown to be capable of contrast enhancing significant input information; sustaining this information in short term memory; producing multistable equilibrium points that normalize, or adapt, the field’s total activity; suppressing noise; and preventing saturation of population response even to input patterns whose intensities are high (Ellias & Grossberg, 1975). In such an on-center off-surround anatomy, a given population excites itself (and possibly near populations) and inhibits populations that are further away (and possibly itself and nearby populations also). And in Cohen and Grossberg (1983), Cohen–Grossberg neural networks were proposed, which can be described by the following differential equations duidt=ai(ui)[bi(ui)j=1ncijhj(uj)],i=1,,n.

In particular, let ai()1,bi(x)=dix, then, the Cohen–Grossberg neural networks reduce to the following Hopfield neural networks dui(t)dt=diui(t)+j=1nwijfj(uj(t))+Ii,i=1,,n, where ui(t) represents the state of the i-th unit at time t; di>0 denotes the rate with which the i-th unit will reset its potential to the resting state in isolation when disconnected from the network and external inputs; wij corresponds to the connection weight of the j-th unit on the i-th unit; fj() is the activation function; and Ii stands for the external input.

There has been a large number of works on the dynamics of neural networks in the literature. Note that in many existing works, the authors mainly focused on the existence of a unique equilibrium and its stability, see Chen (2001), Chen and Amari (2001) and other papers. However, in practice, it is desired that the network has several equilibria, of which each represents an individual pattern. For example, in the Cellular Neural Networks (CNNs) with saturated activation function fj(x)=|x+1||x1|2,j=1,,n see (Fig. 1), a pattern or an associative memory is usually stored as a binary vector in {1,1}n, and the process of the pattern recognition or memory attaining is that the system converges to certain stable equilibrium with all components located in (,1) or (1,+). Also in some neuromorphic analog circuits, multistable dynamics even play an essential role, as revealed in Douglas, Koch, Mahowald, Martin, and Suarez (1995), Hahnloser, Sarpeshkar, Mahowald, Douglas, and Seung (2000), and Wersing, Beyn, and Ritter (2001). Therefore, the study of coexistence and stability of multiple equilibrium points, in particular, the attraction basin, is of great interest in both theory and applications.

In an earlier paper (Chen & Amari, 2001), the authors pointed out that the 1-neuron neural network model du(t)dt=u(t)+(1+ϵ)g(u(t)), where ϵ is a small positive number and g(u)=tanh(u), has three equilibrium points and two of them are locally stable, one is unstable. Recently, for the n-neuron neural networks, many results have been reported in the literature, see Ma and Wu, 2007, Remy et al., 2008, Shayer and Campbell, 2000, Zeng and Wang, 2006, Zhang and Tan, 2004 and Zhang, Yi, and Yu (2008). In Zeng and Wang (2006), by decomposing phase space Rn into 3n subsets, the authors investigated the multiperiodicity of delayed cellular neural networks and showed that the n-neural networks can have 2n stable periodic orbits located in 2n subsets of Rn. The multistability of Cohen–Grossberg neural networks with a general class of piecewise activation functions was also discussed in Cao, Feng, and Wang (2008). It was shown in Cao et al. (2008), Cheng, Lin, and Shih (2007), Zeng and Wang (2006), and other papers that under some conditions, the n-neuron networks can have 2n locally exponentially stable equilibrium points located in 2n saturation regions. But it is still unknown what happens in the remaining 3n2n subsets. In Cheng, Lin, and Shih (2006), the authors indicated that there can be 3n equilibrium points for the n-neuron neural networks. However, they only gave emphasis on 2n equilibrium points which are stable in a class of subsets with positive invariance, never mentioned the stability nor the dynamical behaviors of solutions in other 3n2n subsets. To the best of our knowledge, there are few papers addressing the dynamics in the remaining 3n2n subsets of Rn, nor the attraction basins of all stable equilibrium points.

In this paper, we investigate the neural networks (1) and deal with these issues. To be more general, we present a class of nondecreasing piecewise linear activation functions with 2r corner points, which can be described by: fj(x)={mj1<x<pj1,mj2mj1qj1pj1(xpj1)+mj1pj1xqj1,mj2qj1<x<pj2,mj3mj2qj2pj2(xpj2)+mj2pj2xqj2,mj3qj2<x<pj3,mjr+1mjrqjrpjr(xpjr)+mjrpjrxqjr,mjr+1qjr<x<+, where r1,{mjk}k=1r+1 is an increasing constant series, pjk,qjk,k=1,2,,r are constants with <pj1<qj1<pj2<qj2<<pjr<qjr<+,j=1,2,,n.

The neural networks with activation function (2) can store many more patterns or associative memories than those with saturated function, It is meaningful in applications Fig. 2.

In the following, we will precisely figure out all equilibria for the system (1), and investigate the stability and attraction basin for each equilibrium. Discussions and Simulations are also provided to illustrate and verify theoretical results.

We begin with the multistability for r=1.

Section snippets

Case I: r=1

In this case, the activation function fj reduces to fj(x)={mj<x<pj,Mjmjqjpj(xpj)+mjpjxqj,Mjqj<x<+, where mj,Mj,pj,qj are constants with mj<Mj,pj<qj,j=1,2,,n. And we first investigate 2-neuron neural networks. The n-neuron neural networks can be dealt with similarly.

Case II: r1

Inspired by the discussions above, in this section, we discuss the dynamical system (1) with activation fj described by (2).

Theorem 3

Suppose that{dipik+wiimik+jimax{wijmj1,wijmjr+1}+Ii<0,diqik+wiimik+1+jimin{wijmj1,wijmjr+1}+Ii>0,for i,j=1,2,,n,k=1,2,,r . Then, the dynamical system(1)has (2r+1)n equilibrium points. Among them, (r+1)n are locally stable and others are unstable.

Proof

Obviously, R can be divided into (2r+1) subsets, so that Rn can be divided into (2r+1)n subsets. For example, when r=2, R2

Attraction basins of equilibria

In this section, we investigate attraction basins of equilibria for the system (1) with activation functions (2).

We begin with the case n=2 and r=1. Under conditions (5), Theorem 1 tells us that there are 4 locally stable equilibrium points uS1,uS2,uS3,uS4, in subsets S1,S2,S3,S4, respectively, and 5 unstable equilibrium points uΞ1,uΞ2,uΞ3,uΞ4,uΛ in Ξ1,Ξ2,Ξ3,Ξ4,Λ, respectively.

Take a further look at the dynamics in subsets Ξ1,Ξ2,Ξ3,Ξ4, for example, in Ξ1=[p1,q1]×(q2,), system (1) takes the

Discussions

In Sections 2 Case I:, 3 Case II:, we investigate the multistability of neural networks (1) with the activation function given in (2). The proposed method is also applicable when different neurons have different activation functions.

In fact, let r1,,rn be constants such that the activation fj of the j-th neuron has 2rj corner points. Then, we have

Corollary 1

Suppose that{dipik+wiifi(pik)+jimax{wijmj1,wijmjrj+1}+Ii<0,diqik+wiifi(qik)+jimin{wijmj1,wijmjrj+1}+Ii>0,for k=1,,ri,i,j=1,2,,n . Then, the

Simulations

In the following, we present three more examples to illustrate the effectiveness of the theoretical results.

Example 1

Consider the following neural network with 2-neurons: {du1(t)dt=u1(t)+4f1(u1(t))+f2(u2(t))1,du2(t)dt=2u2(t)+f1(u1(t))+5f2(u2(t))1, where the activation functions are fi(x)=|x+1||x1|2,i=1,2.

It is easy to see that the conditions (5) are satisfied. Therefore, by Theorem 1, there exist 9 equilibria, and 4 of them are locally stable while others are unstable. In fact, the equilibrium

Conclusions

In this paper, we study the neural networks with a class of activation functions, which are nondecreasing piecewise linear with 2r(r1) corner points. We prove that such neural networks have multiple equilibria. Some of them are locally stable and others are unstable. More precisely, such neural networks have (2r+1)n equilibria in all, (r+1)n of which are locally exponentially stable and others are unstable. We also give the attraction region for each locally stable equilibrium. For the

Acknowledgements

Wang Lili is supported by Graduate Innovation Foundation of Fudan University under Grant EYH1411028. Lu Wenlian is supported by the National Natural Sciences Foundation of China under Grant No. 60804044, and sponsored by Shanghai Pujiang Program No. 08PJ14019. Chen Tianping is supported by the National Natural Sciences Foundation of China under Grant No. 60774074, 60974015 and SGST 09DZ2272900.

References (19)

There are more references available in the full text version of this article.

Cited by (102)

  • Multistability of state-dependent switching neural networks with discontinuous nonmonotonic piecewise linear activation functions

    2021, Neurocomputing
    Citation Excerpt :

    In multistability analysis, the activation functions are mainly concentrated on nondecreasing saturated activation functions, sigmoid activation functions, Maxican–Hat-type activation functions and so on. In the existing works, it should be remarkable that the research on multistability basically concentrates on neural networks with continuous activation functions [20–26]. When handling dynamical systems with very high-slope nonlinear elements, we often to model them using discontinuous right sides, rather than using high slopes with limited values [27–29].

View all citing articles on Scopus
View full text