New Results of Global Asymptotical Stability for Impulsive Hopfield Neural Networks with Leakage Time-Varying Delay

Abstract

In this paper, Hopfield neural networks with impulse and leakage time-varying delay are considered. New sufficient conditions for global asymptotical stability of the equilibrium point are derived by using Lyapunov-Kravsovskii functional, model transformation and some analysis techniques. The criterion of stability depends on the impulse and the bounds of the leakage time-varying delay and its derivative, and is presented in terms of a linear matrix inequality (LMI).

Share and Cite:

Xi, Q. (2017) New Results of Global Asymptotical Stability for Impulsive Hopfield Neural Networks with Leakage Time-Varying Delay. Journal of Applied Mathematics and Physics, 5, 2112-2126. doi: 10.4236/jamp.2017.511173.

1. Introduction

As we know, time delay is a common phenomenon that describes the fact that the future state of a system depends not only on the present state but also on the past state, and often encountered in many fields such as automatic control, biological chemistry, physical engineer, neural networks, and so on [1] [2] [3] [4] [5] . Moreover, the existence of time delay in a real system may lead to instability, oscillation, and bad dynamic performance [3] [4] [5] . So, it is significant and necessary to consider the delay effects on stability of dynamical systems. In Recent years, one typical class of neural networks, Hopfield neural networks (HNN) have been successfully applied to associative memory, pattern recognition, automatic control, optimization problems, etc, and HNN with various types of delay have been widely investigated by many authors, some interesting and important results have been reported in the literature, see [6] - [16] and the references therein.

On the other hand, impulsive phenomenon exists universally in a wide variety of evolutionary processes where the state is changed abruptly at certain moments of time, involving such fields as chemical technology, population dynamics, physics and economics [17] [18] [19] . Hopfield neural networks may experience change of the state abruptly, that is, do exhibit impulsive effects. Recently, some results for the stability of HNN with impulse as well as delays are obtained via different approaches [20] - [26] .

In the past several years, a special type of time delay, namely, leakage delay (or forgetting delay), is identified and investigated due to its existence in many real systems such as neural networks, population dynamics and some fuzzy systems [1] [3] . Leakage delay is a time delay that exists in the negative feedback terms of the system which are known as forgetting or leakage terms. It has been shown that such kind of time delay has a tendency to destabilize a system [27] . In [27] , Gopalsamy initially investigated the dynamics of bidirectional associative memory (BAM) network model with leakage delays by using model transformation technique, Lyapunov-Kravsovskii functional and inequalities together with some properties of M-matrices. Based on this work, several papers have considered stability of some kinds of neural networks [28] - [34] . More recently, Li et al. [35] , initially studies the impulsive effects on existence-uniqueness and stability problems of recurrent neural networks with leakage delay via some analysis techniques on impulsive functional differential equations. However, it is worth noting that in those existing results, the leakage delay considered is usually a constant. Stability research on leakage time-varying delay has been hardly considered in the literature. In [36] , Li et al. studied the effect of leakage time-varying delay on stability of nonlinear differential systems, but ignored impulsive effect. It is interesting to consider neural networks with leakage time-varying delay as well as impulse, which describes more realistic models [37] - [40] .

With the above motivation, in this paper, we consider Hopfield neural networks with leakage time-varying delay and impulse. By using Lyapunov-Kravsovskii functional, model transformation and some analysis techniques, New sufficient conditions for global asymptotical stability of the equilibrium point are derived. The criterion depends on the impulse and the bounds or length of the leakage time-varying delay and its derivative, and is given in terms of a linear matrix inequality (LMI). The developed results generalize the corresponding results in reference [36] . The work is organized as follows. In Section 2, we introduce the model, some basic notations and lemmas. In Section 3, we present the main results. Finally, the paper is concluded in Section 4.

2. Preliminaries

Notations. Let denote the set of real numbers, + the set of nonnegative real numbers, + the set of positive integers, n the n-dimensional real space and n × m n × m -dimensional real space equipped with the Euclidean norm | | ,

respectively. For S = ( s i j ) n × n , set S 2 = i = 1 n j = 1 n s i j 2 . A > 0 or A < 0

denotes that the matrix A is a symmetric and positive definite or negative definite matrix. The notation A T and A denote the transpose and the inverse of A , respectively. If A , B are symmetric matrices, A > B means that A B is positive definite matrix. λ m a x ( A ) and λ m i n ( A ) denote the maximum eigenvalue and the minimum eigenvalue of matrix A , respectively. E denotes the identity matrix with appropriate dimensions and Λ = { 1 , 2 , , n } . For any J , S k ( 1 k n ) , set ( J , S ) = { ϕ : J S is continuous } and 1 ( J , S ) = { ϕ : J S is continuously differentiable everywhere except at finite number of points t at which ϕ ( t + ) , ϕ ( t ) , ϕ ˙ ( t + ) , ϕ ˙ ( t ) exist and ϕ ( t + ) = ϕ ( t ) , ϕ ˙ ( t + ) = ϕ ˙ ( t ) where ϕ ˙ denotes the derivative of ϕ } . For any t + , x t is defined by x t = x ( t + s ) , x t = x ( t + s ) , s [ σ , 0 ] . The notation always denotes the symmetric block in one symmetric matrix.

Consider the following impulsive hopfield neural networks with leakage time- varying delay:

{ x ˙ ( t ) = C x ( t σ ( t ) ) + A f ( x ( t ) ) + B g ( x ( t τ ( t ) ) ) + J , t > 0 , t t k , Δ x ( t k ) = x ( t k ) x ( t k ) = J k ( x ( t k ) , x t k ) , k + , x ( t ) = φ ( t ) , t [ η , 0 ] , (1)

where x ( t ) = ( x 1 ( t ) , , x n ( t ) ) T is the neuron state vector of the neural networks; C = d i a g ( c 1 , , c n ) is a diagonal matrix with c i > 0, i Λ ; A and B are the connection weight matrix and the delayed weight matrix, respectively; J is an external input; f and g represent the neuron activation functions. Through-out this paper, we make the following assumptions:

(H1) σ ( t ) and τ ( t ) denote the time-varying leakage delay and time-varying transmission delay, respectively, and satisfies 0 σ ( t ) σ , 0 τ ( t ) τ and | σ ˙ ( t ) | ρ σ < 1 , τ ˙ ( t ) ρ τ < 1 , where σ , τ , ρ σ , ρ τ are some real constants;

(H2) J k ( , ) : n × n n , k + , are some continuous functions;

(H3) The impulsive times t k satisfy 0 = t 0 < t 1 < < t k and i n f k + { t k t k 1 } > 0 .

(H4) φ 1 = ˙ 1 ( [ η ,0 ] , n ) , where η = ˙ max { σ , τ } . For φ 1 , define φ η = sup θ [ η , 0 ] | φ ( θ ) | .

The following Lemmas will be used to derive our main results.

Lemma 2.1. [41] Given any real matrices Σ 1 , Σ 2 , Σ 3 of appropriate dimensions and a scalar ϵ > 0 such that 0 < Σ 3 = Σ 3 T . Then the following inequality holds:

Σ 1 T Σ 2 + Σ 2 T Σ 1 ϵ Σ 1 T Σ 3 Σ 1 + ϵ 1 Σ 2 T Σ 3 1 Σ 2 .

Lemma 2.2. [42] Given any real matrix M = M T > 0 of appropriate dimension and a vector function ω ( ) : [ a , b ] n , such that the integrations concerned are well defined, then

[ a b ω ( s ) d s ] T M [ a b ω ( s ) d s ] ( b a ) a b ω T ( s ) M ω ( s ) d s .

Lemma 2.3. [43] Let X n × n , then

λ m i n ( X ) a T a a T X a λ m a x ( X ) a T a

for any a n if X is a symmetric matrix.

Lemma 2.4. [44] A given matrix S = ( S 11 S 12 S 21 S 22 ) > 0 , where S 11 T = S 11 ,

S 22 T = S 22 , is equivalent to any one of the following conditions:

(1) S 22 > 0 , S 11 S 12 S 22 1 S 12 T > 0 ;

(2) S 11 > 0 , S 22 S 12 T S 11 1 S 12 > 0.

In the following, we assume that some normal conditions, such as Lipschitz continuity of f and g, etc, are satisfied so that the equilibrium point of system (1) does exist, see [13] [21] etc, in which the existence results of equilibrium point are established by employing contraction mapping theorem, Brouwer’s fixed point theorem and some functional method. Note that these results are independent of time delays, so it is easy to extend the results in the literatures to an impulsive neural network with leakage time-varying delays and other delays, we omit the details and investigate the global asymptotic stability of the equilibrium point mainly in next section. As usual, we assume that x = ( x 1 , x 2 , , x n ) T is an equilibrium point of system (1), i.e.

C x * + A f ( x * ) + B g ( x * ) + J = 0 , J k ( x * , x * ) = 0 , k + .

3. Global Asymptotic Stability

In this section, we investigate the global asymptotic stability of the unique equilibrium point of system (1). For this purpose, the impulsive function J k which is viewed as a perturbation of the equilibrium point x * of model (1) without impulses is defined by

J k ( x ( t k ) , x t k ) = D k { x ( t k ) x * C t k σ ( t k ) t k ( x ( s ) x * ) d s } , k + ,

where D k , k + are some n × n real symmetric matrices. It is clear that J k ( x * , x * ) = 0, k + . Such a type of impulse describes the fact that the instantaneous perturbations encountered depend not only on the state of neurons at impulse times t k but also the state of neurons in recent history, which reflects a more realistic dynamics. Similar impulsive perturbations have also been investigated by some researchers recently [22] [23] [25] .

For convenience, we let y ( t ) = x ( t ) x * , then system (1) can be rewritten as

{ y ˙ ( t ) = C y ( t σ ( t ) ) + A Ω ( y ( t ) ) + B Γ ( y ( t τ ( t ) ) ) , t > 0 , t t k , Δ y ( t k ) = y ( t k ) y ( t k ) = D k { y ( t k ) C t k σ ( t k ) t k y ( s ) d s } , k + , y ( t ) = φ ( t ) x * , t [ η , 0 ] , (2)

where

Ω ( y ( t ) ) = [ Ω 1 ( y 1 ( t ) ) , Ω 2 ( y 2 ( t ) ) , , Ω n ( y n ( t ) ) ] T , Ω j ( y j ( t ) ) = f j ( x j * + y j ( t ) ) f j ( x j * ) ,

Γ ( y ( t τ ( t ) ) ) = [ Γ 1 ( y 1 ( t τ ( t ) ) ) , Γ 2 ( y 2 ( t τ ( t ) ) ) , , Γ n ( y n ( t τ ( t ) ) ) ] T ,

Γ j ( y j ( t τ ( t ) ) ) = g j ( x j * + y j ( t τ ( t ) ) ) g j ( x j * ) .

Obviously, y 0 is a solution of system (2). Therefore, to consider the stability of the equilibrium point of system (1), it is equal to consider the stability of zero solution of system (2).

In this paper, we assume that there exist constants M 0, N 0 such that

(H5) Ω T ( y ) Ω ( y ) M y T y , Γ T ( y ) Γ ( y ) N y T y ,

which is a very important assumption for activation functions f and g. Using a model transformation, system (2) has an equivalent form as follows:

{ d d t [ y ( t ) C t σ ( t ) t y ( s ) d s ] = C y ( t ) C y ( t σ ( t ) ) σ ˙ ( t ) + A Ω ( y ( t ) ) + B Γ ( y ( t τ ( t ) ) ) , t > 0 , t t k , Δ y ( t k ) = y ( t k ) y ( t k ) = D k { y ( t k ) C t k σ ( t k ) t k y ( s ) d s } , k + , y ( t ) = φ ( t ) x * , t [ η , 0 ] , (3)

In the following, we shall establish a theorem which provides sufficient conditions for global asymptotical stability of the zero solution of system (3). It implies that, if system (1) has an equilibrium point, then it is unique and globally attractive.

Theorem 3.1. Assume that system (1) has one equilibrium and that assumptions (H1)-(H5) hold. Then the equilibrium of system (1) is unique and is globally asymptotically stable if there exist n × n matrices P > 0 , Q i > 0 , i = 1 , 2 , , 7 such that the following LMI holds:

[ ρ σ P C P A P B σ C T P C ρ σ σ C T P C σ C T P A σ C T P B Q 1 0 0 0 0 0 0 Q 2 0 0 0 0 0 Q 3 0 0 0 0 Q 4 0 0 0 Q 5 0 0 Q 6 0 Q 7 ] < 0 , (4)

and

[ P ( E D k ) T P P ] > 0 , k + , (5)

where

= 2 P C + [ λ max ( Q 2 ) + λ max ( Q 6 ) ] M E + Q 4 + ρ σ 1 ρ σ [ Q 1 + Q 5 ] + 1 1 ρ τ λ max ( Q 3 + Q 7 ) N E .

Proof. Let y ( t ) = y ( t , 0 , φ ) be a solution of system (2) through ( 0, φ ) , where φ . Construct a Lyapunov-Krasovskii functional in the form

V ( t , y ) = V 1 ( t , y ) + V 2 ( t , y ) + V 3 ( t , y ) + V 4 ( t , y ) , (6)

where

V 1 = [ y ( t ) C t σ ( t ) t y ( s ) d s ] T P [ y ( t ) C t σ ( t ) t y ( s ) d s ] ,

V 2 = ρ σ 1 ρ σ t σ ( t ) t y T ( s ) [ Q 1 + Q 5 ] y ( s ) d s ,

V 3 = 1 1 ρ τ t τ ( t ) t Γ T ( y ( s ) ) [ Q 3 + Q 7 ] Γ ( y ( s ) ) d s ,

V 4 = σ t σ t s t y T ( u ) Q 8 y ( u ) d u d s ,

Q 8 = C T P C Q 4 1 C T P C + ρ σ C T P C Q 5 1 C T P C + C T P A Q 6 1 A T P C + C T P B Q 7 1 B T P C .

Calculating the upper right derivative of V ( t , y ) along the solution of system (2) at the continuous interval [ t k 1 , t k ) , k + , and considering the Lemma 2.1-2.3, it can be deduced that

D + V 1 = 2 [ y ( t ) C t σ ( t ) t y ( s ) d s ] T P [ C y ( t ) C y ( t σ ( t ) ) σ ˙ ( t ) + A Ω ( y ( t ) ) + B Γ ( y ( t τ ( t ) ) ) ] = 2 y T ( t ) P C y ( t ) 2 y T ( t ) P C y ( t σ ( t ) ) σ ˙ ( t ) + 2 y T ( t ) P A Ω ( y ( t ) ) + 2 y T ( t ) P B Γ ( y ( t τ ( t ) ) ) + 2 [ t σ ( t ) t y ( s ) d s ] T C T P C y ( t ) + 2 [ t σ ( t ) t y ( s ) d s ] T C T P C y ( t σ ( t ) ) σ ˙ ( t ) 2 [ t σ ( t ) t y ( s ) d s ] T C T P A Ω y (t)

2 [ t σ ( t ) t y ( s ) d s ] T C T P B Γ ( y ( t τ ( t ) ) ) 2 y T ( t ) P C y ( t ) + y T ( t ) P C Q 1 1 C T P y ( t ) ρ σ + ρ σ y T ( t σ ( t ) ) Q 1 y ( t σ ( t ) ) + Ω T ( y ( t ) ) Q 2 Ω ( y ( t ) ) + y T ( t ) P A Q 2 1 A T P y ( t ) + Γ T ( y ( t τ ( t ) ) ) Q 3 Γ ( y ( t τ ( t ) ) ) + y T ( t ) P B Q 3 1 B T P y ( t ) + [ t σ ( t ) t y ( s ) d s ] T C T P C Q 4 1 C T P C [ t σ ( t ) t y ( s ) d s ]

+ y T ( t ) Q 4 y ( t ) + ρ σ [ t σ ( t ) t y ( s ) d s ] T C T P C Q 5 1 C T P C [ t σ ( t ) t y ( s ) d s ] + ρ σ y T ( t σ ( t ) ) Q 5 y ( t σ ( t ) ) + [ t σ ( t ) t y ( s ) d s ] T C T P A Q 6 1 A T P C [ t σ ( t ) t y ( s ) d s ] + Ω T ( y ( t ) ) Q 6 Ω ( y ( t ) ) + [ t σ ( t ) t y ( s ) d s ] T C T P B Q 7 1 B T P C [ t σ ( t ) t y ( s ) d s ] + Γ T ( y ( t τ ( t ) ) ) Q 7 Γ ( y ( t τ ( t ) ) )

2 y T ( t ) P C y ( t ) + y T ( t ) P C Q 1 1 C T P y ( t ) ρ σ + ρ σ y T ( t σ ( t ) ) Q 1 y ( t σ ( t ) ) + y T ( t ) λ max ( Q 2 ) M E y ( t ) + y T ( t ) P A Q 2 1 A T P y ( t ) + Γ T ( y ( t τ ( t ) ) ) Q 3 Γ ( y ( t τ ( t ) ) ) + y T ( t ) P B Q 3 1 B T P y ( t ) + [ t σ ( t ) t y ( s ) d s ] T C T P C Q 4 1 C T P C [ t σ ( t ) t y ( s ) d s ] + y T ( t ) Q 4 y ( t ) + ρ σ [ t σ ( t ) t y ( s ) d s ] T C T P C Q 5 1 C T P C [ t σ ( t ) t y ( s ) d s ] (7)

+ ρ σ y T ( t σ ( t ) ) Q 5 y ( t σ ( t ) ) + [ t σ ( t ) t y ( s ) d s ] T C T P A Q 6 1 A T P C [ t σ ( t ) t y ( s ) d s ] + y T ( t ) λ max ( Q 6 ) N E y ( t ) + [ t σ ( t ) t y ( s ) d s ] T C T P B Q 7 1 B T P C [ t σ ( t ) t y ( s ) d s ] + Γ T ( y ( t τ ( t ) ) ) Q 7 Γ ( y ( t τ ( t ) ) ) ,

D + V 2 = ρ σ 1 ρ σ y T ( t ) [ Q 1 + Q 5 ] y ( t ) y T ( t σ ( t ) ) [ Q 1 + Q 5 ] y ( t σ ( t ) ) ρ σ ( 1 σ ˙ ( t ) ) 1 ρ σ ρ σ 1 ρ σ y T ( t ) [ Q 1 + Q 5 ] y ( t ) y T ( t σ ( t ) ) [ Q 1 + Q 5 ] y ( t σ ( t ) ) ρ σ , (8)

D + V 3 = 1 1 ρ τ Γ T ( y ( t ) ) [ Q 3 + Q 7 ] Γ ( y ( t ) ) Γ T ( y ( t τ ( t ) ) ) [ Q 3 + Q 7 ] Γ ( y ( t τ ( t ) ) ) 1 τ ˙ ( t ) 1 ρ τ 1 1 ρ τ Γ T ( y ( t ) ) [ Q 3 + Q 7 ] Γ ( y ( t ) ) Γ T ( y ( t τ ( t ) ) ) [ Q 3 + Q 7 ] Γ ( y ( t τ ( t ) ) ) , (9)

D + V 4 = σ 2 y T ( t ) Q 8 y ( t ) σ t σ t y T ( s ) Q 8 y ( s ) d s σ 2 y T ( t ) Q 8 y ( t ) σ ( t ) t σ ( t ) t y T ( s ) Q 8 y ( s ) d s σ 2 y T ( t ) Q 8 y ( t ) [ t σ ( t ) t y ( s ) d s ] T Q 8 [ t σ ( t ) t y ( s ) d s ] , (10)

where

Q 8 = C T P C Q 4 1 C T P C + ρ σ C T P C Q 5 1 C T P C + C T P A Q 6 1 A T P C + C T P B Q 7 1 B T P C .

Combining (6)-(10), one may deduce that

D + V y T ( t ) [ 2 P C + ρ σ P C Q 1 1 C T P + λ max ( Q 2 ) M E + P A Q 2 1 A T P + P B Q 3 1 B T P + Q 4 + λ max ( Q 6 ) M E + ρ σ 1 ρ σ ( Q 1 + Q 5 ) + 1 1 ρ τ λ max ( Q 3 + Q 7 ) N E + σ 2 C T P C Q 4 1 C T P C + ρ σ σ 2 C T P C Q 5 1 C T P C + σ 2 C T P A Q 6 1 A T P C + σ 2 C T P B Q 7 1 B T P C ] y ( t ) = y T ( t ) Σ y ( t ) ,

where

Σ = 2 P C + ρ σ P C Q 1 1 C T P + λ max ( Q 2 ) M E + P A Q 2 1 A T P + P B Q 3 1 B T P + Q 4 + λ max ( Q 6 ) M E + ρ σ 1 ρ σ ( Q 1 + Q 5 ) + 1 1 ρ τ λ max ( Q 3 + Q 7 ) N E + σ 2 C T P C Q 4 1 C T P C + ρ σ σ 2 C T P C Q 5 1 C T P C + σ 2 C T P A Q 6 1 A T P C + σ 2 C T P B Q 7 1 B T P C

By the well known Schur complements, we know that Σ < 0 if and only if the LMI (4) holds. Hence, one may derive that

D + V ( t , y ) y T ( t ) Σ * y ( t ) , t [ t k 1 , t k ) , k + , (11)

where Σ * = Σ > 0 .

Suppose that t [ t n 1 , t n ) , for some n + . Then integrating inequality (11) at each interval [ t k 1 , t k ) ,1 k n 1 , we derive that

V ( t 1 ) V ( 0 ) 0 t 1 y T ( s ) Σ * y ( s ) d s ,

V ( t 2 ) V ( t 1 ) t 1 t 2 y T ( s ) Σ * y ( s ) d s ,

V ( t n 1 ) V ( t n 2 ) t n 2 t n 1 y T ( s ) Σ * y ( s ) d s ,

V ( t ) V ( t n 1 ) t n 1 t y T ( s ) Σ * y ( s ) d s ,

which implies that

V ( t ) V ( 0 ) 0 t y T ( s ) Σ * y ( s ) d s + 0 < t k t [ V ( t k ) V ( t k ) ] , t 0. (12)

In order to analyze (12), we need consider the change of V at impulse times.

Firstly, it follows from (5) that

[ P ( E D k ) T P P ] > 0 [ E O O P 1 ] [ P ( E D k ) T P P ] [ E O O P 1 ] > 0

[ P ( E D k ) T P 1 ] > 0

P ( E D k ) T P ( E D k ) > 0 , (13)

in which the last equivalent relation is obtained by Lemma 2.4.

Secondly, from system (3), it can be obtained that

y ( t k ) C t k σ ( t k ) t k y ( s ) d s = y ( t k ) D k [ y ( t k ) C t k σ ( t k ) t k y ( s ) d s C t k σ ( t k ) t k y ( s ) d s ] = ( E D k ) [ y ( t k ) C t k σ ( t k ) t k y ( s ) d s ] ,

which together with (13) yields

V 1 ( t k ) = [ y ( t k ) C t k σ ( t k ) t k y ( s ) d s ] T P [ y ( t k ) C t k σ ( t k ) t k y ( s ) d s ] = [ y ( t k ) C t k σ ( t k ) t k y ( s ) d s ] T ( E D k ) T P ( E D k ) [ y ( t k ) C t k σ ( t k ) t k y ( s ) d s ] < [ y ( t k ) C t k σ ( t k ) t k y ( s ) d s ] T P [ y ( t k ) C t k σ ( t k ) t k y ( s ) d s ] = V 1 ( t k ) .

Obviously, we have V i ( t k ) V i ( t k ) , i = 2 , 3 , 4 , k + .

Thus, we can deduce that

V ( t k ) V ( t k ) , k + .

Substituting the above inequality in (12) yields

V ( t ) + 0 t y T ( s ) Σ * y ( s ) d s V ( 0 ) , t 0. (14)

By simple calculation, it can be deduced that

V ( 0 ) { λ m a x ( P ) ( 1 + C σ ) 2 + ρ σ σ 1 ρ σ λ m a x ( Q 1 + Q 5 ) + τ λ m a x ( Q 3 + Q 7 ) N 1 ρ τ + σ 3 λ m a x ( Q 8 ) } φ η 2 = Δ φ η 2 ,

where Δ = λ m a x ( P ) ( 1 + C σ ) 2 + ρ σ σ 1 ρ σ λ m a x ( Q 1 + Q 5 ) + τ λ m a x ( Q 3 + Q 7 ) N 1 ρ τ + σ 3 λ m a x ( Q 8 ) .

It follows that

λ m i n ( P ) y ( t ) C t σ ( t ) t y ( s ) d s V 1 V V ( 0 ) Δ φ η ,

which implies that

y ( t ) C t σ ( t ) t y ( s ) d s + Δ λ m i n ( P ) φ η .

Employing Gronwall inequality, we get

y ( t ) Δ λ m i n ( P ) φ η e σ ( t ) C Δ λ m i n ( P ) e σ C φ η < ,

which implies that the equilibrium point of system (2) is locally stable, and uniformly bounded on [ 0, ) .

Thus, considering the continuity of the activation function f and g, it can be deduced from system (2) that there exists some constant R > 0 such that y ˙ ( t ) R , t [ t k 1 , t k ) , k + , where y ˙ denotes the right-hand derivative of y at impulse times t k 1 , k + .

In the following, we shall prove that y ( t ) 0 as t .

We first show that

y ( t k ) 0, t k . (15)

It is equivalent to prove that | y i ( t k ) | = 0 as t k , i Λ . Note that

| y ˙ i ( t ) | R , t [ t k 1 , t k ) , k + , then for any ϵ > 0 , there exists a δ = ϵ 2 R > 0

such that, for any t , t [ t k 1 , t k ) , k + , | t t | < δ implies that

| y i ( t ) y i ( t ) | R | t t | = ε 2 , i Λ . (16)

By (H3), we define δ ¯ = min { δ , 1 2 θ } , where θ = inf k + { t k t k 1 } > 0 . From

(14), it can be obtained that

0 t | y i ( s ) | 2 d s 0 t y ( s ) T y ( s ) d s 1 λ min ( Σ * ) 0 t y ( s ) T Σ * y ( s ) d s < , t > 0 ,

which implies that t k t k + δ ¯ | y i ( s ) | 2 d s 0 as t k .

Applying Lemma 2.2, we get

t k t k + δ ¯ | y i ( s ) | d s δ ¯ t k t k + δ ¯ | y i ( s ) | 2 d s 0 , t k . (17)

So for the above-given ϵ , there exists a T = T ( ϵ ) > 0 such that t k > T implies that

t k t k + δ ¯ | y i ( s ) | d s < ϵ 2 δ ¯ .

From the continuity of | y i ( t ) | on [ t k , t k + δ ¯ ] , and using the integral mean value theorem, there exists some constant ξ k [ t k , t k + δ ¯ ] such that

| y i ( ξ k ) | δ ¯ = t k t k + δ ¯ | y i ( s ) | d s < ϵ 2 δ ¯ ,

which leads to

| y i ( ξ k ) | < ϵ 2 . (18)

Combining (16) and (18), one may deduce that, for any ϵ > 0 , there exists a T = T ( ϵ ) > 0 such that t k > T implies that

| y i ( t k ) | | y i ( t k ) y i ( ξ k ) | + | y i ( ξ k ) | ϵ 2 + ϵ 2 = ϵ .

This completes the proof of (15).

Now we are in a position to prove that | y i ( t ) | 0 as t , i Λ . In fact,

it follows from (16) that, for any ϵ > 0 , there exists a δ = ϵ 2 M > 0

such that, for any t , t [ t k 1 , t k ) , k + , | t t | < δ implies that

| y i ( t ) y i ( t ) | ϵ 2 , i Λ . (19)

Since (15) holds, there exists a constant T 1 = T 1 ( ϵ ) > 0 such that

| y i ( t k ) | < ϵ 2 , t k > T 1 . (20)

In addition, applying the same argument as in (17), we can deduce that

t t + δ ¯ | y i ( s ) | d s 0 , t ,

where δ ¯ = min { δ , 1 2 θ } , θ = inf k + { t k t k 1 } > 0.

So, for the above-given ϵ , there exists a constant T 2 = T 2 ( ϵ ) > 0 such that

t δ ¯ t | y i ( s ) | d s < ϵ 2 δ ¯ , t > T 2 . (21)

Set T * = min { t q | t q max { T 1 , T 2 } , q + } . Now we claim that | y i ( t ) | ϵ , t > T * . In fact, for any t > T * and without loss of generality assume that t [ t p , t p + 1 ) , p q . We consider the following two cases.

Case1. t [ t p , t p + δ ¯ ] . In this case, it is obvious from (19) and (20) that

| y i ( t ) | | y i ( t ) y i ( t p ) | + | y i ( t p ) | ϵ 2 + ϵ 2 = ϵ

Case2. t [ t p + δ ¯ , t p + 1 ) . In this case, we know that y i ( s ) is continuous on [ t δ ¯ , t ] [ t p , t p + 1 ) . By the integral mean value theorem, there exists at least one point υ t [ t δ ¯ , t ] such that

t δ ¯ t | y i ( s ) | d s = | y i ( υ t ) | δ ¯ ,

which together with (21) yields | y i ( υ t ) | < ϵ 2 . Then, in view of υ t [ t δ ¯ , t ] , we obtain

| y i ( t ) | | y i ( t ) y i ( υ t ) | + | y i ( υ t ) | ϵ 2 + ϵ 2 = ϵ

So we have proved that | y i ( t ) | ϵ , t > T * . Therefore, the zero solution of system (2) or (3) is globally asymptotically stable, which implies that system (1) has a unique equilibrium point which is globally asymptotically stable. The proof of Theorem 3.1 is therefore complete. W

Remark 3.1. Theorem 3.1 provides some delay-dependent conditions for the global asymptotical stability of the unique equilibrium point of impulsive Hopfield neural networks with leakage time-varying delay. We would like to note that such a result has not been reported in other literatures.

In particular, when the leakage delay and transmission delay are all constants, i.e., σ ( t ) σ , τ ( t ) τ , system (1) becomes

{ x ˙ ( t ) C x ( t σ ) + A f ( x ( t ) ) + B g ( x ( t τ ) ) + J , t > 0, t t k , Δ x ( t k ) = x ( t k ) x ( t k ) = J k ( x ( t k ) , x t k ) , k + , x ( t ) = φ ( t ) , t [ η ,0 ] . (22)

For system (22), we have the following result by Theorem 3.1.

Corollary 3.1. Assume that system (22) has one equilibrium and that assumptions (H2)-(H5) hold. Then the equilibrium of system (22) is unique and is globally asymptotically stable if there exist n × n matrices P > 0 , Q i > 0 , i = 1 , 2 , , 5 such that the following LMI holds:

[ P A P B σ C T P C σ C T P A σ C T P B Q 1 0 0 0 0 Q 2 0 0 0 Q 3 0 0 Q 4 0 Q 5 ] < 0,

and

[ P ( E D k ) T P P ] > 0, k + ,

where

= 2 P C + [ λ m a x ( Q 1 ) + λ m a x ( Q 4 ) ] M E + Q 3 + λ m a x ( Q 2 + Q 5 ) N E .

Remark 3.2. The conditions in Corollary 3.1 are independent on transmission delay and dependent only on leakage delay as ρ τ = 0 in Theorem 3.1. So, based on our results, we would like to say that the stability of system (1) is more sensitive to leakage delay, leakage time-varying delay or leakage constant delay. In other words, we should control not only the bound of leakage delay but also the bound of derivative of leakage delay, to obtain the stability of system (1), while the bound of transmission delay τ or τ ( t ) do not affect the stability of system in our results.

Remark 3.3. So far, there are many papers to study the dynamics of time delay systems and impulsive systems, many effective methods and results have been developed [19] - [26] . But, those results cannot be applied to systems with leakage time-varying delay and impulse which could affect the dynamics of system essentially. In this paper, we investigate the stability of impulsive Hopfield neural networks with leakage time-varying delay by model transformation technique and a certain Lyapunov-Krasovskii functional combined with LMI technique and construct a new criterion. How to improve the dynamics of systems with leakage time-varying delay and impulse may be an interesting problem and requires further research.

4. Conclusion

We have studied the global asymptotic stability of the equilibrium point of impulsive Hopfield neural networks with leakage time-varying delay. Via an appropriate Lyapunov-Krasovskii functional and model transformation technique, a new stability criterion which depends on the impulse and the bounds of leakage time-varying delay and its derivative has been presented in terms of a linear matrix inequality. To the best of our knowledge, so far, few authors have considered the dynamics of systems with leakage time-varying delay and impulse which could affect the dynamics of neural networks essentially. How to further improve the conservation of the developed results is still a difficult problem and need consideration in the future work.

Fund

Supported by National Natural Science Foundation of China (11601269) and Project of Shan-dong Province Higher Educational Science and Technology Program (J15LI02).

Conflicts of Interest

The authors declare no conflicts of interest.

References

[1] Gopalsamy, K. (1992) Stability and Oscillations in Delay Differential Equations of Population Dynamics. Kluwer, Dordrecht.
[2] Hale, J. and Verduyn Lunel, S. (1993) Introduction to Functional Differential Equations. Springer-Verlag, New York.
[3] Haykin, S. (1994) Neural Networks. Prentice-Hall, NJ.
[4] Kolmanovskii, V. and Myshkis, A. (1999) Introduction to the Theory and Applications of Functional Differential Equations. Kluwer, Dordrecht, The Netherlands.
[5] Niculescu, S. (2001) Delay Effects on Stability: A Robust Control Approach. Springer-Verlag, New York.
[6] Van den Driessche, P. and Zou, X. (1998) Global Attractivity in Delayed Hopfield Neural Network Models. SIAM Journal on Applied Mathematics, 58, 1878-1890.
https://doi.org/10.1137/S0036139997321219
[7] Liao, X. and Yu, J. (1998) Robust Stability for Interval Hopfield Neural Networks with Time Delay. IEEE Transactions on Neural Networks, 9, 1042-1045.
https://doi.org/10.1109/72.712187
[8] Liao, X. and Xiao D. (2000) Global Exponential Stability of Hopfield Neural Networks with Time Varying Delays. Acta Electronica Sinica, 28, 87-90.
[9] Chen, T. (2001) Global Exponential Stability of Delayed Hopfield Neural Networks. Neural Networks, 14, 977-980.
[10] Wang, L. and Xu, D. (2002) Stability Analysis of Hopfield Neural Networks with Time Delays. Applied Mathematics and Mechanics (English Edition), 23, 65-70.
https://doi.org/10.1007/BF02437731
[11] Chen, G., Pu, Z. and Zhang, J. (2005) Global Exponential Stability and Global Attractivity for Variably Delayed Hopfield Neural Network Models. Chinese Journal of Engineering Mathematics, 22, 821-826.
[12] Hou, Y., Liao, T. and Yan, J. (2007) Global Asymptotic Stability for a Class of Nonlinear Neural Networks with Multiple Delays. Nonlinear Analysis, Theory, Methods, Applications, 67, 3037-3040.
[13] Zhang, Q., Wei, X. and Xu, J. (2007) Delay-Dependent Global Stability Results for Delayed Hopfield Neural Networks. Chaos, Solitons Fractals, 34, 662-668.
[14] Lou, X. and Cui, B. (2007) Novel Global Stability Criteria for High-Order Hopfield-Type Neural Networks with Time-Varying Delays. Journal of Mathematical Analysis and Applications, 330, 144-158.
[15] Zhou, Q. and Wan, L. (2008) Exponential Stability of Stochastic Delayed Hopfield Neural Networks. Applied Mathematics and Computation, 206, 818-824.
[16] Wu, H. (2009) Global Exponential Stability of Hopfield Neural Networks with Delays and Inverse Lipschitz Neuron Activations. Nonlinear Analysis: Real World Applications, 14, 2776-2783.
[17] Bainov, D. and Simennov, P. (1989) Systems with Impulsive Effect Stability Theory and Applications. Ellis Horwood Limited, New York.
[18] Fu, X., Yan, B. and Liu, Y. (2005) Introduction of Impulsive Differential Systems. Science Press, Beijing.
[19] Liu, X. and Wang, Q. (2007) The Method of Lyapunov Functionals and Exponential Stability of Impulsive Systems with Time Delay. Nonlinear Analysis, Theory, Methods, Applications, 66, 1465-1484.
[20] Liu, B., Liu, X. and Liao, X. (2003) Robust H-Stability of Hopfield Neural Networks with Impulsive Effects and Design of Impulsive Controllers. Control Theory Applications, 20, 168-172.
[21] Yang, Z. and Xu, D. (2006) Global Exponential Stability of Hopfield Neural Networks with Variable Delays and Impulsive Effects. Applied Mathematics and Machanics, 27, 1517-1522.
https://doi.org/10.1007/s10483-006-1109-1
[22] Shen, J., Liu, Y. and Li, J. (2007) Asymptotic Behavior of Solutions of Nonlinear Neural Differential Equations with Impulses. Journal of Mathematical Analysis and Applications, 332, 179-189.
[23] Zhou, Q. (2009) Global Exponential Stability of BAM Neural Networks with Distributed Delays and Impulses. Nonlinear Analysis: Real World Applications, 10, 144-153.
[24] Li, X. and Chen, Z. (2009) Stability Properties for Hopfield Neural Networks with Delays and Impulsive Perturbations. Nonlinear Analysis, Theory, Methods, Applications, 10, 3253-3265.
[25] Fu, X. and Li, X. (2009) Global Exponential Stability and Global Attractivity of Impulsive Hopfield Neural Networks with Time Delays. Journal of Computational and Applied Mathematics, 231, 187-199.
[26] Chen, J., Li, X. and Wang, D. (2013) Asymptotic Stability and Exponential Stability of Impulsive Delayed Hopfield Neural Networks. Abstract and Applied Analysis, 2013, Article ID: 638496.
[27] Gopalsamy, K. (2007) Leakage Delays in BAM. Journal of Mathematical Analysis and Applications, 325, 1117-1132.
[28] Peng, S. (2010) Attractive Periodic Solutions of BAM Neural Networks with Continuously Distributed Delays in the Leakage Terms. Nonlinear Analysis: Real World Applications, 11, 2141-2151.
[29] Li, X. and Cao, J. (2010) Delay-Dependent Stability of Neural Networks of Neutral Type with Time Delay in the Leakage Term. Nonlinearity, 23, 1709-1726.
https://doi.org/10.1088/0951-7715/23/7/010
[30] Balasubramaniam, P., Kalpana, M. and Rakkiyappan, R. (2011) State Estimation for Fuzzy Cellular Neural Networks with Time Delay in the Leakage Term, Discrete and Unbounded Distributed Delays. Computers and Mathematics with Applications, 62, 3959-3972.
[31] Lakshmanan, S., Park, J., Jung, H. and Balasubramaniam, P. (2012) Design of State Estimator for Neural Networks with Leakage, Discrete and Distributed Delays. Applied Mathematics and Computation, 218, 11297-11310.
[32] Li, Z. and Xu, R. (2012) Global Asymptotic Stability of Stochastic Reaction-Diffusion Neural Networks with Time Delays in the Leakage Terms. Communications in Nonlinear Science and Numerical Simulation, 17, 1681-1689.
[33] Lakshmanan, S., Park, J., Lee, T., Jung, H. and Rakkiyappan, R. (2013) Stability Criteria for BAM Neural Networks with Leakage Delays and Probabilistic Time-Varying Delays. Applied Mathematics and Computation, 219, 9408-9423.
[34] Li, X. and Rakkiyappan, R. (2013) Stability Results for Takagi CSugeno Fuzzy Uncertain BAM Neural Networks with Time Delays in the Leakage Term. Neural Computing and Applications, 22, 203-219.
[35] Li, X., Fu, X., Balasubramaniam, P. and Rakkiyappan, R. (2010) Existence, Uniqueness and Stability Analysis of Recurrent Neural Networks with Time Delay in the Leakage Term under Impulsive Perturbations. Nonlinear Analysis: Real World Applications, 11, 4092-4108.
[36] Li, X. and Fu, X. (2013) Effect of Leakage Time-Varying Delay on Stability of Nonlinear Differential Systems. Journal of the Franklin Institute, 350, 1335-1344.
[37] Rakkiyappan, R., Chandrasekar, A. and Lakshmanan, S. (2013) Effects of Leakage Time-Varying Delays in Mar-kovian Jump Neural Networks with Impulse Control. Neurocomputing, 121, 365-378.
[38] Lu, C. and Wang, L. (2014) Robust Exponential Stability of Impulsive Stochastic Neural Networks with Leakage Time-Varying Delay. Abstract and Applied Analysis, 2014, Article ID: 831027.
[39] Suresh Kumar, R., Sugumaran, G., Raja, R., Zhu, Q. and Karthik Raja, U. (2016) New Stability Criterion of Neural Networks with Leakage Delays and Impulses: A Piecewise Delay Method. Cognitive Neurodynamics, 10, 85-98.
https://doi.org/10.1007/s11571-015-9356-y
[40] Balasundaram, K., Raja, R., Zhu, Q., Chandrasekaran, S. and Zhou, H. (2016) New Global Asymptotic Stability of Discrete-Time Recurrent Neural Networks with Multiple Time-Varying Delays in the Leakage Term and Impulsive Effects. Cognitive Neurodynamics, 214, 420-429.
[41] Li, C. and Huang, T. (2009) On the Stability of Nonlinear Systems with Leakage Delay. Journal of the Franklin Institute, 346, 366-377.
[42] Gu, K. (2000) An Integral Inequality in the Stability Problem of Time-Delay Systems. Proceedings of the 39th IEEE Conference on Decision and Control Sydney, December, 2805-2810.
[43] Berman, A. and Plemmons, R.J. (1979) Nonnegative Matrics in Mathematical Sciences. Academic Press, New York.
[44] Boyd, S., Ghaoui, L. E., Feron, E. and Balakrishnan, V. (1994) Linear Matrix Inequalities in System and Control Theory. SIAM, Philadelphia.

Copyright © 2024 by authors and Scientific Research Publishing Inc.

Creative Commons License

This work and the related PDF file are licensed under a Creative Commons Attribution 4.0 International License.