1 Introduction

The theory of impulsive differential equations describes processes which experience a sudden change of their state at certain moments. Processes with such a character arise naturally and often, especially in phenomena studied in physics, chemical technology, population dynamics, biotechnology, and economics. For an introduction of the basic theory of impulsive differential equations in R n , see Lakshmikantham et al. [1], Bainov and Simeonov [2], Samoilenko and Perestyuk [3], and the references therein. The theory of impulsive differential equations has become an important area of investigation in recent years, and it is much richer than the corresponding theory of differential equations (see, for instance, [46] and the references therein).

On the other hand, the theory of boundary value problems with integral boundary conditions for ordinary differential equations arises in different areas of applied mathematics and physics. For example, heat conduction, chemical engineering, underground water flow, thermo-elasticity, and plasma physics can be reduced to nonlocal problems with integral boundary conditions. For boundary value problems with integral boundary conditions and comments on their importance, we refer the reader to the papers in Gallardo [79] and the references therein. For more information about the general theory of integral equations and their relation with boundary value problems, we refer to the book of Corduneanu [10] and Agarwal and O’Regan [11].

Recently, some well-known works, such as Hao et al. [12], Zhang et al. [13] and Ding and Wang [14], deal with impulsive differential equations with integral boundary conditions. However, most of these results are obtained by using the fixed point theorem in cones. It is well known that the method of upper and lower solutions is a powerful tool for proving the existence results for a large class of boundary value problems, see [1517].

In [17], Shen and Wang applied the method of upper and lower solutions to solve impulsive differential equations with nonlinear boundary conditions as follows:

{ x ( t ) = f ( t , x ( t ) , x ( t ) ) , t J , t t k , Δ x ( t k ) = I k ( x ( t k ) ) , k = 1 , 2 , , p , Δ x ( t k ) = J k ( x ( t k ) , x ( t k ) ) , k = 1 , 2 , , p , g ( x ( 0 ) , x ( 0 ) ) = 0 , h ( x ( 1 ) , x ( 1 ) ) = 0 .

Motivated by the works mentioned above, in this paper, we shall employ the method of upper and lower solutions together with Leray-Schauder degree theory to study the existence of a solution of the impulsive BVP

{ x ( t ) + f ( t , x ( t ) , x ( t ) ) = 0 , t J , Δ x ( t k ) = I k ( x ( t k ) ) , k = 1 , 2 , , p , Δ x ( t k ) = J k ( x ( t k ) , x ( t k ) ) , k = 1 , 2 , , p , x ( 0 ) = x ( 1 ) = 0 1 g ( s ) x ( s ) d s ,
(1.1)

where J=[0,1], J =J{ t 1 , t 2 ,, t p }, f:J× R 2 R is continuous, I k , J k C(R) for 1kp, g L 1 [0,1] is nonnegative, 0= t 0 < t 1 < t 2 << t p < t p + 1 =1, Δx( t k )=x( t k + )x( t k ) denotes the jump of x(t) at t= t k , x( t k + ) and x( t k ) represent the right and left limits of x(t) at t= t k , respectively, Δ x ( t k ) has a similar meaning for x (t).

2 Preliminaries

Define PC(J) = {x:JRxC( J ), x( t i + ) and x( t i ) exist, and x( t i )=x( t i ), i=1,2,,p}. P C 1 (J) = {xPC(J):x | ( t i , t i + 1 ) C 1 ( t i , t i + 1 ), x ( t i + ) and x ( t i ) exist, and x ( t i )= x ( t i ), i=1,2,,p}. Note that PC(J) and P C 1 (J) are Banach spaces with the respective norms

x =sup { | x ( t ) | : t J } ,x=max { x , x } .

A function xE=P C 1 (J) C 2 ( J ) is called a solution of (1.1) if it satisfies the differential equation

x (t)+f ( t , x ( t ) , x ( t ) ) =0,t J ,

the impulsive conditions Δx( t k )= I k (x( t k )), Δ x ( t k )= J k (x( t k ), x ( t k )) and the boundary conditions x(0)=x(1)= 0 1 g(s)x(s)ds.

Definition 2.1 The function αE is said to be a lower solution for boundary value problem (1.1) if

α ( t ) + f ( t , α ( t ) , α ( t ) ) 0 , t J , Δ α ( t k ) = I k ( α ( t k ) ) , Δ α ( t k ) = J k ( α ( t k ) , α ( t k ) ) , k = 1 , 2 , , p , α ( 0 ) 0 1 g ( s ) α ( s ) d s , α ( 1 ) 0 1 g ( s ) α ( s ) d s .

The function βE is said to be an upper solution for boundary value problem (1.1) if

β ( t ) + f ( t , β ( t ) , β ( t ) ) 0 , t J , Δ β ( t k ) = I k ( β ( t k ) ) , Δ β ( t k ) = J k ( β ( t k ) , β ( t k ) ) , k = 1 , 2 , , p , β ( 0 ) 0 1 g ( s ) β ( s ) d s , β ( 1 ) 0 1 g ( s ) β ( s ) d s .

For α,βE, we write αβ if α(t)β(t) for all tJ.

Definition 2.2 Let α,βE be such that αβ on J. We say that f satisfies the Nagumo condition relative to α, β if for

ω= max 1 k p { | α ( t k + 1 ) β ( t k + ) | t k + 1 t k , | β ( t k + 1 ) α ( t k + ) | t k + 1 t k } ,

there exists a constant D such that

D>max { ω , α , β } ,

a continuous function Ψ:[0,)(0,), and constants A0, B0 such that

| f ( t , x , y ) | A|y|Ψ ( | y | ) +B,tJ,αxβ,yR

and

ω D 1 Ψ ( s ) ds>A [ max t J β ( t ) min t J α ( t ) ] +B max s ω Ψ 1 (s).

In addition, we assume that the following conditions hold:

(H1) f satisfies the Nagumo condition relative to α, β;

(H2) I k , J k C(R,R). J k (y,z) is nondecreasing in z[D,D] for all 1kp;

(H3) 0< 0 1 g(s)ds<1.

We consider the modified problem

x ( t ) + F ( t , x ( t ) , d d t m α , β ( t , x ) ) = 0 , t J , Δ x ( t k ) = I k ( m α , β ( t k , x ( t k ) ) ) , k = 1 , 2 , , p , Δ x ( t k ) = J k ( m α , β ( t k , x ( t k ) ) , n ( t k , x ( t k ) ) ) , k = 1 , 2 , , p , x ( 0 ) = x ( 1 ) = 0 1 g ( s ) x ( s ) d s ,
(2.1)

where

F(t,x,y)=f ( t , m α , β ( t , x ) , n ( t , y ) ) + m α , β ( t , x ) x 1 + ( x m α , β ( t , x ) ) 2

and

m α , β ( t , x ) = max { α ( t ) , min { x , β ( t ) } } , n ( t , x ) = max { D , min { x , D } } .

One can find the next result, with its proof, in [18].

Lemma 2.1 For each xE, the following two properties hold:

  1. (i)

    d d t m α , β (t,x) exists for a.e. t J ;

  2. (ii)

    if x, x m E and x m x in E, then d d t m α , β (t, x m (t)) d d t m α , β (t,x(t)) for a.e. t J .

Lemma 2.2 For any v(t)C(J), the following boundary value problem

{ x ( t ) + v ( t ) = 0 , t J , Δ x ( t k ) = I k ( x ( t k ) ) , k = 1 , 2 , , p , Δ x ( t k ) = J k ( x ( t k ) , x ( t k ) ) , k = 1 , 2 , , p , x ( 0 ) = x ( 1 ) = 0 1 g ( s ) x ( s ) d s

has a unique solution as follows:

x ( t ) = 0 1 H ( t , s ) v ( s ) d s + 0 < t k < t [ I k ( x ( t k ) ) + ( t t k ) J k ( x ( t k ) , x ( t k ) ) ] t k = 1 p [ I k ( x ( t k ) ) + ( 1 t k ) J k ( x ( t k ) , x ( t k ) ) ] + 1 1 μ 0 1 0 < t k < s [ I k ( x ( t k ) ) + ( s t k ) J k ( x ( t k ) , x ( t k ) ) ] d s ν 1 μ k = 1 p [ I k ( x ( t k ) ) + ( 1 t k ) J k ( x ( t k ) , x ( t k ) ) ] ,
(2.2)

where

H ( t , s ) = G ( t , s ) + 1 1 μ 0 1 G ( s , τ ) g ( τ ) d τ , G ( t , s ) = { t ( 1 s ) , 0 t s 1 , s ( 1 t ) , 0 s t 1 ,

and

μ= 0 1 g(s)ds,ν= 0 1 sg(s)ds.

Lemma 2.3 If x is a solution of BVP (2.1), α(t) and β(t) are lower and upper solutions of (1.1), respectively, αβ, and

I k ( α ( t k ) ) I k (x) I k ( β ( t k ) ) ,k=1,,p,
(2.3)

for

α( t k )xβ( t k ),

then

α(t)x(t)β(t),t[0,1].

Proof Denote y(t)=x(t)β(t), we will only see that x(t)β(t) for every tJ. An analogous reasoning shows that x(t)α(t) for all tJ. Otherwise, if x(t)β(t), t[0,1], does not hold, then sup 0 t 1 (x(t)β(t))>0, there are three cases.

Case 1. Suppose that max 0 t 1 y(t)= sup 0 t 1 (x(t)β(t))=y(0), or max 0 t 1 y(t)=y(1), we only see that max 0 t 1 y(t)=y(0). Easily, it holds that y(0)>0. From Definition 2.1 and (H3), we have

y ( 0 ) = x ( 0 ) β ( 0 ) 0 1 g ( s ) ( x ( s ) β ( s ) ) d s max s [ 0 , 1 ] ( x ( s ) β ( s ) ) 0 1 g ( s ) d s < max s [ 0 , 1 ] ( x ( s ) β ( s ) ) = max s [ 0 , 1 ] y ( s ) ,

which is a contradiction.

Case 2. Suppose that there exist k{0,1,,p} and τ( t k , t k +1) such that

sup t ( t k , t k + 1 ] y(t)=y(τ)=x(τ)β(τ)>0.

Then y (τ)=0 and y (τ)0. On the other hand,

y ( τ ) F ( τ , x ( τ ) , d d t m α , β ( τ , x ) ) + f ( τ , β ( τ ) , β ( τ ) ) = f ( τ , m ( τ , x ( τ ) ) , d d t m α , β ( τ , x ) ) m α , β ( τ , x ( τ ) ) x ( τ ) 1 + ( x ( τ ) m α , β ( τ , x ( τ ) ) ) 2 + f ( τ , β ( τ ) , β ( τ ) ) = f ( τ , β ( τ ) , β ( τ ) ) m α , β ( τ , x ( τ ) ) x ( τ ) 1 + ( x ( τ ) m α , β ( τ , x ( τ ) ) ) 2 + f ( τ , β ( τ ) , β ( τ ) ) = y ( τ ) 1 + y ( τ ) 2 > 0 ,

which is a contradiction. Hence the function y cannot have any positive maximum interior to the interval ( t k , t k +1) for k=1,2,,p.

Case 3. According to Case 2, if sup 0 t 1 y(t)>0, then sup 0 t 1 y(t)=y( t k + ), or sup 0 t 1 y(t)=y( t k )=y( t k ), we only prove that sup 0 t 1 y(t)=y( t k ), k=1,2,,p. Suppose that sup 0 t 1 y(t)=y( t 1 ), easily, y ( t 1 )= y ( t 1 )0.

From (2.1) and (2.3), we have

x ( t 1 + ) x ( t 1 ) = x ( t 1 + ) x ( t 1 ) = I 1 ( m α , β ( t 1 , x ( t 1 ) ) ) = I 1 ( β ( t 1 ) ) = β ( t 1 + ) β ( t 1 ) = β ( t 1 + ) β ( t 1 ) , x ( t 1 + ) x ( t 1 ) = x ( t 1 + ) x ( t 1 ) = J 1 ( m α , β ( t 1 , x ( t 1 ) ) , n ( t 1 , x ( t 1 ) ) ) J 1 ( β ( t 1 ) , β ( t 1 ) ) β ( t 1 + ) β ( t 1 ) .

Hence

y ( t 1 + ) = x ( t 1 + ) β ( t 1 + ) = x ( t 1 ) β ( t 1 ) > 0 , y ( t 1 + ) = x ( t 1 + ) β ( t 1 + ) x ( t 1 ) β ( t 1 ) 0 .

Suppose that y ( t 1 + )=0 and y is nonincreasing on some interval ( t 1 , t 1 +γ)( t 1 , t 2 ), where γ>0 is sufficiently small such that y(t)>0 on t( t 1 , t 1 +γ). For t( t 1 , t 1 +γ),

x ( t ) β ( t ) f ( t , β ( t ) , d d t m α , β ( t , x ) ) m α , β ( t , x ) x 1 + ( x m α , β ( t , x ) ) 2 + f ( t , β ( t ) , β ( t ) ) = m α , β ( t , x ) x 1 + ( x m α , β ( t , x ) ) 2 > 0 ,

which contradicts the assumption of monotonicity of y. Thus, we obtain

0 < y ( t 1 + ) < y ( t 2 ) = y ( t 2 ) , y ( t 2 ) = y ( t 2 ) 0 .

We use the preceding procedure and deduce by induction that

y( t k )>0, y ( t k )0,k=1,2,,p+1,

which contradicts that y(1) cannot be the maximum point. By the same analysis, we can get that y( t k )>0, k=2,3,,p, cannot hold. □

Lemma 2.4 D x (t)D on J, where x(t) is the solution of (2.1).

Proof Here we only show x (t)D. Suppose that there exists ρ( t k , t k + 1 ) with x (ρ)= x ( t k + 1 ) x ( t k ) t k + 1 t k , k{1,2,,p}, and as a result,

D<ω α ( t k + 1 ) β ( t k ) t k + 1 t k x (ρ) β ( t k + 1 ) α ( t k ) t k + 1 t k ω<D.

Therefore, there exist r 1 , r 2 ( t k , t k + 1 ) such that x ( r 1 )=ω, x ( r 2 )=D and either

ω x (t)D,t( r 1 , r 2 ),

or

ω x (t)D,t( r 2 , r 1 ).

We only consider the first case, since the other case can be handled similarly. It follows from the assumption that

x (t)A x (t)Ψ ( x ( t ) ) +Bfor t( r 1 , r 2 ).

This implies that

r 1 r 2 x ( t ) Ψ ( x ( t ) ) d t A r 1 r 2 x ( t ) d t + B r 1 r 2 Ψ 1 ( x ( t ) ) d t = A ( x ( r 2 ) x ( r 1 ) ) + B r 1 r 2 Ψ 1 ( x ( t ) ) d t ,

which yields

ω D d s Ψ ( s ) = x ( r 1 ) x ( r 2 ) d s Ψ ( s ) A ( x ( r 2 ) x ( r 1 ) ) + B ( r 2 r 1 ) max s ω Ψ 1 ( s ) A ( max t J β ( t ) min t J α ( t ) ) + B max s ω Ψ 1 ( s ) .

This, obviously, contradicts the choice of D. The proof is complete. □

Lemma 2.5 (Schauder’s fixed point theorem)

Let K be a convex subset of a normed linear space E. Each continuous, compact map L:KK has a fixed point.

3 Existence results

Theorem 3.1 Suppose that conditions (H1)-(H3) hold. Then BVP (1.1) has at least one solution xE C 2 ( J ) such that

α(t)x(t)β(t),D x (t)D,t[0,1].

Proof Solving (2.1) is equivalent to finding xE which satisfies

x ( t ) = 0 1 H ( t , s ) F ( s , x ( s ) , d d s m α , β ( s , x ) ) d s + 0 < t k < t [ I k ( x ( t k ) ) + ( t t k ) J k ( x ( t k ) , x ( t k ) ) ] t k = 1 p [ I k ( x ( t k ) ) + ( 1 t k ) J k ( x ( t k ) , x ( t k ) ) ] + 1 1 μ 0 1 0 < t k < s [ I k ( x ( t k ) ) + ( s t k ) J k ( x ( t k ) , x ( t k ) ) ] d s ν 1 μ k = 1 p [ I k ( x ( t k ) ) + ( 1 t k ) J k ( x ( t k ) , x ( t k ) ) ] ,

where I k (x( t k ))= I k ( m α , β ( t k ,x( t k ))), J k (x( t k ), x ( t k ))= J k ( m α , β ( t k ,x( t k )),n( t k , x ( t k ))).

Now, define the following operator T:EE by

( T x ) ( t ) = 0 1 H ( t , s ) F ( s , x ( s ) , d d s m α , β ( s , x ) ) d s + 0 < t k < t [ I k ( x ( t k ) ) + ( t t k ) J k ( x ( t k ) , x ( t k ) ) ] t k = 1 p [ I k ( x ( t k ) ) + ( 1 t k ) J k ( x ( t k ) , x ( t k ) ) ] + 1 1 μ 0 1 0 < t k < s [ I k ( x ( t k ) ) + ( s t k ) J k ( x ( t k ) , x ( t k ) ) ] d s ν 1 μ k = 1 p [ I k ( x ( t k ) ) + ( 1 t k ) J k ( x ( t k ) , x ( t k ) ) ] .

It is obvious that T:EE is completely continuous.

By the Schauder fixed point theorem, we can easily obtain that T has a fixed point xE, which is a solution of BVP (2.1). And by Lemma 2.3 and Lemma 2.4, we know that α(t)x(t)β(t), D x (t)D, then BVP (2.1) becomes BVP (1.1), therefore x(t) is a solution of BVP (1.1). The proof is complete. □

Theorem 3.2 Suppose that conditions (H1)-(H3) hold. Assume that there exist two lower solutions α 1 and α 2 and two upper solutions β 1 and β 2 for problem (1.1), satisfying the following:

  1. (i)

    α 1 α 2 β 2 ;

  2. (ii)

    α 1 β 1 β 2 ;

  3. (iii)

    α 2 β 1 , which means that there exists t[0,1] such that α 2 > β 1 ;

  4. (iv)

    if x is a solution of (1.1) with x α 2 , then x> α 2 on (0,1);

  5. (v)

    if x is a solution of (1.1) with x β 1 , then x< β 1 on (0,1).

If f satisfies the Nagumo condition with respect to α 1 , β 2 , then problem (1.1) has at least three solutions x 1 , x 2 and x 3 satisfying

α 1 x 1 β 1 , α 2 x 2 β 2 , x 3 β 1 , x 3 α 2 .
(3.1)

Proof We consider the following modified problem:

x ( t ) + F ( t , x ( t ) , d d t m α 1 , β 2 ( t , x ) ) = 0 , t J , Δ x ( t k ) = I ¯ k ( x ( t k ) ) , k = 1 , 2 , , p , Δ x ( t k ) = J ¯ k ( x ( t k ) , x ( t k ) ) , k = 1 , 2 , , p , x ( 0 ) = x ( 1 ) = 0 1 g ( s ) x ( s ) d s ,
(3.2)

where I ¯ k (x( t k ))= I k ( m α 1 , β 2 ( t k ,x( t k ))), J ¯ k (x( t k ), x ( t k ))= J k ( m α 1 , β 2 ( t k ,x( t k )),n( t k , x ( t k ))).

Now define the following operator T ¯ :EE by

( T ¯ x ) ( t ) = 0 1 H ( t , s ) F ( s , x ( s ) , d d s m α 1 , β 2 ( s , x ) ) d s + 0 < t k < t [ I ¯ k ( x ( t k ) ) + ( t t k ) J ¯ k ( x ( t k ) , x ( t k ) ) ] t k = 1 p [ I ¯ k ( x ( t k ) ) + ( 1 t k ) J ¯ k ( x ( t k ) , x ( t k ) ) ] + 1 1 μ 0 1 0 < t k < s [ I ¯ k ( x ( t k ) ) + ( s t k ) J ¯ k ( x ( t k ) , x ( t k ) ) ] d s ν 1 μ k = 1 p [ I ¯ k ( x ( t k ) ) + ( 1 t k ) J ¯ k ( x ( t k ) , x ( t k ) ) ] .

Since the functions m α 1 , β 2 and n are continuous and bounded, we obtain that there exists M F >0 such that

| F ( t , x ( t ) , x ( t ) ) | M F .

Let

M= 2 M F 1 μ + 3 + ν 2 μ 1 μ k = 1 p [ max x [ d , d ] | I k ( x ) | + max x [ d , d ] , y [ D , D ] | J k ( x , y ) | ] ,

where d=1+ max t J { α 1 , β 2 } and

Ω= { x E , x M } .

It is standard that T ¯ :EE is completely continuous. It is immediate from the argument above that T ¯ ( Ω ¯ )Ω.

Thus,

deg(I T ¯ ,Ω,0)=1.

Let

Ω α 2 = { x Ω : x > α 2  on  ( 0 , 1 ) } , Ω β 1 = { x Ω : x < β 1  on  ( 0 , 1 ) } .

Since α 2 β 1 , α 2 > M ¯ , β 1 < M ¯ (i.e., choose M ¯ such that α 2 , β 1 <M). It follows that Ω β 1 Ω α 2 , Ω β 1 Ω α 2 =, and Ω( Ω β 1 Ω α 2 ¯ ). By assumptions (iv) and (v), there are no solutions in Ω β 1 Ω α 2 . Thus,

deg(I T ¯ ,Ω,0)=deg ( I T ¯ , Ω ( Ω β 1 Ω α 2 ¯ ) , 0 ) +deg ( I T ¯ , Ω β 1 , 0 ) +deg(I T ¯ , Ω α 2 ,0).

We show that deg(I T ¯ , Ω β 1 ,0)=deg(I T ¯ , Ω α 2 ,0)=1, then

deg ( I T ¯ , Ω ( Ω β 1 Ω α 2 ¯ ) , 0 ) =1,

and there are solutions in Ω β 1 Ω α 2 ¯ , Ω β 1 , Ω α 2 , as required.

We now show deg(I T ¯ , Ω α 2 ,0)=1. The proof that deg(I T ¯ , Ω β 1 ,0)=1 is similar and hence omitted. We define IW, the extension to Ω ¯ of the restriction of I T ¯ to Ω ¯ α 2 as follows.

Let

w(t,x,y)=f ( t , m α 2 , β 2 ( t , x ) , n ( t , y ) ) + m α 2 , β 2 ( t , x ) x 1 + ( x m α 2 , β 2 ( t , x ) ) 2 .

Thus, w is a continuous function on [0,1]× R 2 and satisfies

| w ( t , x , y ) | A | y | Ψ ( | y | ) + B for  | y | < D , | w ( t , x , y ) | M 1 for  ( t , x , y ) [ 0 , 1 ] × R 2 ,

for some constants M 1 . Moreover, we may choose M 1 so that α 2 , β 1 < M 1 .

Consider the following problem:

x ( t ) + w ( t , x ( t ) , d d t m α 2 , β 2 ( t , x ) ) = 0 , t J , Δ x ( t k ) = I ˜ k ( x ( t k ) ) , k = 1 , 2 , , p , Δ x ( t k ) = J ˜ k ( x ( t k ) , x ( t k ) ) , k = 1 , 2 , , p , x ( 0 ) = x ( 1 ) = 0 1 g ( s ) x ( s ) d s ,
(3.3)

where I ˜ k (x( t k ))= I k ( m α 2 , β 2 ( t k ,x( t k ))), J ˜ k (x( t k ), x ( t k ))= J k ( m α 2 , β 2 ( t k ,x( t k )),n( t k , x ( t k ))). Now define the following operator:

( W x ) ( t ) = 0 1 H ( t , s ) w ( s , x ( s ) , d d s m α 2 , β 2 ( s , x ) ) d s + 0 < t k < t [ I ˜ k ( x ( t k ) ) + ( t t k ) J ˜ k ( x ( t k ) , x ( t k ) ) ] t k = 1 p [ I ˜ k ( x ( t k ) ) + ( 1 t k ) J ˜ k ( x ( t k ) , x ( t k ) ) ] + 1 1 μ 0 1 0 < t k < s [ I ˜ k ( x ( t k ) ) + ( s t k ) J ˜ k ( x ( t k ) , x ( t k ) ) ] d s ν 1 μ k = 1 p [ I ˜ k ( x ( t k ) ) + ( 1 t k ) J ˜ k ( x ( t k ) , x ( t k ) ) ] .

Again, it is easy to check that x is a solution of (3.3) if x Ω α 2 and Wx=x (note that W:EE is compact). Thus, deg(I T ¯ , Ω α 2 ,0)=0. Moreover, it is easy to see that W( Ω ¯ )Ω. By assumptions (iv) and (v), there are no solutions in Ω β 1 Ω α 2 . So,

deg(I T ¯ , Ω α 2 ,0)=deg(IW,Ω Ω ¯ α 2 ,0)+deg(IW, Ω α 2 ,0)=deg(IW,Ω,0)=1.

Thus there are three solutions, as required. The proof is complete. □

Theorem 3.3 Suppose that conditions (H1)-(H3) hold. Assume that there exist two lower solutions α 1 and α 2 and two upper solutions β 1 and β 2 for problem (1.1), satisfying

  1. (i)

    α 1 α 2 β 2 ;

  2. (ii)

    α 1 β 1 β 2 ;

  3. (iii)

    α 2 β 1 ;

  4. (iv)

    there exists 0< ϵ ¯ < min t [ 0 , 1 ] { α 2 (t) α 1 (t), β 2 (t) β 1 (t)} such that all ϵ(0, ϵ ¯ ], the functions α 2 ϵ and β 1 +ϵ are, respectively, lower and upper solutions of (1.1);

  5. (v)

    α 2 ϵ ¯ β 1 + ϵ ¯ .

If f satisfies the Nagumo condition with respect to α 1 , β 2 , then problem (1.1) has at least three solutions x 1 , x 2 , and x 3 satisfying

α 1 x 1 β 1 , α 2 x 2 β 2 , x 3 β 1 , x 3 α 2 .

Proof In the proof of Theorem 3.3, define

Ω α 2 = { u Ω : u > α 2 ϵ ¯  on  ( 0 , 1 ) } , Ω β 1 = { u Ω : u < β 1 + ϵ ¯  on  ( 0 , 1 ) } ,

where Ω is defined in Theorem 3.2. □