Stochastic Maximum Principle for Optimal Advertising Models with Delay and Non-Convex Control Spaces

Abstract

In this paper we study optimal advertising problems that model the introduction of a new product into the market in the presence of carryover effects of the advertisement and with memory effects in the level of goodwill. In particular, we let the dynamics of the product goodwill to depend on the past, and also on past advertising efforts. We treat the problem by means of the stochastic Pontryagin maximum principle, that here is considered for a class of problems where in the state equation either the state or the control depend on the past. Moreover the control acts on the martingale term and the space of controls U can be chosen to be non-convex but now the space of controls U can be chosen to be non-convex. The maximum principle is thus formulated using a first-order adjoint Backward Stochastic Differential Equations (BSDEs), which can be explicitly computed due to the specific characteristics of the model, and a second-order adjoint relation.

Share and Cite:

Guatteri, G. and Masiero, F. (2024) Stochastic Maximum Principle for Optimal Advertising Models with Delay and Non-Convex Control Spaces. Advances in Pure Mathematics, 14, 442-450. doi: 10.4236/apm.2024.146025.

1. Introduction

We consider a stochastic model for problems of optimal advertising under uncertainty, so we have to study a stochastic version of advertising models. We start from the stochastic model introduced in [1] (see also [2] and [3] ): we consider carryover effects of the advertising, which in the model reads as delay in the control, and with memory effects of the level of goodwill, which in the model reads as delay in the state. We refer also to [4] for optimal advertising problems with memory effects in the level of goodwill.

In our model the delay in the effect of advertising affects the martingale term of the state equation, namely we consider, following [1] , the controlled stochastic differential equation in with pointwise delay in the state and in the control

{ d x ( t ) = [ b 0 u ( t ) a 0 x ( t ) a d x ( t d ) ] d t + σ 1 x ( t ) d W t 1 + σ 2 u ( t d ) d W t 2 , x ( τ ) = x 0 ( τ ) , τ [ d , 0 ] . (1.1)

Following the usual notations, see e.g. [1] , x is the level of goodwill, a 0 and a d are factors related to the image deterioration in case of no advertising, b 0 is a constant representing an advertising effectiveness factor. The diffusion term σ 1 x ( t ) d W t 1 represents for the word of mouth communication, with σ 1 0 the so-called “advertising volatility”; while the second diffusion term keeps track of the delayed effect of the advertising effort u, and the constant σ 2 0 is the effectiveness volatility of the communication. Besides, x 0 ( 0 ) is the initial goodwill level, while x 0 ( τ ) , τ [ d , 0 ) is the history of the goodwill level before the advertising campaign is started.

The functional to maximize, over all controls in U , is the following profit, defined on finite horizon:

J ¯ ( t , x , u ) = E t T ( c ( u ( s ) ) + l ( x ( s ) ) ) d s + E r ( x ( T ) ) , (1.2)

where c is the cost of advertising, l is the current reward, and r is the reward from the final goodwill. Our purpose is to derive a maximum principle for such a problem with non convex control space U, extending the results already present in the literature for the convex case, see e.g. [5] , where the stochastic maximum principle for control problems with pointwise delay in the state and in the control is studied, see also [6] and [7] where more general models are treated but the convexity of the control space U is still required. We underline the fact that stochastic control problems with diffusion depending on the control are difficult to treat; concerning problems related to advertising we mention the recent paper [8] , where the author solves the problem by means of the dynamic programming approach, differently from here the case of pointwise delay cannot be treated, moreover it is taken into account a diffusion depending linearly on the control, but not on the delayed control. We stress that one novelty of this paper is to handle the new difficulty coming form the non convexity of U: also in the non delayed case the non convexity of U makes the approach based on stochastic maximum principle more complicated. First we need to utilize the spike variation of the control and introduce the second variation to handle the control acting on the martingale term. We will therefore follow [9] to introduce this additional step. Thanks to the specific characteristics of the state equation, we can derive a quasi-explicit form for both the first adjoint, which results in an anticipating backward equation, see [10] , and the second adjoint, which is written using the optimal cost and a simple auxiliary process. Note that the specific case we address is not considered in [11] and does not completely fall within the hypotheses of [9] , as we also consider the delay the control appearing in the martingale term.

The structure of the paper is the following: in section 2 we present the notation, the assumptions and the control problem. Section 3 is devoted to collect the results on the first variation of the optimal state and also on the second variation, which we have to study since the space of controls is not convex. Section 4 concerns the first and second adjoint relations necessary to formulate the stochastic maximum principle, which is stated and proved section 4.3.

2. Stochastic Control Problem for Delay Equations Appearing in Advertising Models

2.1. Assumptions and Preliminaries

Let W ( t ) be a 2 dimensional brownian motion defined on a complete probability space ( Ω , F , ) and ( F t ) t 0 the natural filtration generated by W, augmented in the usual way. We fix a finite time horizon T > 0 and we set, for every q 1 and for every 0 u v T , the following spaces:

L F 0 ( Ω × [ u , v ] ; k ) = : { X : Ω × [ u , v ] k , ( F t ) -progressive measurable } (2.1)

L F q ( Ω × [ u , v ] ; k ) = : { X : Ω × [ u , v ] k , ( F t ) -progr .meas ., ( E u v | X ( t ) | q d t ) 1 / q < } (2.2)

L F q ( Ω ; C ( [ u , v ] ; k ) ) = : { X : Ω × [ u , v ] k , ( F t ) -progr .meas ., ( E sup t [ u , v ] | X ( t ) | q ) 1 / q < } (2.3)

2.2. Formulation of the Problem

We recall the state equation we are interested in:

{ d x ( t ) = [ b 0 u ( t ) a 0 x ( t ) a d x ( t d ) ] d t + σ 1 x ( t ) d W t 1 + σ 2 u ( t d ) d W t 2 , x ( τ ) = x 0 ( τ ) , τ [ d , 0 ) , x ( 0 ) = x 0 ( 0 ) . (2.4)

We assume the following on the coefficients:

Hypothesis 2.1

(i) a 0 , b 0 , σ 1 , σ 2 ;

(ii) the control strategy u, the advertisement in this case, belongs to the space

U : = { z L F 0 ( Ω × [ 0, T ] , ) : z ( t ) U , a . s . }

where U is a bounded subset of possibly non convex, in particular U can be a bounded subset of and this represents the realistic situation in which the advertisement is multiple of a given quantity;

(iii) d > 0 is the delay with which the past of the state affects the system;

We recall that the purpose is to maximize, over all controls in U , the following profit, on finite horizon:

J ¯ ( t , x , u ) = E t T ( c ( u ( s ) ) + l ( x ( s ) ) ) d s + E r ( x ( T ) ) , (2.5)

Hypothesis 2.2 We assume the following:

(i) according to the literature, see e.g. [12] , c : U is non linear, convex and locally lipschitz;

(ii) l : [ 0, T ] × represents the current reward, it is twice differentiable with at most linear growth;

(iii) r : represents the foreseen reward from the final goodwill and it is twice differentiable and strictly increasing, with bounded second derivatives.

From now on, since we recall and apply the results in [9] where a mini-mization problem is considered, we focus on the problem (equivalent to the original one of maximizing the profit J ¯ given in (2.5) since J = J ¯ ) of minimizing, over all admissible control in U, the cost functional

J ( t , x , u ) = E t T ( c ( u ( s ) ) l ( x ( s ) ) ) d s E r ( x ( T ) ) , (2.6)

We are going to formulate necessary conditions for optimality.

3. First and Second Order Variations of the Optimal State Equation

Since U is not necessarily convex we use the spike variation method, see [13] , see also [9] for the case of delay equation.

Let ( u , x ) be an optimal couple: that is u is an optimal control and x its corresponding optimal trajectory, that is solution to Equation (2.4). The spike variation works as follows: for ε > 0 we consider an interval V ε [ 0, T ] , with m ( V ε ) = ε , where m is the Lebesgue measure. Let v U we set

u ε ( t ) = { u ( t ) t [ 0, T ] \ V ε v t V ε (3.1)

We are going to derive a maximum principle for the control problem with state Equation (2.4) and cost functional (2.6) in the case of non convex space of controls. We will denote by x ε the solution of (2.4) corresponding to u ε and δ u the spike variation of the control, i.e. δ u ( t ) : = ( u ( t ) v ) I V ε ( t ) .

We can write the equation for the first order variation of the state:

{ d y ε ( t ) = [ a 0 y ε ( t ) a d y ε ( t d ) ] d t + σ 1 y ε ( t ) d W t 1 + σ 2 δ u ( t d ) d W t 2 , t [ 0 , T ] , y ε ( 0 ) = 0 , y ε ( τ ) = 0 , d τ < 0.

(3.2)

where δ u ε ( t d ) = ( u ( t d ) v ) I V ε ( t d ) and the second variation is, for t [ 0, T ] ,

{ d z ε ( t ) = [ a 0 z ε ( t ) a d z ε ( t d ) ] d t + b 0 δ u ( t ) d t + σ 1 z ε ( t ) d W t 1 z ε ( 0 ) = 0 , z ε ( τ ) = 0 , d τ < 0. (3.3)

It is well known that such equations are well posed in L F 2 ( Ω ; C ( [ 0, T ] ; ) ) (see e.g. [14] for a general theory on stochastic delay equations).

The following asymptotic behaviors hold, see ([9, Theorem 3.3]) for the delayed system and the classic result in ([15, Theorem 4.4]),

E sup t [ 0, T ] | y ε ( t ) | 2 = O ( ε ) , E sup t [ 0, T ] | z ε ( t ) | 2 = O ( ε ) , (3.4)

E sup t [ 0, T ] | x ε ( t ) x ( t ) y ε ( t ) | 2 = O ( ε 2 ) , (3.5)

E sup t [ 0, T ] | x ε ( t ) x ( t ) y ε ( t ) z ε ( t ) | 2 = o ( ε 2 ) , (3.6)

Moreover, p 1 ,

E sup t [ d , T ] | y ε ( t ) | p < + , E sup t [ d , T ] | z ε ( t ) | p < + , (3.7)

4. Maximum Principle for the Stochastic Delay Equation

In this Section, first we formulate the first order adjoint equation in 4.1, then we pass to the second order adjoint in 4.2, and finally we formulate the stochastic maximum principle in 4.3.

4.1. First Order Duality Relation

Following [9] we define the first order adjoint equation of (3.2), that is the anticipated backward equation

{ p ( t ) = r x ( x ( T ) ) + t T l x ( x ( s ) ) d s a 0 t T p ( s ) d s a d t T E F s p ( s + d ) d s + σ 1 t T E F s q 1 ( s + d ) d s + t T q 1 ( s ) d W s 1 + t T q 2 ( s ) d W s 2 p ( T τ ) = 0 , for all τ [ d , 0 ) , q 1 ( T τ ) = q 2 ( T τ ) = 0 a .e . τ [ d , 0 ) . (4.1)

that admits a unique solution ( p , q ) L F 2 ( Ω ; C ( [ 0, T ] ; ) ) × L F 2 ( Ω × [ 0, T ] ; 2 ) , see e.g. [10] where we write q = ( q 1 , q 2 ) .

Since Equation (4.1) is linear, in view of the future times conditions we have an explicit, recursive, formulation for the solution ( p , q ) :

Proposition 4.1 Assume Hypotheses 2.1 and 2.2. Define for every

k = 0 , , [ T d ] ,

p k ( t ) : = p | [ ( T ( k + 1 ) d ) 0 , ( T k d ) ] ( t )

and

q k , 1 ( t ) : = q | [ ( T ( k + 1 ) d ) 0 , ( T k d ) ] 1 ( t ) , q k , 2 ( t ) : = q | [ ( T ( k + 1 ) d ) 0 , ( T k d ) ] 2 ( t ) .

Then ( p k , q k ) solve:

{ p 0 ( t ) = e a 0 ( T t ) E F t ( r x ( x ( T ) ) ) + E F t t T e a 0 ( s t ) l x ( x ( s ) ) d s , t [ T d , T ] , q 0 , 1 ( t ) = e a 0 ( T t ) L 1 ( t ) t T e a 0 ( s t ) K 1 ( s , t ) d s , for a .e . t [ T d , T ] , q 0 , 2 ( t ) = e a 0 ( T t ) L 2 ( t ) t T e a 0 ( s t ) K 2 ( s , t ) d s , for a .e . t [ T d , T ] . p k ( t ) = e a 0 ( T k d t ) p k 1 ( T k d ) a d t T k d e a 0 ( s t ) E F s p k 1 ( s + d ) d s + σ 1 t T e a 0 ( s t ) E F s q k 1 , 2 ( s + d ) d s + t T e a 0 ( s t ) q k , 1 ( s ) d W s 1 + t T e a 0 ( s t ) q k , 2 ( s ) d W s 2 , t [ ( T ( k + 1 ) d ) 0 , ( T k d ) ] , k : 1 , , [ T d ] (4.2)

where L 1 and L 2 are given by:

E F t ( r x ( X ( T ) ) ) = E ( r x ( X ( T ) ) ) + t T L 1 ( s ) d W 1 ( s ) + t T L 2 ( s ) d W 1 ( s ) . (4.3)

and K 1 and K 2 are given by:

E F t ( l x ( s ) ) = E ( l x ( s ) ) + T d t K 1 ( s , τ ) d W 1 ( τ ) + T d t K 2 ( s , τ ) d W 2 ( τ ) . (4.4)

Proof. Notice that, to avoid trivialities, we are taking d < T .

Let us consider the first order adjoint Equation (4.1), (, , for t ( T d , T ] : in view of the fact that p ( t ) = q i ( t ) = 0 , i = 1 , 2 for a.e. t [ T , T + d ] , for t [ T d , T ] it can be rewritten as a linear BSDE

{ p ( t ) = r x ( x ( T ) ) + t T l x ( x ( s ) ) d s a 0 t T p ( s ) d s + t T q 1 ( s ) d W s 1 + t T q 2 ( s ) d W s 2 p ( T τ ) = 0 , for all τ [ d , 0 ) , q 1 ( T τ ) = q 2 ( T τ ) = 0 a .e . τ [ d , 0 ) ,

(4.5)

and its solution for t [ T d , T ] , is given by

p 0 ( t ) = e a 0 ( T t ) r x ( x ( T ) ) + t T e a 0 ( s t ) l x ( x ( s ) ) d s + t T e a 0 ( s t ) q 0 , 1 ( s ) d W s 1 + t T e a 0 ( s t ) q 0 , 2 ( s ) d W s 2 (4.6)

Formula (4.2) is just the rearrangement of the variation of constant formula, while (4.3) and (4.4) is an application of the Martingale Representation Theorem, see also [16] . □

Using (3.2), (3.2) and (4.1), we deduce the first duality relation:

Proposition 4.2 Assume 2.1 and 2.2 then:

E p ( T ) y ε ( T ) = E r x ( x ( T ) ) y ε ( T ) = E 0 T [ l x ( x ( s ) ) y ε ( s ) + q 2 ( s ) δ u ( s d ) ] d s (4.7)

E p ( T ) z ε ( T ) = E r x ( x ( T ) ) z ε ( T ) = E 0 T l x ( x ( s ) ) z ε ( s ) d s + E 0 T p ( s ) δ u ( s ) d s (4.8)

Moreover the cost can be written as follows:

J ( u ) J ( u ε ) = E 0 T ( δ l ( t ) + b 0 p ( t ) δ u ( t ) σ 2 q 2 ( t ) δ u ( t d ) ) d t + 1 2 E r x x ( x ( T ) ) ( y ε ( T ) ) 2 + 1 2 E 0 T l x x ( t , x ( t ) ) ( y ε ( t ) ) 2 d t (4.9)

Proof. See ( [9] , Propositions 4.1 and 4.2). □

4.2. Second Order Duality Relation

Following ( [9] , Theorem 4.9]) we introduce the process P, called P 0,0 in [9] , that in this simplified case can be explicitely defined by means of a auxiliary process y ˜ s , 1 (here s indicates the initial time and 1 is the initial condition) that solves the following equation

{ d y ˜ s , 1 ( t ) = ( a 0 y ˜ s , 1 ( t ) a d y ˜ s , 1 ( t d ) ) d t + σ 1 y ˜ s , 1 ( t ) d W ( t ) , t [ s , T ] , y ˜ s , 1 ( s ) = 1 , y ( τ ) = 0 , d τ < s . (4.10)

By means of the solution process y ˜ s , 1 L F 2 ( Ω ; C ( [ 0, T ] ; ) ) , the process P can be now defined by

P ( s ) = E F s r x x ( x ( T ) ) y ˜ s ,1 ( T ) 2 + E F s s T l x x ( t ) y ˜ s ,1 ( t ) 2 d t . (4.11)

Also in this case, the solution y ˜ 1, s of Equation (4.10) can be explicitly defined by recursion:

Proposition 4.3 Assume Hypotheses 2.1 and 2.2. Define

y ˜ k ( t ) : = y ˜ | [ ( s + k d ) , ( s + ( k + 1 ) d ) T ] s ,1 ( t ) , for every k : 0, , [ T d ]

Then y ˜ k solves

{ y ˜ 0 ( t ) = e a 0 σ 1 2 2 ( t s ) + σ 1 ( W 1 ( t ) W 1 ( s ) ) , t [ s , ( s + d ) T ] y ˜ k ( t ) = e a 0 ( t ( s + k d ) ) y ˜ k 1 ( s + k d ) a d s + k d t e a 0 ( t r ) y ˜ k 1 ( r d ) d s + σ 1 s + k d t e a 0 ( t r ) y ˜ k ( s ) d W s 1 , t [ ( s + k d ) , ( s + ( k + 1 ) d ) T ] , k = 1 , , [ T d ] (4.12)

Proof. Thanks to the initial condition in (25), we have that for every t [ s , s + d T ] :

{ d y ˜ s , 1 ( t ) = a 0 y ˜ s , 1 ( t ) d t + σ 1 y ˜ s , 1 ( t ) d W t 1 y ˜ s , 1 ( s ) = 1 , (4.13)

Therefore y ˜ 0 ( t ) = y ˜ | [ s , ( s + d ) T ] s , 1 ( t ) = e a 0 σ 1 2 2 ( t s ) + σ 1 ( W 1 ( t ) W 1 ( s ) ) . Then the general case for t > s + d follows by the application of the constant variation formula on any interval [ ( s + k d ) , ( s + ( k + 1 ) d ) T ] . □

The expansion of the cost then becomes:

Proposition 4.4 Assume 2.1 and 2.2. Then following expansion holds true:

J ( u ) J ( u ε ) = E 0 T ( δ c ( s ) + b 0 p ( s ) δ u ( s ) σ 2 q 2 ( s ) δ u ( s d ) ) ( t ) d s + 1 2 E 0 T σ 2 2 ( δ u ( s d ) ) 2 P ( s ) d s + o ( ε ) (4.14)

Proof. The proof is in [9] [Proposition 4.9, Theorem 4.10, Theorem 4.11] and is a consequence of an approximation procedure based on the regularization of the dirac measure δ d and the well known reformulation of the delayed problem in infinite dimensions. □

4.3. The Stochastic Maximum Principle

We are now in the position to state our main result.

Theorem 4.5 Under assumptions 2.1 and 2.2, any optimal couple ( x , u ) satisfies the following variational inequality:

b 0 ( u ( t ) v ) p ( t ) σ 2 ( u ( t ) v ) E F t q 2 ( t + d ) ( c ( u ( t ) ) c ( v ) ) + 1 2 σ 2 ( u ( t ) v ) 2 E F t P ( t + d ) 0 , v U , a . e . a . s . (4.15)

Where ( p , q ) L F 2 ( Ω ; C ( [ 0, T ] ; ) ) × L F 2 ( Ω × [ 0, T ] ; 2 ) , is the solution of the first order adjoint equation (16), and P is the second adjoint process given by (4.11).

Proof. Let t [ 0, T ] , and V ε = [ t , t + ε ] , then

J ( u ) J ( u ε ) = E 0 T ( δ c ( s ) + b 0 p ( s ) δ u ( s ) σ 2 q 2 ( s ) δ u ( s d ) ) ( t ) d s + 1 2 E 0 T σ 2 2 ( δ u ( s d ) ) 2 P ( s ) d s = E 0 T ( [ ( c ( u ( s ) ) v ) + b 0 p ( s ) ( u ( s ) v ) ] I V ε ( s ) σ 2 q 2 ( s ) ( u ( s d ) v ) I V ε ( s d ) ) d s + 1 2 E 0 T σ 2 2 ( u ( s d ) v ) 2 P ( s ) I V ε ( s d ) d s

= E t t + ε ( c ( u ( s ) v ) I V ε ( s ) b 0 p ( s ) ( u ( s ) v ) ) d s E t t + ε E F t σ 2 q 2 ( s + d ) ( u ( s ) v ) d s + 1 2 E t t + ε E F t σ 2 2 ( u ( s ) v ) 2 P ( s + d ) d s . (4.16)

Letting ε tends to 0, by standard arguments we deduce (4.15).

Acknowledgements

The authors are grateful to the GNAMPA (Gruppo Nazionale per l’Analisi Matematica, la Probabilità e le loro Applicazioni) for the financial support.

Conflicts of Interest

The authors declare no conflicts of interest regarding the publication of this paper.

References

[1] Grosset, L. and Viscolani, B. (2004) Advertising for a New Product Introduction: A Stochastic Approach. Top, 12, 149-167.
https://doi.org/10.1007/bf02578929
[2] Gozzi, F. and Marinelli, C. (2006) Stochastic Optimal Control of Delay Equations Arising in Advertising Models. Stochastic Partial Differential Equations and Applications, VII, Chapman & Hall/CRC, Boca Raton, 133-148.
[3] Gozzi, F., Marinelli, C. and Savin, S. (2009) On Controlled Linear Diffusions with Delay in a Model of Optimal Advertising under Uncertainty with Memory Effects. Journal of Optimization Theory and Applications, 142, 291-321.
https://doi.org/10.1007/s10957-009-9524-5
[4] Hartl, R.F. (1984) Optimal Dynamic Advertising Policies for Hereditary Processes. Journal of Optimization Theory and Applications, 43, 51-72.
https://doi.org/10.1007/bf00934746
[5] Chen, L. and Wu, Z. (2010) Maximum Principle for the Stochastic Optimal Control Problem with Delay and Application. Automatica, 46, 1074-1080.
https://doi.org/10.1016/j.automatica.2010.03.005
[6] Guatteri, G. and Masiero, F. (2021) Stochastic Maximum Principle for Problems with Delay with Dependence on the Past through General Measures. Mathematical Control & Related Fields, 11, 829-855.
https://doi.org/10.3934/mcrf.2020048
[7] Hu, Y. and Peng, S. (1996) Maximum Principle for Optimal Control of Stochastic System of Functional Type. Stochastic Analysis and Applications, 14, 283-301.
https://doi.org/10.1080/07362999608809440.
[8] de Feo, F. (2023) Stochastic Optimal Control Problems with Delays in the State and in the Control via Viscosity Solutions and an Economical Application. arXiv: 2308.14506.
[9] Guatteri, G. and Masiero, F. (2023) Stochastic Maximum Principle for Equations with Delay: Going to Infinite Dimensions to Solve the Non-Convex Case. arXiv: 2306.07422.
[10] Peng, S. and Yang, Z. (2009) Anticipated Backward Stochastic Differential Equations. The Annals of Probability, 37, 877-902.
https://doi.org/10.1214/08-aop423
[11] Meng, W., Shi, J., Wang, T. and Zhang, J.F. (2023) A General Maximum Principle for Optimal Control of Stochastic Differential Delay Systems. arXiv: 2302.03339.
[12] Tapiero, C.S. (1978) Optimum Advertising and Goodwill under Uncertainty. Operations Research, 26, 450-463.
https://doi.org/10.1287/opre.26.3.450
[13] Peng, S. (1990) A General Stochastic Maximum Principle for Optimal Control Problems. SIAM Journal on Control and Optimization, 28, 966-979.
https://doi.org/10.1137/0328054
[14] Mohammed, S.-E.A. (1998) Stochastic Differential Systems with Memory: Theory, Examples and Applications. In: Decreusefond, L., Øksendal, B., Gjerde, J. and Üstünel, A.S., Eds., Stochastic Analysis and Related Topics VI, Birkhäuser Boston, 1-77.
https://doi.org/10.1007/978-1-4612-2022-0_1
[15] Yong, J. and Zhou, X.Y. (1999) Stochastic Controls: Hamiltonian Systems and HJB Equations. Springer-Verlag, New York.
[16] Hu, Y. and Peng, S. (1991) Adapted Solution of a Backward Semilinear Stochastic Evolution Equation. Stochastic Analysis and Applications, 9, 445-459.
https://doi.org/10.1080/07362999108809250

Copyright © 2024 by authors and Scientific Research Publishing Inc.

Creative Commons License

This work and the related PDF file are licensed under a Creative Commons Attribution 4.0 International License.