Stochastic Maximum Principle for Optimal Advertising Models with Delay and Non-Convex Control Spaces ()
1. Introduction
We consider a stochastic model for problems of optimal advertising under uncertainty, so we have to study a stochastic version of advertising models. We start from the stochastic model introduced in [1] (see also [2] and [3] ): we consider carryover effects of the advertising, which in the model reads as delay in the control, and with memory effects of the level of goodwill, which in the model reads as delay in the state. We refer also to [4] for optimal advertising problems with memory effects in the level of goodwill.
In our model the delay in the effect of advertising affects the martingale term of the state equation, namely we consider, following [1] , the controlled stochastic differential equation in
with pointwise delay in the state and in the control
(1.1)
Following the usual notations, see e.g. [1] , x is the level of goodwill,
and
are factors related to the image deterioration in case of no advertising,
is a constant representing an advertising effectiveness factor. The diffusion term
represents for the word of mouth communication, with
the so-called “advertising volatility”; while the second diffusion term keeps track of the delayed effect of the advertising effort u, and the constant
is the effectiveness volatility of the communication. Besides,
is the initial goodwill level, while
is the history of the goodwill level before the advertising campaign is started.
The functional to maximize, over all controls in
, is the following profit, defined on finite horizon:
(1.2)
where c is the cost of advertising, l is the current reward, and r is the reward from the final goodwill. Our purpose is to derive a maximum principle for such a problem with non convex control space U, extending the results already present in the literature for the convex case, see e.g. [5] , where the stochastic maximum principle for control problems with pointwise delay in the state and in the control is studied, see also [6] and [7] where more general models are treated but the convexity of the control space U is still required. We underline the fact that stochastic control problems with diffusion depending on the control are difficult to treat; concerning problems related to advertising we mention the recent paper [8] , where the author solves the problem by means of the dynamic programming approach, differently from here the case of pointwise delay cannot be treated, moreover it is taken into account a diffusion depending linearly on the control, but not on the delayed control. We stress that one novelty of this paper is to handle the new difficulty coming form the non convexity of U: also in the non delayed case the non convexity of U makes the approach based on stochastic maximum principle more complicated. First we need to utilize the spike variation of the control and introduce the second variation to handle the control acting on the martingale term. We will therefore follow [9] to introduce this additional step. Thanks to the specific characteristics of the state equation, we can derive a quasi-explicit form for both the first adjoint, which results in an anticipating backward equation, see [10] , and the second adjoint, which is written using the optimal cost and a simple auxiliary process. Note that the specific case we address is not considered in [11] and does not completely fall within the hypotheses of [9] , as we also consider the delay the control appearing in the martingale term.
The structure of the paper is the following: in section 2 we present the notation, the assumptions and the control problem. Section 3 is devoted to collect the results on the first variation of the optimal state and also on the second variation, which we have to study since the space of controls is not convex. Section 4 concerns the first and second adjoint relations necessary to formulate the stochastic maximum principle, which is stated and proved section 4.3.
2. Stochastic Control Problem for Delay Equations Appearing in Advertising Models
2.1. Assumptions and Preliminaries
Let
be a 2 dimensional brownian motion defined on a complete probability space
and
the natural filtration generated by W, augmented in the usual way. We fix a finite time horizon
and we set, for every
and for every
, the following spaces:
(2.1)
(2.2)
(2.3)
2.2. Formulation of the Problem
We recall the state equation we are interested in:
(2.4)
We assume the following on the coefficients:
Hypothesis 2.1
(i)
;
(ii) the control strategy u, the advertisement in this case, belongs to the space
where U is a bounded subset of
possibly non convex, in particular U can be a bounded subset of
and this represents the realistic situation in which the advertisement is multiple of a given quantity;
(iii)
is the delay with which the past of the state affects the system;
We recall that the purpose is to maximize, over all controls in
, the following profit, on finite horizon:
(2.5)
Hypothesis 2.2 We assume the following:
(i) according to the literature, see e.g. [12] ,
is non linear, convex and locally lipschitz;
(ii)
represents the current reward, it is twice differentiable with at most linear growth;
(iii)
represents the foreseen reward from the final goodwill and it is twice differentiable and strictly increasing, with bounded second derivatives.
From now on, since we recall and apply the results in [9] where a mini-mization problem is considered, we focus on the problem (equivalent to the original one of maximizing the profit
given in (2.5) since
) of minimizing, over all admissible control in U, the cost functional
(2.6)
We are going to formulate necessary conditions for optimality.
3. First and Second Order Variations of the Optimal State Equation
Since U is not necessarily convex we use the spike variation method, see [13] , see also [9] for the case of delay equation.
Let
be an optimal couple: that is u is an optimal control and x its corresponding optimal trajectory, that is solution to Equation (2.4). The spike variation works as follows: for
we consider an interval
, with
, where m is the Lebesgue measure. Let
we set
(3.1)
We are going to derive a maximum principle for the control problem with state Equation (2.4) and cost functional (2.6) in the case of non convex space of controls. We will denote by
the solution of (2.4) corresponding to
and
the spike variation of the control, i.e.
.
We can write the equation for the first order variation of the state:
(3.2)
where
and the second variation is, for
,
(3.3)
It is well known that such equations are well posed in
(see e.g. [14] for a general theory on stochastic delay equations).
The following asymptotic behaviors hold, see ([9, Theorem 3.3]) for the delayed system and the classic result in ([15, Theorem 4.4]),
(3.4)
(3.5)
(3.6)
Moreover,
,
(3.7)
4. Maximum Principle for the Stochastic Delay Equation
In this Section, first we formulate the first order adjoint equation in 4.1, then we pass to the second order adjoint in 4.2, and finally we formulate the stochastic maximum principle in 4.3.
4.1. First Order Duality Relation
Following [9] we define the first order adjoint equation of (3.2), that is the anticipated backward equation
(4.1)
that admits a unique solution
, see e.g. [10] where we write
.
Since Equation (4.1) is linear, in view of the future times conditions we have an explicit, recursive, formulation for the solution
:
Proposition 4.1 Assume Hypotheses 2.1 and 2.2. Define for every
,
and
Then
solve:
(4.2)
where
and
are given by:
(4.3)
and
and
are given by:
(4.4)
Proof. Notice that, to avoid trivialities, we are taking
.
Let us consider the first order adjoint Equation (4.1), (, , for
: in view of the fact that
,
for a.e.
, for
it can be rewritten as a linear BSDE
(4.5)
and its solution for
, is given by
(4.6)
Formula (4.2) is just the rearrangement of the variation of constant formula, while (4.3) and (4.4) is an application of the Martingale Representation Theorem, see also [16] . □
Using (3.2), (3.2) and (4.1), we deduce the first duality relation:
Proposition 4.2 Assume 2.1 and 2.2 then:
(4.7)
(4.8)
Moreover the cost can be written as follows:
(4.9)
Proof. See ( [9] , Propositions 4.1 and 4.2). □
4.2. Second Order Duality Relation
Following ( [9] , Theorem 4.9]) we introduce the process P, called
in [9] , that in this simplified case can be explicitely defined by means of a auxiliary process
(here s indicates the initial time and 1 is the initial condition) that solves the following equation
(4.10)
By means of the solution process
, the process P can be now defined by
(4.11)
Also in this case, the solution
of Equation (4.10) can be explicitly defined by recursion:
Proposition 4.3 Assume Hypotheses 2.1 and 2.2. Define
, for every
Then
solves
(4.12)
Proof. Thanks to the initial condition in (25), we have that for every
:
(4.13)
Therefore
. Then the general case for
follows by the application of the constant variation formula on any interval
. □
The expansion of the cost then becomes:
Proposition 4.4 Assume 2.1 and 2.2. Then following expansion holds true:
(4.14)
Proof. The proof is in [9] [Proposition 4.9, Theorem 4.10, Theorem 4.11] and is a consequence of an approximation procedure based on the regularization of the dirac measure
and the well known reformulation of the delayed problem in infinite dimensions. □
4.3. The Stochastic Maximum Principle
We are now in the position to state our main result.
Theorem 4.5 Under assumptions 2.1 and 2.2, any optimal couple
satisfies the following variational inequality:
(4.15)
Where
, is the solution of the first order adjoint equation (16), and P is the second adjoint process given by (4.11).
Proof. Let
, and
, then
(4.16)
Letting
tends to 0, by standard arguments we deduce (4.15).
Acknowledgements
The authors are grateful to the GNAMPA (Gruppo Nazionale per l’Analisi Matematica, la Probabilità e le loro Applicazioni) for the financial support.