Abstract
We consider a standard Brownian motion whose drift can be increased or decreased in a possibly singular manner. The objective is to minimize an expected functional involving the time-integral of a running cost and the proportional costs of adjusting the drift. The resulting two-dimensional degenerate singular stochastic control problem has interconnected dynamics and it is solved by combining techniques of viscosity theory and free boundary problems. We provide a detailed description of the problem’s value function and of the geometry of the state space, which is split into three regions by two monotone curves. Our main result shows that those curves are continuously differentiable with locally Lipschitz derivative and solve a system of nonlinear ordinary differential equations.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Consider a system whose position or level is subject to random fluctuations and can be corrected by acting on its drift. The latter can be increased or decreased, and the (cumulative) actions affecting the drift’s dynamics are not necessarily absolutely continuous with respect to the Lebesgue measure, as functions of time; also impulses or singularly continuous forces can be applied. The objective of the decision maker is to minimize a total expected functional consisting of the time-integral of a running cost and of the proportional costs of adjusting the drift.
We model this problem as a two-dimensional singular stochastic control problem (see, e.g., [25, 29, 30], and [45] as classical contributions to the theory of singular stochastic control). The system’s position/level X evolves as
for some positive constants \(\alpha ,\eta \) and for a given standard Brownian motion W, and the drift Y is such that
Here, \(\xi ^{+}_t\) (respectively, \(\xi ^{-}_t\)) are the cumulative increase (respectively, decrease) of the drift up to time \(t\ge 0\) and, as such, \(\xi ^{+}\) and \(\xi ^{-}\) are nondecreasing processes, and \(\xi :=\xi ^+ - \xi ^-\) has finite variation. The process X might be thought of as a random demand/level of sales whose instantaneous trend Y can be affected via production, according to supply and demand rules, or through an inventory management policy (see, e.g., the review [44]). Alternatively, X could be the position of a satellite which is subject to random disturbances and can be adjusted by properly acting on its velocity. The decision maker aims at picking a control rule \(\xi \) that minimizes an expected cost functional. This consists of a term measuring the total cost of acting on the system, which is proportional to the total variation of \(\xi \), and of a term involving a running convex cost function f of the current values \((X_t,Y_t)\). For example, if X is a satellite position and Y its velocity, the decision maker might want to keep the satellite as close as possible to a given target level, say 0, while minimizing the system’s kinetic energy; in such a case a possible choice of f might therefore be \(f(x,y)=x^2 + y^2\). The resulting optimization problem then reads as
and it thus takes the form of a two-dimensional degenerate singular stochastic control problem with interconnected dynamics and controls of bounded-variation.
Our problem might be seen as a generalization of the bounded-velocity control of a scalar Brownian motion introduced by Beneš in 1974 [3], which has stimulated a subsequent large literature allowing for different specifications of the performance criterion and incorporating also other features like discretionary stopping and partial observation (see [1, 4, 26,27,28, 34], among many others). However, while in the previous papers the decision maker tracks the position of the Brownian system by choosing the value of its drift within a bounded set (hence the term bounded-velocity control), in our problem the Brownian motion is only indirectly affected by the controller’s actions that, in fact, can unlimitedly increase and decrease the Brownian’s drift at proportional costs. As a result, in our case the optimal control rule is expected to be of singular type (see Sect. 6.1 below), rather than of so-called bang-bang type (cf. [3, 26,27,28], among others).
In [43] the optimal correction problem of a damped random oscillator is studied. Differently to us, in that paper the velocity is subject to random disturbances and it is linearly controlled via a process of bounded variation, while the oscillator’s position is not affected by noise. The authors formulate the problem as a cheap degenerate two-dimensional singular stochastic control problem (i.e. as a problem where the performance criterion does not include the total cost of actions), and a thorough study of the related dynamic programming equation is performed via analytic methods. In [13], it is provided a numerical analysis of the non-cheap linear version of the control problem of [43].
The two papers that are perhaps closest to ours are [32] and [21]. In [32] a singular stochastic control problem with monotone controls and with finite-fuel constraint is considered. The problem is motivated by the issue of irreversible installation of solar panels, where the price of solar electricity is mean-reverting, with drift affected by the cumulative amount of installed solar panels. The authors solve the problem via a guess-and-verify approach and characterize the free boundary as the unique solution to a first-order ordinary differential equation (ODE) complemented by a boundary condition directly implied by the finite-fuel constraint. In [21] a two-dimensional singular stochastic control problem with controls of bounded-variation and interconnected dynamics is studied. The problem’s characteristic is that the mean-reversion level of the diffusive component of the state process is an affine function of the purely controlled second component.
Clearly, there is also a large literature on two-dimensional degenerate bounded-variation stochastic control problems where the two components of the state-process are decoupled (see [2, 16, 20, 23, 24], and [35], among many others). In those works, it is usually shown that the value function solves the associated dynamic programming principle in the classical sense and the free boundaries are characterized in terms of algebraic/integral equations. These results are obtained through various methods ranging from a “guess-and-verify” approach (see, e.g., [2, 16], and [35]), viscosity theory ( [20]), the connection to optimal stopping ([31]), and the link to switching controls ([23] and [24]). However, in our problem, the two dynamics of the state process are interconnected (cf. (1.1) and (1.2)), and this makes the approaches developed in [31] and [23, 24] for the connection to optimal stopping games and optimal switching, respectively, not directly applicable. Moreover, given the complexity of the equations arising in our analysis (see Theorem 6.1 below), the verification of the actual optimality of a smooth solution to the dynamic programming equation is also particularly challenging, thus refraining us from the application of a “guess-and-verify” approach.
For the previous reasons, as in [21], we follow here a direct approach which, by employing techniques from viscosity theory and free-boundary problems, enables us to provide a detailed study of the value function and of the geometry of the problem’s state space. In particular, we show that: (i) the value function V is differentiable with first derivatives that are (locally) Lipschitz, and its y-derivative identifies with the value of an optimal stopping game (see also [31] in the case of decoupled dynamics, and [11] for a coupled non degenerate setting); (ii) the two-dimensional state space is split into three connected regions (continuation and action regions) by two monotone curves (free boundaries); (iii) the expression of the value function in each of those regions is provided; (iii) the second order derivative \(V_{yx}\) is continuous in the whole space (second-order smooth-fit); (iv) the free boundaries solve a system of integral equations.
Furthermore, because the uncontrolled process is a Brownian motion (rather than a more complex Ornstein-Uhlenbeck process as in [21]), in this paper we are able to push the analysis of [21] much further by providing new results (see Sects. 5 and 6 below). In particular, we can show that the free boundaries delineating action and inaction regions are continuously differentiable with locally Lipschitz derivative (see Theorem 6.4). To our knowledge, in the context of a fully degenerate two-dimensional singular stochastic control problem with interconnected dynamics, a similar finding appears here for the first time. The proof of the aforementioned regularity of the free boundaries hinges on a series of intermediate novel results. First of all, we show that the limit of the third derivative \(V_{yxx}\) at the free boundaries along any sequence of points belonging to the (interior of the) continuation region exists and is nonzero (cf. Proposition 5.7). This allows us to apply (a suitable version of) the implicit function theorem and to show that the free boundaries are locally Lipschitz functions of the y coordinate (Proposition 5.8). Then, by exploiting such a property and differentiating the integral equations solved by the free boundaries, we can prove that the latter satisfy a system of (explicitly computable) first-order ODEs. The regularity of the forcing term appearing in the ODEs finally implies that the free boundaries are actually continuously differentiable with locally Lipschitz derivative (see Theorem 6.4). Along with that, we determine explicit expressions for the coefficients A and B appearing in the expression of the value function (cf. Theorem 4.5 and Corollary 6.2). As it is discussed in Remark 6.3, this has not been possible in [21].
Unfortunately, providing boundary conditions complementing the system of ODEs for the free boundaries still remains an open problem. Indeed, it seems hard to identify a relevant value of y for which the values of the free boundaries can be determined, as well as some kind of asymptotic growth that might restrict the functional class where to look for uniqueness of the ODEs’ system. However, in Sect. 6.1 we propose a conjecture about the derivation of a Cauchy problem involving the first derivatives of the free boundaries with respect to the parameter \(\alpha \) (cf. (1.1)), rather than y. A discussion on the structure of the optimal control is also presented in Sect. 6.1.
The rest of the paper is organized as follows. Problem formulation and preliminary results are provided in Sect. 2, while preliminary properties of the free boundaries in Sect. 3. Section 4 contains the structure of the value function and the second-order smooth-fit property, and most of the results of this section follow from them in [21]. Further important properties of the free boundaries - as their (locally) Lipschitz continuity - are proved in Sect. 5, while the system of ODEs for the free boundaries is finally obtained in Sect. 6.
1.1 Notation
In the rest of this paper, we adopt the following notation and functional spaces. We will use \(|\,\cdot \,|\) for the Euclidean norm on any finite-dimensional space, without indicating the dimension each time for simplicity of exposition.
Given a smooth function \(h:\mathbb {R}\rightarrow \mathbb {R}\), we shall write \(h^{\prime }\), \(h^{\prime \prime }\), etc. to denote its derivatives. If the function h admits k continuous derivatives, \(k\ge 1\), we shall write \(h \in C^{k}(\mathbb {R};\mathbb {R})\), while \(h\in C(\mathbb {R};\mathbb {R})\) if such a function is only continuous.
For a smooth function \(h:\mathbb {R}^2\rightarrow \mathbb {R}\), we denote by \(h_x\), \(h_y\), \(h_{xx}\), \(h_{yy}\), etc. its partial derivatives. Given \(k,j\in \mathbb {N}\), we let \(C^{k,j}(\mathbb {R}^2;\mathbb {R})\) be the class of functions \(h:\mathbb {R}^2 \rightarrow \mathbb {R}\) which are k-times continuously differentiable with respect to the first variable and j-times continuously differentiable with respect to the second variable. If \(k=j\), we shall simply write \(C^{k}(\mathbb {R}^2;\mathbb {R})\). Moreover, for a domain \(\mathcal {O} \subseteq \mathbb {R}^d\), \(d\in \{1,2\}\), we shall work with the space \(C^{k,\text {Lip}}_{\text {loc}}(\mathcal {O};\mathbb {R})\), \(k\ge 1\), which consists of all the functions \(h:\mathcal {O}\rightarrow \mathbb {R}\) that are k times continuously differentiable, with locally-Lipschitz kth-derivative(s).
Also, for \(p \ge 1\) we shall denote by \(L^{p}(\mathcal {O};\mathbb {R})\) (resp. \(L^{p}_{\text {loc}}(\mathcal {O};\mathbb {R}))\) the space of real-valued functions \(h:\mathcal {O}\rightarrow \mathbb {R}\) such that \(|h|^p\) is integrable with respect to the Lebesgue measure on \(\mathcal {O}\) (resp. locally integrable on \(\mathcal {O}\)). Finally, for \(k\ge 1\), we shall make use of the space \(W^{k,p}(\mathcal {O};\mathbb {R})\) (resp. \(W^{k,p}_{\text {loc}}(\mathcal {O};\mathbb {R})\)), which is the space of all the functions \(h:\mathcal {O}\rightarrow \mathbb {R}\) that admit kth-order weak derivative(s) in \(L^{p}(\mathcal {O};\mathbb {R})\) (resp. \(L^{p}_{\text {loc}}(\mathcal {O};\mathbb {R}))\)).
2 Problem Formulation and Preliminary Results
Let \((\Omega , \mathcal {F},\mathbb {F}:=(\mathcal {F}_t)_{t\ge 0}, \mathsf {P})\) be a complete filtered probability space rich enough to accommodate an \(\mathbb {F}\)-Brownian motion \(W:=(W_t)_{t\ge 0}\). We assume that the filtration \(\mathbb {F}\) satisfies the usual conditions.
We introduce the (nonempty) set
and for any \(\xi \in \mathcal {A}\) we denote by \(\xi ^+\) and \(\xi ^-\) the two nondecreasing \(\mathbb {F}\)-adapted càdlàg processes providing the minimal decomposition of \(\xi \); that is, such that \(\xi =\xi ^+ - \xi ^-\) and the (random) Borel-measures induced on \([0,\infty )\) by \(\xi ^+\) and \(\xi ^-\) have disjoint supports. In the following, for any \(\xi \in \mathcal {A}\), we set \(\xi ^{\pm }_{0^-}=0\) a.s. and we denote by \(|\xi |_t:=\xi ^+_t + \xi ^-_t\), \(t\ge 0\), its total variation.
For \(\xi \in \mathcal {A}\), \((x,y)\in \mathbb {R}^2\), and \(\alpha >0\), we then consider the purely controlled dynamics
as well as the diffusive
The unique strong solution to (2.3) is given by
The parameter \(\alpha \) measures the strength of the interaction between the processes X and Y. Clearly, for \(\alpha =0\) the two dynamics are decoupled and X is a Brownian motion with volatility \(\eta >0\).
Remark 2.1
It is worth noticing that the restriction \(\alpha >0\) is not necessary for the subsequent analysis; in fact, all the results of this paper (up to obvious modifications) can be still deduced with the same techniques also in the case \(\alpha <0\). We have decided to consider only the case \(\alpha >0\) just in order to simplify the exposition.
Controlling the dynamics (X, Y) gives rise to an instantaneous cost that is proportional – with marginal constant cost \(K>0\) – to the total variation of the exerted control. Moreover, the controller faces also a running cost depending on the current levels \((X_t,Y_t)\). The aim is therefore to choose a control \(\xi \in \mathcal {A}\) such that, for any \((x,y) \in \mathbb {R}^2\), and for a given \(\rho >0\), the cost functional
is minimized; that is, to solve
In (2.5) and in the following, the integrals with respect to \(\mathrm {d}|\xi |\) and \(\mathrm {d}\xi ^{\pm }\) are intended in the Lebesgue-Stieltjes’ sense; in particular, for \(\zeta \in \{|\xi |,\xi ^+,\xi ^-\}\), we set \(\int _0^s (\,\cdot \,) \mathrm {d}\zeta _t := \int _{[0,s]} (\,\cdot \,) \mathrm {d}\zeta _t\) in order to take into account a possible mass at time zero of the Borel (random) measure \(\mathrm {d}\zeta \). The function \(f:\mathbb {R}^2 \rightarrow \mathbb {R}^+\) satisfies the following standing assumption.
Assumption 2.2
There exists constants \(p > 1\), and \(C_0,C_1,C_2>0\) such that the following hold true:
-
(i)
\(0 \le f(z) \le C_0\big (1 + |z|\big )^p\), for every \(z=(x,y)\in \mathbb {R}^2\);
-
(ii)
for every \(z=(x,y),z'=(x',y')\in \mathbb {R}^2\),
$$\begin{aligned} |f(z) - f(z')| \le C_1 \big (1 + f(z)+ f(z')\big )^{1-\frac{1}{p}} |z-z'|; \end{aligned}$$ -
(iii)
for every \(z=(x,y),z'=(x',y')\in \mathbb {R}^2\) and \(\lambda \in (0,1)\),
$$\begin{aligned} 0\le & {} \lambda f(z)+(1-\lambda )f(z')-f(\lambda z + (1-\lambda ) z') \\\le & {} C_2 \lambda (1-\lambda )(1 + f(z) + f(z'))^{\left( 1-\frac{2}{p}\right) ^+}|z-z'|^2; \end{aligned}$$ -
(iv)
\(x \mapsto f_y(x,y)\) is nondecreasing for any \(y \in \mathbb {R}\).
Remark 2.3
-
(i)
By Assumption 2.2-(iii), f is convex and locally semiconcave; then, by [9, Cor. 3.3.8],
$$\begin{aligned} f\in C^{1,{Lip}}_{\text {loc}}(\mathbb {R}^2;\mathbb {R})=W^{2,\infty }_{\text {loc}}(\mathbb {R}^2;\mathbb {R}). \end{aligned}$$ -
(ii)
A function f satisfying Assumption 2.2 is, for example,
$$\begin{aligned} f(x,y)=|x-\hat{x}|^p+|y-\hat{y}|^p, \end{aligned}$$with \(p\ge 2\) ad for some \(\hat{x},\hat{y} \in \mathbb {R}\).
We now provide some preliminary properties of the value function, whose classical proof exploits the linear structure of the state equations.
Proposition 2.4
Let Assumption 2.2 hold and let \(p>1\) be the constant appearing in such assumption. There exist constants \(\widehat{C}_0, \widehat{C}_1,\widehat{C}_2>0\) such that the following hold:
-
(i)
\(0 \le V(z) \le \widehat{C}_0\big (1 + |z|^p\big )\), for every \(z=(x,y)\in \mathbb {R}^2\);
-
(ii)
for every \(z=(x,y),z'=(x',y')\in \mathbb {R}^2\),
$$\begin{aligned} |V(z) - V(z')| \le \widehat{C}_1 \big (1 + |z|+|z'|\big )^{p-1} |z-z'|; \end{aligned}$$ -
(iii)
for every \(z=(x,y),z'=(x',y')\in \mathbb {R}^2\) and \(\lambda \in (0,1)\),
$$\begin{aligned} 0\le & {} \lambda V(z)+(1-\lambda )V(z')-V(\lambda z + (1-\lambda ) z') \\\le & {} \widehat{C}_2 \lambda (1-\lambda ) (1+|z|+|z'|)^{(p-2)^+}|z-z'|^2. \end{aligned}$$
In particular, by (iii), V is convex and locally semiconcave, hence, by Corollary 3.3.8 in [9],
Proof
Due to (2.2) and (2.3), the properties of f required in (ii) and (iii) of Assumption 2.2 are straightly inherited by V (see, e.g., the proof of Theorem 1 of [12], that can easily adapted to our infinite time-horizon setting, or that of Theorem 2.1 in [10]). \(\square \)
3 The Dynkin Game and Preliminary Properties of the Free Boundaries
In this section we show that \(V_y\) identifies with the value function of a suitable Dynkin game (a zero-sum game of optimal stopping), and we derive preliminary properties of the two curves (free boundaries) that delineate the region of the space where the \(|V_y|<K\). In order to simplify the notation, in the following we write \(X^{x,y}\), instead of \(X^{x,y,0}\), to identify the solution to (2.3) for \(\xi \equiv 0\). Most of the results of this section are close to those in Section 3 of [21], and their proof will be therefore omitted for the sake of brevity.
Theorem 3.1
Let \((x,y) \in \mathbb {R}^2\). Denote by \(\mathcal {T}\) the set of all \(\mathbb {F}\)-stopping times, and for \((\sigma , \tau ) \in \mathcal {T}\times \mathcal {T}\) consider the stopping functional
where \(V_x\) is the partial derivative of V with respect to x (which exists and is continuous by Proposition 2.4). One has that the game has a value, i.e.
and such a value is given by
Moreover, the couple of \(\mathbb {F}\)-stopping times \((\tau ^{\star }(x,y),\sigma ^{\star }(x,y)):=(\tau ^{\star },\sigma ^{\star })\) such that
(with the usual convention \(\inf \emptyset = + \infty \)) form a saddle-point; that is,
Notice that the connection between bounded-variation control problems and Dynkin games is established in great generality when either the components of the state process are decoupled (see [31] and [23, 24]), or when the controlled system is one-dimensional (see [5]). In our two-dimensional setting with interconnected dynamics, the approaches followed in the aforementioned works is not applicable. We instead prove Theorem 3.1 by employing Theorems 3.11 and 3.13 in [11], through a suitable (and not immediate) approximation procedure needed to accommodate our degenerate setting. Since the proof of Theorem 3.1 is completely analogous to that developed in order to prove Theorem 3.1 in [21] (see Appendix A therein), we omit the details.
From (3.2) it readily follows that \(-K \le V_y(x,y) \le K\) for any \((x,y)\in \mathbb {R}^2\). Hence, defining
we have that those regions provide a partition of \(\mathbb {R}^2\).
By continuity of \(V_y\) (cf. Proposition 2.4), \(\mathcal {C}\) is an open set, while \(\mathcal {I}\) and \(\mathcal {D}\) are closed sets. Moreover, convexity of V provides the representation
where the functions \(b_1: \mathbb {R} \rightarrow \overline{\mathbb {R}}\) and \(b_2: \mathbb {R} \rightarrow \overline{\mathbb {R}}\) are defined as
(with the usual conventions \(\inf \emptyset = \infty \), \(\inf \mathbb {R} = -\infty \), \(\sup \emptyset = -\infty \), \(\sup \mathbb {R} = \infty \)).
Equation (3.2), together with the fact that \(x \mapsto V_x(x,y)\) is nondecreasing for any \(y \in \mathbb {R}\) by convexity of V (cf. Proposition 2.4) and \(x \mapsto f_y(x,y)\) is nondecreasing by Assumption 2.2-(iv), easily imply the following result.
Lemma 3.2
\(V_y(\cdot ,y)\) is nondecreasing for all \(y \in \mathbb {R}\).
We now move on by obtaining preliminary properties of \(b_1\) and \(b_2\). Its proof can be obtained by exploiting the continuity and the monotonicity of \(V_y\), and easily adjusting the arguments of the proof of Proposition 3.3 of [21] to the present setting in which \(V_y(\cdot ,y)\) is nondecreasing.
Proposition 3.3
The following hold:
-
(i)
\(b_1: \mathbb {R} \rightarrow \mathbb {R} \cup \{-\infty \}\), \(b_2: \mathbb {R} \rightarrow \mathbb {R} \cup \{\infty \}\);
-
(ii)
\(b_1\) and \(b_2\) are nonincreasing;
-
(iii)
\(b_1(x) < b_2(x)\) for all \(x \in \mathbb {R}\);
-
(iv)
\(b_1\) is left-continuous and \(b_2\) is right-continuous.
Let us now define
together with the pseudo-inverses of \(b_1\) and \(b_2\) by
(again, with the usual conventions \(\inf \emptyset = \infty \), \(\inf \mathbb {R} = -\infty \), \(\sup \emptyset = -\infty \), \(\sup \mathbb {R} = \infty \)).
Also the next proposition can be proved by easily adapting to our setting the proof of Proposition 3.4 in [21].
Proposition 3.4
The following holds:
-
(i)
\(g_1(y)= \inf \{ x \in \mathbb {R}:V_y(x,y) > -K\}, \quad g_2(y)= \sup \{ x \in \mathbb {R}:V_y(x,y) < K\};\)
-
(ii)
the functions \(g_1,g_2\) are nonincreasing and \(g_1(y) < g_2(y)\) for any \(y \in \mathbb {R}\);
-
(iii)
If \(\bar{b}_2<\infty \), then \(g_2(y)=-\infty \) for all \(y \ge \bar{b}_2\) and if \(\underline{b}_1>-\infty \), then \(g_1(r)=\infty \) for all \(y \le \underline{b}_1\).
4 The Structure of the Value Function and the Second-Order Smooth Fit
In this section, we exploit the results of the previous section in order to determine the structure of the value function V, and to show that \(V_{yx}\) is continuous on the whole state space (second-order smooth-fit property).
For any given and fixed \(y\in \mathbb {R}\), denote by \(\mathcal {L}^y\) the infinitesimal generator associated to the uncontrolled process \(X^{x,y}\). Acting on \(g \in C^2(\mathbb {R};\mathbb {R})\) it yields
Any solution \(\beta (\cdot ,y)\) to the second-order ordinary differential equation (ODE)
can be written as
where the strictly positive functions \(\psi \) and \(\varphi \) are given, for any \(z\in \mathbb {R}\), by
with
Notice that \(\psi (\cdot ,z)\) is strictly increasing while \(\varphi (\cdot ,z)\) is strictly decreasing for any \(z\in \mathbb {R}\).
By the dynamic programming principle, we expect that V identifies with a suitable solution to the following variational inequality
We now show that V is a viscosity solution to (4.4). Later, this will enable us to determine the structure of V (see Proposition 4.5 below) and then to upgrade its regularity (cf. Proposition 4.7) in order to derive necessary optimality conditions for the boundaries splitting the state space (cf. Theorem 6.1).
Definition 4.1
-
(i)
A function \(v\in C^0(\mathbb {R}^2; \mathbb {R})\) is called a viscosity subsolution to (4.4) if, for every \((x,y) \in \mathbb {R}^2\) and every \(\beta \in C^{2,1}(\mathbb {R}^2;\mathbb {R})\) such that \(v-\beta \) attains a local maximum at (x, y), it holds
$$\begin{aligned} \max \bigg \{ -\beta _y(x,y) - K,\ \beta _y(x,y)- K ,\ \rho \beta (x,y) -[\mathcal {L}^y\beta (\cdot ,y)](x) - f(x,y) \bigg \} \le 0. \end{aligned}$$ -
(ii)
A function \(v\in C^0(\mathbb {R}^2; \mathbb {R})\) is called a viscosity supersolution to (4.4) if, for every \((x,y) \in \mathbb {R}^2\) and every \(\beta \in C^{2,1}(\mathbb {R}^2;\mathbb {R})\) such that \(v-\beta \) attains a local minimum at (x, y), it holds
$$\begin{aligned} \max \bigg \{ -\beta _y(x,y) - K,\ \beta _y(x,y)- K,\ \rho \beta (x,y) - [\mathcal {L}^y\beta (\cdot ,y)](x) - f(x,y) \bigg \} \ge 0. \end{aligned}$$ -
(iii)
A function \(v\in C^0(\mathbb {R}^2; \mathbb {R})\) is called a viscosity solution to (4.4) if it is both a viscosity subsolution and supersolution.
Following the arguments developed in Theorem 5.1 in Section VIII.5 of [22], one can show the following result (see also Proposition 4.2 in [21])
Proposition 4.2
The value function V is a viscosity solution to (4.4).
Remark 4.3
Recall that by Proposition 2.4-(iii) our value function V lies in the class \(W^{2,\infty }_{\text {loc}}(\mathbb {R}^2;\mathbb {R})\). Hence, by Lemma 5.4 in Chapter 4 of [46] it is also a strong solution to (4.4) (in the sense, e.g., of [7]; see the same reference also for relations between these notions of solutions); that is, it solves (4.4) in the pointwise sense almost everywhere.
We have decided to employ the concept of viscosity solution since our analysis will later make use of the variational inequality (4.4) on sets of null Lebesgue measure (regular lines) (see Proposition 4.4 and Proposition 4.7 below). Because the viscosity property holds for all (and not merely for a.e.) points of the state space \(\mathbb {R}^2\), the concept of viscosity solution is still able to provide information on V on regular lines.
For later use, notice that the function
is finite by Assumption 2.2-(i) and standard estimates, and continuously differentiable with respect to y and x, given the assumed regularity of \(f_x\) and \(f_y\) in Assumption 2.2-(iii). Moreover, for any given and fixed \(y\in \mathbb {R}\), we introduce the scale function density of the process \(X^{x,y}\)
the density of the speed measure
as well as the positive constant (normalized) Wronskian between \(\psi \) and \(\varphi \)
Then, letting
be the Green function, we have that \(\widehat{V}\) admits the representation (cf., e.g., Ch. 2 of [6])
that is, using (4.1) and (4.7),
By direct calculations, it thus follows from (4.11) that \(\widehat{V}\) identifies with a classical particular solution to the inhomogeneous linear ODE
Recall now the regions \(\mathcal {C}\), \(\mathcal {I}\) and \(\mathcal {D}\) from (3.4), and that \(V_y=-K\) on \(\mathcal {I}\), while \(V_y=K\) on \(\mathcal {D}\). The next proposition provides the structure of V inside \(\mathcal {C}\). Its proof can be obtained by arguing exactly as in the proof of Proposition 4.4 of [21] (see also Remarks 4.3 and 4.5 therein), and it is therefore omitted.
Proposition 4.4
Recall (3.7) and let \(y_o \in (\underline{b}_1,\bar{b}_2)\).
-
(i)
The function \(V(\cdot ,y_o)\) is a viscosity solution to
$$\begin{aligned} \rho \beta (x,y_o)-[\mathcal {L}^{y_o}\beta (\cdot ,y_o)](x) - f(x,y_o) =0, \quad x \in (g_1(y_o),g_2(y_o)).\nonumber \\ \end{aligned}$$(4.13) -
(ii)
\(V(\cdot ,y_o) \in C^{3,Lip}_{\text {loc}}((g_1(y_o),g_2(y_o)); \mathbb {R})\).
-
(iii)
There exist constants \(A(y_o)\) and \(B(y_o)\) such that for all \(x \in (g_1(y_o),g_2(y_o))\)
$$\begin{aligned} V(x,y_o)=A(y_o)\psi (x, \alpha y_o) + B(y_o) \varphi (x, \alpha y_o) + \widehat{V}(x,y_o), \end{aligned}$$where the functions \(\psi \) and \(\varphi \) have been defined in (4.1) and \(\widehat{V}\) is as in (4.5).
We can now determine the structure of the value function V. The proof of the next proposition is completely analogous to that of Theorem 4.6 in [21]; however, we provide it here since it will be useful in the proof of a subsequent result (cf. Proposition 5.8).
Proposition 4.5
Define the sets
There exist functions
such that the value function defined in (2.6) can be written as
where \(\bar{\mathcal {C}}\) denotes the closure of \(\mathcal {C}\),
and
Proof
We start by deriving the structure of V within \(\mathcal {C}\). Using Lemma 4.4, we already know the existence of functions \(A,B:(\underline{b}_1,\bar{b}_2) \rightarrow \mathbb {R}\) such that
Take now \(y_o \in (\underline{b}_1,\bar{b}_2)\). Since \(g_1(y) < g_2(y)\) for any \(y \in \mathbb {R}\) (cf. Proposition 3.4-(ii)), we can find x and \(\tilde{x}\), \(x \ne \tilde{x}\), such that \((x,y),(\tilde{x},y) \in \mathcal {C}\) for any given \(y \in (y_o-\varepsilon ,y_o+\varepsilon )\), for a suitably small \(\varepsilon >0\). Now, by evaluating (4.18) at the points (x, y) and \((\tilde{x},y)\), we obtain a linear algebraic system that we can solve with respect to A(y) and B(y) so to obtain
The denominators of the last two expressions do not vanish due to the strict monotonicity of \(\psi \) and \(\varphi \), and to the fact that \(x \ne \tilde{x}\). Since \(y_o\) was arbitrary and V, \(\widehat{V}\), \(V_y\), and \(\widehat{V}_y\) are continuous with respect to y, we therefore obtain that A and B belong to \(W^{2,\infty }_{\text {loc}}((\underline{b}_1,\bar{b}_2);\mathbb {R})={C}_{\text {loc}}^{1,\text {Lip}}((\underline{b}_1,\bar{b}_2);\mathbb {R})\). The structure of V in the closure of \(\mathcal {C}\), denoted by \(\overline{\mathcal {C}}\), is then obtained by Proposition 4.4 and by recalling that V is continuous on \(\mathbb {R}^2\) and that A, B, and \(\widehat{V}\) are also continuous.
Given the definition of \(z_1\) and \(z_2\), the structure of V inside the regions \(\mathcal {I}\) and \(\mathcal {D}\) follow by (3.4) and the continuity of V. \(\square \)
Remark 4.6
Actually, by (4.19) and (4.20) one has that A and B belong to \(W^{2,\infty }\) up to \(\underline{b}_1\) (resp. \(\bar{b}_2\)) if \(\underline{b}_1\) (resp. \(\bar{b}_2\)) is finite (cf. also Remark 4.7 in [21]).
Notice that
The next result shows that one actually has continuity of \(V_{yx}\) on the whole \(\mathbb {R}^2\). Its proof can be obtained by following that of Theorem 5.1 in [21] (see also Proposition 5.3 in [20]), upon recalling that in our setting \(V_y(\cdot ,y)\) is nondecreasing (cf. Lemma 3.2).
Proposition 4.7
One has that
Hence, \(V_{yx} \in C(\mathbb {R}^2;\mathbb {R})\).
Lemma 4.8
It holds \(V_{yxx} \in L^{\infty }_{\text {loc}}(\mathbb {R}\times (\underline{b}_1,\bar{b}_2);\mathbb {R})\).
Proof
Notice that by (4.1) one has \(\psi _{xx}(x, \alpha y)=r^2_1(\alpha y)\psi (x, \alpha y)\), \(\varphi _{xx}(x, \alpha y)=r^2_2(\alpha y)\varphi (x, \alpha y)\), and \(\psi _{zxx}(x, \alpha y)=r'_1(\alpha y) r_1(\alpha y)\psi (x, \alpha y)(2 +x r_1(\alpha y))\), \(\varphi _{zxx}(x, \alpha y)=r'_2(\alpha y) r_2(\alpha y)\varphi (x, \alpha y)(2 + x r_2(\alpha y))\). Moreover, \(\widehat{V}_{yxx} \in L^{\infty }_{\text {loc}}(\mathbb {R}^2)\) by direct calculations on (4.11), and \(A_y,B_y \in W_{\text {loc}}^{1,\infty }((\underline{b}_1,\bar{b}_2);\mathbb {R})\) by Proposition 4.5. Hence, \(V_{yxx} \in L^{\infty }_{\text {loc}}(\mathbb {R}\times (\underline{b}_1,\bar{b}_2);\mathbb {R})\) by (4.15). \(\square \)
5 Further Properties of the Free Boundaries
In this section we move on by proving further properties of the free boundaries under additional mild requirements on f. The main aim of the following analysis is to determine a local Lipschitz regularity of the free boundaries \(g_1\) and \(g_2\) (cf. Proposition 5.8). Exploiting this regularity, in Sect. 6 we will show that \(g_1\) and \(g_2\) must solve a system of first-order ordinary differential equations in the classical sense.
Throughout the rest of this paper, the following assumption is in place.
Assumption 5.1
-
(i)
\(\lim _{x \rightarrow \pm \infty }f_x(x,y)=\pm \infty \).
-
(ii)
\(f_{yx}\) exists and it is continuous.
-
(iii)
One of the following holds true:
-
(a)
\(x\mapsto f_y(x,y)\) is strictly increasing for any \(y\in \mathbb {R}\);
-
(b)
\(f_{yx}\equiv 0\) and \(f(\cdot ,y)\) is strictly convex for any \(y\in \mathbb {R}\).
-
(a)
Remark 5.2
The functions f discussed in Remark 2.3 satisfy the previous assumptions.
We start by studying the limiting behavior of the functions \(b_i\) and some natural strict bounds for \(g_i\).
Proposition 5.3
-
(i)
Let Assumption 5.1-(i) hold. Then
$$\begin{aligned} \bar{b}_1 = \lim _{x \downarrow - \infty } b_1(x) = \infty , \quad \underline{b}_2 = \lim _{x \uparrow \infty } b_2(x)=-\infty ; \end{aligned}$$hence, by Proposition 3.3-(iii), one also has \(\underline{b}_1=-\infty \) and \(\bar{b}_2= \infty \).
-
(ii)
Define
$$\begin{aligned} \zeta _1(y):= & {} \sup \{x \in \mathbb {R}: -\alpha V_x(x,y) - f_y(x,y) - \rho K \ge 0\}, \quad y\in \mathbb {R}, \\ \zeta _2(y):= & {} \inf \{x \in \mathbb {R}: -\alpha V_x(x,y) - f_y(x,y) + \rho K \le 0\}, \quad y\in \mathbb {R}. \end{aligned}$$Then, for any \(y\in \mathbb {R}\), we have
$$\begin{aligned} g_1(y)< \zeta _1(y)< \zeta _2(y) < g_2(y). \end{aligned}$$
Proof
Proof of (i). Here we show that \( \lim _{x \downarrow -\infty } b_1(x)=\infty \). The fact that \( \lim _{x \uparrow \infty } b_2(x)=-\infty \) can be proved by similar arguments. We argue by contradiction assuming \(\bar{b}_1:=\lim _{x \downarrow -\infty } b_1(x)< \infty \). Take \(y_o> \bar{b}_1\), so that \(\tau ^\star =\tau ^{\star }(x,y_o)=\infty \) for all \(x\in \mathbb {R}\), the latter being the stopping time defined in (3.3). Then, take \(x_o < g_2(y_o)\) such that \((x_o,y_o)\in \mathcal {C}\). Clearly, every \(x<x_o\) belongs to \(\mathcal {C}\), and therefore, by the representation (4.15), we obtain that it must be \(B(y_o)=0\); indeed, otherwise, by taking limits as \(x\rightarrow -\infty \) and using (4.1), we would contradict Proposition 2.4. Moreover, since for any \(y\in \mathbb {R}\) one has \(\psi _x(x, \alpha y)\rightarrow 0\) when \(x\rightarrow -\infty \) (cf. (4.1)), we then have by dominated convergence
Now, setting
for \(x < x_o\), we have by monotonicity of \(f_y(\cdot ,y)\) (cf. Assumption 2.2-(iv))
The latter implies
Hence, letting \(x \downarrow -\infty \), using (5.1), and invoking the dominated convergence theorem we get a contradiction.
Proof of (ii). The fact that for any \(y\in \mathbb {R}\) we have \(g_1(y) \le \zeta _1(y)\) and \(g_2(y) \ge \zeta _2(y)\) can be obtained as in the proof of item (ii) of Proposition 6.1 in [21], by employing the proved regularity of \(V_y(\cdot ,y)\) and the semiharmonic characterization of [38] (see Eqs. (2.27)–(2.29) therein, suitably adjusted to take care of the integral term appearing in (3.2)). Moreover, \(\zeta _1(y)< \zeta _2(y)\) for any \(y \in \mathbb {R}\) by definition. It thus remains to show that one actually has \(g_1(y) < \zeta _1(y)\) and \(g_2(y) > \zeta _2(y)\) for any \(y\in \mathbb {R}\).
We only prove that \(g_2(y) > \zeta _2(y)\) for any \(y\in \mathbb {R}\), as the other case can be treated similarly. Suppose that there exists some \(y_o\) such that \(g_2(y_o)=\zeta _2(y_o)\). Then \(V_y(\zeta _2(y_o),y_o)=K\). Let now \(\tau ^{\star }:=\tau ^{\star }(\zeta _2(y_o),y_o)\) be the optimal stopping time for the sup player when the Dynkin game (3.2) starts at time zero from the point \((\zeta _2(y_o),y_o)\), and for \(\varepsilon >0\) define
Then by using that \(f_{y}(\cdot ,y_o) + \alpha V_x(\cdot ,y_o)\) is nondecreasing and locally Lipschitz by Assumption 2.2-(iii) and Proposition 2.4(iii), we have from (3.2) for some constant \(C(y_o)>0\)
Using now that, by definition of \(\zeta _2\), it must be \(\big (f_y + \alpha V_x\big )(\zeta _2(y_o),y_o)=\rho K\), and rearranging terms, we get that
Notice now that (cf. eq. (4.3) in [15], among others)
and
Then, because
using the last two formulas in (5.4) and performing a first-order Taylor’s expansion around \(\varepsilon =0\) of the terms on the right-hand side of (5.4), one finds that the first term on the right-hand side of (5.4) is positive and converges to zero as \(\varepsilon \downarrow 0\) with order \(\varepsilon ^2\), while the second term is negative and converges to zero with order \(\varepsilon \). We thus reach a contradiction in (5.4) for \(\varepsilon \) small enough, and therefore it cannot exist \(y_o\) at which \(g_2(y_o)=\zeta _2(y_o)\). \(\square \)
The next result readily follows from Proposition 5.3-(i).
Corollary 5.4
Let Assumption 5.1-(i) hold. Then the functions \(g_1, g_2\) defined in (3.8) are finite.
We now move on by proving that the boundaries \(b_1\) and \(b_2\) are strictly monotone. This will in turn imply that their inverses \(g_1\) and \(g_2\) are actually continuous.
Proposition 5.5
Let Assumption 5.1 hold. Then the functions \(b_1, b_2\) are strictly decreasing.
Proof
We prove the claim only for \(b_1\), since analogous arguments apply to prove it for \(b_2\).
Case (a). We assume here that item (a) of Assumption 5.1-(iii) holds, i.e. that \(x\mapsto f_y(x,y)\) is strictly increasing for any \(y\in \mathbb {R}\). By Proposition 4.5, we can differentiate the first line of (4.15) with respect to y and get by Proposition 4.4-(i) that \(V_y\) solves inside \(\mathcal {C}\) the equation
By continuity, (5.5) also holds on \(\overline{\mathcal {C}}\), i.e.
In particular it holds on \(\partial ^1 \mathcal {C}:=\overline{\mathcal {C}}\cap \mathcal {I}\). Assume now, by contradiction, that the boundary \(b_1\) is constant on \((x_o,x_o + \varepsilon )\), for some \(x_o \in \mathbb {R}\) and some \(\varepsilon >0\). Then, setting \(y_o:=b_1(x_o)\), we will have \(V_{yxx}(\cdot ,y_o)=V_{yx}(\cdot ,y_o)=0\) and \(V_y(\cdot ,y_o)=-K\) on \((x_o,x_o + \varepsilon )\). Hence, we obtain from (5.5) that
and thus
But now \(\alpha V_{xx}(x,y_o) \ge 0\) for any \(x\in (x_o,x_o+\varepsilon )\) by convexity of \(V(\cdot ,y_o)\), while, by assumption, \(f_{yx}\) must be strictly positive on a subset of \((x_o,x_o+\varepsilon )\) with positive measure. Hence a contradiction is reached.
Case (b). We assume here that item (b) of Assumption 5.1-(iii) holds, i.e. that \(f_{yx}\equiv 0\) and that \(f(\cdot ,y)\) is strictly convex for any \(y\in \mathbb {R}\). In such a case the claim can be proved by employing the same arguments of the proof of Proposition 6.3 in [21]. \(\square \)
From the above result, the following corollary is immediate.
Corollary 5.6
Let Assumption 5.1 hold. Then the functions \(g_1, g_2\) defined in (3.8) are continuous.
The next result will be of fundamental importance to show the local Lipschitz property of \(g_i\), \(i=1,2\) via a suitable application of the implicit function theorem (cf. Proposition 5.8 below).
Proposition 5.7
Let Assumption 5.1 hold. Then there exists
and one has \(\ell (x_o,y_o) \ne 0\) for any \((x_o,y_o) \in \partial \mathcal {C}.\)
Proof
We provide the proof only for any \((x_o,y_o) \in \partial ^2\mathcal {C}:=\overline{\mathcal {C}}\cap \mathcal {D}\), as the other case can be treated similarly.
First of all, we notice that the limit in (5.9) exists since, by Proposition 4.5, the function \(V:\mathcal {C}\rightarrow \mathbb {R}\) can be differentiated twice with respect to x and once with respect to y with continuity up to the boundary \(\partial \mathcal {C}\).
Case (a). We assume here that item (a) of Assumption 5.1-(iii) holds, i.e. that \(x\mapsto f_y(x,y)\) is strictly increasing for any \(y\in \mathbb {R}\). Suppose, by contradiction, that for some \(y_o \in \mathbb {R}\) one has
Then taking limits as \((x,y) \rightarrow \ (g_2(y_o),y_o)\) for \((x,y) \in \mathcal {C}\) in (4.13) we find, using that \(V_{yx}(g_2(y_o),y_o)=0\) by Proposition 4.7 and that \(V_y(g_2(y_o),y_o)=K\),
Since \(g_2(y_o) > \zeta _2(y_o)\) by Proposition 5.3, and by definition of \(\zeta _2\), it must be
which also implies that \(-\alpha V_{xx}(x,y_o) = f_{yx}(x,y_o)\) for any \(x \in (\zeta _2(y_o), g_2(y_o)).\) We then conclude as in Case (a) of the proof of Proposition 5.5.
Case (b). We assume here that item (b) of Assumption 5.1-(iii) holds, which implies that there exists q such that \(f_{y}(x,y)=q(y)\) for any \((x,y) \in \mathbb {R}^2\). Suppose again, with the aim of reaching a contradiction, that for some \(y_o \in \mathbb {R}\) one has (5.10). Then taking limits as \((x,y) \rightarrow \ (g_2(y_o),y_o)\) for \((x,y) \in \mathcal {C}\) in (4.13) we find, using that \(V_{yx}(g_2(y_o),y_o)=0\) by Proposition 4.7 and that \(V_y(g_2(y_o),y_o)=K\),
As before, because \(g_2(y_o) > \zeta _2(y_o)\) by Proposition 5.3, and by definition of \(\zeta _2\), it must be
that is, V is an affine function of x in that interval. However, using the latter and (4.13), we also have
and we reach a contradiction since \(f(\cdot ,y_o)\) is strictly convex by assumption, while \(V(\cdot ,y_o)\) is affine. \(\square \)
Thanks to the previous analysis we are finally able to prove the aforementioned LIpschitz regularity of \(g_1\) and \(g_2\).
Proposition 5.8
Let Assumption 5.1 hold. Then the functions \(g_1,g_2\) are locally Lipschitz.
Proof
Define the function
where A, B are the functions of Proposition 4.5. Then, one clearly has that \(\bar{V} \in C^{2,1}(\mathbb {R}^2;\mathbb {R})\) (since, by Sobolev embedding, A and B belong to \(C^{1,\text {Lip}}_{\text {loc}}(\mathbb {R};\mathbb {R})\)), and \(\bar{V} = V\) in \(\mathbb {R}^2 \cap \bar{\mathcal {C}}\). Moreover, the mixed derivative \(\bar{V}_{yx}\) exists and is continuous, and standard differentiation yield
Since \(A_y\) and \(B_y\) are locally Lipschitz by Proposition 4.5, and \(\psi \) and \(\varphi \) are smooth (cf. (4.1)), we deduce that \(\bar{V}_{yx}(x,\cdot )\) is locally Lipschitz.
Let now \(y_o\in \mathbb {R}\). Then, for any given \(x_o \in \mathbb {R}\) such that \((x_o,y_o) \in \partial \mathcal {C}\), we know by Proposition 5.7 that \(\bar{V}_{yxx}(x_o,y_o) \ne 0\), while \(\bar{V}_{yx}(x_o,y_o) = 0\). Therefore, the implicit function theorem (see, e.g., the Corollary at p. 256 in [14] or Theorem 3.1 in [36]) implies that, for any \(i=1,2\) and for suitable \(\delta ,\delta '>0\), there exists a unique continuous function \(\bar{g}_i: (y_o - \delta , y_o + \delta ) \rightarrow (x_o-\delta ', x_o + \delta ')\) such that \(\bar{V}_{yx}(\bar{g}_i(y),y) = 0\) in \((y_o - \delta , y_o + \delta )\). Also, the aforementioned properties of \(\bar{V}_{yxy}\) and \(\bar{V}_{yxx}\) imply that there exists \(C(y_o)>0\) such that
Recalling now that \(\bar{V}_{yx}(g_i(y),y)=0\), we can identify \(\bar{g}_i={g}_i\), \(i=1,2\), in \((y_o - \delta , y_o + \delta )\) and therefore \({g}_i\) is locally Lipschitz therein. Given the arbitrariness of the point \((x_o,y_o)\) the proof is complete. \(\square \)
6 A System of Differential Equations for the Free Boundaries
In this section we derive a first-order system of nonlinear differential equations for the free boundaries \(g_1\) and \(g_2\), i.e. we will be able to write
for some explicitly determined maps \(G_1,G_2\), whose regularity will allow also to establish a \(C^{1,\text {Lip}}\) regularity for \(g_1,g_2\). To the best of our knowledge, for a two-dimensional degenerate singular stochastic control problem with interconnected dynamics as ours, a similar result appears here for the first time.
We first move on by establishing four equations relating \(g_1,g_2\) and A, B. Recall (4.6), (4.7), and (4.9). We also denote by p the transition density of \(X^{x,y}\) with respect to the speed measure; then, letting \(A \mapsto \mathsf {P}_t(x,A,y)\), \(A\in \mathcal {B}(\mathbb {R})\), \(t>0\) and \(y\in \mathbb {R}\), be the probability of starting at time 0 from level \(x \in \mathbb {R}\) and reaching the set \(A \in \mathcal {B}(\mathbb {R})\) in t units of time, we have (cf., e.g., p. 13 in [6])
The density p can be taken positive, jointly continuous in all variables and symmetric (i.e. \(p(t,x,z,y)=p(t,z,x,y)\)).
Theorem 6.1
Let Assumption 5.1 hold. Recall (4.1), (4.15), and for any \((x,y) \in \mathbb {R}^2 \) define
Then, the free boundaries \(g_1\) and \(g_2\) as in (3.8), and the coefficients A and B are such that
Proof
To obtain Eqs. (6.3) and (6.4) we exploit the proved regularity of \(V(\cdot ,y)\) (cf. Propositions 2.4, 4.7, and Lemma 4.8) in order to follow the proof of Theorem 6.5 in [21]. This is based on an application of the local time-space calculus of [37] to the process \((e^{-\rho s}V_y(X_s, y))_{s\ge 0}\) and the use of the Green function (4.9), the transition probability (6.1), and Fubini’s theorem. Alternatively (and equivalently) they can be derived from (4.15) by imposing that \(V_{y}(\cdot ,y)\) and \(V_{yx}(\cdot ,y)\) are continuous on \(\partial \mathcal {C}\) and proceeding via the more analytical direct approach of the proof of Proposition 5.5 in [20]. In particular, the second-order smooth fit \(V_{yx}(g_i(y),y)=0\), \(i=1,2\), easily gives (6.5) and (6.6). \(\square \)
Corollary 6.2
Let Assumption 5.1 hold. For \(z \in \mathbb {R}\) define
and for \(y \in \mathbb {R}\), \(i,j=1,2\), \(j\ne i\),
Furthermore, for any \((x_1,x_2,y) \in \{(x_1,x_2,y)\in \mathbb {R}^3: x_1\ne x_2\}\), define
and
Then, one has
as well as
Proof
In order to derive (6.9) and (6.10), notice that, given \(g_1\) and \(g_2\), and exploiting (4.15), one has from (6.3) and (6.4) that A and B solve the linear system
By using expressions for \(\psi \), \(\varphi \), \(S_x\) and \(m_x\) (cf. (4.1), (4.6) and (4.7)) one can explicitly evaluate the integrals appearing on the left-hand sides of (6.13) and (6.14). Then, solving the latter two equations with respect to A and B one finds after some simple but tedious algebra (6.9) and (6.10). Notice indeed that the denominator appearing in (6.9) and (6.10) is nonzero since \(g_1 \ne g_2\) and one has \(\sinh ^2(z) - z^2 >0\) for any \(z\ne 0\).
In order to find (6.11) and (6.12) we solve (6.5) and (6.6) with respect to \(A'(y)\) and \(B'(y)\), and use (4.1), (4.2), (4.3), and (6.7). \(\square \)
Remark 6.3
In [21] a system of Eqs. like (6.3) and (6.4) has also been obtained (see eqs. (6.11) and (6.12) therein). However, in [21] the uncontrolled process is of Ornstein-Uhlenbeck type and this made it not possible to determine explicit expressions for A(y) and B(y) as in (6.9) and (6.10) above. Indeed, the complex form of the functions \(\psi \) and \(\varphi \) associated to the Ornstein-Uhlenbeck process does not allow to conclude that the determinant of the coefficients’ matrix arising when one tries to solve (the analogous of) (6.13) and (6.13) with respect to A(y) and B(y) is nonzero.
We can now state the main result of this paper.
Theorem 6.4
Let \(D:=\{(x_1,x_2,y)\in \mathbb {R}^3: x_1\ne x_2\}\). There exist explicitly computableFootnote 1 functions \(G_i\in C^{0,\text {Lip}}_{\text {loc}}(D;\mathbb {R})\), \(i=1,2\) such that
In particular, \(g_i\in C^{1,\text {Lip}}_{\text {loc}}(\mathbb {R};\mathbb {R})\) for \(i=1,2\).
Proof
Recall Theorem 4.5 and (6.2). In particular, for any (x, y) such that \(g_1(y) \le x \le g_2(y)\) – i.e. for any \((x,y) \in \bar{\mathcal {C}}\) – we have by (4.15)
with A, B belonging to \(W^{2,\infty }_{\text {loc}}(\mathbb {R};\mathbb {R})\). Defining then the function
one has \(\bar{H} = H\) on \(\bar{\mathcal {C}}\).
Introduce now \(\Phi _i: D \rightarrow \mathbb {R}\) defined as
Observing that (cf. Chapter II in [6])
one can readily see that, by (6.3)–(6.4), for any \(y \in \mathbb {R}\) one has
Thanks to Assumption 2.2 and Theorem 4.5, one has that \(\bar{H}\in C^{1,\text {Lip}}_{\text {loc}}(\mathbb {R}^2;\mathbb {R})\). Hence, for any \(i=1,2\), the map \((x_1,x_2) \mapsto \Phi _i(x_1,x_2,y)\) belongs to \(C^{2}(D; \mathbb {R})\) for each \(y\in \mathbb {R}\) and the map \(y \mapsto \Phi _i(x_1,x_2,y)\) belongs to \(C^{1,\text {Lip}}_{\text {loc}}(D; \mathbb {R})\) for each \((x_1,x_2)\in \mathbb {R}^2\). Recalling Proposition 5.8 we can take the total derivative on both terms appearing in (6.18) we obtain for a.e. \(y\in \mathbb {R}\) that
The determinant of the matrix \(\Lambda \), denoted by \(|\Lambda |\), is given by
We now aim at showing that \(|\Lambda |(g_1(y),g_2(y),y)\) does not vanish for any \(y\in \mathbb {R}\) under Assumption 5.1-(iii). On the one hand, if item (a) of that assumption holds, i.e. \(x \mapsto f_y(x,y)\) is strictly increasing, then we have that \(x \mapsto H(x,y)\) is such as well. Since \(\bar{H} = H\) on \(\bar{\mathcal {C}}\) and \(g_2(y)> \zeta _2(y)> \zeta _1(y) > g_1(y)\) by Proposition 5.3-(ii), we have
and
therefore, \(|\Lambda |(g_1(y),g_2(y),y)<0\). On the other hand, if item (b) of Assumption 5.1-(iii) holds, i.e. if \(f_{yx}\equiv 0\) and \(f(\cdot ,y)\) is strictly convex for any \(y\in \mathbb {R}\), we can argue by contradiction as in Case (b) of the proof of Corollary 5.6. To this end, suppose, for example, that \(\bar{H}(g_1(y_o),y_o) + K \rho = H(g_1(y_o),y_o) + K \rho =0\), for some \(y_o\in \mathbb {R}\). Denoting \(f_y(x,y)=q(y)\) it then follows that
by definition of \(\zeta _1\) (cf. Proposition 5.3); that is, V is an affine function of x in that interval. However, using the latter and (4.13), we also have
and we reach a contradiction since f is strictly convex in x by assumption while V is affine. The same argument also implies that \(\bar{H}(g_1(y_o),y_o) + K \rho \ne 0\). We have then proved that in any case one has \(|\Lambda |(g_1(y),g_2(y),y) \ne 0\) under Assumption 5.1-(iii).
We can therefore invert the matrix \(\Lambda \) appearing in (6.19) and obtain that for a.e. \(y \in \mathbb {R}\)
Observe now that, given the aforementioned regularity of \(\frac{\partial \Phi _i}{\partial x_j}\), \(i,j=1,2\), and of \(\frac{\partial \Phi _i}{\partial y}\), \(i=1,2\), we have \(G_i\in C^{0,\text {Lip}}_{\text {loc}}(D;\mathbb {R})\); hence, \(g_i\in C^{1,\text {Lip}}_{\text {loc}}(\mathbb {R};\mathbb {R})\). \(\square \)
Remark 6.5
Notice that the right-hand sides of (6.21) are indeed functions only of \((g_1(y),g_2(y),y)\). To see that, it is enough to feed (6.9) and (6.10), and (6.11) and (6.12) in the right-hand sides of (6.21), upon noticing that for any \(i,j=1,2\), \(\frac{\partial \Phi _i}{\partial x_j}\) depend on A(y), B(y), while, for any \(i=1,2\), \(\frac{\partial \Phi _i}{\partial y}\) depend on \(A'(y), B'(y)\).
Remark 6.6
In the proof of Proposition 5.6 of [20] (see page 2213 therein; see also Step 4 in the proof of Lemma 7 in [35] and the proof of Proposition 6 in [17]), a system of ODEs for the free boundaries is determined with the aim of proving that the free boundaries belong to \(C^1\) and are strictly monotone. In our problem, proving strict monotonicity of \(g_1\) and \(g_2\) would require to establish a strict sign for \(G_1\) and \(G_2\) (cf. (6.21)). However, the interaction between our dynamics – and the consequent dependency of \(\psi \), \(\varphi \), and \(m_x\) on y – makes the partial derivatives \(\frac{\partial \Phi _i}{\partial y}\) appearing in (6.21) much more complex than the analogous quantities in [20] or [35], and this in turn makes it unclear that \(G_i<0\), \(i=1,2\) (although expected).
6.1 A Discussion on Theorem 6.4 and the Optimal Control
6.1.1 On Theorem 6.4
Given the full degeneracy of our setting, the fact that the free boundaries \(g_i\), \(i=1,2\), belong to the class \(C^{1,\text {Lip}}_{\text {loc}}(\mathbb {R};\mathbb {R})\) is, to the best of our knowledge, a remarkable result. Indeed, the lack of uniform ellipticity of the diffusion coefficient makes it already difficult to obtain a preliminary (locally) Lipschitz property of \(g_i\)s by invoking results from PDE theory ( [8] and [39], among others) or techniques as those in [40, 41], and [42]. Also the probabilistic approach developed in [18] is not directly applicable since our free boundaries are associated with a Dynkin game rather than to an optimal stopping problem.
It is also worth stressing that Theorem 6.4 not only provides regularity of the free boundaries, but also a system of ODEs. To the best of our knowledge, a similar result appears here for the first time. Clearly, in order to provide a complete characterization of \(g_i\)s, (6.15) should be complemented by boundary conditions. The determination of those is a non trivial task. As a matter of fact, we have not been able to identify a relevant value of y for which the values of the free boundaries can be determined. The only information available is that the free boundaries diverge for large (in absolute value) levels of y; but this is clearly not enough. Even enforcing a finite-fuel constraint like \(\underline{y} \le Y^{y,\xi }_t \le \overline{y}\) a.s. for any \(t\ge 0\) would not help in order to obtain boundary conditions. Indeed, differently to the case with monotone controls (see [32]), here the drift process Y can be pushed back into \((\underline{y},\overline{y})\) once any of the boundary points of that interval is reached. Also, it is not clear to us how to obtain some kind of asymptotic growth of the free boundaries in order to restrict the functional class where to look for uniqueness of (6.15).
A possible way to obtain a complete implementable characterization of the free boundaries might be the following. Instead of thinking of \(g_1\) and \(g_2\) as functions of y, for a fixed parameter \(\alpha \), one could look at those as functions of \(\alpha \), for any given and fixed y. Bearing this in mind, one might try to prove that \(\alpha \mapsto g_i(\alpha ;y)\) are (at least) locally Lipschitz on \([0,\infty )\), and then follow the approach developed in this section in order to obtain a system of ODEs involving \(\partial _{\alpha }g_i(\alpha ;y)\), \(i=1,2\), rather than \(\partial _{y}g_i(\alpha ;y)\). Those ODEs would then be complemented by a natural boundary condition since, by continuity, \(g_1(0+;y)\) and \(g_2(0+;y)\) would coincide with the free boundaries uniquely determined in Proposition 5.5 of [20]. However, it is not straightforward to prove the aforementioned Lipschitz regularity of \(\alpha \mapsto g_i(\alpha ;y)\); indeed, a preliminary analysis shows that this is related to that of \(\alpha \mapsto V_x(x, y; \alpha )\), and how to prove the latter is not clear to us. The investigation of such an interesting conjecture is therefore left for future research.
6.1.2 On the Optimal Control
Figure 1 provides an illustrative description of the geometry of the state space. The horizontal lines in Fig. 1 represent the directions of the actions induced by the optimal control rule \(\xi ^{\star }\). This should be such that the jumps of the two-dimensional process \((X^{x, y,\xi ^{\star }}_t, Y^{y,\xi ^{\star }}_t)_{t\ge 0}\) are induced by the optimal control only at initial time, if the initial data (x, y) lie in the interior of \(\mathcal {I}\) or \(\mathcal {D}\), or at those times at which the process meets jumps of the free boundaries. The size of those interventions should be such that the process is immediately brought to the closest point on \(\partial \mathcal {C}\), from where it evolves according to (2.2) and (2.3) and in such a way that it is kept inside the closure of \(\mathcal {C}\) in a minimal way. Mathematically, this amounts to construct \((X^{x, y,\xi ^{\star }}_t, Y^{y,\xi ^{\star }}_t)_{t\ge 0}\) as a (degenerate) diffusion that is reflected at \(\partial \mathcal {C}\).
The latter is per se an interesting and not trivial problem, whose solution in multi-dimensional settings strongly hinges on the smoothness of the reflection boundary itself; sufficient conditions can be found in the seminal papers [19] and [33]. Unfortunately, our information on \(\partial \mathcal {C}\) do not suffice to apply the results of the aforementioned works since we are not able to exclude horizontal segments of the free boundaries \(g_1\) and \(g_2\) (cf. Case (1) and Case (2) in [19]). Indeed, although we can provide explicit formulas for the maps \(G_1\) and \(G_2\) appearing in (6.15), their complex expressions makes it hard to show that they are strictly negative (see also Remark 6.6). On the other hand, also the more constructive approach followed in Section 5 of [11] seems not to apply in general to our case, unless we assume (as the authors of [11] do) a linear growth of the free boundaries \(b_i\), \(i=1,2\), or further requirements on f leading to a weak solution to the reflection problem as in Proposition 7.3 of [21]. We therefore leave for future research the general study of the intricate and intriguing problem of constructing the optimal control.
Notes
Cf. Remark 6.5.
References
Ata, B., Harrison, J.M., Shepp, L.A.: Drift control of a Brownian processing system. Ann. Appl. Probab. 15, 1145–1160 (2005)
Alvarez, L.H.R.: Optimal capital accumulation under price uncertainty and costly reversibility. J. Econ. Dyn. Control 35, 1769–1788 (2011)
Beneš, V.E.: Girsanov functionals and optimal bang-bang laws for final value stochastic control. Stoch. Process. Appl. 2, 127–140 (1973)
Beneš, V.E., Shepp, L.A., Witsenhausen, H.S.: Some solvable stochastic control problems. Stochastics 4, 38–83 (1980)
Boetius, F.: Bounded variation singular stochastic control and Dynkin games. SIAM J. Control Optim. 44, 1289–1321 (2005)
Borodin, A.N., Salminen, P.: Handbook of Brownian Motion- Facts and Formulae, 2nd Corrected Springer, Berlin (2014)
Caffarelli, L.A., Crandall, M.G., Kocan, M., Šwiech, A.: On viscosity solutions of fully nonlinear equations with measurable ingredients. Commun. Pure Appl. Math. 49(4), 365–398 (1996)
Caffarelli, L.A., Salsa, S.: A Geometric Approach to Free Boundary Problems. Graduate Studies in Mathematics, vol. 68. American Mathematical Society, Providence, RI (2005)
Cannarsa, P., Sinestrari, C.: Semiconcave Functions, Hamilton-Jacobi Equations, and Optimal Control. Progress in Nonlinear Differential Equations and Their Applications, vol. 58. Birkhäuser, Basel (2014)
Chiarolla, M.B., Haussmann, U.G.: Geometric Approach to Monotone Stochastic Control. Ph.D. Thesis, The University of British Columbia (1992)
Chiarolla, M.B., Haussmann, U.G.: Controlling inflation: the infinite horizon case. Appl. Math. Optim. 41, 25–50 (2000)
Chow, P.-L., Menaldi, J.-L., Robin, M.: Additive control of stochastic linear systems with finite horizon. SIAM J. Control Optim. 23(6), 858–899 (1985)
Chow, P.-L., Menaldi, J.-L.: On the Numerical Solution of a Stochastic Optimal Correction Problem. Transactions of the third army conference on applied mathematics and computing, pp. 531–546 (1986)
Clarke, F.H.: Optimization and Nonsmooth Analysis. SIAM, Philadelphia, PA (1990)
Dayanik, S., Karatzas, I.: On the optimal stopping problem for one-dimensional diffusions. Stoch. Process. Appl. 107, 173–212 (2003)
De Angelis, T., Ferrari, G., Moriarty, J.: A non convex singular stochastic control problem and its related optimal stopping boundaries. SIAM J. Control Optim. 53(3), 1199–1223 (2015)
De Angelis, T., Ferrari, G., Moriarty, J.: A solvable two-dimensional degenerate singular stochastic control problem with non convex costs. Math. Oper. Res. 44(2), 512–531 (2019)
De Angelis, T., Stabile, G.: On lipschitz continuous optimal stopping boundaries. SIAM J. Control Optim. 57(1), 402–436 (2019)
Dupuis, P., Ishii, H.: SDEs with oblique reflection on nonsmooth domains. Ann. Probab. 21(1), 554–580 (1993)
Federico, S., Pham, H.: Characterization of the optimal boundaries in reversible investment problems. SIAM J. Control Optim. 52(4), 2180–2223 (2014)
Federico, S., Ferrari, G., Schuhmann, P.: A singular stochastic control problem with interconnected dynamics. SIAM J. Control Optim. 58, 2821–2853 (2020)
Fleming, W..H., Soner, H..M.: Controlled Markov Processes and Viscosity Solutions, 2nd edn. Springer, Berlin (2005)
Guo, X., Tomecek, P.: Connections between singular control and optimal switching. SIAM J. Control Optim. 47, 421–443 (2008)
Guo, X., Tomecek, P.: A class of singular control problems and the smooth fit principle. SIAM J. Control Optim. 47, 3076–3099 (2009)
Harrison, J.M., Taksar, M.I.: Instantaneous control of brownian motion. Math. Oper. Res. 8(3), 439–453 (1983)
Karatzas, I., Ocone, D.: The resolvent of a degenerate diffusion on the plane, with application to partially observed stochastic control. Ann. Appl. Probab. 2, 629–668 (1992)
Karatzas, I., Ocone, D.: The finite-horizon version for a partially observed stochastic control problem of Beneš and Rishel. Stoch. Anal. Appl. 11, 569–605 (1993)
Karatzas, I., Ocone, D.: A leavable bounded-velocity stochastic control problem. Stoch. Process. Appl. 99, 31–55 (2002)
Karatzas, I.: A class of singular stochastic control problems. Adv. Appl. Prob. 15, 225–254 (1983)
Karatzas, I., Shreve, S.E.: Connections between optimal stopping and singular stochastic control I. Monotone follower problems. SIAM J. Control Optim 22(6), 856–877 (1984)
Karatzas, I., Wang, H.: Connections between bounded-variation control and Dynkin games. In: J.L. Menaldi, A. Sulem and E. Rofman (eds.) Optimal Control and Partial Differential Equations; Volume in Honor of Professor Alain Bensoussan’s 60th Birthday, pp. 353–362. IOS Press, Amsterdam (2005)
Koch, T., Vargiolu, T.: Optimal Installation of Solar Panels with Price Impact: a Solvable Singular Stochastic Control Problem. (2019) Preprint on arXiv:1911.04223
Lions, P.L., Sznitman, A.S.: Stochastic Differential Equations with Reflecting Boundary Conditions. Commun. Pur. Appl. Math. 37, 511–537 (1984)
Matoglu, M.O., Vate, J.V., Wang, H.: Solving the drift control problem. Stoch. Syst. 5(2), 324–371 (2015)
Merhi, A., Zervos, M.: A model for reversible investment capacity expansion. SIAM J. Control Optim. 46(3), 839–876 (2007)
Papi, M.: On the domain of the implicit function and application. J. Ineq. Appl. 3, 221–234 (2005)
Peskir, G.: A change-of-variable formula with local time on curves. J. Theor. Probab. 18(3), 499–535 (2005)
Peskir, G.: Optimal stopping games and nash equilibrium. Theory Probab. Appl. 53, 558–571 (2008)
Petrosyan, A., Shahgholian, H., Uraltseva, N.: Regularity of Free Boundaries in Obstacle-type Problems. Graduate Studies in Mathematics, vol. 136. American Mathematical Society, Providence, RI (2012)
Soner, H.M., Shreve, S.E.: Regularity of the value function for a two-dimensional singular stochastic control problem. SIAM J. Control Optim. 27(4), 876–907 (1989)
Soner, H.M., Shreve, S.E.: A Free Boundary Problem Related to Singular Stochastic Control. Applied Stochastic Analysis (London, 1989) 265–301 (1991)
Soner, H.M., Shreve, S.E.: A free boundary problem related to singular stochastic control: the parabolic case. Commun. Part. Differ. Equ. 16(2–3), 373–424 (1991)
Sun, M., Menaldi, J.-L.: Monotone control of a damped oscillator under random perturbations. IMA J. Math. Control Inform. 5, 169–186 (1988)
Urban, T.L.: Inventory models with inventory-level-dependent demand: a comprehensive review and unifying theory. Eur. J. Oper. Res. 162, 792–804 (2005)
Taksar, M.I.: Average optimal singular control and a related stopping problem. Math. Oper. Res. 10(1), 63–81 (1985)
Yong, J., Zhou, X.Y.: Stochastic Control—Hamiltonian Systems and HJB Equations. Springer, Berlin (1999)
Acknowledgements
Financial support by the German Research Foundation (DFG) through the Collaborative Research Centre 1283 is gratefully acknowledged by the authors.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Federico, S., Ferrari, G. & Schuhmann, P. Singular Control of the Drift of a Brownian System. Appl Math Optim 84 (Suppl 1), 561–590 (2021). https://doi.org/10.1007/s00245-021-09779-3
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00245-021-09779-3
Keywords
- Singular stochastic control
- Dynkin game
- Viscosity solution
- Free boundary
- Smooth-fit
- Brownian motion
- Ordinary differential equation