Abstract
In this paper, we investigate the following problem: For a given upper bound for the ruin probability, maximize the expected discounted consumption of an investor in finite time. The endowment of the agent is modeled by a Brownian motion with positive drift. We give an iterative algorithm for the solution of the problem, where in each step an unconstrained, but penalized problem is solved. For the discontinuous value function \(V(t,x)\) of the penalized problem, we show that it is the unique viscosity solution of the corresponding Hamilton–Jacobi–Bellman equation. Moreover, we characterize the optimal strategy as a barrier strategy with continuous barrier function.
Similar content being viewed by others
References
Asmussen, S., Taksar, M.: Controlled diffusion models for optimal dividend pay-out. Insur. Math. Econ. 20, 1–15 (1997)
Avanzi, B.: Strategies for dividend distribution: a review. N. Am. Actuar. J. 13, 217–251 (2009)
Bayraktar, E., Young, V.: Maximizing utility of consumption subject to a constraint on the probability of lifetime ruin. Finance Res. Lett. 5, 204–212 (2008)
Beneš, V.E., Shepp, L.A., Witsenhausen, H.S.: Some solvable stochastic control problems. Stochastics 4, 39–83 (1980)
Billingsley, P.: Convergence of Probability Measures. Wiley Series in Probability and Statistics: Probability and Statistics. Wiley, New York (1999)
Bouchard, B.: Introduction to stochastic control of mixed diffusion processes, viscosity solutions and applications in finance and insurance. Lecture notes, Université Paris-IX Dauphine, Ceremade and CREST (2007). Available online www.ceremade.dauphine.fr/~bouchard/pdf/PolyContSto.pdf
Burdzy, K., Chen, Z., Sylvester, J.: The heat equation and reflected Brownian motion in time-dependent domains. II. Singularities of solutions. J. Funct. Anal. 204, 1–34 (2003)
Burdzy, K., Chen, Z., Sylvester, J.: The heat equation and reflected Brownian motion in time-dependent domains. Ann. Probab. 32, 775–804 (2004)
de Finetti, B.: Su un’ impostazione alternativa della teoria collettiva del rischio. In: Transactions of the XVth International Congress of Actuaries 2, pp. 433–443 (1957)
Doetsch, G.: Handbuch der Laplace-Transformation. Band I: Theorie der Laplace-Transformation. Birkhäuser, Basel (1971)
Elie, R.: Finite time Merton strategy under drawdown constraint: a viscosity solution approach. Appl. Math. Optim. 58, 411–431 (2008)
Fleming, W.H., Soner, H.M.: Controlled Markov Processes and Viscosity Solutions. Springer, New York (1993)
Frolova, A., Kabanov, Y., Pergamenshchikov, S.: In the insurance business risky investments are dangerous. Finance Stoch. 6, 227–235 (2002)
Gaier, J., Grandits, P., Schachermayer, W.: Asymptotic ruin probabilities and optimal investment. Ann. Appl. Probab. 13, 1054–1076 (2003)
Grandits, P.: Optimal consumption in a Brownian model with absorption and finite time horizon. Appl. Math. Optim. 67, 197–241 (2013)
Grandits, P.: Existence and asymptotic behavior of an optimal barrier for an optimal consumption problem in a Brownian model with absorption and finite time horizon. Appl. Math. Optim. 69, 233–271 (2014)
Grandits, P.: Asymptotic behavior of an optimal barrier in a constrained optimal consumption problem. Preprint, TU Vienna (2014). Available online www.fam.tuwien.ac.at/~pgrand/constraint_asymp.pdf
Itô, K., McKean, H.P.: Diffusion Processes and Their Sample Paths. Grundlehren der Mathematischen Wissenschaften, vol. 125. Springer, Berlin (1974)
Karatzas, I., Shreve, S.: Brownian Motion and Stochastic Calculus. Springer, New York (1991)
Karatzas, I., Lehoczky, J.P., Sethi, S.S., Shreve, S.: Explicit solution of a general consumption/investment problem. Math. Oper. Res. 11, 261–294 (1986)
Lundberg, F.: I. Approximerad Framställning av Sannolikhetsfunktionen, II. Återförsäkring av Kollektivrisker. Almqvist and Wiksell, Uppsala (1903)
Paulsen, J.: Optimal dividend payouts for diffusions with solvency constraints. Finance Stoch. 7, 457–473 (2003)
Paulsen, J., Gjessing, H.K.: Ruin theory with stochastic return on investments. Adv. Appl. Probab. 29, 965–985 (1997)
Pham, H.: Continuous-Time Stochastic Control and Optimization with Financial Applications. Springer, Berlin (2009)
Schmidli, H.: Stochastic Control in Insurance. Springer, Berlin (2008)
Shreve, S.E., Lehoczky, J.P., Gaver, D.P.: Optimal consumption for general diffusions with absorbing and reflecting barriers. SIAM J. Control Optim. 22, 55–75 (1984)
Taksar, M.: Optimal risk and dividend distribution control models for an insurance company. Math. Methods Oper. Res. 51, 1–42 (2000)
van Moerbeke, P.: On optimal stopping and free boundary problems. Arch. Ration. Mech. Anal. 60, 101–148 (1976)
Williams, R.: Asymptotic variance parameters for the boundary local times of reflected Brownian motion on a compact interval. J. Appl. Probab. 29, 996–1002 (1992)
Acknowledgements
I want to thank an anonymous referee for carefully reading the manuscript and for numerous suggestions, leading to an improvement of the paper.
Author information
Authors and Affiliations
Corresponding author
Additional information
Support by the “Austrian Science Foundation” (Fonds zur Förderung der wissenschaftlichen Forschung), Project nr. P26487, is gratefully acknowledged.
Appendix
Appendix
1.1 9.1 Auxiliary results for Sect. 3
Proof of Lemma 3.1
In the following proof, we use the notation \(X^{C}\) for the underlying process if the starting point is clear from the context.
(a) Let us assume first \(0 \leq t < T,0<x\). With \(u(t):= \max(t,\hat{t})\), define the stopping time
We provide now the definition of the new strategy \(\hat{C}\) as
Hence, after time \(\hat{t}\) we run our endowment without paying out anything, until the first moment our trajectory is above or equal to the level of the old trajectory. Then we jump down to this old trajectory and use the given strategy \(C\).
Let \(B^{(\hat{t}, \hat{x})}:= \{0 \leq-X^{\hat{C}}_{u(\hat{t})+} + X^{C}_{u(\hat{t})+} \leq|\Delta t|^{\frac{1}{3}}+\vert\Delta x \vert\} \cap\left\{ \tau =T\right\}\). By construction,
We distinguish 3 cases.
Case 1: \(\omega\in B_{1}:= \{ \omega\in B^{(\hat{t}, \hat{x})} | C_{T}-C_{u+}>0\}\). Let now \(|\Delta t|^{\frac{1}{3}}+\vert\Delta x \vert\leq \Delta(\omega)\). Because \(C\) is left-continuous, there exists a \(\delta(\omega)>0\) with \(C_{T-s}-C_{u+} \geq \frac{C_{T}-C_{u+}}{2}\) for \(0 \leq s \leq\delta\). Hence, we arrive at
Furthermore, since \(\tau(\omega)=T\) for \(\omega \in B\), we know that for all \(\epsilon>0\),
Choosing now \(\epsilon\) as \(\delta(\omega)/2\) gives
which implies for \(\Delta<\nu\) that
Hence,
Therefore (9.2) gives \(X^{\hat{C}}_{\kappa+}>0\) for \(\Delta\) small enough. Finally, we arrive at
for \((\Delta t, \Delta x)\) small enough on the set \(B_{1}\).
Case 2: \(\omega\in B_{2}:= \{ \omega\in B | C_{T}-C_{u+}=0, X^{C}_{u+}+\mu(T-u)+ \sigma(W_{T}-W_{u})>0\}\). In this case, we choose \(\hat{C}_{s}=\hat{C}_{u+}\) for \(s>u\) and get \(X^{\hat{C}}_{T}\geq X^{C}_{T}-\Delta\), so that we have \(\hat{\tau}=T\) for all \(\omega\) in \(B_{2}\), for \(\Delta\) small enough.
Case 3: \(\omega\in B_{3}:= \{ \omega\in B | C_{T}-C_{u+}=0, X^{C}_{u+}+\mu(T-u)+ \sigma(W_{T}-W_{u})=0\}\). We first note that \(\mathbf{P}[B_{3}]=0\), and hence (9.1) gives
So by Cases 1 and 2, we conclude that
Finally, since \(\mathbf{P}[\hat{\tau}=T] \leq \mathbf{P}[\tau=T]\) by construction, we get
The cases \((t=T,x>0)\) and \((0\leq t < T,x=0)\) are trivial, hence part a is proved.
Let us finally remark that for \((t=T,x=0)\), part a is wrong since \(\mathbf{P}[\tau=T]=1\). But if the point \((\hat{t},\hat{x})\) approaches \((T,0)\) in a flat enough way, say linearly, e.g., as \(\hat{x}=T-\hat{t}\), then \(\mathbf{P}[\hat{\tau}< T]\) is clearly not \(o(1)\) for \(\hat{t} \to T\).
(b) We first consider the case \(0< t< T,0< x\). Take \(\hat{C},u\) from part a and let \(G:= \{X^{C}_{u+}-X^{\hat{C}}_{u+} \leq|\Delta t|^{\frac{1}{3}}+\vert\Delta x \vert\}\). Hence
for \((\Delta t,\Delta x) \to0\), uniformly in \((t,x)\). We split the probability space \(\varOmega\) and arrive at
The second term tends to zero by (9.3), and for the first term, it is sufficient that \(\Vert\nu_{1}-\nu_{2}\Vert_{L^{2}} \to0\) for \(\Delta x \searrow0\), uniformly in \(x\), where we use
But the above \(L^{2}\)-convergence is implied by Lemma 9.1, so that we get part b for \(0< t< T,0< x\). The case \(t=T\) is trivial, and for \(t< T\), \(x=0\), one can again use Lemma 9.1. □
Lemma 9.1
Let \(X_{t}:= \Delta x+\mu t + \sigma W_{t}\), with small positive \(\Delta x\), and let further \(\tau := \inf\{t>0 | X_{t}=0\}\wedge T\). Then we have \(\Vert\tau\Vert_{L^{2}} \leq\mathrm{const.}\Delta x\), where the constant depends only on the model parameters.
Proof
By (3.5.12) in [19], \(\tau\) has the defective density
and a point mass \(\mathbf{P}[\tau=T] = \int_{T}^{\infty}f(s)\, ds\). Hence, we get
In the following, \(D\) denotes a generic constant which may vary from place to place, and which depends only on the model parameters. So we get
which concludes the proof. □
1.2 9.2 Auxiliary results for Sect. 4
Proof of Lemma 4.2
We define our partition as
for \(i=2,3,\dots,M\), \(j \in\mathbf{N}_{0}\), and \(Q^{1,j}:= [0,\frac{T}{M}] \times[\frac{j}{M}, \frac{j+1}{M})\) for \(j \in\mathbf{N}_{0}\). Let \(C^{(i,j)}\) be an admissible strategy for \(X^{t_{i},x_{j}}\), \((t_{i},x_{j}):= (\frac{T(i-1)}{M},\frac{j}{M})\), \((i,j) \in {\mathcal{I}}\). By our definition of the half-open rectangles \(Q^{i,j}\), this strategy is also admissible for all \((t,x) \in Q^{i,j}\), for all \((i,j)\in{\mathcal{I}}\). Similarly as in the proof of Proposition 3.2 (cf. (3.4) there), one finds that for all \(\epsilon>0\), there exists an \(M>0\) large enough such that
for all \((t,x) \in Q^{i,j}\) and all \((i,j)\in{\mathcal{I}}\). Since \(V_{n}(t,x)-e^{-\beta t}x\) is by Proposition 3.2 uniformly continuous, we have also
for all \((t,x) \in Q^{i,j}\) and all \((i,j)\in{\mathcal{I}}\). We now choose \(C^{(i,j)}\) such that one has
By (9.4), (9.6) and (9.5), one concludes that
for all \((t,x) \in Q^{i,j}\) and all \((i,j)\in{\mathcal{I}}\). □
Proof of Lemma 4.5
Instead of the estimate (9.4) in the proof of Lemma 4.2, we claim that we have now
for all \((t,x) \in Q^{i,j}\), all \((i,j)\in{\mathcal{I}}\) and \(M\) large enough. Indeed, to get (9.4), we have used (3.4), and the point where the function \(f_{n}\) entered in this estimate was via an estimate of \(|\mathbf{E}[ f_{n}(\tau)-f_{n}(\hat{\tau})]|\). This is now replaced by \(\mathbf{E}[f(\hat{\tau})-f(\tau)] \geq0\), by the definition of our partition and the points \((t_{i},x_{j})\). Moreover, we can replace, using the uniform upper semicontinuity of \(W\) from Proposition 3.3(a), the estimate (9.5) by
for all \((t,x) \in Q^{i,j}\), all \((i,j)\in{\mathcal{I}}\) and \(M\) large enough. Finally, we can take (9.6) without alterations, and this gives, together with (9.7) and (9.8), our assertion. □
1.3 9.3 Auxiliary results for Sect. 5
Proof of Proposition 5.2
Since we know by Proposition 3.3 that the value function \(V\) is upper semicontinuous, we have \(V^{*}=V\). Now let \((\bar{t},\bar{x}) \in Q\) be a maximizer of \(V-\phi\) with
where \(\phi\in C^{1,2}(Q)\). We have to show that
Let \((t_{m},x_{m})\to(\bar{t},\bar{x})\) for \(m\to\infty\), with \(V(t_{m},x_{m}) \to V(\bar{t},\bar{x})\). Moreover, let
for \(m \to\infty\). We argue by contradiction. So assume that we have
By continuity, this implies the existence of \(\epsilon_{0},\eta>0\) with
on a sphere \(K((\bar{t},\bar{x});\eta) \subset Q\) with radius \(\eta\).
Let \(\theta_{m}\) be the exit time of \(X^{(t_{m},x_{m})}\) from \(K_{m}:= K((t_{m},x_{m});\eta/2)\). We have \(K_{m} \subset K((\bar{t},\bar{x});\eta)\), where the inclusion holds for \(m\) large enough. For \(h>0\), \(\theta_{m}+h\) is a stopping time, and we apply our DPP in Proposition 4.4(b) for this stopping time. Hence for all \(\delta>0\), there exists an admissible strategy \(C^{m}\) such that
Let now \(\lambda_{m}:= [X^{(t_{m},x_{m})}_{(\theta _{m})+},X^{(t_{m},x_{m})}_{(\theta_{m})}] \cap\partial K_{m}\), where \(\partial K_{m} \) denotes the boundary of \(K_{m}\). Using this definition, we have
Hence,
For the first inequality, we have used the upper semicontinuity of \(V\) from Proposition 3.3. The argument for the second one is exactly as in [12, Lemma VIII.3.1] (moreover we use \(\phi\geq V\) to estimate the third term). Finally, for the third inequality we use
with \(\hat{X}_{\theta_{m}} \in[\lambda_{m},X_{\theta_{m}}]\), employing (9.9). For notational convenience, we now omit the upper index \((t_{m},x_{m})\) of the state process. Itô’s formula implies
Here we have used (9.9), the notation \(C^{m,c}\) for the continuous part of \(C^{m}\), \(J_{X}\) for the jump times of \(X\), and \(\tilde{X}_{s} \in[X_{s+},X_{s}]\). Moreover, \(\mathcal{S}\) denotes a stochastic integral with respect to Brownian motion with an integrand integrable enough to guarantee that this term vanishes in the expectation. Hence, we do not specify this term explicitly. Combining (9.11) and (9.12) gives
where the first integral includes a possible jump at \(\theta_{m}\), too. Applying \(\limsup_{h\to0}\) in (9.10) gives
Using (9.13) and taking the expected value, we end up with
Let \(\hat{\theta}_{m}:= \inf\{s> t_{m} | X_{s} \notin\hat{K}_{m}:= K((t_{m},x_{m});\eta/4) \subset K_{m} \}\) and define the set \(A_{m}:= \{ \theta_{m} =\hat{\theta}_{m}\}\), which means that if we leave \(\hat{K}_{m}\) on \(A_{m}\), then we leave also \(K_{m}\). Obviously, we have \(|X_{\theta_{m}}-\lambda_{m} | \geq\eta/4\) on the set \(A_{m}\). We distinguish now several cases.
Case 1: \(\mathbf{P}[A_{m}] \geq1/2\). In this case, we can give for the right-hand side of (9.14) the upper estimate \(-\frac{\epsilon_{0} \eta}{8}-\gamma_{m}\).
Case 2: \(\mathbf{P}[A_{m}^{c}] \geq1/2\). Set now \(B_{m}:= \{ \int_{t_{m}}^{\theta_{m}} dC^{m}_{s} \geq\eta/10\}\) and distinguish two subcases:
Case 2.1: \(\mathbf{P}[A_{m}^{c} \cap B_{m}] \geq1/4\). This is the case where we leave only the smaller sphere and not the larger one, and where the integral term of the right-hand side of (9.14) gives a significant contribution. We find \(-\frac{\epsilon_{0} \eta}{40}-\gamma_{m}\) as an upper estimate for the right-hand side of (9.14).
Case 2.2: \(\mathbf{P}[A_{m}^{c} \cap B_{m}^{c}] \geq1/4\). We define
and observe that we have \(\tilde{\theta}_{m} \leq\theta_{m}\) on \(B^{c}_{m}\). Moreover, one has \(\mathbf{E} [\tilde{\theta}_{m}-t_{m}] \geq D\) for some generic constant \(D\) depending on \(\eta,\mu\) and \(\sigma\). Hence \(\mathbf{E}[\theta_{m}-t_{m}] \geq D\), and for the upper bound for the right-hand side of (9.14), we find \(-\epsilon_{0} D-\gamma_{m}\).
Therefore, we have in all cases
for all \(m\). Hence, by (9.14), \(-\delta\leq-\epsilon_{0} D-\gamma_{m} \) for all \(m\), which is a contradiction for \(m\) large and \(\delta\) small enough. □
Proof of Proposition 5.4
We first transform the HJB equation to an equivalent one, where we use \(V=e^{\lambda_{1} t+\lambda_{2} x}\tilde{V}\) for some real \(\lambda_{i}\). This yields (we replace now for convenience \(\tilde{V}\) again by \(V\))
with \(\tilde{\mathcal{L}}V:= \nu_{1} V+V_{t} + \tilde{\mu}V_{x} +\frac {\sigma ^{2}}{2}V_{xx}\), \(\nu_{1}:= \lambda_{1} +\lambda_{2}\mu+\frac{\sigma^{2}}{2}\lambda_{2}^{2}\), \(\tilde{\mu}:= \mu+ \sigma^{2}\lambda_{2}\), \(\nu_{2}:= \beta+\lambda_{1}\), \(\nu_{3}:= \lambda_{2}\). In the sequel, we choose \(\lambda_{2}>0\) arbitrarily large and \(\lambda_{1} := -\lambda_{2}^{3}\). All together, we have the choice of parameters
Our proof goes by contradiction. So assume that
for some open bounded set \(\mathcal{O}\) in \(\mathbf{R}_{+}\). Here we have used the method of penalized supersolution (\(u,v\) increase only linearly) for the last equality above (see [24, Th. 4.4.3, Step 2 and Th. 4.4.4, Step 1]).
Let now \(w:= 10K e^{-\nu_{3} x -\nu_{3} (T-t)}\) with \(\nu_{3}>0\) from above, and let \(U\) be a small open rectangle at \((T,0)\), i.e., \(U:= \{ (t,x) | T- \frac{1}{\nu_{3}}< t <T, 0<x < \frac{1}{\nu_{3}} \}\). We want to show that
holds on \(\bar{Q} \setminus U\). To do this, we first show that \(v+w\) is a supersolution, too. Indeed, we get \(-\tilde{\mathcal{L}}w =(-\nu_{1}-\nu_{3} +\tilde{\mu}\nu_{3} -\frac {\sigma ^{2}}{2}\nu_{3}^{2})w >0\). Moreover, one has \(w_{x} +\nu_{3}w=0\), and both together imply
hence the asserted supersolution property. Indeed, on the sets \(\{(t,0) | 0 \leq t \leq T-\frac{1}{\nu_{3}} \}\) (resp., \(\{ (T,x) | x \geq\frac{1}{\nu_{3}} \}\)) we have \(u-v-w =-w <0\). Moreover, on the rest of the boundary one gets \(K-\frac{10}{e^{2}}K\) as an upper bound for \(u-v-w\).
We now show (9.16) by contradiction. So assume
where \(A^{o}\) denotes the interior of \(A\) and \(L\) is the compact set \(L:= [0,T] \times[0,z]\) for some \(z>0\). (Here we have again used the method of penalized supersolution.) We now apply the well-known dedoubling variable technique and define
where \(\phi_{\epsilon}(t,s,x,y):= \frac{(t-s)^{2}+(x-y)^{2}}{2\epsilon}\). As \(\varPhi_{\epsilon}\) is continuous on \(((\bar{Q} \setminus U)\cap L)^{2}\), let us denote its maximum by \(M_{\epsilon}\) and its maximizer by \((t_{\epsilon},s_{\epsilon},x_{\epsilon},y_{\epsilon})\). By going to subsequences, we can assume that \((t_{\epsilon},s_{\epsilon},x_{\epsilon},y_{\epsilon})\to(\bar{t},\bar {s},\bar {x},\bar{y})\) holds for \(\epsilon\to0\). As in [24, Theorem 4.4.4], we infer that
with \((\bar{t},\bar{x}) \in((\bar{Q} \setminus U)\cap L)^{o}\), because of (9.17). Note that the form of the HJB equation does not enter into this argument.
As in [24], we introduce now the concept of a super(sub)jet \(\mathcal{P}^{2,+}u(t_{0},x_{0})\) (resp., \(\mathcal{P}^{2,-}v(t_{0},x_{0})\)) of an upper (lower) semicontinuous function \(u\) (resp., \(v\)) at a point \((t_{0},x_{0}) \in[0,T) \times\mathbf{R}\) as the set of elements \((q_{0},p_{0},K_{0})\) in \(\mathbf{R}^{3}\) satisfying
(or
respectively). For technical reasons, one also needs the limiting super- and subjets; we define \(\overline{\mathcal{P}}^{2,+}u(t,x)\) as the set of elements \((q,p,K) \in\mathbf{R}^{3}\) for which there exists a sequence \((t_{\epsilon},x_{\epsilon},q_{\epsilon},p_{\epsilon},K_{\epsilon})\) in \([0,T)\times\mathbf{R} \times\mathcal{P}^{2,+}u(t_{\epsilon},x_{\epsilon})\) satisfying
The set \(\overline{\mathcal{P}}^{2,-}v(t,x)\) is defined similarly.
Now, Ishii’s lemma in the form of [24], Remark 4.4.9, gives the existence of real numbers \(K,N\) such that
with \(c^{2}K-d^{2} N \leq\frac{3}{\epsilon}(c-d)^{2}\) for any real numbers \(c,d\). Hence setting \(c=d\) gives
The superjet (resp., subjet) characterization of supersolutions (resp., subsolutions) provides
Now (9.19) is equivalent to
and the second equation of (9.20) implies
We distinguish two cases:
Case 1: \(\nu_{1} u(t_{\epsilon},x_{\epsilon})+\frac{1}{\epsilon }(t_{\epsilon}-s_{\epsilon})+ \frac{\tilde{\mu}}{\epsilon}(x_{\epsilon}-y_{\epsilon})+\frac{\sigma^{2}}{2}K \geq0\) for some subsequence, again denoted by \(\epsilon\) , tending to zero. The assumption, as well as the first equation of (9.21), implies the inequality \(\nu_{1}(u(t_{\epsilon},x_{\epsilon})-(v+w)(s_{\epsilon},y_{\epsilon}))+\frac {\sigma^{2}}{2}(K-N)\geq0\), from which one gets for \(\epsilon\to0\) that \(\nu_{1} M+\frac{\sigma^{2}}{2}(K-N)\geq0\). By (9.15), (9.17) and (9.18), this is a contradiction.
Case 2: \(e^{-\nu_{2} t_{\epsilon}-\nu_{3} x_{\epsilon}}-\frac {1}{\epsilon }(x_{\epsilon}-y_{\epsilon})-\nu_{3} u(t_{\epsilon},x_{\epsilon}) \geq0\) for some subsequence, again denoted by \(\epsilon\) , tending to zero. The assumption, as well as the second equation of (9.21), implies the inequality \(-\nu_{3}(u(t_{\epsilon},x_{\epsilon})-(v+w)(s_{\epsilon},y_{\epsilon}))+e^{-\nu _{2} t_{\epsilon}-\nu_{3} x_{\epsilon}}-e^{-\nu_{2} s_{\epsilon}-\nu_{3} y_{\epsilon}} \geq0\) and in the limit \(-\nu_{3}M \geq0\), which is by (9.15) and (9.17) a contradiction.
Hence, our assumption of a positive supremum of \(u-v-w\) on \(\bar{Q} \setminus U\) is wrong. Therefore, we have \(u-v-w \leq0\) on \(\bar{Q} \setminus U\), and also on \(\bar{Q} \setminus V\) if we define \(V:= \{ (t,x) | T- \frac{1}{\sqrt{\nu_{3}}}< t <T, 0<x < \frac{1}{\sqrt{\nu_{3}}} \}\) (remember that \(\nu_{3}\) is large). Since \(w\) can be made arbitrarily small on \(\bar{Q} \setminus V\) and as the same holds for \(V\), we conclude
Because we have assumed \(u=v\) on \(( \{ (t,0) | t \in[0,T] \} \cup \{ (T,x) | x \in\mathbf{R}_{+} \} )\), we get finally \(u \leq v\) on \(\bar{Q}\). □
1.4 9.4 Auxiliary results for Sect. 6
Proof of Lemma 6.2
In an analogous manner as in [15, Lemma 3.1], one can show that the following construction of \(H(x)\) is successful. We set \(H(x):= \underline{H}(x) +\overline{H}(x)\), and give first the definitions of \(\underline{H} (x)\) and \(\overline{H}(x)\) for \(x \in [0,\epsilon]\). They are
where \(\alpha\) depends on the model parameters and has the asymptotic (for \(\epsilon\to0\)) behavior \(\alpha\sim\frac{120 (2\hat{\mu}+\hat{\mu}^{2} \bar{\gamma })}{\epsilon^{5}}\). For \(\overline{H}(x)\) one can take
with
and
For negative \(x\), we use the definition \(H(-x)=-H(x)e^{2\hat{\mu}x}\), \(x>0\), which stems from the fact that the transformed terminal gain function \(h(x)=H(x)e^{\hat{\mu}x -\lambda T}\), with \(\lambda:= \beta+\frac{\mu^{2}}{2\sigma^{2}}\), is assumed to be an odd function. □
We describe now the sequence of transformations of BVPs needed in Sect. 6. Our first lemma transforms the original problem into a new one, where we replace the former region \(I\) by \(I\cup\hat{I}\), with \(\hat{I}\) the mirror image of \(I\) with respect to the line \(x=0\). Moreover, it shows that the correct inequalities assumed in region I (resp., II) in Theorem 6.5 (see Definition 6.4) follow if we assume \(C^{4}\)-regularity (except at the point \((T,0)\)) up to and including the boundary.
Lemma 9.2
A function \(V\) and a boundary function \(b\) solving the system
solve also (6.1) of Theorem 6.5.
Proof
What we have actually to show is that we have the respective inequalities
and
The proof of (9.23) is completely analogous to the proof of Lemma 3.2 in [15], with the only alteration that we have now \(d(T) =H_{0}+\bar{\gamma}\), which implies
We proceed with inequality (9.24) and start with the calculation of the left-hand boundary condition of \(V\), i.e., \(V(t,-b(t))\). Using the right-hand boundary conditions \(V(t,b(t))=e^{-\beta t}(b(t)+d(t))\), \(V_{x}(t,b(t))=e^{-\beta t}\), \(V_{xx}(t,b(t))=0\), the transformation \(v(t,x)=V(t,x)e^{\hat{\mu}x -\frac{\mu ^{2}}{2\sigma^{2}}t}\) and the fact that \(v\) is an odd function in \(x\) give after some elementary algebra
For the transformed function \(v\), one gets
Moreover, \(v\) fulfils the boundary value problem
The function \(h\) has a discontinuity at \(x=0\), and using the Standing Assumption on \(H\), one finds \(h(0\pm)=\pm\tilde{\gamma}\). Subtracting out this jump, we define \(\tilde{h}\) as \(\tilde{h}(x):= h(x)-\tilde{\gamma} \mathrm{sgn}(x)\). Our definition of \(H\) in the Standing Assumption is chosen in such a way that we have \(h(-x)=-h(x)\) for \(x \in[-\epsilon,\epsilon]\), \(h^{\prime\prime}(0+)=0\) and \(h^{\prime \prime \prime\prime}(0+)=0\). This implies that \(\tilde{h}\in C^{4}([-\epsilon,\epsilon])\).
We now decompose the function \(v\) into a discontinuous part fulfilling a Cauchy problem and a regular part fulfilling a BVP, i.e., \(v =: v^{(1)}+v^{(2)}\) with
respectively,
Note that the function on the right-hand side of the second equation in (9.26) is in \(C^{1}([0,T])\), since away from the point \((T,0)\), the function \(v^{(1)}\) is smooth, and \(g \in C^{1}([0,T])\) by the assumptions of our lemma. Since the boundary data and the final data are obviously compatible, we get by Gevrey’s result, see [15, Lemma 3.5], that \(v^{(2)}_{x} \in C(\overline{G_{1}})\). Hence, \(v^{(2)}_{x}\) fulfils the BVP
Again the function on the right-hand side of the second equation in (9.27) is in \(C^{1}([0,T])\), and since the boundary and final data are compatible, we conclude by Gevrey’s result that \(v^{(2)}_{xx} \in C(\overline{G_{1}})\).
We now approximate the function \(v^{(1)}\) by smooth functions \(v^{(1,\delta)}\) which fulfil
where \(\phi^{(\delta)}\in C^{\infty}(\mathbf{R})\) fulfils \(\phi ^{(\delta)}(x) \to\tilde{\gamma}\mathrm{sgn}(x)\) pointwise for \(\delta\to0\), except at \(x=0\). Moreover, we impose that \(\phi^{(\delta)}\) are odd functions, fulfil \(\phi^{(\delta)'}(x)\geq0\), \(\phi^{(\delta)''}(x)\leq0\) for \(x\geq0\), and \(\phi^{(\delta)}(x):= \tilde{\gamma}\mathrm{sgn}(x)\) for \(|x|\geq \delta/2\). The solution is well known and given by
Obviously, one has
for \(\delta\to0\), pointwise in \(\overline{G_{1}} \setminus\{(T,0)\}\) and uniformly on each compact \(K\) with \(K \subset\overline{G_{1}} \setminus\{(T,0)\}\). Let \(v^{(\delta)}:= v^{(1,\delta)}+v^{(2)}\), which gives for \(v^{(\delta)}\) the system
Because of (9.29), we have
for \(\delta\to0\), pointwise in \(\overline{G_{1}} \setminus\{(T,0)\}\) and uniformly on each compact \(K\) with \(K \subset\overline{G_{1}} \setminus\{(T,0)\}\), or (with \(V^{(\delta)}:= v^{(\delta)}e^{-\hat{\mu}x +\frac{\mu ^{2}}{\sigma^{2}}t}\))
for \(\delta\to0\), pointwise in \(\overline{G_{1}} \setminus\{(T,0)\}\) and uniformly on each compact \(K\) with \(K \subset\overline{G_{1}} \setminus\{(T,0)\}\). Since \(v^{(1,\delta)}\in C^{\infty}\), we have \(v^{(\delta)}_{xx} \in C(\overline{G_{1}})\) and \(v^{(\delta)}_{x}\) fulfils
Furthermore, we have \(\tilde{h}'(x)=h'(x)\) for \(x \neq0\), and \(\tilde{h} '(0)=h'(0+)=h'(0-)\). Our Standing Assumption on \(H\), the transformation formula between \(H\) and \(h\), as well as the oddness of \(h\), implies \(\tilde{h}'(x) >0\) for \(x \in[-\epsilon,\epsilon]\), which in turn gives
Since we know that the convergence \(k^{(\delta)}(t) \to k(t) >0\) for \(\delta\to 0\) is uniform on \([0,T]\) by (9.31), we conclude that \(k^{(\delta)}(t)>0\) on \([0,T]\) for \(\delta\) small enough. The maximum principle implies by (9.33) and (9.34) that \(v^{(\delta)}_{x}>0\) on \(G_{1}\) for small enough \(\delta\), especially \(v^{(\delta)}_{x}(t,0) >0\) for \(0\leq t \leq T\). Hence, transforming to \(V^{(\delta)}\) gives by the oddness of \(v^{(\delta)}\) that
for \(\delta\) small enough. Defining \(g^{(\delta)}\) and \(m^{(\delta )}\) analogously to \(k^{(\delta)} \) and exploiting (9.29) gives \(g^{(\delta)}(t) \to g(t)\), \(k^{(\delta)}(t) \to k(t)\), \(m^{(\delta )}(t) \to m(t)\), uniformly on \([0,T]\), hence \(V^{(\delta)}_{xx}(t,b(t)) \to V_{xx}(t,b(t))\), uniformly on \([0,T]\), i.e., there exists a positive function \(\nu (\delta ) \to0\) with \(\delta\to0\) such that
for \(t \in[0,T]\).
For the terminal value, we claim that
By our transformation law, this is equivalent to \((v^{(\delta )}_{xx}-2\hat{\mu}v^{(\delta)} _{x}+\hat{\mu}^{2}v^{(\delta)})(T,x) \leq0\) or, if we use our terminal condition (9.30), equivalent to
Now, the sum of the second, fourth and \(\hat{\mu}^{2} h\) from the sixth term are less than or equal to zero by our Standing Assumption (viii). Hence it remains to show that we have \(\phi^{(\delta )''}(x)-2\hat{\mu}\phi^{(\delta)'} (x)+\hat{\mu}^{2}(\phi^{(\delta)}(x)-\tilde{\gamma}) \leq0\), which is true by our assumptions on the function \(\phi^{(\delta)}\) formulated after (9.28). Therefore, (9.37) is true.
Since \(V^{(\delta)}_{xx} \in C(\overline{G_{1}})\), the maximum principle, together with (9.35)–(9.37), implies \(V^{(\delta)}_{xx}(t,x) \leq\nu(\delta)\) on \(\overline{G_{1}}\). Employing (9.32) and going to the limit \(\delta\to0\) gives \(V_{xx}(t,x) \leq0\) on \(\overline{G_{1}} \setminus\{(T,0)\}\), which by \(V_{x}(t,b(t))=e^{-\beta t}\) finishes the proof. □
Since we want to apply the Green’s function method for the backward heat equation, we remove the drift of the problem by a simple exponential transformation, and we formulate the problem now for a new dependent variable \(v\).
Lemma 9.3
(i) If a function \(v= v(t,x)\) and a boundary function \(b\) solve the system
then \(V\) defined by \(V(t,x):= e^{-\hat{\mu}x +\frac{\mu^{2}}{2\sigma ^{2}}t}v(t,x)\) and \(b\) solve system (9.22).
(ii) The transformed terminal condition \(h\) fulfils
Proof
The proof consists of simple calculations. Note that the asserted values of \(h\) and its derivatives at \(\epsilon\) are the same as in [15], with the only difference that we now have a different value for \(d(T)\). □
Remark 9.4
Note that in comparison to system (9.38) in [15, Lemma 3.3], we have skipped the fifth condition in (9.38) there. It was used there to prove the fifth property in (9.22) which we do not have any more in our new system (9.22).
In the following lemma, we define a function we use repeatedly in the sequel.
Lemma 9.5
Let \(F\) be the solution of the Cauchy problem
Then \(F\) has the representation
and \(\frac{\partial^{k+\ell}}{\partial t^{k} \partial x^{\ell}}F(t,x)\) is continuous on \(\overline{G_{1}} \setminus\{(T,0)\}\) (resp., uniformly continuous on each compact set \(K \subset\overline{G_{1}} \setminus\{(T,0)\}\)), for \(k,\,\ell\in\mathbf{N}_{0}\). Moreover,
on \(\overline{G_{1}} \setminus\{(T,0)\}\). Finally, we note that
is the fundamental solution of our backward heat equation.
By subtracting the function \(F\) from \(v\), we get a regularized version of \(v\), namely \(\tilde{v}(t,x):= v(t,x)-F(t,x)\). In the sequel, we consider \(\tilde{v}\) and derive a sequence of BVPs for it. The proof of the following lemma follows obviously from Lemma 9.3.
Lemma 9.6
(i) If a function \(\tilde{v}\) and a boundary function \(b\) solve the system
then \(v\), defined by \(v(t,x):= \tilde{v}(t,x)+F(t,x)\) and \(b\) solve (9.38).
(ii) The transformed terminal condition \(\tilde{h}\) fulfils
Our next result shows that if we choose proper boundary conditions for \(\tilde{v}\), then \(\tilde{v}\) has automatically the smoothness asserted in Lemma 9.6.
Lemma 9.7
If a function \(\tilde{v}\) and a boundary function \(b\) solve the system
then \(\tilde{v}\) and \(b\) solve (9.39), especially the smoothness conditions formulated there. Moreover, we have the explicit representations
Proof
The proof is analogous to the proof of Lemma 3.4 in [15]. We only have to modify the boundary conditions, using our function \(F(t,x)\). □
We now start to derive similar systems for the derivatives of \(\tilde{v}\).
Lemma 9.8
If a function \(\tilde{u}\) and a boundary function \(b\) solve the system
then \(\tilde{v}\) defined by \(\tilde{v}(t,x):= -\tilde{M}(t)+\int_{-b(t)}^{x} \tilde{u}(t,z) \, dz\), where we define the function \(\tilde{M}(t):= e^{\hat{\mu}b(t)-\lambda t}\left (b(t)+d(t)\right)-F(t,b(t))\), and \(b\) solve (9.40) of Lemma 9.7.
Proof
The proof is analogous to the proof of Lemma 3.7 in [15]. We just have to use the relation \(\hat{\mathcal{L}}F(t,b(t))=0\) at one instance. □
Proof of Lemma 6.6
Again, the proof is analogous to the proof of Lemma 3.8 in [15]. This time we just have to use the relations \(\hat{\mathcal{L}}F_{x}(t,b(t))=0\) and \(F_{xx}(T,b(T))=0\). □
Proof of Lemma 6.10, part a
The proof is identical to the proof of Lemma 5.1 in [15], since the only point where the form of the terminal function \(H\) is used is inequality (46) there, where one has to estimate \(\mathbf{E} [H^{2}(X_{T})]\). So exchanging the old terminal function by the new one gives just a different constant. □
Proof of Lemma 6.10, part b
Let \((r_{n})\) be a sequence such that \(r_{n} \downarrow T-\eta\) and define \(\tilde{V}(t,x):= e^{-\hat{\mu}x +\frac{\mu^{2}}{2\sigma^{2}}t} \tilde{v}(t,x)\). Then one has \(\mathcal{L}\tilde{V}_{xt}=0\) in \(\overline{G}_{1}\) and after some elementary transformations,
where \(F^{(i)}\) denotes the \(i\)th \(x\)-derivative of \(F\) and \(c_{i}\) some generic, model-dependent constants which may vary from place to place, and where \(c_{0}\) in the second equation of (9.43) fulfils \(c_{0} <0\). Define now \(\hat{V}:= \tilde{V}_{xt}-\rho\) for some positive constant \(\rho\) to be chosen later. This gives for \(\hat{V}\) the system
Since we know by Lemma 6.10(a) that \(b(t) \geq C\) for some generic (\(\epsilon\)-dependent) constant \(C\), one concludes that \(|\sum_{i=0}^{3} c_{i}F^{(i)}(t,b(t))| \leq C\). Note also that for \(b(t)\) large, \(|F^{(i)}(t,b(t))|\), \(i=1,2,3\), becomes small. Hence we can choose a model-dependent positive constant \(\rho\) such that we have by the maximum principle \(\hat{V} <0\) on \(\overline{G}_{1}\). Introducing \(w:= \frac{\hat {V}}{-\beta e^{-\beta t}}\) gives \(w>0\) on \(\overline{G}_{1}\), on the one hand, and the system
with \(c_{0} >0\) in the third equation of (9.44), on the other hand. Since we know that \(w\) is positive, it fulfils the maximum principle.
Moreover, \(M \in C^{1}((T-\eta,T])\) and \(M^{*}(t):= \sup_{s \in[t,T]} M(s)\) is absolutely continuous with respect to Lebesgue measure on \([r_{n},T]\) for all \(n\). Now let us define the set \(I:= \{ t \in [r_{n},T] | M^{*}(t)=M(t), M(t) \geq\varGamma\}\) for some positive constant \(\varGamma> \Vert\frac{K(x)}{-\beta e^{-\beta T}}\Vert_{C([-\epsilon,\epsilon])}\). We also choose \(\varGamma\) large enough such that \(t \in I\) implies \(w(t,-b(t))=M^{*}(t)>w(s,b(s))\), for all \(s \in[t,T]\). By the maximum principle, this implies \(w_{x}(t,-b(t)) \leq0\), or \(\hat {V}_{x}(t,-b(t)+) \geq0\), or \(\tilde{V}_{xxt}(t,-b(t)+)\geq0\), for \(t \in I\). Using the boundary condition for \(\tilde{V}_{xxt}\), one arrives at
or
for a generic positive (\(\epsilon\)-dependent) constant \(C\).
We set now \(C^{+}:= 1+\frac{2\mu^{2}}{\beta\sigma^{2}}\), view the functions \(M\) and \(b\) as functions of the backward time \(\nu := T-t\) and calculate the derivative of \(M\) for \(\nu\) with \(t(\nu) \in I\). Note that since with \(\varGamma\) large, also \(M(\nu)\) becomes large, we can infer that \(b(\nu)\) becomes large, hence \(|F^{(i)}(t,b(\nu))|\), \(i=1,2,3\), become small, whereas \(F^{(0)}(t,b(\nu ))\) is approximately \(\tilde{\gamma}>0\). We get
By the explicit formula for \(M\) in the third equation of (9.44), the fact that \(M(\nu)\) and \(b(\nu)\) are large, \(c_{0} >0\), \(F^{(0)}(t,b(\nu))>0\) and \(| F^{(i)}(t,b(\nu))|\), \(i=1,2,3\), are small, we have \(M(\nu) > \frac{C^{+} e^{2 \hat{\mu}b(\nu)}}{2}> -e^{ \hat {\mu}b(\nu )+\lambda t} \sum_{i=0}^{3} c_{i}F_{t}^{(i+1)}(t,b(\nu))\). Hence we can estimate the right-hand side above by \(b'(\nu) ( \dots )+M(\nu)\). Moreover, we have
which gives us in the case where \(b'(\nu) \geq0\) the upper estimate
for some generic positive constant \(C\). If \(b'(\nu) <0\), we get the upper estimate \(M(\nu)\). Summing up, we have, for \(t(\nu) \in I\),
This is obviously sufficient for \(M(t) \leq C\), \(t \in[r_{n},T]\), uniformly in \(n\). By the formula for \(M(t)\), the same holds true for \(b(t)\). □
Proof of Lemma 6.10, part c
Let \((t_{n})\) be a decreasing sequence in the interval \((T-\eta,T]\) such that \(b'(t_{n})\) is a maximizer of \(b'\) in the interval \([t_{n},T]\). If the sequence \((b'(t_{n}))\) is bounded, there is nothing to show. So let us assume that \(b'(t_{n})\) becomes large with increasing \(n\). We fix a large \(n\) and define \(\overline{b}:= b'(t_{n})\). The goal is to construct an upper bound for \(\overline{b}\) independent of \(n\).
Let \(t_{0}:= t_{n}+\frac{1}{\sqrt{\overline{b}}}\). We further define \(w\) as the solution of the boundary value problem
where \(x_{0}\) solves \(\frac{-b(t_{n})+x_{0}}{t_{0}-t_{n}}=\overline{b}\). So we have a boundary value problem in a wedge with small (\(\overline{b}\) is large) angle. The maximum principle for \(w\) implies
for \(t \in[t_{n},t_{0}]\). The maximum principle applied for the function \(w-V_{xt}\) on the set \(G_{0}:= \{ (t,x)| -b(t) \leq x \leq b(t),t_{n} \leq t\leq t_{0} \}\) gives \(w-V_{xt} \geq0\) on \(G_{0}\) (for the boundary values of \(V_{xt}\) see [15, Eq. (51)]). Using the fact that \(w\) and \(V_{xt}\) agree at \((t_{n},b(t_{n}))\), we conclude \(\left(w_{x}-V_{xtx} \right)(t_{n},b(t_{n})) \leq 0\), or
Now we split \(w\) into three functions
and define \(w^{(1)}:= -\beta e^{-\beta t_{0}}\). For \(w^{(2)}\) and \(w^{(3)}\), we define proper BVPs such that (9.46) holds. The function \(w^{(2)}\) is defined as the solution of the BVP
The proof of
is exactly as in [15, Lemma 5.3], and we are left with the problem for \(w^{(3)}\), where the new terminal condition enters. We define \(w^{(3)}\) as solution of the BVP
Let now
This constant depends only on the quantities mentioned in the formulation of the lemma. (We show in the next section, Lemma 7.3(a), that for a special sequence \(\epsilon_{k} \to0\), \(M\) depends only on the model parameters.) Define now \(w^{(4)}\) as the solution of
The maximum principle implies \(w^{(4)}\geq w^{(3)}\), and as
one concludes that
Introducing the new coordinates
and the new dependent variable \(u:= w^{(4)}e^{\frac{\tilde{b}}{\sigma ^{2}}z+\frac{\tilde{b}^{2}}{2\sigma^{2}}\tau}\), where \(\tilde{b}:= \overline{b}-\mu\), gives for \(u(\tau,z)\) the system
which has the solution \(u(\tau,z)= \frac{M}{\sqrt{2\pi\tau}\sigma } \int _{x_{0}-M}^{x_{0}+M} e^{\frac{\tilde{b}}{\sigma^{2}}z} (e^{-\frac{(z-\xi)^{2}}{2\sigma^{2} \tau}}- e^{-\frac{(z+\xi)^{2}}{2\sigma^{2} \tau}})\, d\xi\). By elementary calculation, one gets the value of \(u_{z}(\tau,0)\), and transforming back to the original variables gives
hence
for a generic positive constant \(c\), depending on the model parameters and \(M\). We conclude by (9.49) and (9.47)–(9.45) that \(V_{xtx}(t_{n},b(t_{n})) \geq-c\sqrt{\overline{b}}-c\) holds, i.e., by the boundary condition for \(V_{xtx}\),
which implies immediately \(\overline{b}\leq c\). □
Proof of Lemma 6.10, part d
The proof works analogously to the proof of Lemma 5.4 in [15]. Indeed, instead of the integral equation (11) there, we work with our new integral equation (6.2), i.e., we have to replace \(B', D\) and \(h\) by \(\tilde{B}', \tilde{D}\) and \(\tilde{h}\). By this replacement, we generate a new term in (77), stemming from \(\frac{d}{dt}\left(-F_{xxx}(t,b(t))\right)\). But the new term can be incorporated in \(-c+cg'(T-s)\) there (with an \(\epsilon\)-dependent constant \(c\)). The replacement of \(D\) by \(\tilde{D}\) in (78) leads to a term which can be incorporated in \(cb(T-s)+c\) there. Hence, (79) and (80) are true as well. (81) can be dealt with as (78) before, and finally, the replacement of \(h\) by \(\tilde{h}\) does not generate alterations at all. □
1.5 9.5 Auxiliary results for Sect. 7
Proof of Lemma 7.3 a
Checking the proof of Lemma 6.10(c), one sees that we have to check only that the constant \(M\) defined in (9.48) is independent of \(k\). Sufficient for this is \(\sup_{x \in[-b^{(k)}(t_{0}),b^{(k)}(t_{0})]} V^{(k)}_{xt}(t,x) \leq C\), for a generic \(k\)-independent constant \(C\) depending on \(t_{0}\) and the model parameters. By \(\mathcal{L}V^{(k)}_{x}=0\), this is equivalent to \(\inf_{x \in[-b^{(k)} (t_{0}),b^{(k)}(t_{0})]} V^{(k)}_{xxx}(t,x) \geq-C\) or, after transforming to the \(v\)-variable, \(\inf_{x \in[-b^{(k)}(t_{0}),b^{(k)}(t_{0})]} v^{(k)}_{xxx}(t,x) \geq-C \). By definition we have \(v^{(k)}=: \tilde{v}^{(k)}+F(t,x)\), which implies that it is sufficient that
for a generic \(k\)-independent constant \(C\) depending on \(t_{0}\) and the model parameters. We recall (see Lemma 9.7) that \(\tilde{v}^{(k)}_{xxx}\) solves the system
We now assert the following:
Claim 1: \(\Vert\tilde{B}^{(k)}(t)\Vert_{L^{p}(0,T)} \leq C\) , uniformly in \(k\) , for all \(1 \leq p < \infty\). For the first term of \(\tilde{B}^{(k)}\), this is obvious since the \(b^{(k)}\) are uniformly bounded, and for \(F_{xxx}(t,b^{(k)}(t))\), this follows from the explicit formula
and
Claim 2: Let \(0 <\delta_{0} <\sqrt{3}\) be an arbitrary fixed number. Then there exists \(\Delta>0\) independent of \(k\) such that for \(t\in[T-\Delta,T]\) ,
We prove Claim 2 by contradiction. So let \(f^{(k)}(\nu):= \frac {b^{(k)}(\nu )}{g(\nu)}\) with \(\nu:= T-t\) and \(g(\nu):= \sigma\sqrt{-\nu\ln{\nu}}\). Furthermore, let \(\underline{f}(\nu):= \inf_{k} f^{(k)}(\nu)\) and assume that there exists a sequence \(\nu_{\ell}\to0\) such that \(\underline{f}(\nu_{\ell}) < \delta_{0}\) for all \(\ell\in\mathbf{N}\). Hence there exist numbers \(K(\ell)\) such that
for some \(\delta_{0} < \delta_{1} < \sqrt{3}\) and for all \(k \geq K(\ell)\).
Note that we use alternatingly in the following the physical time \(t\) and the backward running time \(\nu\), which is a slight abuse of notation. Since the barrier strategy with barrier function \(b^{(k)}\) is the optimal strategy, using the notation \(V_{k}\) for the value function with terminal gain \(H^{(k)}\) and using \(\hat{H}^{(k)}(x):= H^{(k)}(x)-\bar {\gamma}\) gives for all \(k \geq K(\ell)\)
Here we have used \(dX_{s}=\mu ds+\sigma dW_{s} -dC^{(k)}_{s}\) and integration by parts in the first equality, \(\tau^{(k)}_{\ell}\leq T\) and the non-negativity of \(X\) in the first inequality, and finally \(\hat{H}^{(k)}(X_{\tau^{(k)}_{\ell}})-X_{\tau^{(k)}_{\ell}} \leq 2 \epsilon_{k}\) in the last inequality. Note that the \(O\) above is uniform in \(k\). We are now looking for an upper estimate of \(\mathbf{P}[\tau^{(k)}_{\ell}=T]\) and denote by \(\hat{\tau}^{(k)}_{\ell}\) the ruin time if we do not consume anything after the lump sum at \(t_{\ell}\). Clearly, \(\mathbf{P}[\tau^{(k)}_{\ell}=T] \leq\mathbf{P}[\hat{\tau}^{(k)}_{\ell}=T]\), and we estimate the latter probability by
with \(r(\nu_{\ell}):= \frac{\mu}{\sigma}\sqrt{\frac{\nu_{\ell}}{2}}+\delta_{1} \frac{g(\nu_{\ell})}{\sigma\sqrt{2\nu_{\ell}}}\). Here we used for the first inequality standard results (see, e.g., (2.8.3) in [19]), and the second holds for \(r(\nu_{\ell}) \geq1/2\), which we assume without loss of generality. Noting that \(r(\nu_{\ell}) \leq\delta_{2} \sqrt{\frac{-\ln\nu_{\ell}}{2}}\) for some \(\delta_{2}\) with \(\delta_{0} < \delta_{1} < \delta_{2} <\sqrt{3}\), we end up, using (9.51) and (9.52), with
for a constant \(\delta_{3} < \frac{3}{2}\) arbitrarily close to \(3/2\), for all \(\ell\) and for all \(k \geq K(\ell)\), with the \(O\) uniform in \(k\).
We now consider a barrier strategy with barrier \(\sqrt{3}g(\nu)\) for \(0 \leq\nu\leq\nu_{\ell}\) and denote it by \(C^{\sqrt{3}}\). We get for the target functional, using integration by parts as before,
For the first term, we use \(-\mathbf{E}[e^{-\beta\tau_{\ell}}] \geq -e^{-\beta T} \mathbf{P}[\tau_{\ell}=T]-e^{-\beta t_{\ell}} \mathbf{P}[\tau_{\ell}< T]\) and arrive at
for some generic positive constant \(C\). We now use from Lemma 9.9 below the upper estimate \(\mathbf{P}[\tau_{\ell}< T] \leq C \nu_{\ell}^{\frac {3}{2}}\sqrt{-\ln\nu_{\ell}}\) for \(\ell\) large enough, say \(\ell\geq L\). All together, and observing that in the last integral of (9.54), \(X_{s} \leq\sqrt{3}g(\nu)\) holds, we conclude that
for all \(\ell\geq L\). Since the target functional should be less than or equal to the value function, this gives via (9.53)
for all \(\ell\geq\bar{L}\) and \(k \geq K(\ell)\). Letting \(k\) tend to infinity, this is an obvious contradiction, proving Claim 2, and therefore also Claim 1.
One checks that \(\Vert\tilde{h}^{(k)\prime\prime\prime}(x)\Vert _{L^{1}(-\epsilon_{k},\epsilon_{k})}\) is uniformly bounded, and the fact that \(\tilde{B}^{(k)}\) is uniformly bounded above implies that \(\Vert\tilde{v}^{(k)}_{xxx+}\Vert_{L^{1}(-b^{(k)}(t),b^{(k)}(t))}\) is uniformly bounded in \(t \in[0,T]\) and \(k\). Hence, in order to show
Claim 3: \(\Vert\tilde{v}^{(k)}_{xxx}\Vert _{L^{1}(-b^{(k)}(t),b^{(k)}(t))}\) is uniformly bounded in \(t \in[0,T]\) and \(k\) , it suffices to show that \(\int_{t}^{T} (\frac{d}{ds} \int_{-b^{(k)}(s)}^{b^{(k)} (s)} \tilde{v}^{(k)}_{xxx}(s,x)\,dx)\,ds\) is uniformly bounded in \(t \in[0,T]\) and \(k\). We calculate
where the first equality holds because \(\tilde{v}^{(k)}\in C(\overline {G_{1}^{(k)}})\), the second follows from the first equation in (9.41), the third follows from (9.41), the fact that \(F_{xx}\) solves the backward heat equation (away from \((T,0)\)) and \(d^{(k)\prime}(s) = \beta( b^{(k)}(s)+d^{(k)}(s) ) -\mu\), which is itself implied by (9.25). Finally, the fourth equality defines \(B^{(k)}\) (resp., \(D^{(k)}\)) as the boundary values of \(\tilde{v}^{(k)}_{xxx}\) (resp., \(\tilde{v}^{(k)}_{xxxx}\)), but without the \(F\)-term.
Using now the backward heat equation for \(F_{xx}\), we get for the time derivative that \(\frac {d}{ds}(2F_{xx}(s,b^{(k)}(s)))=-J_{2}(s)\). Hence, we can estimate, with a generic constant \(C\),
because of Claim 2 and the explicit formula for \(F_{xx}(t,x)=Ce^{-\frac{x^{2}}{2\sigma^{2}(T-t)}}\frac{x}{(T-t)^{3/2}}\).
Concerning \(J_{1}\), one calculates
with some generic positive constants \(C\). We claim that
for some positive constant \(C\), for \(i=3,4,\dots,7\), uniformly in \(t \in[0,T]\) and \(k\). We show this only for \(J_{6}\), because the other cases work analogously, or are even simpler. One calculates
From this, (9.56) for \(i=6\) is immediate. (For the estimate of the last term on the right-hand side, one can use the result for \(J_{5}\), which can obviously be proved without using the case \(i=6\)). Equation (9.55) and bound (9.56) provide Claim 3.
Let now \(t_{0}\) be given with \(0 \leq t_{0} < T\). Set \(t_{1}:= t_{0}+ \frac {T-t_{0}}{2}\) and consider the BVP for \(\tilde{v}^{(k)}_{xxx}\) on the time interval \([0,t_{1}]\). By Claim 2, one has obviously \(\tilde{B}^{(k)}(t) \geq-C(t_{0})\) on \([0,t_{1}]\), with the constant independent of \(k\). Moreover, Claim 3 provides the estimate \(\Vert\tilde{v}^{(k)} _{xxx}(t_{1},x)\Vert_{L^{1}(-b^{(k)}(t_{1}),b^{(k)}(t_{1}))} \leq C\). Both together give \(\tilde{v}^{(k)}_{xxx}(t,x) \geq-C(t_{0})\), for \(t\in[0,t_{0}]\) and \(x \in[-b^{(k)}(t),b^{(k)}(t)]\), with the constant independent of \(k\), which concludes our proof (see (9.50)). □
Lemma 9.9
Let \(b(t):= \sigma\sqrt{-3 (T-t) \ln(T-t)}\), \(t>T-1\), and let \(X_{s}\), \(s \in[t,T]\), be the process \(x+\mu(s-t)+\sigma(W_{s}-W_{t})\) reflected on the barrier \(b(t)\). Then we have
and
for some positive constant \(C\) and for \(t\to T\).
Proof
In the following, we use the abbreviation \(\nu=T-t\) for the time to maturity \(T\) and \(\mathbf{P}_{t}[\,\cdot\,]\) as shortcut for \(\mathbf{P} [\,\cdot\,| X_{t}=b(t)]\). We start with the proof of (9.57).
One simply estimates the survival probability of our reflected process \(X\) by the survival probability of an agent who consumes nothing, which is evidently larger. So
for \(\nu\to0\). Here we have used in the last equality a well-known formula for the distribution of the supremum of Brownian motion (see, e.g., (2.8.4) in [19]), and for the last asymptotic equivalence the asymptotic behavior of the standard normal distribution.
In order to prove (9.58), we first note that setting \(\mu =0\) makes the probability in question certainly smaller; so we stick in the proof to this case. Let \(\hat{X}\) be our process \(X\), but not stopped at the ruin time \(\tau\). Hence, \(\hat{X}_{T} {\in} (-\infty,0]\). Define \(D {:=} \{ \tau {<} T \}\) and consider on the set \(D\) the stopping time \(\tau_{1} {:=} \inf\{ t>\tau| \hat{X}_{t} =\pm b(t)\}\). Set
which implies by the obvious symmetry \(\mathbf{P}_{t}[D_{1}|D]=\mathbf{P}_{t}[D_{2}|D]=1/2\). Since the process \(\hat {X}_{t}\) is not reflected on the lower branch \(-b(t)\), it is clear that the probability \(\mathbf{P}_{t} [\hat{X}_{T} <0 | \hat{X}_{\tau _{1}}=-b(\tau_{1})]\) is very high by the proof of (9.57), and in any case we have
for \(\nu\) small enough, which implies
The process \(\hat{X}\) is considered in [7], and if one has \(\mathbf{P} _{t}[\hat{X}_{T}=0]>0\), the authors call this phenomenon a heat atom. Moreover, they show that this occurs if and only if the reflecting boundary is a so-called upper function of Brownian motion. In the second part of the proof of their Theorem 2.2, the authors show (in our notation) that \(\mathbf{P}_{t}[\hat{X}_{T}=b(T)] \geq\mathbf{P}[A]>0\), where the set \(A\) is defined as
Now, we obviously have \(\mathbf{P}[A_{1}] \sim1-C\frac{\nu^{3/2}}{\sqrt {-\ln \nu}}\) for \(\nu\to0\), and by time reversal and the classical considerations concerning the Kolmogorov test as in [18, Sect. 1.8], we get \(\mathbf{P}[A_{2}] \geq1-\int_{0}^{\nu}\frac{b(s)}{\sigma\sqrt{2\pi s^{3}}}e^{-\frac{b(s)^{2}}{2\sigma^{2} s}}\, ds \sim1-C(-\ln\nu)\nu^{3/2}\). Therefore, we obtain \(\mathbf{P}[A] \geq L_{1}(\nu) \sim1-C(-\ln \nu)\nu ^{3/2}\), and finally, using (9.59),
which concludes the proof. □
Rights and permissions
About this article
Cite this article
Grandits, P. An optimal consumption problem in finite time with a constraint on the ruin probability. Finance Stoch 19, 791–847 (2015). https://doi.org/10.1007/s00780-015-0275-x
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00780-015-0275-x