Abstract
This article studies sufficient conditions for near-optimal stochastic control for systems governed by general linear controlled forward–backward stochastic differential equations (FBSDEs in short). The control is allowed to enter into both drift and diffusion coefficients. We prove that under certain additional conditions on the Hamiltonian, the near-maximum condition on the Hamiltonian function in the integral form is sufficient for near-optimality. As an applications, an example is given to illustrate our theoretical results.
Similar content being viewed by others
References
Zhou XY (1998) Stochastic near-optimal controls: necessary and sufficient conditions for near-optimality. SIAM Control Optim 36(3):929–947
Huang J, Li X, Wang G (2010) Near-optimal control problems for linear forward-backward stochastic systems. Automatica 46:397–404
Hui E, Huang J, Li X, Wang G (2011) Near-optimal control for stochasic recursive problems. Syst Control Lett 60:161–168
Hafayed M, Abbas S (2014) On near-optimal mean-field stochastic singular controls: necessary and sufficient conditions for near-optimality. J Optim Theory Appl 160(3):778–808
Hafayed M, Abba A, Abbas S (2014) On mean-field stochastic maximum principle for near-optimal controls for poisson jump diffusion with applications. Int J Dynam Control 2:262–284
Hafayed M, Abbas S (2013) Stochastic near-optimal singular controls for jump diffusions: necessary and sufficient conditions. J Dyn Control Syst 19(4):503–517
Hafayed M, Abbas S, Veverka P (2013) On necessary and sufficient conditions for near-optimal singular stochastic controls. Optim Lett 7(5):949–966
Hafayed M, Veverka P, Abbas S (2012) On maximum principle of near-optimality for diffusions with jumps, with application to consumption-investment problem. Differ Equ Dyn Syst 20(2):111–125
Hafayed M, Veverka P, Abbas A (2014) On Near-optimal necessary and sufficient conditions for forward-backward stochastic systems with Jumps, with applications to finance. Appl Math 59(4):407–440
Zhang L, Huang J, Li X (2015) Necessary condition for near optimal control of linear forward-backward stochastic differential equations. Int J Control 1–29. doi:10.1080/00207179
Zhou XY, Sethi S (1994) A sufficient condition for near optimal stochastic controls and its applications to manufacturing systems. Appl Math Optim 29:67–92
Wu Z (2013) A general maximum principle for optimal control of forward-backward stochastic systems. Automatica 49:1473–1480
Yong J (2010) Optimality variational principle for controlled forward–backward stochastic differential equations with mixed initial-terminal conditions. SIAM J Control Optim 48:4119–4156
Ma J, Yong J (1999) Forward-backward stochastic differential equations and their applications. Springer, New York
Ekeland I (1974) On the variational principle. J Math Anal Appl 47:324–353
Clarke FH (1983) Optimization and nonsmooth analysis. Wiley, New York
Author information
Authors and Affiliations
Corresponding author
Additional information
This work was partially supported by the National Nature Science Foundation of China (Nos. 11201263, 11201264), the Nature Science Foundation of Shandong Province (No. ZR2012AQ004) and the ITEA MODRIO project of INRIA.
Appendix
Appendix
The following result gives the definition and some basic properties of the Clarke’s generalized gradient.
Definition 7
Let F be a convex set in \({\mathbb {R}}^{n}\) and let \(f:F\rightarrow {\mathbb {R}}\) be a locally Lipschitz function. The generalized gradient of f at \( \widehat{x}\in F\), denoted by \(\partial _{x}f\left( \widehat{x}\right) \), is a set defined by
where \(f^{\circ }\left( \widehat{x},\upsilon \right) =\lim \sup _{y\rightarrow \widehat{x},t\rightarrow 0}\frac{1}{t}\left( f\left( y+t\upsilon \right) -f\left( y\right) \right) \).
Proposition 8
If \(f:{\mathbb {R}}^{n}\rightarrow {\mathbb {R}}\) is locally Lipschitz at \(x\in {\mathbb {R}}^{n}\), then the following statements holds
-
1.
\(\partial _{x}f\left( x\right) \) is nonempty, compact and convex set in \({\mathbb {R}}^{n}\).
-
2.
\(\partial _{x}\left( -f\right) \left( x\right) =-\partial _{x}\left( f\right) \left( x\right) \).
-
3.
\(\partial _{x}f\left( x\right) \ni 0\) if f attains a local minimum or maximum at x.
-
4.
If f is continuously differentiable at x, then \(\partial _{x}f\left( x\right) =\left\{ f^{\prime }\left( x\right) \right\} \).
-
5.
If f, \(g:{\mathbb {R}}^{n}\rightarrow {\mathbb {R}}\) are locally Lipschitz functions at \(x\in {\mathbb {R}}^{d}\), then \(\partial _{x}\left( f+g\right) \left( x\right) \subset \partial _{x}f\left( x\right) +\partial _{x}g\left( x\right) \).
For the detailed proof of the above Proposition see Clarke [16] or the book by Yong and Zhou ([13] Lemma 2.3).
Lemma 9
(Ekeland’s Variational Principle [15]) Let (F, \(d_{F})\) be a complete metric space and \(f:F\rightarrow \overline{{\mathbb {R}}} \) be a lower semi-continuous function which is bounded from below. For a given \(\varepsilon >0\), suppose that \(u^{\varepsilon }\in F\) satisfying \( f\left( u^{\varepsilon }\right) \le \inf _{u\in F}f(u)+\varepsilon \). Then for any \(\delta >0\), there exists \(u^{\delta }\in F\) such that
-
1.
\(f\left( u^{\delta }\right) \le f\left( u^{\varepsilon }\right) \).
-
2.
\(d_{F}\left( u^{\delta },u^{\varepsilon }\right) \le \delta \).
-
3.
\(f\left( u^{\delta }\right) \le f\left( u\right) +\dfrac{ \varepsilon }{\delta }d_{F}\left( u,u^{\delta }\right) ,\,\) for all \(u\in F\).
Rights and permissions
About this article
Cite this article
Zhang, L. Sufficient condition for near-optimal control of general controlled linear forward–backward stochastic differential equations. Int. J. Dynam. Control 5, 306–313 (2017). https://doi.org/10.1007/s40435-015-0179-9
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s40435-015-0179-9
Keywords
- Sufficient conditions for near-optimal control
- Forward–backward stochastic differential equations
- Second-order adjoint equations
- Ekeland’s variational principle
- Clarke’s generalized gradient