Abstract

This paper analyzes one kind of optimal control problem which is described by forward-backward stochastic differential equations with Lévy process (FBSDEL). We derive a necessary condition for the existence of the optimal control by means of spike variational technique, while the control domain is not necessarily convex. Simultaneously, we also get the maximum principle for this control system when there are some initial and terminal state constraints. Finally, a financial example is discussed to illustrate the application of our result.

1. Introduction

Stochastic optimal control is an important matter that cannot be neglected in modern control theory in long days. As is known to all, Pontryagin’s [1] maximum principle is one of the main ways to settle the stochastic optimal control problem. By introducing the Hamiltonian function, a necessary condition for the optimal control of stochastic control systems was given by him, which was called the maximum condition. From that time, plenty of works on this issue have been done. Peng [2] was the first one to prove the general maximum principle of the forward-backward stochastic control system with diffusion coefficient containing the control variable by the technique of the second-order Taylor expansion and the second-order duality. He [3] was also the first one to demonstrate the maximum principle of forward-backward stochastic control systems from the view of backward stochastic differential equations (BSDE). In Peng’s paper [3], the control domain was convex (in local form); Xu [4] extended this conclusion to the case of the nonconvex control domain (in global form), but the control variables were not included in the diffusion coefficient. And these results were extended to the fully coupled case in the form of local and global by Shi and Wu [5, 6] in 1998 and in 2006, respectively. On the basis of these works, Situ [7] was the first to obtain the maximum principle of the forward stochastic control system with global form of random jumps in 1991. Shi and Wu [8] and Shi [9] acquired the maximum principle for a kind of forward-backward stochastic control system with Poisson jumps in the form of local and global, respectively. The fully coupled forward-backward stochastic control system was extended by Liu et al. [10] at the base of Shi and Wu [8], and in the meanwhile, they also obtained the maximum principle with the control system be constrained about initial-terminal state constraints. Considering that in real life, the decision makers could only get partial information but not complete information in most cases; many scholars have paid attention to the partial observable stochastic optimal control problem and have achieved many results (see, for example, [1113]). Traditionally, when using a stochastic partial differential equation called the Zakai equation to transform a full-information optimal control problem to the partially observable case, scholars will encounter a difficult problem: an infinite-dimensional optimal control problem. Wang and Wu [14] proposed a backward separation approach and replaced the original state and observation equation with the Zakai equation, and lots of complicated stochastic calculi in infinite-dimensional spaces were avoided in this way. Based on this approach, Xiao [15] studied a partially observed optimal control of forward-backward stochastic systems with random jumps and obtained the maximum principle and sufficient conditions of an optimal control under some certain convexity assumptions. Wang et al. [16] proved the maximum principles for forward-backward stochastic control systems with correlated state and observation noises. More recent conclusions of the partially observed stochastic control problem can be seen from the studies conducted by Wang et al. [17], Zhang et al. [18], and Xiong et al. [19].

In these years, through the study of mathematical economics and mathematical finance, many scholars turn their attention to the stochastic control system driven by Lévy process. In 2000, Nualart and Schoutens [20] built a pair of pairwise strongly orthonormal martingales which was called the Teugels martingale. Meanwhile, under some exponential moment conditions, they also obtained a martingale representation in that paper. Under these two important conclusions, for BSDE driven by the Teugels martingale, they [21] proved the existence and uniqueness theorem of its solution in the next year. From then on, a number of important results were proved: Meng and Tang [22] obtained the maximum principle of the forward stochastic control system driven by Lévy process. A necessary and sufficient condition for the existence of the optimal control of backward stochastic control systems driven by Lévy process was deduced by Tang and Zhang [23] through convex variation methods and duality techniques. For the forward-backward stochastic control system driven by Lévy process, there are also a lot of achievements: based on the existence and uniqueness theorem of FBSDEL [24], Zhang et al. [25] obtained a necessary condition of the optimal control and verification theorem, but in their control system, the backward state variables and did not enter the forward part. Wang and Huang [26] extended this result to the fully coupled control system and obtained the continuity result depending on parameters about FBSDEL and the local form maximum principle. Subsequently, Huang et al. [27] studied this control system with terminal state constraints and obtained the corresponding necessary maximum principle using Ekeland’s variational. For more recent conclusions about the stochastic control problem driven by Lévy process, please refer to [2830].

In this paper, we will study the optimal control problem for forward-backward stochastic control systems driven by Lévy process, which could be considered as a nonconvex control domain case that is extended from the result of [25]. With the technique of spike variation and Ekeland’s variational principle, the maximum principle of this type of control system and the control system with initial and final state constraints are obtained.

The structure of this paper is as follows. Section 2 describes some of the preparations used in this paper. The maximum principle and the one with initial and terminal state constraints as the major results of this paper will be shown in Sections 3 and 4. As an application of the maximum principle, Section 5 gives an optimal consumption problem in the financial market. Section 6 is the summary of this article.

2. Preliminary Statement

Let be a complete probability space which satisfied the usual conditions, and the information structure is given by which is generated by two processes: a standard Brownian motion valued in and an independent 1-dimensional Lévy process of the form ; here, is a pure jump process. And assume that Lévy measure satisfies the following two conditions; thereby, Lévy process has moments in all orders.(i).(ii), and some .

Denote , , and for . And let be the compensated power jump process of order ; then, Teugels martingale is defined by ; the coefficients correspond to orthonormalization of the polynomials with respect to the measure .

Then, are pairwise strongly orthogonal martingales, and their predictable quadratic variation processes are , is an indicator function here. And is an martingale; for more details of the Teugels martingale, see Nualart and Schoutens [20].

In the following of this section, we shall assume some notations: for a Hilbert space ,....

For the following FBSDEL,where take the value in and mappings , and take the value in , , , and , respectively. Convenient for writing, set column vector and , where is a full rank matrix.

Assumption 1. (i)All mappings in equation (1) are uniformly Lipschitz continuous in their own arguments, respectively.(ii)For all , for , respectively, and .(iii).(iv), where . , , and are nonnegative constants which satisfied , , and (resp. ) when (resp. ).Then, the following existence and uniqueness of the solution conclusion holds.

Lemma 1. There exists a unique solution in satisfying FBSDEL (1) under Assumption 1.
The detailed certification process of this conclusion can be seen in [24].

3. Stochastic Maximum Principle

In this section, for any given admissible control , we consider the following stochastic control system:where is given. An admissible control is an -predictable process which takes values in a nonempty subset of ; is the set of all admissible controls.

And the performance criterion iswhere is a given Frechet differential function.

Our optimal control problem amounts to determining an admissible control such that

In order to get the necessary conditions for the optimal control, we assume is the optimal control, and the corresponding solution of (2) is recorded as and introduce the “spike variational control” as follows:and are the state trajectories of ; here, is an arbitrary admissible control and is a sufficiently small constant.

We also need the following assumption and variational equation (6).

Assumption 2. (i), , and are continuously differentiable with respect to , and the derivatives are all bounded.(ii)There exists a constant , and it holds that .Here, , , , , , , , , and .

Lemma 2. Suppose Assumptions 1 and 2 hold; for the first-order variation , we have the following estimations:

Proof. We first prove inequations (7) and (8). For the forward part of the first-order variation equation, we haveApplying Gronwall’s inequation, we haveSimilarly, (8) holds.
We next estimate , , and ; the backward part of the first-order variation equation can be rewritten asSquaring both sides of (17) and using the fact ofwe getWhen with we haveApplying Gronwall’s inequation, we haveConsider the BSDE of the first-order variation equation in the interval :Thus,So, when with , we haveAfter a finite number of iterations, (9), (11), and (13) are obtained. And (10), (12), and (14) can be proved by using a similar method and the following inequalities: ; .

Lemma 3. Under hypothesis, Assumptions 1 and 2, it holds the following four estimations:

Proof. To prove (25), we observe thatwhereIt follows easily from Lemma 2 thatSincethenwith , , and .
By Gronwall’s inequation, we haveNext, we prove (26), (27),, (28); it can be easily checked thatHere,SincethenSquaring both sides of the equation above, we getFrom Lemma 2 and equation (25), we haveThen, we can get (26), (27), and (28) by applying the iterative method to the above relations.

Lemma 4. (variational inequality). Under the conditions that Assumptions 1 and 2 are established, we can get the following variational inequality:

Proof. From the four estimations in Lemma 3, we have the following estimation:Therefore,We introduce the following Hamiltonian function asand the following adjoint equationwhereIt is easily check that adjoint equation (45) has a unique solution quartet .
Then, we get the main result of this section.

Theorem 1. Let hypothesis, Assumptions 1 and2, hold; is an optimal control, and the corresponding optimal state trajectories are ; let be the solution of adjoint equation (45); and then, for each admissible control , we have

Proof. Applying Its formula to and , it follows from (6), (45), and the variational inequality thatBy the definition of , we know that, for any , the following inequation holds:Then, (47) can be easily checked.

4. Stochastic Control Problem with State Constraints

In this part, we are going to discuss stochastic control problems with state constraints in control system (2). Specifically, the initial state constraints and final state constraints are as follows:where . Our optimal control problem is to find such thatsubject to the state constraints (50). In the following, we will apply Ekeland’s variational principle to solve this optimal control problem. Firstly, we need the following assumptions.

Assumption 3. Control domain is assumed to be closed, the mappings in state constraints are continuously differentiable, and are bounded.
For , letSame as Section 3, we also assume be the optimal control, and the corresponding optimal state trajectories are . In order to solve the constraint problem, we need the following penalty cost functional, for any :It can be checked that is continuous, and for any ,It can be obtained by Ekeland’s variational principle that there exists such thatFor fixed and admissible control , we define the spike variation as follows:for any , and it is easy to check from of (55) thatLet be the trajectories corresponding to and be the trajectories corresponding to . The variational equation we used in this section is the same as the one in Section 3, with and . And we also assume that the solution of this variational equation is . Similar to the approach in Lemmas 2 and 3, it can be shown thatThen, by (57), the following variational inequality holds:whereNow, let be the solution ofHere, , , , etc. Applying s formula to , variational inequality (59) can be rewritten aswhere the Hamiltonian function is defined as (44). Sincethere exists a convergent subsequence, still denoted by such that when with .
Let be the solution ofFrom the continuous dependence of the solutions of FBSDEs with Lévy process on the parameters, we can prove that the following convergence holds: ; then, in equation (62), it implies thatSimilarly, there exists a convergent subsequence such that when with . Since , as , we have and , which is the solution of the variational equation as same as (6).
The following adjoint equation is introduced:Similarly, it can be proved that ; then, inequation (65) impliesThen, we get the following theorem:

Theorem 2. Let Assumptions 13 hold; is an optimal control, and the corresponding optimal state trajectories are , and is the solution of adjoint equation (66); then, there exists nonzero constant with such that, for any admissible control , the maximum condition (67) holds.
This conclusion can be drawn from the above analysis directly.

5. A Financial Example

In this section, we will study the problem of optimal consumption rate selection in the financial market, which will naturally inspire our research to the forward-backward stochastic optimal control problem in Section 3.

Assume the investor’s asset process in the financial market is described by the following stochastic differential equation with the Teugels martingale:where and are the expected return and volatility of the value process at time , respectively, and is the consumption rate process. Assume , , , and are all uniformly bounded measurable random processes. The purpose of investors is to select the optimal consumption strategy at time , minimizing the following recursive utility:where is the following backward stochastic process:with constant , the discount factor , and the Arrow–Pratt measure of risk aversion .

If the consumption process is regarded as a control variable, then combining (68) with (70), we encounter the following control system:which is obviously a special case of stochastic control system (2) with and .

We can check that both Assumptions 1 and 2 are satisfied. Then, we can use our maximum principle (Theorem 1) to solve the above optimization problem. Let be an optimal consumption rate and be the corresponding wealth process and recursive utility process. In this case, the Hamiltonian function H reduces toand is the solution of the following adjoint equation:

According to the maximum principle (47), we haveand the optimal consumption rate

By solving FBSDE (73), we get , and . Then, we get the optimal consumption rate of the investor which is

6. Conclusions

In this paper, a nonconvex control domain case of the forward-backward stochastic control driven by Lévy process is considered, and we obtain the global stochastic maximum principle for this stochastic control problem. And then, the problem of stochastic control with initial and final state constraints on the state variables is discussed, and a necessary condition about existence of the optimal control is also acquired. A financial example of optimal consumption is discussed to illustrate the application of the stochastic maximum principle.

Data Availability

No data were used to support this study.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

This work was supported by the NNSF (11701335), Shandong NSF (ZR2019BG015), NSSF (17BGL058), Shandong Provincial Higher Education Science and Technology Plan Project (J18KA236), and Shandong Women’s College High-level Research Project Cultivation Foundation (2018GSPGJ07).