Abstract

We consider a nonzero-sum stochastic differential game which involves two players, a controller and a stopper. The controller chooses a control process, and the stopper selects the stopping rule which halts the game. This game is studied in a jump diffusions setting within Markov control limit. By a dynamic programming approach, we give a verification theorem in terms of variational inequality-Hamilton-Jacobi-Bellman (VIHJB) equations for the solutions of the game. Furthermore, we apply the verification theorem to characterize Nash equilibrium of the game in a specific example.

1. Introduction

In this paper we study a nonzero-sum stochastic differential game with two players: a controller and a stopper. The state in this game evolves according to a stochastic differential equation driven by jump diffusions. The controller affects the control process in the drift and volatility of at time , and the stopper decides the duration of the game, in the form of a stopping rule for the process . The objectives of the two players are to maximize their own expected payoff.

In order to illustrate the motivation and the background of application for this game, we show a model in finance.

Example 1. Let be a filtered probability space, let be a -dimensional Brownian Motion, and let be -independent compensated Poisson random measures independent of , . For , , where is the Lévy measure of a Lévy process with jump measure such that for all . is the filtration generated by and (as usual augmented with all the -null sets). We refer to [1, 2] for more information about Lévy processes.
We firstly define a financial market model as follows. Suppose that there are two investment possibilities: (1)a risk-free asset (e.g., a bond), with unit price at time given by (2)a risky asset (e.g., a stock), with unit price at time given by where is -adapted with a.s., is a fixed given constant, , , is -predictable processes satisfying for a.a. , a.s. and
Assume that an investor hires a portfolio manager to manage his wealth from investments. The manager (controller) can choose a portfolio , which represents the proportion of the total wealth invested in the stock at time . And the investor (stopper) can halt the wealth process by selecting a stop-rule . Then the dynamics of the corresponding wealth process is (see, e.g., [35]). We require that for a.a. , a.s. and that
At terminal time , the stopper gives the controller a payoff , where is a deterministic mapping. Therefore, the controller aims to maximize his utility of the following form: where is the discounting rate, is a cost function, and is the controller’s utility. We denote the expectation with respect to and the probability laws of starting at .
Meanwhile it is the stopper’s objective to choose the stopping time such that his own utility is maximized, where is the stopper’s utility.
As this game is typically a nonzero sum, we seek a Nash equilibrium, namely, a pair such that This means that the choice is optimal for the controller when the stopper uses and vice verse.
The game (4) and (8) are a nonzero-sum stochastic differential game between a controller and a stopper. The existence of Nash equilibrium shows that, by an appropriate stopping rule design , the stopper can induce the controller to choose the best portfolio he can. Similarly, by applying a suitable portfolio , the controller can force the stopper to stop the employment relationship at a time of the controller’s choosing.

There have been significant advances in the research of stochastic differential games of control and stopping. For example, in [69] the authors considered the zero-sum stochastic differential games of mixed type with both controls and stopping between two players. In these games, each of the players chooses an optimal strategy, which is composed by a control and a stopping . Under appropriate conditions, they constructed a saddle pair of optimal strategies. For the nonsum case, the games of mixed type were discussed in [6, 10]. The authors presented Nash equilibria, rather than saddle pairs, of strategies [10]. Moreover, the papers [1117] considered a zero-sum stochastic differential game between controller and stopper, where one player (controller) chooses a control process and the other (stopper) chooses a stopping . One player tries to maximize the reward and the other to minimize it. They presented a saddle pair for the game.

In this paper, we study a nonzero-sum stochastic differential game between a controller and a stopper. The controller and the stopper have different payoffs. The objectives of them are to maximize their own payoffs. This game is considered in a jump diffusion context under the Markov control condition. We prove a verification theorem in terms of VIHJB equations for the game to characterize Nash equilibrium.

Our setup and approach are related to [3, 17]. However, their games are different from ours. In [3], the authors studied the stochastic differential games between two controllers. The work in [17] was carried out for a zero-sum stochastic differential game between a controller and a stopper.

The paper is organized as follows: in the next section, we formulate the nonzero-sum stochastic differential game between a controller and a stopper and prove a general verification theorem. In Section 3, we apply the general results obtained in Section 2 to characterize the solutions of a special game. Finally, we conclude this paper in Section 4.

2. A Verification Theorem for Nonzero-Sum Stochastic Differential Game between Controller and Stopper

Suppose the state at time is given by the following stochastic differential equation: where , , are given functions, and denotes a given subset of .

We regard and as the control processes, assumed to be càdlàg, -adapted and with values in , for a.a. , , . And we put . Then is a controlled jump diffusion (see [18] for more information about stochastic control of jump diffusion).

Fix an open solvency set . Let be the bankruptcy time. is the first time at which the stochastic process exits the solvency set . Similar optimal control problems in which the terminal time is governed by a stopping criterion are considered in [1921] in the deterministic case.

Let and be given functions, for . Let be a family of admissible controls, contained in the set of such that (9) has a unique strong solution and for all , where denotes expectation given that . Denote by the set of all stopping times . Moreover, we assume that

Then for and we define the performance functionals as follows: We interpret as if . We may regard as the payoff to the controller who controls and as the payoff to the stopper who decides .

Definition 2 (nash equilibrium). A pair is called a Nash equilibrium for the stochastic differential game (9) and (13), if the following holds:

Condition (14) states that if the stopper chooses , it is optimal for the controller to use the control . Similarly, condition (15) states that if the controller uses , it is optimal for the stopper to decide . Thus, is an equilibrium point in the sense that there is no reason for each individual player to deviate from it, as long as the other player does not.

We restrict ourselves to Markov controls; that is, we assume that , . As customary we do not distinguish between and , and . Then the controls and can simply be identified with functions and , where and .

When the control is Markovian, the corresponding process becomes a Markov process with the generator of given by where is the gradient of , and is the th column of the matrix .

Now we can state the main result of this section.

Theorem 3 (verification theorem for game (9) and (13)). Suppose there exist two functions ; such that(i), , where is the closure of and is the interior of ; (ii) on , .Define the following continuation regions: Suppose spends time on a.s., , that is, (iii) for all , , , (iv) is a Lipschitz surface, , (v). The second-order derivatives of are locally bounded near , respectively, , (vi), (vii) there exists such that, for , (viii); , for all , .For define and, in particular, Suppose that (ix)the family is uniformly integrable, for all and , . Then is a Nash equilibrium for game (9), (13), and

Proof. From (i), (iv), and (v) we may assume by an approximation theorem (see Theorem  2.1 in [18]) that , .
For a given , we define, with , In particular, let be as in (vii). Then,
We first prove that (21) holds. Let be as in (24). For arbitrary , by (vii) and the Dynkin’s formula for jump diffusion (see Theorem  1.24 in [18]) we have where . Therefore, by (11), (12), (i), (ii), (viii), and the Fatou lemma, Since this holds for all , we have In particular, applying the Dynkin’s formula to we get an equality, that is, where and . Hence we deduce that Since we always have we conclude by combining (27), (29), and (30) that which is (21).
Next we prove that (22) holds. Let be as in (vii). For , by the Dynkin’s formula and (vii), we have where ; .
Letting gives, by (11), (12), (i), (ii), (viii), and the Fatou Lemma, The inequality (33) holds for all . Then we have Similarly, applying the above argument to the pair we get an equality in (34), that is, We always have Therefore, combining (34), (35), and (36) we get which is (22). The proof is completed.

3. An Example

In this section we come back to Example 1 and use Theorem 3 to study the solutions of game (4) and (8). Here and in the following, all the processes are assumed to be one-dimension for simplicity.

To put game (4) and (8) into the framework of Section 2, we define the process ; by Then the performance functionals to the controller (6) and the stopper (7) can be formulated as follows: In this case the generator in (16) has the form

To obtain a possible Nash equilibrium for game (4) and (8), according to Theorem 3, it is necessary to find a subset of and ; , such that (i) and , for all ; (ii) and , for all ; (iii) and , for all and ; (iv)there exists such that , for all .

Imposing the first-order condition on and , we get the following equations for the optimal control processes : With as in (41), we put

Thus, we may reduce game (4) and (8) to the problem of solving a family of nonlinear variational-integro inequalities. We summarize as follows.

Theorem 4. Suppose there exist satisfying (41) and two -functions ; such that (1)(2), ; (3) and for all ; (4) and for all and for all ; (5) for all , where is given by (42); (6) for all , where is given by (42). Then the pair is a Nash equilibrium of the stochastic differential game (4) and (8), where Moreover, the corresponding equilibrium performances are

In this paper we will not discuss general solutions of this family of nonlinear variational-integro inequalities. Instead we discuss a solution in special case when

Let us try the functions , , of the form and a continuation region of the form Then we have where

By conditions (1) and (3) in Theorem 4, we get

From conditions (4), (5), and (6) of Theorem 4, we get the candidate for the optimal control as follows:

Let on , . By condition (5) in Theorem 4, we have for . Substituting (52) into , we obtain Similarly, we obtain for by condition (6) in Theorem 4. And we substitute (53) in to get

Therefore, we conclude that where and are the solutions of (56) and (57), respectively.

According to Theorem 4, we use the continuity and differentiability of at to determine , , that is,

At the end of this section, we summarize the above results in the following theorem.

Theorem 5. Let , and let be the solutions of equations (56)–(58). Then the pair given by is a Nash equilibrium of game (4) and (8). The corresponding equilibrium performances are

4. Conclusion

A verification theorem is obtained for the general stochastic differential game between a controller and a stopper. In the special case of quadratic cost, we use this theorem to characterize the Nash equilibrium. However, the question of the existence and uniqueness of Nash equilibrium for the game remains open. It will be considered in our subsequent work.

Acknowledgment

This work was supported by the National Natural Science Foundation of China under Grant 61273022 and 11171050.