Abstract

Traditional guidance algorithms for hypersonic glide vehicles face the challenge of real-time requirements and robustness to multiple deviations or tasks. In this paper, an intelligent online multiconstrained reentry guidance is proposed to strikingly reduce computational burden and enhance the effectiveness with multiple constraints. First, the simulation environment of reentry including dynamics, multiconstraints, and control variables is built. Different from traditional decoupling methods, the bank angle command including its magnitude and sign is designed as the sole guidance variable. Secondly, a policy neural network is designed to output end-to-end guidance commands. By transforming the reentry process into a Markov Decision Process (MDP), the policy network can be trained by deep reinforcement learning (DRL). To address the sparse reward issue caused by multiconstraints, the improved Hindsight Experience Replay (HER) method is adaptively combined with Deep Deterministic Policy Gradient (DDPG) algorithm by transforming multiconstraints into multigoals. As a result, the novel training algorithm can realize higher utilization of failed data and improve the rate of convergence. Finally, simulations for typical scenes show that the policy network in the proposed guidance can output effective commands in much less time than the traditional method. The guidance is robust to initial bias, different targets, and online aerodynamic deviation.

1. Introduction

From the last few decades until now, hypersonic glide vehicles (HGV) have attracted the attention of researchers [15] due to their high velocity and wide flight airspace. Usually, cooperative guidance has been widely studied as an effective threat to valuable targets represented by hypersonic vehicles [610]. HGV shows its superiority against cooperation for its large flight envelope, consisting of the boost phase, initial descend phase, glide phase, and dive phase. To achieve long-range penetration, the glide phase plays an important role. For different mission requirements, the guidance law of the glide phase presents different results. This paper will focus on the flight mission in that HGV flies to the neighborhood of the target at a certain height, and the total time of the glide phase is constrained.

Undergoing years of development, the most popular methods of reentry guidance can be summarized in reference tracking guidance [1122] and predictor-corrector guidance [2329]. As far back as 1978, Harpold and Graves [11] described the theoretical basis and guidance architecture of reference tracking guidance, which has been successfully applied in the guidance of space shuttles and other reentry vehicles. The reference tracking guidance is composed of offline trajectory optimization and online trajectory tracking. Commonly, the optimal or suboptimal trajectory is calculated by planning or optimization and stored in an onboard computer before the flight. Once entering the glide phase, the trajectory tracker generates guidance commands according to errors between the actual flight and the stored standard trajectory. Not taking deviation and disturbance into account, the trajectory planning problem can be seen as an optimal control problem, which can be solved by the Legendre pseudospectral method [13, 14], improved Gauss pseudospectral method [15, 16], etc. Other commonly used optimization methods in reentry flight mainly include particle swarm optimization [17] and multiphase convex programming [18]. The trajectory tracking methods mainly include linear quadratic regulator [19], sliding mode control [20], nonlinear geometric method [21], and fuzzy approximation method [22]. Reference tracking guidance is easy to realize, and the method does not consume much onboard computing resource. Nevertheless, the method is easily affected by initial state changes and kinds of disturbance such as aerodynamic errors, which means its robustness is not good in some situations.

Oppositely, predictor-corrector guidance requires no reference trajectory, which makes it flexible in more flight missions with initial errors and onboard disturbance. The task information stored in the onboard computer is merely terminal constraints. The predictor computes the terminal states by integrating dynamics equations. The corrector calculates the command of AOA (angle of attack) and bank angle to correct the errors between the calculated terminal states and terminal constraints. By repeating the guidance flow, the final errors of terminal constraints are limited to a very small range. Based on different prediction methods, predictor-corrector guidance is classified as analytical predictor-corrector guidance (APCG) and numerical predictor-corrector guidance (NPCG). Analytical prediction reduces the computational work by making some assumptions and simplifying equations of flight states. Kluever [23] proposed a guidance method for the skip-entry phase based on closed-form expressions that are derived from a matched asymptotic expansion analysis. Yu and Chen [24] innovatively presented a method for the derivation of analytical solutions to hypersonic gliding problems by transforming the equations into a linear system with variable coefficients and solving it based on spectral decomposition. Zhang et al. [25] established an auxiliary geocentric rotation frame, simplified the complex dynamics, and got a series of analytical expressions taking energy as the argument. Though the calculation of this kind of prediction is rapid, it faces the problem of insufficient accuracy. Numerical prediction is widely used due to its high accuracy. Joshi et al. [27] adopted NPCG at each guidance cycle to generate a feasible trajectory and presented a guidance algorithm that can satisfy path and terminal constraints. Lu [28] summarized the methods of predecessors and developed a unified predictor-corrector guidance method, based on a simple and robust numerical predictor with an altitude-rate-dependent, which is suitable for kinds of reentry vehicles. Cheng et al. [29] designed a deep neural network in place of the numerical integration and provided an intelligent multiconstrained predictor-corrector guidance, which can be seen as an improvement of NPCG. Methods based on NPCG improve the accuracy, meanwhile causing the problem of intensive computation. Limited by the computing power of the onboard computer, NPCG is not operative in practical use, especially in tasks with a short guidance cycle. Both the reference tracking guidance and predictor-corrector guidance have advantages and disadvantages. If there is no deviation or disturbance, the reentry problem can be solved under the framework of trajectory planning, which is widely studied [3033]. However, the complex online deviation and rapidity requirement call for an intelligent online method.

In recent years, deep reinforcement learning (DRL) has made a great breakthrough [3436]. Deep Q-learning (DQN) [35] was first proposed in 2013, which shows great performance in dealing with continuous states in the game field. Then, the Deep Deterministic Policy Gradient (DDPG) [36] further expands the action space from discrete to continuous, which aroused the rapid development of DRL [3739]. With the strong representation ability of deep learning (DL) and the natural decision-making ability of reinforcement learning (RL), DRL is widely applied in various fields [4042]. The sparse reward problem is common in DRL tasks, which describes that the agent may only receive a reward signal at the terminal time. To improve the efficiency of training and avoid the sparse reward, based on universal value function approximators [43], Andrychowicz et al. [44] innovatively proposed the Hindsight Experience Replay (HER) method. The HER method replaces the initial goal of the agent with an achieved goal, which greatly leads to abundant reward signals. The original intention of this design decides that the HER method is suitable for those tasks with multigoals. Prianto et al. [45] applied HER to soft actor-critic used for path planning for multiarm manipulators, and the trained agent can generate the shortest path for arbitrary scenarios. Manela and Biess [46] combined HER with curriculum learning on three challenging throwing tasks, which shows wonderful performance in multiple goals and sparse reward functions. Considering that the terminal constraints in the reentry task are similar to multigoals, we can explore the HER method in the DRL guidance frame.

This paper is aimed at finding a real-time plan and guidance scheme for reentry with the help of deep reinforcement learning. Compared with previous studies, the main contributions of this paper are as follows: (1)The problem of reentry with multiple (path and terminal) constraints is transformed into a Markov Decision Process, which is the precondition of solving it in DRL. Different from the decoupling method in most previous studies, the guidance command of bank angle including magnitude and sign is combined as an action. This makes the guidance logic concise(2)The HER method is introduced and improved in the training of guidance commands. The HER condition is presented creatively according to the peculiarities of the reentry problem. The flow of this method is easy to generalize to similar guidance problems with other additional constraints(3)The guidance network is trained offline and called online. The guidance scheme provides an end-to-end guidance command which is calculated by the well-trained network rather than complex guidance schemes. The engineering implementation of the proposed algorithm is simple, and the real-time performance is excellent

The remainder of the paper is organized as follows. Section 2 describes the reentry problem with multiconstraints and the bounds of control variables. Section 3 introduces the basic theory of DRL and the benefit of the HER method. Section 4 proposes the intelligent multiconstrained guidance algorithm and gives the architecture of two kinds of neural networks. Section 5 gives the simulations and analysis to verify the efficiency of the method. Finally, the conclusions of this paper are in Section 6.

2. Problem Formulation

2.1. Reentry Dynamics Equations

Considering the earth’s rotation, the dynamics equations of the hypersonic glide vehicle are given as follows: where represents the velocity, represents the flight path angle, represents the velocity heading angle, represents the distance between the Earth center and the vehicle, represents the longitude, represents the latitude, represents the vehicle mass, represents the gravitational acceleration, represents the self-rotation rate of the Earth, and represents the bank angle. The expressions of lift and drag force are where represents the atmospheric density, represents the reference area, and and represent the drag coefficient and lift coefficient, respectively, which can be obtained by two interpolation functions for AOA and the Mach number.

is a function of height (, is the radius of the Earth):

Given an initial state, the trajectory can be obtained by integration based on Equation (1), command of AOA and bank angle.

2.2. Multiple Constraints

To complete the flight mission, the HGV must satisfy all the path constraints and terminal constraints.

2.2.1. Path Constraints

During the reentry flight, the path constraints can be expressed as where is the heating rate of the stagnation point, is the dynamic pressure, is the aerodynamic load, and is a constant. , , and denote the maximum limits, respectively.

Apart from the three hard constraints, there is a soft QEGC (quasiequilibrium glide condition) constraint, which can make the reentry trajectory smooth:

According to Equations (2)–(5), we can obtain the three lower bounds and one upper bound of vs. , in which the constraint of the upper bound is soft.

According to Equation (6), the - corridor in the glide phase is obtained. Only if the flight trajectory of HGV is limited in the corridor, the path constraints can be satisfied. Intuitively, the - corridor is shown in Figure 1.

2.2.2. Terminal Constraints

Assume that HGV is required to reach the vicinity of the target at a specified height, the terminal constraints (height error , distance error ) can be summarized as where , , and represent the final height, latitude, and longitude at the terminal time, is the required height, and represent the latitude and longitude of the target, and and are the thresholds of height and distance.

In many practical application scenes, the terminal constraint on velocity is not so tight, which means that the velocity constraint takes the form of where is the terminal velocity and is the required minimum velocity.

Compared to traditional reentry tasks, we consider more terminal constraints in different tasks. For example, to arrive at the required time , there is another constraint: where is the flight time and is the threshold of time.

After this, we give the criterion for the success of a reentry flight. In cases where all the path constraints are satisfied, if the terminal states , , , and , then the flight mission is successful.

Similarly, if the reentry task requires HGV to arrive at a desired terminal approach angle, the relevant constraint condition is adaptively formulated: where is the LOS (line of sight) angle, is the desired approach angle, and is the angle threshold.

2.2.3. Switch Condition

As is shown in Figure 2, HGV cannot begin sliding from an arbitrary initial status so there is a switch condition. The switch condition refers to the constraint condition at the junction of the initial descent phase and glide phase.

Referring to [1], the switch condition is defined as that the absolute value of the difference between the slope of the QEGC at the current point and the actual slope of - is less than a small preselected value. where is a small value.

The first slope can be derived from Equation (1):

The QEGC equation is rewritten as

According to Equation (13), the derivate of to is obtained:

Both sides of Equation (13) take a derivative to velocity, and then, we can obtain

Substituting Equation (13) into (15), then we can obtain

Suppose that , Equation (16) is simplified as

Then, the second slope is obtained:

Substituting Equations (12) and (18) into (11), the switch condition is obtained.

2.3. Control Variables

During the reentry process, the trajectory is decided by the AOA and the bank angle . Commonly, the AOA profile is set as the following piecewise linear function: where is the max AOA limited by the aerodynamic and is the best AOA when the ratio of lift-to-drag takes the maximum value, which is helpful to increase the flight range of HGV.

Once the AOA profile is determined, the only control variable is the bank angle . According to the QEGC constraint and the three hard path constraints, the upper magnitude bound of can be derived as follows:

The traditional approach to generating guidance command is to calculate the magnitude of iteratively to satisfy longitudinal constraint and change the sign of in the horizontal plane in the heading angle corridor to fly HGV toward the target. However, the application of neural networks allows us to generate both the magnitude and sign of at the same time. If the output of a policy network is , the command of is as follows: where , .

3. Deep Deterministic Policy Gradient with Hindsight Experience Replay

3.1. Deep Reinforcement Learning

Based on Markov Decision Process (MDP), standard reinforcement learning consists of an environment, an agent, and their interactions. The environment is initiated as a state , which can describe its initial features. The agent observes the state of the environment and chooses an action . The action affects the environment to transfer to the next state with a certain probability and generates a reward to the agent. Then, the agent interacts with the environment repeatedly, so there is a sequence of tuples . The agent’s goal is to increase its accumulated rewards. where is the discount factor, which is used to balance the current and future reward.

Action-value function is set to evaluate how good the action is: . And if the policy is optimal, is the optimal action-value function, which satisfies the Bellman equation:

In DRL, if the policy is deterministic, it can be considered as a mapping from states to actions, commonly in a form of neural network with the parameter .

3.2. Deep Deterministic Policy Gradient (DDPG)

The baseline algorithm used for the training of neural networks is Deep Deterministic Policy Gradient (DDPG), which belongs to the family of model-free policy gradient DRL methods. There are four neural networks in DDPG, which are called the online actor network , the target actor network , the online critic network , and the target critic network . The online actor network parameterized by exports actions of the agent, and the online critic network parameterized by is designed to approximate action-value function . The actual behavior policy of the agent is presented from with an added noise .

The parameter is optimized by decreasing the loss:

The actor is updated by the theory of policy gradient: where is the distribution of the state.

The existence of the target actor network and the target critic network can make the process of training more stable. Besides, the data is stored in a replay buffer to reduce the relevance between experiences and deal with the experiences effectively.

3.3. Hindsight Experience Replay (HER)

The key idea behind Hindsight Experience Replay (HER) is derived from universal value function approximators (UVFA) [43]. In UVFA, there is a goal space . For any goal , there is a pseudoreward . The first state-goal pair in every episode is sampled from a distribution , and in the same episode, the goal is fixed. At every step in an episode, the agent gets an action from the current state and goal . On this premise, the traditional -function is expanded to .

In a task where the agent needs to achieve multiple goals, the designed reward is quite likely to be sparse. To deal with sparse rewards, an effective trick is to exploit the failed data samples too. In the method of HER, the states of the agent are expanded to combinations of states and goals. The essence of HER rests with designing a virtual goal , which is achieved in a failed trajectory. For example, there is an RL task in Figure 3, where the agent has a goal position. Only if the agent arrives at the goal it gets a reward 1, otherwise 0. In most episodes, the agent gets 0 rewards. It is quite difficult for conventional RL algorithms to achieve good results for the sparse reward problem especially when and are large numbers. In the HER method, if the final position is , we can define the position as a virtual goal , and then, the final reward , but . By doing this, the algorithm can learn from failed experiences, and the reward is changed from sparse to dense.

With the help of virtual goals, the efficiency of experience is improved, which makes the convergence of neural networks significantly ameliorated.

4. Design of Intelligent Guidance and Training of Neural Networks

4.1. Overall Design

In this section, the HER-DDPG-based guidance is proposed. First of all, we normalize the reentry simulation into two segments (scenario initialization and guidance step) and open two corresponding interfaces to the intelligent algorithm. In the segment of scenario initialization, the position and velocity of HGV are randomly initialized within a certain range. The simulation starts from the specified point at the initial phase and continues to the switch point decided by Equation (11). After entering the glide phase, at each decision step later, HGV receives the guidance command of bank angle via the interface to the intelligent algorithm and calculates AOA according to the profile given by Equation (19). The simulation proceeds until distance error or the velocity or any path constraint is not satisfied.

After constructing the two segments, the transformation from simulation to MDP is accomplished. The kinematical parameters of HGV are mapped to states, and the guidance command of bank angle is designed as the action. Based on the theory of HER, the terminal constraints are transformed into multigoals and the expanded states are defined. The reward function is designed as a combination of several linear functions of terminal height, terminal distance error, and flight time error, which is used for generating dense signals of policy gradient. Finally, based on HER and DDPG, the algorithm goes into operation as is shown in Figure 4. When the neural networks converge, the training process is terminated.

4.2. Markov Decision Process Variables

The fundamental variables in MDP are defined as follows.

4.2.1. States

The basic state of the HGV agent is defined as , which can uniquely express the features of the motion.

4.2.2. Goals

Goals describe the terminal constraints of HGV with a certain fixed tolerance of position and time. The initial goal is defined as . In this paper, , if the final distance error then the terminal distance constraint is satisfied. If the loose constraint , we decide that the HER condition is met.

The HER condition is shown in Figure 5, which means that despite its failure of arriving at the given target, HGV has arrived at the terminal position successfully. Here, we define , , , and as target longitude, target latitude, required height, and required time on the HER condition. The intention of HER is to efficiently increase the sampled data of failed episodes, which means that they may be successful in other target tasks. Therefore, the required height and time are modified as the actual values at the terminal time when the errors concerning initial goals are in a specifically designed range. The concrete values of them in this paper are given as

Then, the new goal is defined as .

In this case, the new terminal errors can be expressed as

By doing this, lots of failed data is transformed into successful data with new terminal constraints. The reward signal is more abundant in the training process, and the average reward can increase. In DRL training, the more abundant the reward signal, the faster the rate of convergence.

4.2.3. Expanded States

The expanded state can be defined as

On the HER condition, the expanded state is modified accordingly:

4.2.4. Rewards

In the paper where HER is proposed [44], the success of the task is merely relative to the position of a robotic arm, so it is feasible to learn from rewards that are sparse and binary. However, the number of terminal constraints in the reentry task is so up to 4 that the binary reward is no longer suitable. For the reentry task, the priorities of satisfying different terminal constraints are not at the same level. Intuitively, the range constraint is the most important, for it requires the agent to adjust both the magnitude and the sign of the bank angle to arrive in the vicinity of the target. Secondly, the height constraint is also important. After the two constraints are under consideration, the time-related reward is added. Considering the following dive phase after the glide phase, the constraint thresholds are defined: , , and .

Piecewise linear functions are applied to design the range-relative reward , height-relative reward , and time-relative reward :

The 3 subrewards are shown in Figure 6.

Then, the total reward is designed as follows:

is designed to reduce the terminal distance error in the range of , which means that if the distance error then the subreward , and if , then the total reward is determined by and . is designed to reduce the terminal height error, so the smaller the , the larger the . And in the whole training process, the number of samples with larger is more than smaller , so as decreases, the slopes of piecewise linear functions increase. Finally, only if the height error is the subreward added in . The design of is similar, so the smaller the , the larger the . And as decreases, the slopes of piecewise linear functions increase too.

If the terminal states , , , and , then the flight mission is successful. According to Equations (30)–(33), the reward of a successful episode is in the range of . For some successful episodes, the average reward value will be in the range of 10 to 140, which depends on the final data distribution of and .

4.2.5. Actions

Different from the widely used decoupling methods in current guidance algorithms, the magnitude and sign command of the bank angle are provided at the same time in this paper. Commonly action is calculated by a policy network. In DDPG, the output of the policy network is a continuous value . The action is mapped to the bank angle command in the form of Equation (21).

4.3. Algorithm Flow for Training

Based on HER, the training algorithm of multiconstrained reentry guidance is given in Algorithm 1.

Randomly initialize parameters of actor network and critic network .
Initialize target actor and target critic with , .
Initialize basic replay buffer
for, do
 Initialize an HER replay buffer and a random process noise for exploration
 Initialize an initial goal randomly
 Run the Scenario Initialization Function, sample a basic state
for, do
  Combine basic state and goal to expanded state
  Sample an action from actor and noise:
  Execute the action in the Policy Step Function and observe a new state
  Combine basic new state and goal to expand new state
  Store the transition in and
  if the episode is done
   Judge whether the HER condition is met and record it
  end if
end for
if the HER condition is met
   for, in do
    Calculate and , ,
    Recalculate reward according to Equations (30)–(33)
    Combine basic states and goal to expand states:
    
    Store the transition in
   end for
   Clear data in
end if
for, do
   Sample a minibatch from the replay buffer
   Update critic by Equation (24), update actor by Equation (25)
   Update target network periodically: ,
end for
end for
4.4. Neural Network Design

The designed actor network model and critic network model are shown in Figures 7 and 8. And the details of every layer are listed in Tables 1 and 2.

5. Simulations and Analysis

5.1. Simulation Settings

The HGV model used in this paper is the common aero vehicle (CAV-H). In Equation (1), and . In Equation (2), . In Equation (3), , , and . In Equation (4), , , , and  kW/m2. In Equation (8), . In Equation (19), , , , and . The initial longitude and latitude of HGV are 0 deg. The size of the integral step is 0.01 s, and the size of the guidance (policy) step is 150 s. All used random variables in Table 3 are subject to the uniform distribution.

After setting the above parameters, the networks are trained by Algorithm 1.

5.2. Training Process of Neural Networks

Comparative experiments are made at the same time to verify the superiority of DDPG+HER method. DDPG algorithm [36] and Proximal Policy Optimization (PPO) algorithm [37] are selected as control groups. The hyperparameters of the three algorithms used in the training process are set in Table 4.

After the training, the average returns and success count in the latest 100 episodes are shown in Figures 9 and 10. The total episode number is 23161.

It can be seen from Figure 9 that the average return of the DDPG+HER method begins rising first and gets more rewards than other methods. The sparse reward problem is effectively alleviated. At the end time of training, the success rate of the DDPG+HER method in Figure 10 converges to 100%, which implies that the policy given by the actor is operative. The best success count of DDPG is less than 75 and the best success count of PPO is less than 60, which powerfully illustrates that the proposed DDPG+HER method is more effective in the training of the policy network.

In Figure 11, the loss of actor and critic in the DDPG+HER method is given.

It can be seen from Figure 11 that the loss of the critic is decreasing to zero along with the progress of training. Meanwhile, the magnitude of actor loss is increasing in the process, which means that the score of the agent is on a trend of improvement.

5.3. Random Single Trajectory Analysis

After the convergence of neural networks, the typical simulation is performed randomly. To evaluate the performance of the proposed guidance, a contrast simulation based on NPCG is also performed. The proposed method in this paper is called HERG for short.

The terminal constraints are set as target longitude , latitude , required height , and required time . The errors of aerodynamics are set as and . For the trajectory of HERG, the terminal height is 30.755 km, and the terminal time is 1359.7 s. For the trajectory of NPCG, the terminal height is 32.916 km, and the terminal time is 1386.6 s.

The results are shown in Figures 1214.

Figure 12 shows that the HGV can arrive at the required target area. Figure 13 shows that the policy decision is executed 9 times, and terminal height and time are in the required range. The reverse number of the bank angle is reduced greatly for HERG than NPCG. And as is seen in Figure 14, all of the path constraints are satisfied.

The simulations are performed in a system with a 12th Gen Intel Core i5-12600KF CPU and 16 GB RAM. In the NPCG method, the corrector works a total of 9 times. As the time of integration decreases, the computing time decreases. The computing time of NPCG is listed in Table 5, and the total computing time is 6.431 s. In the HERG method, the policy network is called 9 times and the total computing time is 0.0937 s.

The contrast simulation results are shown in Table 6. It is obvious that with similar terminal errors, the calculation time of the proposed method is less than 1/68 of NPCG. By contrast with NPCG, the proposed method makes a breakthrough in real-time performance.

5.4. Verifications on Monte Carlo Simulations

To evaluate the effectiveness of the proposed guidance, 1000 Monte Carlo simulations are performed, whose initial states and errors of aerodynamic coefficients are set according to Table 3. The results are shown in Figures 1522.

As is seen in Figure 15, all of the 1000 trajectories of the glide phase are in the range of the - corridor, and naturally, in Figures 1619, all of the path constraints are satisfied. Moreover, the AOA and bank angle in different trajectories take on a continuous trend as is shown in Figures 17 and 18, which indicates that the output of the policy network is steady and robust.

The statistical data of terminal constraints is shown in Table 7.

The detailed data is shown in Figures 2325.

Figure 23 shows the terminal time error and height error, which implies that all the time errors are in the required range of 20 s and all the height errors are in the required range of 2 km. Figures 24 and 25 show histograms of the terminal time error and height error, which implies that the proposed method can satisfy all the terminal constraints.

6. Conclusions

In this paper, an intelligent multiconstrained reentry guidance is developed to generate super real-time guidance commands with no need for a decoupling method. First, the magnitude and sign of the bank angle are combined as the sole guidance command rather than the form of longitudinal guidance and lateral guidance. Then, a policy network is constructed to output the guidance command with the help of DRL. Based on DDPG and HER, the terminal constraints are transformed into multigoals. The improved HER method shows more excellent performance in the training simulation. In simulations, the policy network in the proposed guidance can give commands online in less than 1/68 of NPCG time. The robustness to initial state bias and online aerodynamic errors is excellent. However, the deflection of bias and errors is limited in the range of the training phase. If online parameters are beyond the range, the policy network should be retrained. The relation between the scale of the network and the deflection range deserves further study, and the method proposed in this paper can be used as a reference.

Data Availability

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

This work was supported by the National Natural Science Foundation of China under Grant No. U20B2005.