Abstract
The rapid development of hypersonic vehicles has motivated the related research dramatically while the evasion of the hypersonic vehicles becomes one of the challenging issues. Different from the work based on the premise that the pursuers’ information is fully known, in this paper the evasion guidance for air-breathing hypersonic vehicles (AHVs) against unknown pursuer dynamics is studied. The gradient descent is employed for parameter estimation of the unknown dynamics of the pursuer. The energy-optimized evasion guidance algorithm is further developed by taking the acceleration constraint and energy optimization into consideration. Under the proposed algorithm, the system can deal with the unknown pursuer dynamics effectively and provide more practical guidance for the evasion process. The simulation results show that the proposed method can enable the AHV to achieve successful evasion.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
The rapid development of hypersonic vehicles has motivated the relate research dramatically [1, 2]. In the past decades, the advances of the hypersonic weapons drive the revolution of the abilities and strategies of pursuers. Therefore, the hypersonic vehicles will be confronted with rigorous pursuit threat, and it is significant to carry out the research upon the evasion strategies of the hypersonic vehicles.
As an emerging research area, the research on the evasion strategies of hypersonic vehicles is few. The few existing studies [3,4,5] are all aimed at known waypoints and no-fly zones, utilizing trajectory planning to achieve evasion. But the research on pursuit-evasion (PE) games of hypersonic vehicles is almost blank. However, the evasion guidance laws in PE games of some other offensive weapons (such as ballistic missiles and cruise missiles) have been extensively studied, mainly based on optimal control theory and differential game method.
The optimal control theory has been widely applied in the research of PE games [6,7,8,9,10,11,12]. In Ref. [6], the optimal evasion strategy from proportionally guided missiles is proposed under the assumption of two-dimensional linearized kinematics. The analytical expression of the specific maneuvering moment is provided, demonstrating that the maneuvering switching times are related to the pursuer's proportional guidance coefficient. And the three-dimensional optimal evasion strategy in the case of linear kinematics model is investigated in Ref. [7]. Reference [8] puts forward the optimal evasion strategy with a path-angle constraint and against two pursuers. The optimal acceleration command is a bang-bang non-singular one governed by a switching function. And the number of switching points and their location is dependent on the dynamics of the evader and the pursuers, and the constraints of the problem. Considering the physical constraints of angular velocity and angle of attack, Ref. [9] carries out the research of an optimal three-dimensional avoidance trajectory. Similarly, in [10,11,12], the optimal control theory are used to construct the guidance laws.
There are also some evasion strategies derived by the differential game method [13,14,15,16,17,18,19,20]. The linear quadratic PE games with terminal velocity constraints is discussed by differential game method in Ref. [13]. A large family of feedback solutions has been obtained from a simple-looking performance index that contains a weighted combination of the pursuer's control effort, the evader's control effort, the miss distance, and the terminal lateral velocity. In Ref. [14], the problem of capturing a maneuvering target is formulated as a non-cooperative zero-sum differential game. State dependent riccati equation technique is used to obtain the guidance law directly from the nonlinear state equations, without any requirement of linearizing them or making restrictive assumptions. And implementing the proposed guidance law does not require time-to-go estimate. Reference [15] poses the problem of intercepting a maneuvering target at a prespecified impact angle in nonlinear zero-sum differential games framework. A feedback form solution is proposed by extending state-dependent riccati equation method to nonlinear zero-sum differential games. In addition, Ref. [16,17,18] studied multiplayer PE games from other perspectives based on differential games.
In the aforementioned studies [3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18], the evasion tasks are all achieved by maximizing the miss distance or the line-of-sight angle rate. However, for the purpose of energy optimization and the subsequent combat missions after the evasion, maximizing the miss distance may cause the undesired energy consumption and deviation from the original course, may not be the best choice. In recent years, by introducing the concept of evasion with specified miss distance (SMD), some evasion strategies about PE games has attempted to evasion with minimum energy consumption [19,20,21]. In Ref. [19,20,21], to obtain the analytical expression of the evasion guidance command, the miss distance constraint is regarded as an equality constraint. As a result, the miss distance can be controlled to a precise value while placing excessive requirements on evader’s maneuverability. Therefore, the existing evasion strategies with SMD have limited scope of application.
On the other hand, in the above-mentioned studies [3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21], all the evasion strategies are developed on the basis of the perfect information of the pursuers. However, the dynamic information of the pursuers is unknown to the evader in the real scene. This problem can be solved by introducing machine learning. In recent years, computation intelligence have been applied in theoretical research on hypersonic vehicles [22,23,24] and multiplayer PE games [25,26,27,28]. However, Ref [22,23,24] are all about the intelligent control of hypersonic vehicles, while Ref. [25,26,27,28] studied the cooperative pursuit strategy of multiple pursuers. The above-mentioned studies do not deal with the evasion guidance of hypersonic vehicles. Regarding the dynamic uncertainty estimation problem mentioned above, Ref. [29,30,31] employ machine learning to estimate the model dynamics. However, Ref. [29,30,31] are not the evasion-faced information estimation research and are difficult to use in the PE games of hypersonic vehicles.
Based on the studies mentioned above, it is obvious that, on the premise that the dynamics of the pursuer is unknown and the evader’s maneuver ability is not dominant, the existing strategies cannot achieve the evasion while considering the acceleration constraints and energy optimization simultaneously. It has to be pointed out that, Ref. [3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18] all try to achieve the evasion by maximizing the miss distance or the line-of-sight angle rate. However, it is difficult to balance acceleration constraints and energy optimization in this case. Reference [19,20,21] derived the evasion strategies with SMD, but it has excessive requirements for evader’s maneuverability, which is not suitable for the AHV for the reason of acceleration constraint. In addition, Ref. [3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21] assumes that the pursuers’ information is fully known, but does not take into account the uncertainty of pursuer dynamics. And in the studies [22,23,24,25,26,27,28,29,30,31] of computational intelligence, the estimation of the uncertainty of pursuer dynamics is not addressed.
In view of the above-mentioned discussion, while considering the flight characteristics of the AHVs studied in this paper, the main problems inside the work for evasion guidance are as follows:
-
(i)
The unknown pursuer dynamics will affect the results of PE game.
-
(ii)
The acceleration constraint must be guaranteed during the whole process of evasion.
-
(iii)
In order to better perform the subsequent combat missions after evasion, it is necessary to perform energy optimization during the process of evasion.
Compared with the existing studies (assuming that the enemy’s dynamics are known), in this paper the gradient descent method is applied in this article to estimate the unknown dynamics parameters instead of directly assuming as known. At the same time, the flight capabilities of both aircrafts are taken into account during the design process of the guidance law.
2 Problem formulation and preliminaries
2.1 Pursuit-Evasion game model
In this paper, it is assumed that both the AHV and the pursuer are within the detection range of each other, and the evasion-pursuit relationship and the definition of relative angles in horizontal plane are shown in Fig. 1.
The AHV and the pursuer are regarded as mass points, and the subscripts E & P indicate the state of the AHV and the pursuer, respectively. Both flight speeds VE & VP are assumed to be constant, and the flight path varies with lateral accelerations \(a_{E}\, \& \, a_{P}\) which are normal to velocity direction. \(a_{{E,c}} \, \& \,a_{{P,c}}\) are guidance commands and \(\theta _{E} \, \& \,\theta _{P}\) are the trajectory deflection angles, respectively. r denotes the relative distance between the AHV and the pursuer.
The combat scenario is set as that the AHV under head-on pursuit situation (HOPS). Therefore, it is assumed that the AHV and the pursuer are both on the X axis at initial time \(t_{0}\), with almost the opposite moving directions.
It is assumed [19, 20, 32, 33] that the dynamic equations of both sides can be linearized near the initial line of sight, and can be established:
where \(a_{i}^{ \bot } = a_{i} \cos \theta _{i} ,i = E,P\). The coefficients \(A_{i} \, \& \,b_{i}\) in Eq. (1) can be regard as constants. According to Fig. 1, the state variable can be chosen as \({\mathbf{x}} = \left[ {\begin{array}{*{20}c} y & {\dot{y}} & {{\mathbf{x}}_{E} } & {{\mathbf{x}}_{P} } \\ \end{array} } \right]^{{\text{T}}}\), where \(y = y_{E} - y_{P}\) and the position deviation normal to the initial line of sight. \({\mathbf{x}}_{E}\) and \({\mathbf{x}}_{P}\) denote the state variable vectors of the AHV and the pursuer, respectively. The relative motion equations can be expressed as:
where
The initial state \({\mathbf{x}}\) at \(t_{0}\) can be given by
To facilitate the analysis of the problem, the time to go is defined as
where \(t_{f}\) is the terminal engagement time point and \(V_{r}\) is the closing speed between the two sides.
In order to simplify expressions, Eq. (2) can be further written as
where \(u\, \& \,v\) are the guidance commands \(a_{{E,c}} \, \& \,a_{{P,c}}\) of both sides, respectively. The position deviation is expressed by scalar output:
where \({\mathbf{c}}^{{\mathbf{T}}} = [\begin{array}{*{20}c} 1 & 0 & {0_{{1 \times n_{E} }} } & {0_{{1 \times n_{P} }} } \\ \end{array} ]\). The absolute value \(\left| {y(t_{f} )} \right|\) is regarded as the miss distance.
In this paper, it is assumed that the guidance law of the pursuer is known, which is given as
On the basis of Eq. (1), in order to simplify the derivation process, the dynamic equation of the pursuer can be expressed as
where the coefficients \(A_{P} \, \& \,b_{P}\) in Eq. (9) are unknown to the evader.
Remark 1
The scramjet is used as the power system of the AHV. If the evasion maneuver occurs in the longitudinal plane, the sharp change of attitude angle and angular velocity will affect the working state of scramjet [34, 35]. Lateral maneuver can be carried out at a fixed altitude and fixed speed to avoid the impact on the attitude angle and angular velocity due to changes in speed and altitude during evasion. In some existing literature about the evasion of hypersonic vehicles, the modelling and simulations are conducted in 2-dimension [32, 33].
Remark 2
In the PE (pursuit-escape) game model, in order to simplify the derivation process, the dynamic equations of aircrafts can be expressed by first-order equations, and the dynamic parameters in the equations are constants. This way of expression can be seen in many studies [19, 20, 32, 33, 36, 37] about PE (pursuit-escape) games.
Remark 3
Since the values of \(A_{P} \, \& \,b_{P}\) in Eq. (9) depend on the acceleration response speed of the pursuer, in the real scene, the dynamic information of the pursuer is unknown to the evader.
2.2 Design goals
\(M\) is chosen as the lower bound of miss distance for successful evasion, and the condition for successful evasion can be shown as:
The acceleration command for the AHV is limited and can be given by
The energy consumption of the AHV during the evasion process is given by
In summary, the main problem in this paper can be summarized as Problem 1.
Problem 1
Consider the PE game model given by Eq. (6) with unknown pursuer dynamics given by Eq. (9), the evasion strategy should be derived to minimize the energy consumption given by Eq. (12), while the miss distance subject to Eq. (10) and the control constraint subject to Eq. (11).
3 Main results
3.1 The structure of evasion strategy
The design of evasion strategy in this article is divided into two steps. In details, we firstly collected the flight data of the pursuer and utilized the gradient descent method to train a estimation model of the unknown pursuer dynamics. Then the energy-optimized evasion guidance algorithm while considering system constraints is derived on the basis of the estimated parameters.
The framework of the evasion strategy design in this paper is shown as Fig. 2.
3.2 The estimation of unknown pursuer dynamics
According to Eq. (9), the pursuer’s dynamic coefficients \(A_{P} \, \& \,b_{P}\) need to be estimated.
The key to the estimation of unknown coefficients in the equation is how to determine the influence of two or more related inputs on the corresponding output.
In this paper, for the two inputs \(a_{P} \, \& \,a_{{Pc}}\) in Eq. (9), the influence of \(a_{P} \, \& \,a_{{Pc}}\) on the output is determined through multiple iterations of a large amount of data. That is, the machine learning is used to obtain the coefficients \(A_{P} \, \& \,b_{P}\) in Eq. (9).
As one of the basic algorithms of machine learning, the gradient descent method has the following advantages: the standard deviation is low, the number of iterations required for convergence is small, and the learning efficiency will not decrease [38, 39]. The basic idea of the gradient descent algorithm is to find the minimum value of the loss function according to the direction of the gradient descent of the loss function. Therefore, the gradient descent algorithm is employed to learn the pursuer dynamics in this work.
In this paper, for calculating the unknown pursuer dynamics in Eq. (9), set \(h_{\theta } (x_{0} ,x_{1} ) = \dot{a}_{P} = A_{P} a_{P} + b_{P} a_{{Pc}}\), then we have
where \(x_{0} \, \& \,x_{1}\) correspond to \(a_{P} \, \& \,a_{{Pc}}\), respectively, \(\theta _{0} \, \& \,\theta _{1}\) correspond to the unknown parameters \(A_{P} \, \& \,b_{P}\), respectively.
Equation (13) is the multiple regression equation. In Eq. (13), the response variable \(h_{\theta } (x)\) is shown with independent variables \(x_{i} (i = 0,1)\), \(\theta _{i} (i = 0,1)\) is a constant value. And, for model building process to minimize the differences between the observed and the predicted response values, coefficients \(\theta _{0} \&\theta _{1}\) are calculated.
For the multiple regression equation Eq. (13), when the number of samples is determined, there are \((x_{0}^{{(0)}} ,x_{1}^{{(0)}} ,y_{0} )\), \((x_{0}^{{(1)}} ,x_{1}^{{(1)}} ,y_{1} )\), …, \((x_{0}^{{(m)}} ,x_{1}^{{(m)}} ,y_{m} )\). Then the loss function can be expressed as
Calculate the partial derivative of the current loss function
Define \(\alpha\) as the update step size, representing the learning rate. When the learning rate is determined, the expression of \(\theta _{i}\) is updated to
Based on Eqs. (13–16), the steps to estimate \(A_{P} \, \& \,b_{P}\) can be expressed as follows.
-
1.
Establish a pursuit-evasion game simulation system, conduct pursuit-evasion tests under unknown dynamic parameters, and collect sample data through the historical game process.
-
2.
Take Eq. (13) as the multiple regression equation, according to the calculation method in Eqs. (14–16), calculate the loss function’s partial derivative \(\frac{{\partial J(\theta _{0} ,\theta _{1} )}}{{\partial \theta _{i} }}\) and the update expression of \(\theta _{i}\).
-
3.
The minimum value of the loss function \(J(\theta _{0} ,\theta _{1} )\) and the corresponding coefficients \(\theta _{0} \, \& \,\theta _{1}\) can be obtained through iterative calculation(data comes from step 1). And \(\theta _{0} \, \& \,\theta _{1}\) are equivalent to \(A_{P} \, \& \,b_{P}\) in Eq. (9).
3.3 The evasion guidance command
Considering the fact that AHVs’ maximum acceleration is inferior to the pursuer’s, it poses a great challenge to the evasion of the AHV. As a result, effective evasion algorithms are required for achieving successful evasion while giving consideration to the acceleration constraints and energy optimization simultaneously.
For the purpose of energy optimization, the evasion command \(u_{1}\) is designed by introducing the SMD. Giving consideration to the acceleration constraints, the evasion command \(u_{2}\) based on NDG is designed. When the pursuer's capability is strong enough to cause the AHV's evasion command \(u_{1}\) to saturate, the evasion command is switched from \(u_{1}\) to \(u_{2}\). The composition structure of evasion command is shown as Fig. 3.
To makes the proposed framework more clear, the following explanations are provided. For the goal of the evasion with minimum energy consumption, the guidance command \(u_{1}\) should be given priority. However, when the pursuer’s capability is strong enough that the AHV cannot achieve successful evasion with \(u_{1}\), the evasion command \(u_{2}\) is employed.
Accordingly, the composite evasion command can be given by
where \(t_{1}\) is the moment when \(u_{1}\) reaches saturation for the first time(\(\left| {u_{1} (t_{1} )} \right| \ge u_{{\max }}\)).
This section is divided into two parts. Firstly, the derivation of SMD—based evasion command \(u_{1}\) is presented. Secondly, the derivation of NDG—based evasion command \(u_{2}\) is presented. With Eq. (17) and the expression of \(u_{1}\) and \(u_{2}\), the composite guidance command \(u_{c}\) is generated.
3.3.1 Derivation of SMD–based evasion command in evasion strategy
In this subsection, the process of deriving \(u_{1}\) is given.
Substituting Eq. (8) into Eq. (6) yields
To facilitate the derivation process, the zero effort miss distance \(Z_{1} (t)\) is introduced here[40]. According to Eq. (18), the expression of \(Z_{1} (t)\) and its derivative can be given by
with the boundary conditions
where \(\phi _{1} ( \cdot , \cdot )\) is the state transition matrix, corresponding to the state matrix \(\left( {A + B_{v} F} \right)\).
After the transition, the miss distance can be expressed as \(\left| {Z_{1} (t_{f} )} \right|\) and the condition for successful evasion of the AHV is given by
On this basis, Problem 1 in Sect. 2.2 can be convert to Problem 2.
Problem 2
In line of the PE game model given by Eq. (20), the evasion strategy should be derived to minimize the energy consumption given by Eq. (12), while the miss distance subject to Eq. (23) and the control constraint subject to Eq. (11).
For solving the Problem2, in consideration of bounded acceleration of the AHV, the performance index is given by
subject to
where \(Q\left( {Q > 0} \right)\) is a weighted matrix to be designed. By adjusting \(Q\), the miss distance is larger than the boundary value \(M\) while \(u_{1}\) meets the acceleration constraint.
As mentioned in Sect. 2.1, the AHV and the pursuer are both on the \(X\) axis at initial time \(t_{0}\), with opposite moving directions. As a result, there is \(Z_{1} (t) = 0\). On this basis, considering the evasion condition Eq. (23), the condition that \(Z_{1} (t_{f} ) > M\) is chosen as the criteria of successful evasion.
According to Eq. (24), the analytic expression of \(u_{1}\) can be derived and expressed as
Substituting Eq. (25) into Eq. (20), and integrate the new equation, then we have
For calculating \(Z_{1} (t)\) and \(b(t)\), the vector \({\mathbf{Y}}\) is introduced to simplify computation.
Easy to prove that \({\mathbf{Y}}\) satisfies
Equation (28) is the adjoint equation of the model in Eq. (18). According to Eq. (27) and Eqs. (19, 21, 28) can be expressed as
In summary, the energy-optimized evasion command \(u_{1}\) is given by
where
3.3.2 Derivation of NDG-based evasion command in evasion strategy
In this subsection, the process of deriving \(u_{2}\) is given, based on the NDG method [41].
During the derivation, according to the principle of the NDG, the expression of \(v(t)\) cannot be assumed in advance. Therefore, different from \(Z_{1} (t)\), the zero effort miss distance \(Z_{2} (t)\) is given by
with the boundary conditions
where \(\phi _{2} ( \cdot , \cdot )\) is the state transition matrix, corresponding to the state matrix \(A\).
After the transition, the condition for successful evasion of the AHV is given by
On this basis, Problem 1 in Sect. 2.2 can be convert to Problem 3.
Problem 3
In line of the PE game model given by Eq. (33), the evasion strategy should be derived to minimize the energy consumption given by Eq. (12), while the miss distance subject to Eq. (36) and the control constraint subject to Eq. (11).
According to Fig. 3, when the NDG-based command \(u_{2}\) is employed, the AHV try to evade with maximum miss distance, and the performance index is given by
subject to
The Hamiltonian of the problem is
The adjoint equations are
On the basis of extremum principle, then we have
Substituting Eq. (40) into Eq. (33) yields
In the NDG, both sides should adopt the optimal strategy. On this basis, for the reason that the maximum acceleration of the AHV is weaker than the pursuer’s, there is
Integral on both sides of Eq. (41), we obtain
For \(\int_{t}^{{t_{f} }} {L{\text{d}}t} < 0\), there is
Substituting Eq. (44) into Eq. (40), the NDG-based evasion command \(u_{2}\) is given by
where \(Z_{2} (t)\) can be calculated by Eq. (32).
To sum up, by substituting Eqs. (3145) into Eq. (17), the composite evasion command \(u_{c}\) is given by
where
Remark 4
Compared with the existing studies [3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21], the major innovations of the proposed method are highlighted as follows.
-
(i)
In a lot of previous work [3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18], either the acceleration constraint is not taken into account, or the bang-bang control is directly adopted and the energy optimization is ignored. In contrast, the evasion strategy introduces the boundary value \(M\)(lower bound of miss distance for successful evasion) and combines the concept of SMD and NDG theory, so as to solve the energy optimization problem under the premise of acceleration constraints.
-
(ii)
In some previous work [19,20,21], the evader tries to evasion with SMD(\(Z(t_{f} ) = M\),\(Z(t_{f} )\) is the zero effort miss distance). But the equation constraint \(Z(t_{f} ) = M\) has excessive requirements for the evader’s maneuverability. In this paper, the conditions of successful evasion are seen as an inequality \(Z(t_{f} ) > M\). And the inequality constraint \(Z(t_{f} ) > M\) does not have excessive requirements for the evader’s maneuverability. Therefore, the evasion command in this work is more practical and has wider applicability.
-
(iii)
In all the above-mentioned previous work [3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21], the evasion strategies are developed on the basis of the perfect information of the pursuers. In this paper, by introducing the gradient descent method, the dynamic information of the pursuer is estimated. With the estimation, a more realistic dynamic model of pursuer is obtained. On the basis of the estimated model, the evasion strategy is designed. As far as authors know, the evasion strategy in this work is the first one which can achieve the successful evasion against the pursuer with unknown dynamics.
4 Results and discussion
4.1 Simulation of pursuer dynamics estimation
In the process of pursuer dynamics estimation, the relevant simulation parameters are set as follows:
-
1.
Without prior information, the initial values of coefficients \(\theta _{i}\) are randomly selected as \(\theta _{0} {\text{ = }}7\),\(\theta _{1} {\text{ = }}2\).
-
2.
The learning rate \(\alpha\) needs to be carefully selected. If \(\alpha\) is too small, the function \(J(\theta _{0} ,\theta _{1} )\) will decrease slowly and consume a lot of resources. If \(\alpha\) is too large, the function \(J(\theta _{0} ,\theta _{1} )\) may not converge. In this paper, \(\alpha = 0.001\).
The estimation results of pursuer dynamics are shown in Figs. 4, 5 and 6 And the analysis is as follows:
-
1.
It can be seen from Fig. 4 that the loss function converges to the order of 0.005, so the result of parameter estimation can be considered to be with high accuracy.
-
2.
According to Figs. 5 and 6, after the 6380th and 6144th iterations, the parameters \(\theta _{0} \, \& \,\theta _{1}\) converged to -0.5077 & 0.5073, respectively.
-
3.
As can be seen in Figs. 4, 5 and 6, in the first 10 times iterative calculation, the numerical curves of all parameters have severe oscillation. The reason is that the initial values of \(\theta _{0} \, \& \,\theta _{1}\) are randomly selected. However, the oscillation will not affect the accuracy of the final result. It can be seen that after several iterations, the oscillation has completely disappeared.
-
4.
In summary, there is \(\hat{A}_{P} = \theta _{0} = - 0.5077\), \(\hat{b}_{P} = \theta _{1} = 0.5073\). Substituting the value of \(\hat{A}_{P} \, \& \,\hat{b}_{P}\) into Eq. (9), then the estimation result of the pursuer dynamic model can be expressed as: \(\dot{a}_{P} = A_{P} a_{P} + b_{P} a_{{Pc}} \approx - 0.5077a_{P} + 0.5073a_{{Pc}}\)
4.2 Simulation of evasion guidance of the AHV
In the process of evasion guidance, the relevant assumption and simulation parameters are set as follows:
-
1.
The augmented proportional navigation (APN) guidance law is selected as the guidance law of the pursuer.
-
2.
The basic parameters of both sides and simulation parameters of the engagement scenario in this paper are shown in Tables 1 and 2, respectively.
In Table 1, \(a\) is the speed of sound. The value of \(\left| {a_{{P\max }} /g} \right|\) is set as 8.0 for the sake of guaranteeing the robustness of evasion strategies. In the existing literature, in order to meet the hit-to-kill condition for the pursuit, the maneuverability of pursuers should be more than three times the AHV’s acceleration [42]. For the sake of guaranteeing the robustness of evasion strategies in this paper, the value of \(\left| {a_{{P\max }} /g} \right|\) is set as 8.0. In addition, in order to meet the conditions of HOP, the initial value of trajectory deflection angle \(\theta _{{E_{0} }}\) is set between \(\left[ {0,10} \right]\)°.
In Table 2, \(M\) is set as 1 m which is the maximum distance to meet the hit-to-kill condition. \(r_{0}\) is set as \(\left[ {60,100} \right]\) km due to the fact that pursuers can detect and track the evader within 100 km.
In this section, the composite evasion command \(u_{c}\) (Eq. 46) is simulated. (The value of \(u_{c}\) is based on the estimation result in Sect. 4.1.) The key index parameters \(\left| {y(t_{f} )} \right|\) (miss distance) is shown below for evaluating the success of evasion. As a comparison, the simulation results of \(u_{1}\) (Eq. (31)) and \(u_{2}\) (Eq. (45)) are also given below.
where \(u_{c}\): the composite evasion command. \(u_{1}\): the SMD-based evasion command. \(u_{2}\): the NDG-based evasion command.
The simulation results are given in Figs. 7, 8, 9, 10, 11 and 12 and Table 3. The analysis is as follows:
-
1.
According to Eq. (10), the criterion for successful evasion is \(\left| {y(t_{f} )} \right| > M = 1m\).The flight trajectories with different evasion commands are respectively given in Figs. 7, 8, 9, from which we can see that the AHV can achieve successful evasion with \(u_{2}\) or \(u_{c}\),but not with \(u_{1}\) (even if \(u_{1}\) does not meet the acceleration constraint in Fig. 10. Therefore, it can be concluded that \(u_{2}\) and \(u_{c}\) have stronger evasion capabilities than \(u_{1}\).
-
2.
It can be seen from Fig. 11 that different evasion commands have a greater impact on the AHV’s trajectory deflection angle at the end of the evasion. Compared with \(u_{1}\) and \(u_{c}\), the evasion command \(u_{2}\) produces a large trajectory deflection angle. Considering the AHV's fast flight speed, a large trajectory deflection angle will cause the AHV to deviate from the original route to a great extent. Moreover, combining the energy consumption in Table 3, it can be seen that \(u_{2}\) can enable the AHV to achieve successful evasion while paying great cost, which is not conducive to the combat mission after evasion.
-
3.
As can be seen in Fig. 12, as the initial trajectory deflection angle \(\theta _{{E_{0} }}\) and initial distance \(r_{0}\) change, the AHV can still achieve successful evasion with \(u_{c}\). Therefore, the evasion command \(u_{c}\) can enable the AHV to complete the evasion for different initial conditions.
In summary, compared to \(u_{1}\) and \(u_{2}\), the proposed command \(u_{c}\) is the only guidance command that can enable the AHV to achieve the successful evasion while considering acceleration constraints and energy optimization simultaneously.
5 Conclusions
In this paper, considering the unknown dynamics of the pursuer, a novel evasion strategy based on the estimation of the pursuer dynamics is proposed for the AHV. The structure of proposed evasion algorithm mainly combines the estimation of unknown pursuer dynamics based on gradient descent and an composite evasion command based on SMD and NDG. Through the gradient descent method, the dynamic model of the pursuer was obtained. On the basis of the estimated model, an evasion strategy that satisfies the acceleration constraint and energy optimization at the same time was designed. The simulation results show that, with the proposed evasion algorithm, the AHV can achieve successful evasion against the pursuer with unknown dynamics, and take into account the energy optimization and acceleration constraint simultaneously. The future direction for this work can be the strategies in the multiplayer PE games. Specifically, the cooperative strategy of multiple teammates and pursuit-evasion strategies against multiple enemies can be studied with the aid of computation intelligence in the future.
Data Availability
The data used to support the findings of this study are included within the article.
References
Qiu YF, Li HX, Li XQ (2021) Research status and technical research trend of hypersonic missile. J Innov Soc Sci Res 8(1):172–175
Wang LP, Zhao S, Jiang CJ et al (2020) Developments of hypersonic weapon abroad. Aerosp Electron Warfare 36(5):61–64
Zhang KN, Zhou H, Chen WC (2012) Trajectory planning for hypersonic vehicle with multiple constraints and multiple maneuvering penetration strategies. J Ballist 24(3):85–90
Wang Q, Muo HD, Wu ZD et al (2015) Predictive reentry guidance for hypersonic vehicles considering no-fly zone. J Harbin Inst Technol 47(2):104–109
Gao CS, Chen EK, Jing WX (2017) Maneuver evasion trajectory optimization for hypersonic vehicles. J Harbin Inst Technol 49(4):16–20
Shinar J, Steinberg D (1977) Analysis of optimal evasive maneuvers based on a linearized two-dimensional kinematic model. J Aircr 14(8):795–802
Shinar J, Rotsztein Y, Bezner E (1979) Analysis of three-dimensional optimal evasion with linearized kinematics. J Guidan Control 2(05):353–360
Cliff EM, Kelley HJ, Ben-Asher J (1988) Optimal evasion with a path-angle constraint and against two pursuers. J Guidance 11(4):300–304
Raivio T, Ranta J (2002) Optimal missile avoidance trajectory synthesis in the endgame. AIAA Guidance, Navigation, and Control Conference and Exhibit, Monterey, California, 5–8 August 2002.
Remzi AKDAG, Turgay ALTILAR (2006) Modeling evasion tactics of a fighter against missiles in three dimensions. AIAA guidance, navigation, and control conference, Keystone, Colorado, 21–24 August, 2006
Karelahti J, Virtane K, Raivio T (2007) Near-optimal missile avoidance trajectories via receding horizon control. J Guid Control Dyn 30(05):1287–1298
Sun AK, Liu HHT (2008) Multi-Pursuer Evasion. AIAA Guidance, Navigation and Control Conference, Honolulu, Hawaii, 18–21 August 2008.
Ben-Asher JZ (1996) Linear quadratic pursuit-evasion games with terminal velocity constraints. J Guid Control Dyn 19(2):499–501
Bardhan R, Ghose D (2015) An SDRE based differential game approach for maneuvering target interception. AIAA Guidance, Navigation, and Control Conference, Kissimmee, Florida, 5–9 January 2015
Bardhan R, Ghose D (2015) Nonlinear differential games-based impact-angle-constrained guidance law. J Guid Control Dyn 38(3):384–402
Exarchos I, Tsiotras P, Pachter M (2016) UAV collision avoidance based on the solution of the suicidal pedestrian differential game. AIAA Guidance, navigation, and control conference, San Diego, California, USA, 4–8 January 2016
Shaferman V, Shima T (2017) A cooperative differential game for imposing a relative intercept angle. AIAA Guidance, Navigation, and Control Conference, Grapevine, Texas, 9–13 January 2017
Garcia E, Fuchs ZE, Milutinovic D et al. A Geometric Approach for the Cooperative Two-Pursuer One-Evader Differential Game. IFAC-Papers OnLine, vol. 50, no. 1, pp. 15209–15214, 2017, ISSN 2405–8963
Weiss M, Shima T (2016) Minimum Effort pursuit/evasion guidance with specified miss distance. J Guid Control Dyn 39(5):1069–1079
Weiss M, Shima T, Castaneda D, Rusnak I (2017) Combined and cooperative minimum-effort guidance algorithms in an active aircraft defense scenario. J Guid Control Dyn 40(5):1241–1254
Guo H, Fu WX, Fu B (2019) Smart homing guidance strategy with control saturation against a cooperative target-defender team. J Syst Eng Electron 30(2):366–383
Xu B, Wang X, Shi ZK (2021) Robust Adaptive Neural Control of Non-Minimum Phase Hypersonic Vehicle Model. IEEE Trans Syst Man Cybern Syst 51(2):1107–1115
Xu B, Shi ZK, Sun FC et al (2019) Barrier lyapunov function based learning control of hypersonic flight vehicle with AOA constraint and actuator faults. IEEE Trans Cybern 49(3):1047–1057
Xu B, Wang DW, Zhang YM et al (2017) DOB-based neural control of flexible hypersonic flight vehicle considering wind effects. IEEE Trans Industr Electron 64(11):8676–8685
Wang Y, Dong L, Sun C (2020) Cooperative control for multi-player pursuit-evasion games with reinforcement learning. Neurocomputing 412:101–114
Singh G, Lofaro D, Sofge D (2020) Pursuit-evasion with decentralized robotic swarm in continuous state space and action space via deep reinforcement learning. 12th International conference on agents and artificial intelligence, pp 226–233
Tan L, Gong QH, Wang HX (2018) Pursuit-evasion game algorithm based on deep reinforcement learning. Aerospace Control 36:3–8
Al-Talabi, Ahmad A (2017) Fuzzy reinforcement learning algorithm for the pursuit-evasion differential games with superior evader. 2017 International automatic control conference (CACS)
Zhang J, Wang H, Zheng J et al (2020) Adaptive sliding mode-based lateral stability control of Steer-by-Wire vehicles with experimental validations. IEEE Trans Veh Technol 69(9):9589–9600
Ye M, Wang H (2020) Robust adaptive integral terminal sliding mode control for steer-by-wire systems based on extreme learning machine. Comput Elect Eng 86
Hu Y, Wang H, Cao Z et al (2019) Extreme-learning- machine-based FNTSM control strategy for electronic throttle. Neural Comput Appl 32:14507–14518
Yan T, Cai Y, Xu B (2020) Evasion guidance algorithms for air-breathing hypersonic vehicles in three-player pursuit-evasion games. Chin J Aeronaut. https://doi.org/10.1016/j.cja.2020.03.026
Yan T, Cai Y (2019) General evasion guidance for air-breathing hypersonic vehicles with game theory and specified miss distance. IEEE 9th Annual International Conference on CYBER Technology in Automation, Control, and Intelligent Systems
Dalle DJ, Frendreisy SGV, Driscoll JF, et al. (2010) Hypersonic vehicle flight dynamics with coupled aerodynamics and reduced-order propulsive models. AIAA Atmospheric Flight Mechanics Conference. 2010, August, 2010.
Brocanelli M, Gunbatar Y, Serrani A et al. (2012) Robust control for unstart recovery in hypersonic vehicles. AIAA Guidance, Navigation and Control Conference. 2012, no. August 2012, pp. 1–20.
Prokopov O, Shima T (2013) Linear quadratic optimal cooperative strategies for active aircraft protection. J Guid Control Dyn 36(3):753–764
Zheng WT, Fonod R, Shima T et al (2018) Cooperative guidance law for target pair to lure two pursuers into collision. J Guidance Control Dyn 41(8):1687–1699
Chen J, Gan M, Zhu Q et al (2021) Robust standard gradient descent algorithm for arx models using aitken acceleration technique. IEEE Trans Cybern. https://doi.org/10.1109/TCYB.2021.3063113
Lei Y, Tang K (2021) Learning rates for stochastic gradient descent with nonconvex objectives. IEEE Trans Pattern Anal Mach Intell. https://doi.org/10.1109/TPAMI.2021.3068154
Shaferman V, Shima T (2008) Linear quadratic differential games guidance law for imposing a terminal intercept angle. AIAA Guidance, Navigation and Control Conference Exhibition August, pp. 1–19.
Turetsky V, Shinar J (2003) Missile guidance laws based on pursuit-evasion game formulations. Automatica 39(4):607–618
Sun L, Huang KX, Chang XF et al (2015) Researching interception of near-space hypersonic cruise missile. J Northwest Polytechn Univer 33(4):615–620
Acknowledgments
This work was supported by National Natural Science Foundation of China(Nos.90816027, 61503302, 61933010); Aeronautical Science Foundation of China (No. 20160153002).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflicts of Interest
The authors declare that there is no conflict of interest regarding the publication of this paper.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Yan, T., Cai, Y. & Xu, B. Evasion guidance for air-breathing hypersonic vehicles against unknown pursuer dynamics. Neural Comput & Applic 34, 5213–5224 (2022). https://doi.org/10.1007/s00521-021-06250-w
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00521-021-06250-w