Keywords

1 Introduction

The appearance of new system radar brings new requirements to the UCAV on electronic warfare, thus the UCAV path planning has become an urgent problem. Good path planning can improve the safety performance of the UCAV and help UCAV accomplish their tasks well.

Currently, the main methods of UCAV path planning include intelligent algorithms, and neural networks [1]. Sun proposes quantum genetic algorithm for mobile robot path planning [2]. He guides and realizes path optimization by introducing genetic operators including quantum crossover operator and quantum gate mutation operator with the essential characteristics of quantum. However, the algorithm is easy to fall into local extreme points and the convergence speed is slow. To achieve a fast search of the path, Wang uses the fuzzy neural network to plan the path of the mobile robot [3]. But in electronic warfare, there is a lack of training samples, resulting in poor applicability. In order to overcome the shortcomings of small samples, Peng proposes the 3-D path planning with Multi-constrains [4]. He takes multi-constrains into account in the planning scheme. A path is generated by searching in the azimuth space using genetic algorithm and geometry computation. It causes the algorithm to converge slowly.

With the development of artificial intelligence, more path planning algorithms have been proposed [5,6,7,8]. In [9], Q-learning-based path planning algorithm is presented to find a target in the maps which are obtained by mobile robots. Q-learning is a kind of reinforcement learning algorithm that detects its environment. It shows a system which makes decisions itself that how it can learn to make true decisions about reaching its target. However, because the limitations of the Q table of the algorithm [10], its calculation accuracy is poor. Although he uses the reward criteria, the reward values is too single, which makes it less accurate.

Under the UCAV and radar confrontation model, this paper proposes a UCAV path planning algorithm based on deep reinforcement learning to solve the UCAV signal recognition and path planning problem. In order to enable the UCAV to complete the task of identifying the radar signal while reaching the target area, the proposed method combines the neural network of deep learning with the reward criteria of reinforcement learning. We set the reasonable reward values, state values and action values of the UCAV. Then, we use ResNet migration learning to improve the recognition rate of radar signals. The Deep Q-Network is trained to realize adaptive generation of UCAV path. Finally, simulation experiments verify the effectiveness of the proposed method.

The structure of this paper is organized as follows: The model of deep reinforcement learning is described in the Sect. 2. The Sect. 3 introduces the system structure of the path planning. It includes confrontation model and training constraints. The Sect. 4 shows the simulation results. The conclusion will be discussed in Sect. 5.

2 Deep Reinforcement Learning

On one hand, the traditional reinforcement learning algorithms rely on human-involved feature design [11], on the other hand they rely on approximations of values functions and strategy functions. Deep learning, especially Convolutional Neural Networks (CNN), can extract high-dimensional features of images [12]. The Google technical team combines the CNN in deep learning with the Q-learning algorithm in reinforcement learning. They propose the Deep Q-Network (DQN) algorithm [13]. As a pioneering work of deep reinforcement learning, the DQN can finish end-to-end learning from perception to action.

Figure 1 shows the structure of the DQN algorithm. DQN effectively removes the instability and divergence caused by neural network nonlinear action values approximator. It greatly improves the applicability of reinforcement learning. First, the experience replay in the figure randomizes the data. Thereby, it can remove the correlation between the observed data, smooth the data distribution and increase the utilization of historical data. Secondly, the CNN is used to replace the traditional reinforcement learning table mechanism. By using two networks to iteratively update, the algorithm uses the DQN loss function to adjust the direction of the current network values toward the target values network. It is also periodically updated to reduce the correlation with the target network. In addition, through truncating the rewards and regularizing the network parameters, the gradient is limited to the appropriate range, resulting in a more robust training process.

Fig. 1.
figure 1

The structure of DQN model.

3 UCAV Path Planning Research Program

Aiming at the problem that the UCAV avoids the ground radar detection to effectively break through the defense area and accurately identify radar signals, the deep reinforcement learning algorithm is used to train the model to realize the automatic generation of the UCAV path. This paper considers how to plan a reasonable path for a UCAV to arrive at the target area, which makes the UCAV not detected by the two radar.

3.1 Path Planning System Structure

The structure of the path planning system of the UCAV is shown in Fig. 2. Under the background of complex electromagnetic environment, we build a model of the UCAV and radar confrontation. On the one hand, we use the radar signal-to-noise ratio to calculate the radar detection probability. On the other hand, we analyze the UCAV kinematic constraints to study the state values and action values of the UCAV in intelligent decision. These two factors are combined as the reward criteria of the UCAV path planning, which is used to train the intelligent decision network parameters of the UCAV. Finally, a real and reliable path of the UCAV is obtained according to the result.

Fig. 2.
figure 2

The structure of path planning system.

3.2 Confrontation Model and Training Constraints

In order to simulate the real confrontation environment, it is assumed that there are two random search radars. The distance between the radars satisfies:

$$ 0.5R_{\hbox{max} } \le d < 3R_{\hbox{max} } $$
(1)

where \( R_{\hbox{max} } \) is the maximum working distance of the radar.

The initial position of UCAV is fixed, and the target area is the ground hemisphere with a radius of \( d_{0} \). After establishing the battlefield environment, the training constraints for the path planning include the reward values, the state values, and the action values constraints. These three types of constraints are the basis for training the parameters of the UCAV intelligent decision network.

Kinematic constraints of the UCAV

Set the random speed vector of the UCAV to:

$$ A = [v_{x} ,v_{y} ,v_{z} ] $$
(2)

where: \( v_{x} \) is the velocity component of the north direction of the UCAV; \( v_{y} \) is the velocity component of the east direction of the UCAV; \( v_{z} \) is the velocity component of the UCAV facing upward.

The UCAV uses uniform motion:

$$ \sqrt {v_{x}^{2} + v_{y}^{2} + v_{z}^{2} } = C_{0} $$
(3)

where \( C_{0} \) is the UCAV speed values.

The coordinate position \( [x,y,z] \) of the UCAV is taken as the state values \( S \), where \( x \) represents the coordinates of the north of the UCAV, \( y \) represents the coordinates of the east, and \( z \) represents the height of the UCAV, and satisfies \( 0 \le z \le 2.5\,{\text{km}} \).

After moving for \( \Delta t \) time, according to the action constraints and the current state values, the next state values of the UCAV is:

$$ S^{'} = [x + v_{x} \Delta t,y + v_{y} \Delta t,z + v_{z} \Delta t] $$
(4)

The action values of UCAV after \( \Delta t \) time is:

$$ A^{\prime} = [v_{x}^{\prime } ,v_{y}^{\prime } ,v_{z}^{\prime } ] $$
(5)

The action values also needs to satisfy (6):

$$ \sqrt {v_{x}^{\prime2} + v_{y}^{\prime2} + v_{z}^{\prime2} } = C_{0} $$
(6)

When the state values of the UCAV satisfies (7), it can be determined that the UCAV has reached the target area:

$$ \sqrt {(x - x_{0} )^{2} + (y - y_{0} )^{2} + (z - z_{0} )^{2} } \le d_{0} $$
(7)

where:

\( [x_{0} ,y_{0} ,z_{0} ] \) is the target point coordinate;

\( d_{0} \) is the target area radius.

Signal recognition

The traditional signal recognition based on radar characteristic parameters can not meet the recognition requirements in complex electromagnetic environment in modern electronic countermeasures. In recent years, the development of deep learning, especially the widespread use of neural networks, has provided new ideas and methods for signal processing and recognition. The recognition based on neural network image features can be well applied in the field of radar signal recognition. However, the reasonable conversion of radar signals into images is the key to reliable identification.

First, generating a reliable signal is the basis for signal processing:

$$ x(t) = ae^{j\varphi (t)} + n(t) $$
(8)

where \( a \) is the amplitude, we suppose \( a = 1 \) in this paper; \( \varphi (t) \) is the instantaneous phase of the radar signal; \( n(t) \) is the white Gaussian noise. The radar communication signal are BPSK, QPSK, 8PSK, ASK, OQPSK, QAM16, QAM32, QAM64, QAM256.

The theory of Cyclic Spectrum(CS) is established based on the cyclic and stationary characteristics. The signal processing of the communication signal by the cyclic spectrum can obtain good results. Therefore, after receiving the communication modulation signal, the signal processing of the cyclic spectrum analysis is performed:

$$ S_{x}^{\alpha } (f) = \int_{ - \infty }^{\infty } {R_{x}^{\alpha } (\tau )} e^{ - j2\pi f\tau } d\tau $$
(9)

where \( \alpha \) is the cyclic frequency and \( R_{x}^{\alpha } (\tau ) \) is the cyclic autocorrelation of the signal \( x(t) \):

$$ R_{x}^{\alpha } (\tau ) = \mathop {\lim }\limits_{T \to \infty } \frac{1}{T}\int_{{ - \frac{T}{2}}}^{{\frac{T}{2}}} {x(t + \frac{\tau }{2})} x^{*} (t - \frac{\tau }{2})e^{ - j2\pi \alpha t} dt $$
(10)

Figure 3 shows nine types of radar communication signals according formula (10).

Fig. 3.
figure 3

Nine types of radar communication signals

Compared with the common convolutional neural network, the ResNet network mainly adds a shortcut connection between the input and the output, so that the network can make the subsequent layer can directly learn the residual. When the traditional convolutional layer or fully connected layer is used for information transmission, there will be problems of information loss due to inconvenient connection between input and output. ResNet solves this problem to some extent, and the ResNet network passes the input information. Therefore, we use the ResNet network to extract features from the three-dimensional spectrum of the cyclic spectrum of the communication signal. Migration learning is a new machine learning method. It has good adaptability and can improve the quality of feature extraction. Therefore, the above network application migration learning is adapted to the communication field. Figure 4 shows the migration process of the ResNet. We use pre-training model ResNet to process the signal. Finally, the system uses the Support Vector Machine(SVM) to reach the purpose of classification and identification and get the communication signal recognition rate \( \eta \).

Fig. 4.
figure 4

The migration process of the ResNet.

3.3 Reward Criteria

In order to obtain a reasonable UCAV trajectory in the network training of the model, we set the appropriate radar detection probability and the distance between the UCAV and the center of the target area as the reward values of the intelligent decision network. On this basis, we will use the processed signal recognition rate as a supplement to the reward mechanism to achieve the purpose of the multitasking of the UCAV.

The detection probability \( P_{d} \) of the radar is an important indicator for the effective penetration of the UCAV. When \( P_{d} \le 0.1 \), the radar does not find the UCAV; when \( P_{d} > 0.1 \), the radar finds the UCAV. In order to train the UCAV to reach the target area in the shortest time without being detected by the radar, the reward values of the UCAV path planning is:

$$ R({\varvec{\upomega}}) = P_{d} (\omega_{1} ) + D(\omega_{2} ) $$
(11)

where: \( D \) is the distance between the UCAV and the center of the target area; \( {\varvec{\upomega}}{ = [}\omega_{1} ,\omega_{2} ] \) are the weights of the detection probability and the flight time of the UCAV respectively. Different weights can get different reward trends. In this paper, we set the reward as:

$$ R = \frac{{\omega_{1} }}{{\log_{2} P_{d}^{ - 1} }} - \omega_{2} (D - 250)^{3} { + }2 $$
(12)

where: \( \omega_{1} { = }1,\omega_{2} { = 0} . 0 0 1 \).

However, in order to enable the UCAV to better identify the radar communication signal. We improve the reward with communication signal recognition rate \( \eta \):

$$ R = \eta (\frac{{\omega_{1} }}{{\log_{2} P_{d}^{ - 1} }} - \omega_{2} (D - 250)^{3} { + }2) $$
(13)

3.4 Path Planning Algorithm

Intelligent decision making is the core of the UCAV path planning algorithm. In the traditional reinforcement learning algorithm, the state and action space are discrete and the dimension is low. Q table can be used to store the Q values of each state action. In solving the problem of UCAV path planning, the state and action of the UCAV are highly dimensionally continuous, and the data is also very large. Due to the limitations of the Q table, it is very difficult to store data, which makes the traditional reinforcement learning algorithm cannot solve the problem of path planning. The DQN algorithm replaces the Q table by fitting a loss function so that similar states get similar action outputs. We propose that the path planning algorithm based on deep reinforcement learning, which can effectively solve the problem that the data is too large to be trained. Besides, it can break the correlation between data and improve the training efficiency. We have discussed the three elements of the DQN algorithm: state, action, and reward. Figure 4 shows the DQN algorithm flow.

Fig. 5.
figure 5

DQN algorithm flow.

In Fig. 5, the experience replay is used to learn the previous experience. We send \( (S,A,R,S^{\prime}) \) to the experience replay for learning. At the same time, we send the output reward values \( R \) to the DQN loss function. The current state action pair is sent to the current values network, and the next state \( S^{\prime} \) is sent to the target values network. The DQN loss function can be expressed by (14):

$$ L(\omega ) = E[(r + \gamma \mathop {\hbox{max} }\limits_{{A^{\prime}}} Q(S^{\prime},A^{\prime},\omega ) - Q(S,A,\omega ))^{2} ] $$
(14)

where: \( \gamma \) is attenuation coefficient of the reward values; \( \omega \) is the weight of the loss function; \( r + \gamma \mathop {\hbox{max} }\limits_{{A^{\prime}}} Q(S^{\prime},A^{\prime},\omega ) \) is the values of the target values network; \( Q(S,A,\omega ) \) is the Q values of the current values network.

It can be seen that the loss function is calculated by the mean square error of the difference between the target values and the current values. After the loss function is derived, we can calculate the gradient of the loss function:

$$ \frac{\partial L(\omega )}{\partial \omega } = E[(r + \gamma \mathop {\hbox{max} }\limits_{{A^{\prime}}} Q(S^{\prime},A^{\prime},\omega ) - Q(S,A,\omega ))\frac{\partial Q(S,A,\omega )}{\partial \omega }] $$
(15)

Therefore, we use stochastic gradient descent to update parameters to obtain an optimal values \( Q(S,A,\omega ) \). The current values network in DQN uses the latest parameters, which can be used to evaluate the values function of the current state action pair. But the target values network parameters are a long time ago. After the current values network is iterated, the UCAV takes action on the environment to update the target values network parameters according to (15). In this way, a learning process is completed. The optimal Q values is stored in the network to realize the practical application of the optimal track (Table 1).

Table 1. The process described by algorithm 1

4 Simulation Experiment and Results

After analyzing the UCAV motion constraints and the path planning algorithm, UCAV and radar confrontation model parameters are set in Table 2. The simulation experiment parameters are shown in Table 3.

Table 2. UCAV and Radar Confrontation Platform parameters

Combined with the UCAV and radar confrontation model parameters in Table 3 and the simulation experiment parameters in Table 3, we use the DQN algorithm to achieve simulation training. Then we normalize the data of the reward values. The normalization function is:

$$ f(x) = 1/(1 + e^{ - x} ) $$
(16)
Table 3. Simulation experiment parameters

On this basis, in order to observe the relation between the reward values and the episodes, we take the reward values of 10 times to do an average. Finally, we can obtain the following simulation results by combining the above parameters and conditions.

Fig. 6.
figure 6

The reward values of 300 episodes conversion curve.

Figure 6 shows the variation of the overall reward values for each episode of the UCAV. The abscissa represents the number of episode and the ordinate is the reward values after normalization. The reward values 0 and 1 represent the minimum and maximum values after normalization respectively. Figure 6 shows that compared with Q-learning, the DQN overall reward values gradually increases, and eventually stabilizes after the episodes reaches 80, and the average reward values is close to 0.9. This is because the DQN network uses the gradient descent method to correct the loss function, so that the Q values of the intelligent decision network is optimized. When the episode is 60, the reward values is abrupt. This is because after a certain number of trainings, the algorithm will re-randomly search for the optimal solution to avoid falling into local optimum.

When the episodes are 190 and 240, due to the random sampling of the experience replay in the DQN algorithm, the correlation between data is cut off. The simulation shows that the accuracy of the algorithm reaches 90%.

In each modulation category, 100 tests are implemented. It is clear from Fig. 7(a) that the classification accuracy of BPSK, QAM16, QAM64, QPSK and 8PSK is 100%. The classification accuracy of OQPSK, ASK, QAM256 and QAM64 becomes worse, especially for QAM32 and QAM256. Figure 7(b) shows the recognition rate curve based on the CS method signal recognition and the recognition rate curve based on the traditional signal recognition. This indicates that the residual neural network signal recognition based on cyclic spectrum has a good recognition rate. The proposed method is lower than the traditional method at -6 dB, but the proposed method is superior to the traditional method with the improvement of signal-to-noise ratio. This is because by transmitting the input information directly to the output, the ResNet network only needs to learn the difference between input and output, which simplifies the learning goal and difficulty, protects the information integrity to a certain extent.

Fig. 7.
figure 7

(a) Signal recognition confusion matrix under 0 dB condition; (b) The recognition rate of proposed algorithm.

Fig. 8.
figure 8

(a) The 3D renderings of the UCAV single-shot to the target; (b) The 2D renderings of the UCAV multiple times to the target. (Color figure online)

Figure 8(a) shows the effect of the UCAV reaching the target area. The black point represents the radar and the red curve represents the path of the UCAV. This shows that the UCAV can independently plan a path to the target area without being detected by the radar.

In Fig. 8(b), a light blue circle represents the target area, a red star represents the center ground projection of the target area, and a red projected point represents a of the UCAV approaching the target area. Figure 8(b) shows that after 300 episodes, the success rate of the UCAV reaching the target reaches 90%. This is because the deep reinforcement learning algorithm has good self-learning and correction capabilities. The simulation results show that the UCAV path planning algorithm based on deep reinforcement learning can independently plan a reasonable path in the unknown space.

5 Conclusion

This paper proposes a UCAV path planning algorithm based on deep reinforcement learning, which solves the problem of poor adaptability and slow calculation speed in traditional track planning algorithm. Besides, it realizes the aircraft signal recognition. The simulation results show that the proposed algorithm achieves 90% accuracy under the conditions of compromise planning time and flight quality. Besides, the signal recognition rate reaches 92.59% under 0 dB condition. It has good convergence, which can be applied in the field of modern electronic warfare.