1 Introduction

In recent years, with the advantages of large coverage and long communication distance, satellite communication technology has been widely used in global business, education and entertainment [1]. High data rates and ubiquitous services result in a large amount of energy consumption for transmitters and receivers. At the same time, the demand for fixed and mobile satellite broadband access is also growing. In order to improve the transmission efficiency of satellite communication, people put forward the idea of integrated satellite terrestrial networks [2], that is, both satellite and terrestrial network work in the S-band to save the cost of facilities and spectrum resources. In addition, the signal fading of higher frequency bands (such as Ka band) in satellite communication cannot be ignored [4]. Therefore, in the satellite communication system, the growing demand for high-speed multimedia services and the transmission characteristics of signals put forward challenges to the limited spectrum resources. More efficient and reasonable utilization of limited spectrum resources in satellite communication systems has gradually become a hot issue.

Although frequency resources are becoming more and more precious, the spectrum utilization rate of satellite communication systems is low [5]. The frequency band occupied by most fixed satellite communication systems is above 4 GHz, but the spectrum allocation strategy is unreasonable, and the test results show that the spectrum utilization rate is very low [6]. The dynamic spectrum technology based on cognitive radio can effectively solve the problem of spectrum resource utilization, which has been studied by many scholars [7]. In 2003, dynamic spectrum allocation technology based on cognitive radio is the main research direction of "next generation program" (XG) proposed by Defense Advanced Research Projects Agency [8]. Reference [9] describes the statistical interference model of terrestrial wireless communication system to satellite communication system in the satellite cognitive radio network, and evaluates the interference between two coexisting systems. Reference[10] proposed a multi beam satellite cognitive system based on beam hopping, which can significantly improve the spectrum utilization of satellite communication system. Reference [11] analyzes the future development prospects of cognitive wireless network technology in satellite communications, and points out the new application of cognitive wireless network technology and multimedia-related technologies in satellite communications.

The problem of low spectrum utilization has become a prerequisite for the introduction of cognitive radio technology. However, satellite communication systems have high requirements for the stability of signal transmission, and interference from cognitive users to authorized users should be avoided [12]. In satellite cognitive wireless networks, the signal transmission delay will cause the time lag for cognitive users to sensing authorized users. Therefore, the focus is on the behavior prediction technology of authorized users in the satellite cognitive radio network, the artificial neural network is used for behavior prediction, and a multi-step prediction algorithm based on the cascaded forward neural network is proposed.

The paper is organized as follows. In Sect. 2, we introduce the behavior prediction model of the satellite cognitive radio networks. In Sect. 3, the user behavior prediction algorithms are investigated, and a new multi-step behavior predicting algorithm based on cascade forward artificial neural network is proposed. The performance of the behavior prediction algorithms is verified in Sect. 4. Finally, the conclusion is provided in Sect. 5s.

2 Behavior Prediction Model

2.1 Behavior Prediction System Model

Assume that the cognitive network sensing the bandwidth B of the frequency band, where fmax and fmin are the upper and lower boundaries of the target frequency band, respectively. According to the transmission performance requirements of cognitive users, the target frequency band is divided into N sub-bands with bandwidth B, and B × N ≤ W. Therefore, the cognitive network base station can obtain the sensing results of N sub-bands in each sensing process. If the authorized user is occupying the sub-band, then si = 1(i = 1,2,…,N), otherwise si = 0. In addition, assume that the length of the sensing time slot of the cognitive network is ∆t. The process of establishing a spectrum database by a cognitive base station is shown in Fig. 1. The process of establishing a spectrum database by a cognitive base station needs to weigh the processing complexity, the effectiveness of spectrum resource allocation, the length of the sub-band sensing time slot ∆tand the bandwidth B.

Fig. 1
figure 1

Cognitive base station spectrum database formation process

In the dynamic spectrum access scenario of the satellite cognitive radio network, the cognitive base station uses the information sensed by the sensing node to obtain the spectrum occupancy of the sub-band, and stores it in the spectrum database. Then the user behavior prediction system uses the established spectrum database and the behavior of authorized users at the current time to predict the behavior in the next several time slots. The prediction system performs information extraction and self-learning based on the historical behavior of authorized users. The ground area covered by non-geostationary orbit satellites is periodic, and the data flow of geostationary orbit satellites also has a certain distribution law [13]. The user behavior prediction system uses the distribution law of authorized users' spectrum occupancy behavior in the time dimension as a prediction basis. In order to reduce the prediction time and improve the real-time performance of the prediction results, the prediction system needs to learn and extract features in advance, and use the current spectrum sensing results to make predictions. Therefore, for the non-geostationary satellites in the satellite cognitive radio network, the ground-based cognitive network can use the time when the satellite is not visible for feature extraction. For the geostationary satellites, the ground-based cognitive network can only use the historical information in the spectrum database.

2.2 User Behavior Model

The behavior of authorized users in cognitive radio networks is not completely random, but is affected by human activities and follows certain mathematical models [14]. Therefore, cognitive users can predict the behavior of authorized users. In traditional work, modeling the behavior of authorized users with deterministic traffic is relatively mature, while random traffic modeling usually adopts exponential Poisson distribution model and Pareto distribution model [[[15,16,17]]]. In this paper, the exponential ON/OFF model is used to model the behavior of authorized users, and on this basis, the behavior prediction algorithm is studied.

The behavior of authorized users using any sub-band i (i = 1,2,…,N) can be represented by a time continuous semi-Markov process [16]. For the sub-band i, the ON model indicate the occupied state \({T}_{ON}^{i}\) and the OFF model indicate the unoccupied state \({T}_{OFF}^{i}\) Therefore, the state of each sub-band is constantly changing between ON and OFF, and the cognitive users' sensing results can be represented by a sequence of 1 and 0 accordingly. Assumed that \({T}_{ON}^{i}\) and \({T}_{OFF}^{i}\) are described by independent and identically distributed random variables Yi and Xi respectively.

Firstly, we assume that \({T}_{ON}^{i}\) obeys the exponential distribution with the expectation of \({e}_{1}=\frac{1}{{\lambda }_{1}}:\)

$${T}_{ON}^{i}\sim {f}_{{Y}_{i}}\left(y\right)=\left\{\begin{array}{cc}{\lambda }_{1}{e}^{-{\lambda }_{1}y}& ,y\ge 0\\ 0& ,y<0\end{array}\right.$$
(1)

Similarly, suppose that \({T}_{OFF}^{i}\) obeys exponential distribution with expectation \({e}_{2}=\frac{1}{{\lambda }_{2}},\) that is:

$${T}_{OFF}^{i}\sim {f}_{{Y}_{i}}\left(x\right)=\left\{\begin{array}{cc}{\lambda }_{1}{e}^{-{\lambda }_{2}y}& ,y\ge 0\\ 0& ,y<0\end{array}\right.$$
(2)

The value \({e}_{1}\) and \({e}_{2}\) are the service parameter of authorized user, which respectively reflect the average time of sub-band in the ON and OFF state. The greater the value \({T}_{ON}^{i}\) and \({T}_{OFF}^{i},\) the longer the authorized user stays in a certain state, and vice versa, the higher the frequency of state transitions. Therefore, the process of predicting the behavior of authorized users is to use the sensing results of the binary sequence at the current moment and the past moment to estimate the value of the sequence in the next few time slots.

3 User behavior Prediction Algorithm

3.1 User Behavior Prediction Algorithm Based on Update Theory

For the exponential ON/OFF model, the update theory [18] can be used to predict the behavior of authorized user. The prediction algorithm based on update theory uses the current state of sensing result to predict the behavior in the next time slot. For sub-band i (i = 1,2,…,N), assuming that its state in the current sensing slot is si, si ϵ (ON, OFF). In the next sensing time slot, the probability that the sub-band is not occupied is Pi.

$${P}_{i}=\left\{\begin{array}{cc}\frac{{\lambda }_{1}+{\lambda }_{2}\bullet {e}^{-\left({\lambda }_{1}+{\lambda }_{2}\right)\Delta t}}{{\lambda }_{1}+{\lambda }_{2}},& {S}_{i}=ON\\ \frac{{\lambda }_{1}+{\lambda }_{2}\bullet {e}^{-\left({\lambda }_{1}+{\lambda }_{2}\right)\Delta t}}{{\lambda }_{1}+{\lambda }_{2}},& {S}_{i}=OFF\end{array}\right.$$
(3)

However, the shortcoming of this algorithm is also obvious. In order to predict user behavior, the cognitive base station must know the ON/OFF model parameters of the user behavior in advance, that is, the value of \({\lambda }_{1}\) and \({\lambda }_{2}\) must be obtained in advance. This is very difficult for base stations in satellite cognitive wireless networks, because authorized users are not obligated to inform cognitive users of their behavior models, and the authorized users 'model may change over time. In actual situations, cognitive base stations can only make predictions based on historical information in the spectrum database.

3.2 One Step Prediction Algorithm Based on Feed-forward Neural Network (FFN)

In the cognitive wireless network, the cognitive base station usually does not know the law of authorized users to occupy the spectrum, and can only predict it through the binary sequence stored in the spectrum database. The one-step prediction algorithm based on the feedforward neural network is based on this idea [19], and uses the powerful nonlinear mapping ability of the feedforward artificial neural network to use the binary sequence of the current and previous moments as the input of the network. After training and learning under a certain number of samples, the feedforward network can predict state in the next time slot.

Assume that the binary sequence S = {s1,s2,…st}, si ϵ {0,1} represents the history data stored in the spectrum database, and st is the sensing result at current moment. FNN takes the binary data [St-τ+1,St-τ+2,…St-1,St]T with the time window of length τ as input, and the output \({\widehat{s}}_{t+1}\) is the predicted state in the next time slot. The time window length τ is also called prediction order. FFN uses a sliding time window to continuously predict user behavior. As time goes by, when FFN obtains the St+1 at the next moment, it will be used to update the time window, that is, the neural network uses [St-τ+2,St-τ+3,…St,St+1]T as input to predict St+2.

Before using FFN to predict user behavior, the network needs to be trained first. The authorized user status information in the spectrum database is divided into training samples, and each group of training samples includes an input of length τ and an expected output St+2 FFN continuously adjusts the network parameters through the error back propagation (BP) algorithm to narrow the gap et between its output \({\widehat{s}}_{t+1}\) and the expected valueSt+1.

Artificial neural network is the realization of user feature extraction and nonlinear mapping in the prediction process. It is a massively parallel computing system connected by a large number of simple processors, which realizes information processing through the interaction between neurons. The basic artificial neural network consists of at least three layers, namely an input layer, an output layer and at least one hidden layer. The front and back layers are fully connected, and there is no connection between neurons in each layer.

The model of an artificial neural network with a hidden layer is shown in Fig. 2. The input layer has I neurons, the output layer has K neurons, and the hidden layer has J neurons, \({w}_{ji}^{JI}\left(j=\mathrm{1,2},\dots J; i=\mathrm{1,2},\dots ,I\right)\) represents the weight between the input layer and the hidden layer.\({b}_{j}^{J}\left(j=\mathrm{1,2},\dots ,J\right)\) and \({b}_{k}^{K}\left(k=\mathrm{1,2},\dots ,K\right)\) represent the bias of the hidden layer and the output layer, respectively.

Fig. 2
figure 2

Artificial Neural Network Model

Each input layer neuron is connected to the hidden layer, and the input signal is assigned to the hidden layer through a weight matrix. Any node j in the hidden layer performs a summation operation on the input signal and adds the offset \({b}_{j}^{J}\) to form the net input \({v}_{j}^{J}\) of the node, which is then processed with a specific transfer function. Usually the hyperbolic tangent S function φ(•) is used as the transfer function, then the output \({o}_{j}^{J}\) of the j-th hidden layer neuron can be expressed as:

$${o}_{j}^{J}=\varphi {v}_{j}^{J}=\frac{1-\mathrm{exp}\left(-{v}_{j}^{J}\right)}{1+\mathrm{exp}\left(-{v}_{j}^{J}\right)}$$
(4)

where \({v}_{j}^{J}=\sum_{i=1}^{I}{w}_{ji}^{JI}{o}_{j}^{I}+{b}_{j}^{J},\) and \({o}_{j}^{I}\) is the output of i-th neuron in the input layer. Then \({o}_{j}^{I}(j=\mathrm{1,2},\dots ,J)\) undergoes the same process in the output layer, the final output signal \({o}_{k}^{K}=\varphi \left({v}_{k}^{K}\right),(k=\mathrm{1,2},\dots ,K)\) of the neural network can be obtained.

In order to make the actual output \({o}_{k}^{K}\) close to the expected value dk, the weight of the network needs to be adjusted through the BP algorithm. Define the cost function ε as the energy of the error between the actual output and the expected value.

$$\varepsilon =E=\frac{1}{2}\left[\sum_{k=1}^{K}{\left({d}_{k}-{o}_{k}^{K}\right)}^{2}\right]$$
(5)

The neural network gradually reduces the cost function \(\varepsilon\) by adjusting the weight of each layer. The BP learning algorithm uses a gradient descent method to modify the weights layer by layer through continuous iteration. For the output layer, the process of BP algorithm to modify its weight \({v}_{kj}^{KJ}\) is as follows:

$${w}_{kj}^{KJ}\left(n+1\right)={w}_{kj}^{KJ}\left(n\right)+{\Delta w}_{kj}^{KJ}$$
(6)

where \({w}_{kj}^{KJ}\left(n\right)\) is the weight of the layer at the n-th iteration. \({\Delta w}_{kj}^{KJ}\) is the correction amount of the weight, which satisfies:

$${\Delta w}_{kj}^{KJ}={\eta }_{k}\left(-\frac{\partial \varepsilon }{{\partial w}_{kj}^{KJ}}\right)={\eta }_{k}\left(-\frac{\partial \varepsilon }{{\partial v}_{k}^{K}}\right)\times \left(\frac{{\partial v}_{k}^{K}}{{\partial w}_{kj}^{KJ}}\right)$$
(7)

where \({\eta }_{k}\) is the learning rate of the k-th neuron, and the net input of this neuron is \({v}_{k}^{K}=\sum_{j=1}^{J}{w}_{kj}^{KJ}{o}_{j}^{J}+{b}_{k}^{K},\) then we have:

$$\left(\frac{\partial {v}_{k}^{K}}{\partial {w}_{kj}^{KJ}}\right)={o}_{j}^{J}$$
(8)

The term \(\left(-\frac{\partial \varepsilon }{{\partial v}_{k}^{K}}\right)\) in Eq. (7) is defined as the local gradient of the output layer, expressed as δk.

$$\begin{array}{c}{\delta }_{k}=\left(-\frac{{\partial }_{\varepsilon }}{{\partial c}_{k}^{K}}\right)=\left(-\frac{{\partial }_{\varepsilon }}{{\partial O}_{k}^{K}}\right)\times \left(-\frac{{\partial o}_{k}^{K}}{{\partial v}_{k}^{K}}\right)\\ =\left({d}_{k}-{O}_{k}^{K}\right)\times \left(\frac{{\partial }_{\varphi }\left({v}_{k}^{K}\right)}{{\partial v}_{k}^{K}}\right)\\ \begin{array}{c}=\left({d}_{k}-{O}_{k}^{K}\right)\times {\varphi }^{^{\prime}}\left({v}_{k}^{K}\right)\\ =\frac{1}{2}\left({d}_{k}-{O}_{k}^{K}\right)\left[1-{\left({O}_{k}^{K}\right)}^{2}\right]\end{array}\end{array}$$
(9)

Combining Eqs. (6) and (9), the update formula for the weights of neurons in the output layer is obtained:

$${w}_{kj}^{KJ}\left(n+1\right)={w}_{kj}^{KJ}\left(n\right)+{\eta }_{k}\left\{\frac{1}{2}\left({d}_{k}-{o}_{k}^{K}\right)\left[1-{\left({o}_{k}^{K}\right)}^{2}\right]\right\}{o}_{j}^{J}$$
(10)

The BP algorithm also updates the hidden layer weights in a gradient descent manner. In each iteration, the correction amount of the hidden layer weight is:

$$\begin{array}{c}{\Delta w}_{ji}^{JI}={\eta }_{j}\left(-\frac{\partial \varepsilon }{{\partial w}_{ji}^{JI}}\right)\\ ={\eta }_{j}\left(-\frac{\partial \varepsilon }{{\partial v}_{j}^{J}}\right)\times {\eta }_{j}\left(-\frac{{\partial v}_{j}^{J}}{{\partial w}_{ji}^{JI}}\right)\\ \triangleq {\eta }_{j}{\delta }_{j}\times \left(-\frac{{\partial v}_{j}^{J}}{{\partial w}_{ji}^{JI}}\right)\end{array}$$
(11)

where \({\eta }_{j}\) is the learning rate of the j-th neuron in the hidden layer, and \({\delta }_{j}\) is the local gradient of the hidden layer. Because of \({v}_{j}^{J}=\sum_{i=1}^{I}{w}_{ji}^{JI}{o}_{j}^{J}+{b}_{j}^{J},\) there is \(\left(-\frac{{\partial v}_{j}^{J}}{{\partial w}_{ji}^{JI}}\right)={o}_{i}^{I}.\) The local gradient \({\delta }_{j}\) of the hidden layer satisfies:

$$\begin{array}{c}{\delta }_{j}=\left(-\frac{\partial \varepsilon }{{\partial v}_{j}^{J}}\right)=-\left(\frac{\partial \varepsilon }{{\partial O}_{k}^{K}}\right)\times \left(\frac{{\partial O}_{k}^{K}}{{\partial v}_{k}^{K}}\right)\times \left(\frac{{\partial v}_{k}^{K}}{{\partial O}_{j}^{J}}\right)\times \left(\frac{{\partial O}_{j}^{J}}{{\partial v}_{j}^{J}}\right)\\ =-\sum_{k=2}^{K}\left(\frac{\partial \varepsilon }{\partial {O}_{k}^{K}}\right)\times {\varphi }^{^{\prime}}\left({v}_{k}^{K}\right)\times \left(\frac{{av}_{k}^{K}}{{\partial O}_{j}^{J}}\right){\varphi }^{^{\prime}}\left({v}_{j}^{J}\right)\\ \begin{array}{c}=\sum_{k=1}^{K}\left({d}_{k}-{O}_{k}^{K}\right)\times \frac{1}{2}\left[1-{\left({O}_{k}^{K}\right)}^{2}\right]\times {w}_{kj}^{KJ}\times \frac{1}{2}\left[1-{\left({O}_{j}^{J}\right)}^{2}\right]\\ =\left(\sum_{k=1}^{K}{\delta }_{k}{w}_{kj}^{KJ}\right)\times \frac{1}{2}\left[1-{\left({O}_{j}^{J}\right)}^{2}\right]\end{array}\end{array}$$
(12)

From Eq. (12), the local gradient \({\delta }_{k}\) of the output layer has an effect on the local gradient \({\delta }_{j}\) of the hidden layer, that is, the error function flows to the hidden layer through the output layer. Therefore, the update formula for the weights of hidden layer neurons is as follows:

$$\begin{array}{c}{w}_{ji}^{JI}\left(n+1\right)={w}_{ji}^{JI}\left(n\right)+{\Delta w}_{ji}^{JI}={w}_{ji}^{JI}\left(n\right)+{\eta }_{j}{\delta }_{j}{o}_{i}^{I}\\ ={w}_{ji}^{JI}\left(n\right)+{\eta }_{j}\left(\sum_{k=1}^{K}{\delta }_{k}{w}_{kj}^{KJ}\right)\times \frac{1}{2}{\delta }_{k}\left[1-{\left({o}_{j}^{J}\right)}^{2}\right]{o}_{i}^{I}\end{array}$$
(13)

In the training process of the neural network, the error between the actual output signal and the expected signal constitutes a cost function, which is the driving force for the continuous training and adjustment of the network. Therefore, in the neural network model shown in Fig. 2, the input signal is transmitted layer by layer from left to right, and the network modifies the weights layer by layer from right to left through the BP learning algorithm.

3.3 Multi Step Prediction Algorithm Based on Cascaded Neural Network (CFN)

The structure of CFN is similar to FFN, but the input signal in the network is connected to each hidden layer behind it through a weight matrix. The structure of CFN is shown in Fig. 3.

Fig. 3
figure 3

Structure of cascade neural network

In CFN, all hidden layers except the first hidden layer contain two weight matrices W1i and W2i, which are used to control the output signal of the upper layer and the input signal of the network, respectively. This structure can give more degrees of freedom to the training process, thereby improving the non-linear mapping ability of the network. Therefore, the cognitive base station can perceive the behavior of authorized users in the next few time slots, and perform access control or spectrum switching for cognitive users in time to minimize interference to authorized users. For CFN, the BP learning algorithm shown in Fig. 4 is used for training to optimize the weight matrix W1i,W2i and bias matrix bi of the network. Its goal is to make the actual output of the network closest to the expected output in the sense of the smallest mean square error.

Fig. 4
figure 4

Flow chart of BP learning algorithm

However, the change of the CFN structure increases the complexity of the training process, because each iteration of BP training needs to modify more weight matrices. In addition, if too many training samples are provided to the network, it will also lead to overfitting. The multi-step prediction algorithm based on CFN makes use of the strong nonlinear mapping ability of CFN to predict the authorized user behavior in the next several sensing time slots. CFN also uses [St-τ+1,St-τ+2,…St-1,St]T as the network input. The difference from FFN is that the network outputs the predicted value \({\left[{\widehat{S}}_{t+1},{\widehat{S}}_{t+2},\dots {\widehat{S}}_{t+M}\right]}^{T}\) of user behavior in the M (M ≥ 2) sensing slots in the future. In addition, the expected value of the network during the training process is also a column vector \({\left[{S}_{t+1},{S}_{t+2},\dots {\widehat{S}}_{t+M}\right]}^{T}\) of length M.

With the prediction of the neural network, the behavior st+1 (1 ≤ i ≤ M) of authorized users in each sensing slot in the future will be predicted M times. The cognitive base station can combine the prediction results of M times to schedule cognitive users on this basis. However, with the increase of the time interval between the target sensing time slot and the current time window, the correlation between st+1 (1 ≤ i ≤ M) in the future sensing time slot and historical data in the time window is decreasing. That is, as the number of sensing time slots i increases, the deviation between the estimated result \({\widehat{S}}_{t+1} (1\le i \le M)\) and the actual result st+1 (1 ≤ i ≤ M) of the authorized user will increase.

In order to reduce the interference to authorized users, the cognitive base station in the network adopts the hard decision mode of "or" combination. For the M prediction results of the behavior of the authorized user in the next perception time slot, if there is an authorized user occupying the sub-band, the cognitive base station considers that the authorized user will occupy the sub-band.

4 Simulation Results and Analysis

FFN and CFN use a hidden layer, the number of neurons in the hidden layer is 10, and the transfer function of the hidden layer is φ(•). The number of neurons in the input layer and output layer is the same as the length τ of training time window. FFN and CFN use the same training samples for training, and the simulation results are shown in Fig. 5.

Fig. 5
figure 5

Training results of artificial neural networks

Figure 5 shows that the convergence speed of CFN is faster than FFN. In addition, after 40 iterations, the mean square error of CFN is less than FFN, but all remains unchanged. Therefore, CFN has stronger learning and generalization capabilities than FFN, but due to the high complexity of the network structure, relatively long training time is required.

For user behavior prediction algorithms based on FFN and CFN, the prediction order will have a great impact on the predicted conflict probability. A larger prediction order increases the number of neurons in the input layer and increases the complexity of the network, thereby weakening the training effect of the network. On the other hand, a smaller prediction order cannot satisfy the non-linear mapping relationship from input to output. Simulate the exponential ON/OFF model of authorized users and compare the conflict probability of the algorithm under different prediction orders. The parameters used are shown in Table 1.

The user behavior prediction algorithms studied in this paper all use Monte Carlo simulation for 10,000 times. The curve of the algorithm's spectrum conflict probability with the network prediction order is shown in Fig. 6.

Fig. 6
figure 6

spectrum conflict probability with prediction order.

When the prediction order is greater than 4, the conflict probability of the multi-step prediction algorithm based on CFN is significantly lower than that one step prediction algorithm based on FFN. When the prediction order is set to 2 or 3, the conflict probability is almost the same. On the other hand, the conflict probability of the two algorithms is the smallest when the prediction order is 5. Therefore, in the authorized user exponential ON/OFF model studied in this paper, two prediction algorithms based on neural networks set the prediction order to 5.

Set service parameter e1 ∈ [2,10] and e1 = 5, and other simulation parameters are shown in Table 1. The error prediction probability in Fig. 7 shows that with the increase of e1, the error prediction probability \({P}_{overall}^{e}\) of the three prediction algorithms gradually decreases. Because the increase in e1 means that the average time for authorized users to stay in the ON state increases, while e2 remains the same, resulting in a decrease in the frequency of authorized users to switch between ON/OFF states. Therefore, the correlation of the behavior of authorized users in adjacent sensing time slots is increased, making it easier for the cognitive base station to predict its behavior.

Table 1 prediction algorithm simulation parameters.
Fig. 7
figure 7

Error prediction probability with business parameter e1

In addition, it can be found that the error prediction probability of the Multi-step prediction algorithm based on CFN is higher than the other two algorithms. This is because the algorithm adopts the "or" combination hard decision method when judging the results of M predictions. Although this method can well avoid the spectrum conflict between cognitive users and authorized users, some spectral holes are also missed. On the other hand, when e1 is less than 5, the error prediction probability of one step prediction algorithm based on FFN is lower than that based on update theory. When e is greater than 5, the performance of the two algorithms is almost the same. Therefore, compared with the prediction algorithm based on the update theory, the algorithm based on the neural network has better prediction accuracy in the environment where the status of authorized users changes rapidly.

Figure 8 shows the conflict probability of the algorithm under different authorized user service parameters. Comparing Figs. 7 and 8, the conflict probability in the prediction process is lower than the error prediction probability. This is because there are two types of mispredictions: spectrum conflicts and cognitive base stations mistakenly believe that spectrum holes are occupied. The former measures the interference risk of cognitive users to authorized users from the perspective of authorized users, and the latter measures the utilization of spectrum resources from the perspective of cognitive users.

Fig. 8
figure 8

Spectrum conflict probability with business parameter e1

The conflict probability of the two prediction algorithms based on neural networks is lower than that of the traditional prediction algorithms based on update theory, indicating that it is feasible to use artificial neural networks. Compared with the prediction method based on the update theory, the neural network avoids the complicated description of the characteristics of authorized users, and uses the powerful learning ability and nonlinear mapping ability to effectively describe the relationship between input and output. Therefore, the prediction method based on neural network can realize prediction by learning and extracting historical information from the spectrum database without knowing the service parameters of authorized users. On the other hand, the cognitive base station adopts the "offline" mode to predict the trained network, only a small amount of addition and multiplication operations are needed to realize the prediction function, and the prediction speed of the two algorithms is relatively fast.

Compared with the one step prediction algorithm based on FFN, the multi-step prediction algorithm based on CFN can effectively reduce the probability of spectrum conflict in the network. In addition, the multi-step prediction algorithm based on CFN further reduces the probability of conflict by increasing the number of prediction steps. However, it also increases the probability of prediction errors, leading to a decrease in spectrum utilization in cognitive wireless networks. Therefore, in practical applications, cognitive base stations need to balance the utilization of spectrum resources and interference to authorized users.

5 Conclusion

This article discusses the behavior prediction algorithm of authorized users in satellite cognitive wireless networks. First, it analyzes the exponential ON/OFF model of user behavior prediction in satellite cognitive wireless network and the learning mechanism of artificial neural network. Then study the traditional prediction algorithm based on update theory and FFN, and propose a prediction algorithm based on CFN. The algorithm uses CFN's powerful learning ability and non-linear mapping ability to realize multi-step prediction of authorized user behavior, judge the results of multiple predictions of user behavior in a certain sensing time slot, and reduce the spectrum conflict between cognitive users and authorized users. Finally, the simulation results show that the CFN-based multi-step prediction algorithm proposed in this paper can effectively reduce the probability of spectrum conflicts in cognitive wireless networks. Cognitive base stations using predictive algorithms can effectively reduce the interference of cognitive users to authorized users through reasonable access management and scheduling methods.