Abstract
Radar working state recognition is the basis of cognitive electronic countermeasures. Aiming at the problem that the traditional supervised recognition technology is difficult to obtain prior information and process the incremental signal data stream, an unsupervised and incremental recognition method is proposed. This method is based on a backpropagation (BP) neural network to construct a recognition model. Firstly, the particle swarm optimization (PSO) algorithm is used to optimize the preference parameter and damping factor of affinity propagation (AP) clustering. Then, the PSO-AP algorithm is used to cluster unlabeled samples to obtain the best initial clustering results. The clustering results are input as training samples into the BP neural network to train the recognition model, which realizes the unsupervised recognition. Secondly, the incremental AP (IAP) algorithm based on the -nearest neighbor (KNN) idea is used to divide the incremental samples by calculating the closeness between samples. The incremental samples are added to the BP recognition model as a new known state to complete the model update, which realizes incremental recognition. The simulation experiments on three types of radar data sets show that the recognition accuracy of the proposed model can reach more than 83%, which verifies the feasibility and effectiveness of the method. In addition, compared with the AP algorithm and -means algorithm, the improved AP method improves 59.4%, 17.6%, and 53.5% in purity, rand index (RI), and -measure indexes, respectively, and the running time is at least 34.8% shorter than the AP algorithm. The time of processing incremental data is greatly reduced, and the clustering efficiency is improved. Experimental results show that this method can quickly and accurately identify radar working state and play an important role in giving full play to the adaptability and timeliness of the cognitive electronic countermeasures.
1. Introduction
With the diversity of radar systems, the complexity of electromagnetic signals, the intelligence of the countermeasure targets, and the development application of new antijamming technologies, the combat effectiveness of traditional electronic countermeasure systems has gradually declined. There is an urgent need to transform from “artificialization” to “intelligence.” The United States military first proposed the concept of “cognitive electronic warfare” in 2010. It successively released project announcements such as “Adaptive Electronic Warfare Behavioral Learning (BLADE)” and “Adaptive Radar Countermeasures (ARC)” [1]. Aiming at new multifunctional radars, a radar working state recognition model is built through the machine learning algorithm to infer the behavior intention of the system. Thus, the capability of real-time attack and adaptive to the environment is improved.
Radar working state recognition is the crucial link in cognitive electronic warfare. It specifically refers to acquiring radar signal characteristic parameters and using the data characteristics to explore signal parameter rules and identify its working state. The rapid and accurate recognition of the radar working state is essential for evaluating the radar threat level, distributing the jamming resources, and giving full play to the electronic combat effectiveness.
Radar working state recognition is essentially a pattern classification problem, so it can be solved using the classification algorithm in machine learning theory. At present, most scholars are mainly based on prior information using supervised learning methods such as Bayesian classifier [2, 3], decision tree [4, 5], and support vector machine [6, 7] to identify the working state. The BP neural network has strong self-learning and nonlinear mapping capabilities, which is particularly suitable for solving problems with complex internal mechanisms and can establish the mapping relationship between characteristic signal parameters and radar working conditions. It is the most widely used in working state recognition [8–11]. Still, there are two problems with the BP neural network in the virtual electronic countermeasure environment: (1) This method requires labeled sample data before classification. Still, noncooperative radar’s signal characteristic parameters and corresponding working state are hard to get, so the problem of the lack of prior knowledge identification cannot be solved. (2) Above methods are off-line batch learning methods. Aiming at the incremental signal data stream generated in the engagement process, it cannot judge online whether it belongs to a new state and update the recognition model in time. Because the unsupervised learning method does not need sample labels, the clustering method can be used to complete sample prelabeling to build a knowledge base. The data in the knowledge base can be used as a training set and input into the BP neural network to improve the recognition accuracy. Therefore, to solve the problems of adaptability and timeliness of recognition, the research on the unsupervised incremental recognition method based on the combination of clustering and classification has more practical application significance.
Common unsupervised clustering algorithms include partition-based clustering, density-based clustering, and fuzzy-based clustering. In studies [12, 13], the -means algorithm has been used to complete radar signal sorting and recognition. The process is simple and efficient. However, this algorithm needs to set the value in advance and is very sensitive to noise and outliers. Some studies [14, 15] have proposed an adaptive density clustering algorithm for radar signal features, which can cluster the arbitrary shape distribution. However, the clustering results are related to the algorithm parameters, and the parameter tuning process is complex. Other studies [16, 17] have introduced a radar signal classification method based on fuzzy -means clustering, which overcomes the influence of noise and missing data. Still, the algorithm’s performance depends on the initial clustering center. AP algorithm is a relatively new clustering algorithm [18], which does not have to predetermine cluster number and initial cluster centers. It can efficiently solve the problem of the non-Euler distribution of data. Therefore, it can be used to solve the problem of noncooperative target state recognition without prior information. However, the preference parameter and damping factor are two important parameters affecting the clustering algorithm’s effects [19]. The former affects the number of clusters, while the latter affects the convergence speed and accuracy. Parameter values are usually based on experience. The process is complicated and difficult to obtain the optimal parameters. Moreover, when new training data arrives, the new data needs to be merged into the original data set and recluster based on the whole new data set [20]. Therefore, the recognition model cannot be updated in time and increases the space and time.
An unsupervised and incremental radar working state recognition method is proposed in this paper to solve the above problems. The innovations and main contributions include the following:(1)An unsupervised and incremental recognition model is proposed, mainly divided into three steps: the initial sample prelabeling based on PSO-AP, the recognition model building based on BP, and the incremental sample clustering based on IAP(2)The parameter optimization ability of the PSO algorithm [21] is used to obtain the best preference parameter and damping factor parameter of the AP algorithm, automatically eliminate the AP algorithm oscillation, converge to the optimal clustering result, and improve clustering accuracy(3)The BP neural network recognition model is built. The clustering results of PSO-AP are used as a prior knowledge base to input into the recognition model and improve recognition accuracy(4)The idea of incremental learning is introduced, and the IAP algorithm based on KNN is proposed. Based on the PSO-AP clustering results, samples are classified online by calculating the compactness between incremental samples and sample data in KNN instead of the whole samples. Thus, a lot of space and time costs are saved(5)The three radars provided in studies [22, 23] are taken as the research objects. Simulation experiments show that the recognition algorithm combining IAP clustering with PSO and BP neural network improves the recognition accuracy and efficiency. This method achieves good classification and recognition for known and unknown states, which proves the feasibility and effectiveness of the proposed method
The rest of this paper is organized as follows. The radar working state recognition model based on the unsupervised and incremental method is presented in Section 2. A detailed introduction to the key algorithms of the recognition model is presented in Section 3. The simulation experiment and analysis of the results are presented in Section 4. Finally, the conclusions drawn from this study are discussed in Section 5.
2. Radar Working State Recognition Based on the Unsupervised and Incremental Method
2.1. Radar Working State Recognition Process
Due to the military sensitivity and complexity of radar signals, it is challenging to obtain real-time signal parameters and greatly increase the difficulty of radar working state recognition. Therefore, this paper proposes the radar working state recognition process, as shown in Figure 1. The process is divided into four steps: Firstly, the signal features that identify the radar working state are extracted. Secondly, the data set is randomly generated according to the range of characteristic signal parameters. Then, the data set is preprocessed. Finally, the data is input into the radar working state recognizer, and the recognition result is output.(1)Extraction of signal characteristic parameters: the radar working state characteristic parameters include replay frequency, pulse width, duty cycle, and carrier frequency. When the radar is in different working states, the signal’s characteristic parameters also have significant differences(2)Data set simulation: it is difficult to obtain the measured radar data due to the constraint of experimental environment conditions. According to the radar state characteristic parameter range table, samples for each state are randomly generated to construct the simulation experiment data set(3)Data preprocessing: data preprocessing mainly normalizes the data set to eliminate the influence of parameters’ dimensional difference and improve the recognition accuracy. After normalization, according to the two cases of radar known and unknown working state recognition, the data set is randomly divided into initial samples and incremental samples by proportion(4)Radar working state recognizer: the function of the recognizer is to output the corresponding radar working state recognition results according to the input signal characteristic parameters. The recognizer adopted in this paper is based on the unsupervised and incremental recognition method to realize the radar working state recognition

2.2. Recognition Model Based on the Unsupervised and Incremental Method
To quickly and accurately complete the prelabeling of initial samples without prior information and the online division of incremental samples, achieve the known and unknown working state recognition of incremental signal samples, and further improve cognitive electronic countermeasure system efficiency, this paper uses label samples generated by improved AP clustering as a training set of the BP model. It generates an unsupervised and incremental radar working state recognition model. The model’s step flow is shown in Figure 2.

The specific steps are as follows:
Step 1. The data set is generated randomly, and the input parameters are normalized to make it in [0,1]. The data set is divided into initial samples and incremental samples.
Step 2. PSO-AP algorithm is used to prelabel the initial sample data to get the training samples with labels, and the number of samples in the clustering center and each cluster is recorded.
Step 3. Input AP initial clustering results into BP neural network as prior knowledge to build the recognition model.
Step 4. Adopt the KNN idea to IAP clustering for incremental samples. If a known state is judged, then return to Step 3 to complete recognition of the known state. If it is unknown, then update the clustering results and perform Step 5.
Step 5. When the incremental sample reaches a certain size, the clustering cluster is input into the BP neural network as a new training sample. The recognition model is updated to complete the unknown state detection. Otherwise, return to Step 4.
3. Algorithm Analysis
3.1. PSO-AP Algorithm
3.1.1. AP Algorithm
AP algorithm is a new clustering algorithm based on “information transfer” between data points, proposed in Science magazine in 2007. Unlike other clustering algorithms such as -means, AP does not need to set the number and center of clustering before clustering, which is widely used in face classification and image segmentation. The AP clustering process is shown in Figure 3. The principle is to regard all the data points as potential clustering center, and the connection lines between data points form a network, that is, similarity matrix. Then, constantly update and pass the availability and responsibility information between points through similarity matrix, and calculate the cluster center of each sample for automatic clustering.

The algorithm can be summarized as the following steps:
Step 1. According to equation (1), the Euclidean distance is used to calculate the similarity between data points and points , and the similarity matrix is obtained. Elements on the diagonal are defined as the preference .
where the smaller the is, the lower the ability of the point as the center of point clustering is.
Step 2. Set and . Calculate the availability and responsibility according to
where represents the degree to which the point is suitable for the clustering center of the point and represents the degree to which the point selects the point as its clustering center.
Step 3. To avoid the oscillation of the AP algorithm in the iterative process, a damping factor is added in each iteration, and the information transmission process is updated as follows:
Step 4. Determine the cluster center point according to
Step 5. Iterate and update and . When the maximum number of iterations is reached, or the clustering center remains unchanged after multiple iterations, the iteration is stopped, and the remaining data objects are allocated to the corresponding clustering. Otherwise, continue with Step 2-Step 4.
3.1.2. AP Algorithm Parameter Optimization Based on PSO
The clustering efficiency of the AP algorithm is mainly affected by preference and damping factor . Among them, the larger the is, the greater the clustering number is. The is generally the mean of the similarity matrix [18]. However, it cannot always obtain the optimal clustering result due to the difference in data sets. The smaller the is, the greater the change of and is, and the number of iterations will be reduced. In general, is debugged in (0, 1). However, the process is complex, and it is difficult to obtain the best parameters. As a swarm intelligence optimization algorithm, PSO finds the optimal solution through the simple behavior of personal particles and the information interaction with the global particles. It has the characteristics of simple operation, fast convergence, and strong global searchability. In recent years, scholars have increasingly applied it to data processing and proposed some new PSO methods. For example, facing the challenge of the “curse of dimensionality” when dealing with high-dimensional data, Song et al. [24] proposed a variable-size cooperative coevolutionary PSO algorithm for feature selection (FS). Focusing on the FS problem with fuzzy cost, Hu et al. [25] proposed a fuzzy multiobjective FS method with PSO. Xue et al. [26] proposed an adaptive PSO algorithm to solve the local optimal stagnation in large-scale FS. In this paper, we use PSO to optimize the parameters of AP clustering. The specific implementation steps of the PSO-AP algorithm are as follows:
Step 1. is in [, 0.1], where is the mean of the similarity matrix. is in [0.5, 0.95]. Initialize the particle swarm parameters. Set the population size and maximum number of iterations . This paper selects purity as the particle fitness function. Purity is one of the clustering evaluation validation indexes, and its overall idea is to divide the correct number of samples by the total number of samples. The calculation formula is
where is the total number of samples, is the sample set of clustering, is the clustering cluster th, is the sample set, and is the sample th.
Study [27] mentioned that the Silhouette (clustering evaluation index) is taken as the fitness function. Because the Silhouette introduces distance calculation, and different data sets are suitable for different distance calculation rules, it is difficult to calculate the Silhouette in a unified way. Moreover, purity is the evaluation index with label, and the Silhouette is the evaluation index without label. Many experiments show that the evaluation index with label is better than that without label, so it is more appropriate to take purity as the fitness function.
Step 2. The particles’ dimension is determined by and , and the position and velocity are initialized randomly.
Step 3. Calculate fitness value, personal extremum, and global extremum according to equations (1)–(6).
Step 4. In each iteration, a particle updates its position by learning the optimal position found by the particle itself, that is, the personal leader (Pbest) and the particle’s neighborhoods, that is, the global leader (Gbest). Without loss of generality, assuming that the personal and global leaders of the th particle are and , respectively, this particle is updated by using equations (7) and (8) [28].
where , , , and are the random numbers between [0,1], is the current iteration number, is the inertia factor, and and are the learning factors.
Step 5. Update the particle swarm fitness value, individual extremum, and population extremum, and regard the particle with the largest fitness value in each generation as the current optimal particle.
Step 6. When the maximum number of particle swarm iteration is reached, the optimal values are output as the optimal and . Otherwise, repeat Step 2-Step 5.
Step 7. Substitute the optimal and into the AP algorithm to complete clustering.
3.2. BP Algorithm
The BP neural network’s basic idea is to realize the nonlinear mapping relationship between input samples and output samples using the forward propagation of signals and the backpropagation of errors. Figure 4 shows a typical structure of a three-layer BP neural network model. The sample characteristic signal is firstly transmitted from the input layer to the hidden layer and then reaches the output layer after calculation. Next, the output result is calculated as the radar working state. When the actual output result is inconsistent with the expected output result, the error will be transmitted back. The weight of each element will be corrected by the gradient descent method. Repeat the sample until the network output error is reduced to an acceptable level or reaches a default number of learning times.

3.3. IAP Algorithm Based on KNN
In cognitive electronic warfare, the jammer usually receives the new signal samples in batches or streams. The traditional AP algorithm needs to combine the new training data into the original data set to retrain the whole model, lacking adaptability and timeliness. The KNN algorithm is based on the idea that the unknown sample should be consistent with most known samples in its neighbors. It divides the incremental data by calculating the -nearest neighbors of the unknown sample. It has the characteristics of mature theory, low time complexity, and incremental learning ability and can achieve efficient incremental clustering combined with the AP algorithm. To complete the online recognition of the unknown signal sample, this paper proposes an IAP algorithm with the idea of KNN. The algorithm is based on the existing AP clustering results to update the model with the new sample data and improve the clustering efficiency. The incremental processing process is shown in Figures 5 and 6.


The specific steps of the algorithm are as follows:
Step 1. The PSO-AP algorithm is used to cluster the samples to obtain the clustering center and the number of samples in each cluster.
Step 2. Calculate the distance between the incremental sample and the initial sample, use the cross-validation method to determine the value of KNN, and get the cluster to which each nearest neighbor belongs.
Step 3. Calculate the average distance between nearest neighbors, and calculate the average distance between incremental data and each cluster in the nearby neighbors, where is the number of clusters.
Step 4. If , then the increment belongs to the cluster of . Otherwise, it belongs to the new cluster.
Step 5. When the incremental data scale is reached, stop incremental clustering. Otherwise, loop through Steps 2-4.
4. Simulation Experiment and Analysis of Results
The experimental simulation platform of this paper is as follows:(i)The operating system is windows 10(ii)The CPU is Intel(R) Core (TM) i7 2.6 GHz(iii)The memory is 8 GB(iv)The programming tool is MATLAB R2016b
The algorithm is verified, and the experimental results are processed by the MATLAB R2016b tool. Taking the airborne phased array radar (APAR) in the air-air combat mode in reference [22] and the radar A and B provided in reference [23] as examples, three experiments were designed to verify the performance of the model algorithm: (1) validation of PSO-AP algorithm, (2) value optimization of IAP clustering, and (3) validation of radar known and unknown working state recognition.
4.1. Simulation and Preprocessing of Data Sets
Due to the military confidentiality of the radar, the actual signal parameters of the radar working state are difficult to obtain. Many studies [22, 23] generate data randomly based on the working parameter ranges. Therefore, according to the parameter range in Tables 1 and 2, this paper randomly generated 100 samples for each state to form a data matrix of , , and dimensions, respectively. The simulation data set is shown in Table 3.
This paper uses the improved AP algorithm to analyse signal characteristic parameters, and the AP algorithm is based on the similarity between the data points for clustering. If the data is an order of magnitude difference, the error will be great when similarity is obtained by using the Euclidean distance. Therefore, the data needs to be normalized processing to improve the clustering effect of the AP algorithm. The commonly used normalization methods include deviation normalization and -score normalization. However, the former is affected by the maximum and minimum sample, leading to unstable normalization results, while the latter requires the distribution of data to be approximately Gaussian. To effectively narrow the gap between the data, the nonlinear normalization method was adopted in this paper, and the calculation formula is
where is the sample data after processing the original data .
4.2. Algorithm Evaluation Index
In this paper, common clustering validation indexes (such as purity, RI, and -measure) and classification effect evaluation indexes (such as confusion matrix and accuracy) were selected to conduct quantitative analysis on the experimental results of the improved AP algorithm and BP model, respectively. Among them, the calculation formula of purity is shown in equation (6), and the calculation methods of other indexes are shown in equations (10)–(12). The above quantitative evaluation indicators are all positive indicators.
4.2.1. RI
where is the number of two similar sample points in the same cluster, is the number of two nonsimilar sample points in the same cluster, is the number of two nonsimilar sample points in different clusters, and is the number of two similar sample points in different clusters.
4.2.2. -Measure
where , , and the general default parameter .
4.2.3. Confusion Matrix
A confusion matrix is a table for evaluating a classification model’s performance, with rows representing actual values and columns representing predicted values. The confusion matrix’s each unit represents the statistics corresponding to different predicted states and actual states, as shown in Table 4.
4.2.4. Accuracy (ACC)
where is the number of samples, is the number of sample classes, and is the number of correctly identified samples of class th.
4.3. Experimental Results and Analysis
4.3.1. Validation of PSO-AP Algorithm
(1) Algorithm Parameter Setting. In this experiment, the AP clustering purity was selected as the fitness function, and the preference and damping factor were selected as the particle dimensions. The maximum number of iterations of the AP clustering algorithm was 100. According to the commonly used range of other parameters, the PSO-AP algorithm’s parameters are set as presented in Table 5.
(2) Result Analysis. Taking the radar APAR data set as an example, Figures 7–9 show the relationship among particle fitness value, preference, damping factor, and iteration number. As can be seen from the figures, the maximum number of particle iterations is 50. When the iteration reaches the 18th generation, the fitness function tends to converge. At this point, the fitness is maximum, and the preference and damping factor are optimal. The values are as follows: , , and .



To compare and analyse algorithms, this experiment selected PSO-AP and AP algorithms to carry out clustering tests on three data sets, respectively. To avoid the randomness of the algorithms, a maximum value and a minimum value were removed after running each algorithm 20 times and then took the average value. The results are shown in Table 6.
The comparison and analysis of purity, RI, and -measure show that the PSO-AP algorithm has a better clustering effect on the three data sets than the AP algorithm. Purity, RI, and -measure indicators have increased by at least 11.9%, 6.8%, and 41.3%, respectively. The PSO-AP algorithm can obtain a better clustering effect because the PSO algorithm’s good global search optimization ability is used to adjust the AP algorithm’s parameters automatically. Firstly, the PSO algorithm rapidly obtains the optimal preference parameters and damping factors through continuously updating iterative optimization calculation. Then, according to the optimal preference parameters and damping factors, the AP algorithm completes clustering to obtain the best clustering effect. The PSO-AP algorithm can solve the setting value problem of the preference parameter and damping factor in the original AP algorithm and improve the clustering effect of the algorithm. Therefore, it is feasible to apply the PSO algorithm to the parameter optimization of the AP algorithm.
4.3.2. Value Optimization of IAP Clustering
The key of the IAP algorithm based on KNN is the selection of value. If the value is too small, the model is too complex and prone to overfitting. If the value is too large, it is easy to cause fuzzy clustering. According to the rule of thumb, the value is usually odd and does not exceed the square root of the training sample. Considering the size of the sample data set in this paper, the range of values was set to all odd numbers between 1 and 19.
In this experiment, according to the change law between the verification set’s accuracy and the value, the 10-fold cross-validation method was used to select the algorithm’s optimal value. The specific method was to randomly divide the data set into ten groups of disjoint subsets. In each experiment, one group was successively selected as the verification set and the other nine groups as the training set. The verification set was taken as the incremental set. The average accuracy obtained through 10 times of IAP clustering experiment was used as the estimation of the algorithm accuracy, and the accuracy of all values was finally compared. The classification accuracy of each validation with the changing of value is shown in Figure 10.

(a)

(b)

(c)

(d)

(e)

(f)

(g)

(h)

(i)

(j)
Taking group 1 as an example, it can be seen from the figure that under this kind of grouping, when for the A data set, the classification accuracy of the verification set is the highest, which is 97.5%. When for the B data set, the classification accuracy of the verification set is the highest, which is 82.5%. When or for the APAR data set, the classification accuracy of the verification set is the highest, which is 93.33%. Similarly, for other groups, the value at the highest accuracy of the verification set was selected as the best value in the current group. The accuracy of verification set at the corresponding value was statistically analysed. The statistical results show that the average classification accuracy of A, B, and APAR data sets is, respectively, 94.75%, 79%, and 95%.
To observe the influence of the value on the classification accuracy, the accuracy of the verification set with the same value under different groups was counted and averaged. The curve of the variation rule is shown in Figure 11. When the average accuracy of the verification set is the highest, the value was taken as the optimal value. As can be seen from the figure, A data set’s optimal value is 3, B data set’s optimal value is 5, and APAR data set’s optimal value is 7. The value in the KNN has a great influence on the clustering result. When the value is set improperly, the clustering effect will be poor.

4.3.3. Validation of Radar Known and Unknown Working State Recognition
(1) Data Set Division. To verify the effectiveness of the radar working state recognition model proposed in this paper for known working state recognition and unknown status detection, the original data set constructed in Section 4.1 was divided into two types of data sets in this experiment:(1)Assuming that all working states of the radar are known, 20% of the data from each class in the original data set were randomly selected as the incremental set and the remaining as the training set. The data division is shown in Table 7(2)Assuming that a working state of the radar is unknown, a class in the original data set was taken as the incremental set of the unknown state and the other classes as the training set of the known state. 20% of the data from each class were randomly selected to form a test set for the experiment. The data division is shown in Table 8
(2) Effectiveness Analysis of IAP Clustering. Firstly, the training set and the incremental set were used to conduct the IAP clustering analysis for the known and unknown states, respectively. AP parameters were set using the method described in Section 4.3.1, and values of the data set were set using the method described in Section 4.3.2. Purity, RI, -measure, and time were selected to measure the clustering effect. To facilitate the comparative analysis of the model, in this experiment, AP and -means [29] algorithms are selected to verify the effectiveness of the IAP algorithm. The clustering number of the -means algorithm is set as shown in Table 9.
The IAP, AP, and -means algorithms were repeatedly run 20 times on A, B, and APAR data sets, respectively. After removing one maximum and one minimum value, take the average as the result. The clustering results are shown in Table 10.
According to the comparison of simulation indicators of purity, RI, -measure, and time, the following conclusions are drawn:(1)From the perspectives of purity, RI, and -measure, for the known and unknown working states of the three radar data sets, the clustering effect of the -means algorithm depends on the number of clusters and the selection of the initial clustering center. It is also sensitive to “noise” and “outliers.” Because the iterative method often converges to a local minimum, the clustering effect is not ideal. The AP algorithm overcomes the disadvantage of the -means algorithm. The algorithm does not need to specify the number of clusters and cluster centers and is not sensitive to the initial value of the data. However, the clustering result is still affected by parameters. The IAP algorithm firstly takes the clustering result of the PSO-AP algorithm as the initial clustering center, which overcomes the influence of the AP algorithm parameter values. Then, introduce the idea of the KNN algorithm to increment clustering, which has a high tolerance for outliers and noise. Therefore, the IAP algorithm can achieve a better clustering effect in most cases. When appropriate parameters are selected, the IAP algorithm has the highest clustering purity, the AP algorithm is the second, and the -means algorithm is the worst. For the IAP algorithm, the RI is lower than that of the -means algorithm only on the unknown working state clustering of A and B data sets, the RI is lower than that of the AP algorithm and -means algorithm on the unknown working state clustering of APAR data sets, and the -measure is lower than that of the -means algorithm on the known working state clustering of the B data set. The IAP algorithm can always obtain good clustering results due to the proper and fixed initial center point(2)From the perspective of time, the -means algorithm is simple and efficient with the fastest convergence rate and only needs to calculate the distance between data points and centers. The time complexity is ( is the number of samples; is the number of clustering). Due to the need to calculate the similarity between all sample points, the AP algorithm has a slow convergence speed and time complexity is . When the sample size is large, the AP algorithm has low efficiency. Based on the current AP clustering results, the IAP algorithm uses the KNN idea to cluster incremental samples. Because the KNN algorithm is online technology, incremental samples can be directly added to the data set without retraining, and the algorithm’s running time is greatly shortened. The time complexity is (, is the number of samples, and is the number of incremental samples). The AP algorithm’s efficiency is improved, especially for large-scale data, and the clustering speed is obvious
In summary, the -means algorithm runs efficiently, but the clustering accuracy rate is low. The AP algorithm has a better clustering effect, but the algorithm runs inefficiently for incremental clustering. The IAP algorithm overcomes the shortcomings of the AP algorithm and -means algorithm, improves the clustering accuracy and convergence speed, and proves the effectiveness and feasibility of IAP clustering in radar known and unknown working state recognition.
(3) Validity Analysis of the BP Recognition Model. (a)Model parameter setting
The input layer and output layer node numbers are, respectively, determined by the dimension of the input feature vector and the desired output dimensions. The hidden layer node number is determined by the empirical formula and trial and error method. Assuming that the output result of corresponding each state is represented by numbers 1, 2, etc., the recognition model network structure parameters of three radars are set as shown in Table 11.
Other parameters were set as follows:(i)Activation function of output layer: Purelin(ii)Training function of neural network: Trainlm(iii)Maximum iteration step of neural network: 1000(iv)Training target error of neural network: 0.001(v)Learning rate factor of neural network: 0.01(b)Result analysis
Samples generated by IAP clustering of the three data sets in Section (2) of 4.3.3 were, respectively, input into BP neural network as training sets for model training. After the training is completed, the test set samples in Table 8 were input. Assuming that S1-S3 of A and B data sets are known states, S4 are unknown states, S1-S5 of APAR data sets are known states, and S6 are unknown states. The recognition results are shown in Figures 12–14.



As can be seen from the confusion matrix, the radar working state recognition model based on IAP clustering and BP neural network has a self-learning function, which correctly identifies the known and unknown working state. Because the improved AP algorithm uses the PSO algorithm to overcome the influence of the parameter value of the algorithm itself and introduces the KNN algorithm idea to divide the incremental samples, the accuracy of clustering is improved. The initial samples prelabeled by improved AP clustering are input into the BP neural network model for training. Using the good classification ability of the BP neural network, the recognition of radar working status is well completed. The overall recognition accuracy of the model for the three data sets is high, that is, respectively, 0.875, 0.838, and 0.917.
5. Conclusions
For the traditional supervised working state recognition methods, the recognition efficiency is low due to the lack of timeliness and adaptability to the environment. This paper proposes an unsupervised incremental radar working state recognition model to quickly and accurately identify the category to which the state belongs and determine whether there is a new category when facing the continuous incoming signal data stream. At the same time, it can timely update existing classification models to adapt to changes in the external environment. The conclusions are as follows:(1)Adopting the idea of “first cluster and then sort,” the incremental PSO-AP algorithm and BP classification algorithm are combined to build an unsupervised and incremental radar working state recognition model and complete known state recognition and unknown state detection. The efficiency of the cognitive electronic countermeasure system can be further improved(2)The improved AP algorithm can meet the prelabeling of the unsupervised initial training set and ensure the incremental recognition requirement. It automatically adjusts the parameters of the AP algorithm through the parameter optimization ability of the PSO algorithm, which improves the clustering effect. Moreover, the idea of KNN is introduced to realize IAP clustering. When processing incremental data, it avoids the traversal of all data and improves the clustering speed(3)This model overcomes the disadvantages of the low efficiency of the traditional AP algorithm for incremental data clustering and low accuracy of the -means algorithm and has higher accuracy and faster convergence speed for different radar state clustering. The training samples obtained by the improved AP clustering make the BP model have a higher recognition rate
However, this paper also has some shortcomings. For example, the incremental processing of the IAP algorithm is carried out one by one, and when the incremental data is too large, the algorithm’s efficiency may be greatly reduced. Therefore, we will study the incremental algorithm of batch processing in the future to further improve the efficiency of radar working state recognition.
Data Availability
The data used to support the findings of this study are included within the article.
Conflicts of Interest
The authors declare that they have no conflicts of interest.
Acknowledgments
This study was supported by the National Natural Science Foundation of China (61773389), Natural Science Foundation of Shaanxi Province (2020JQ-298), and Postdoctoral Science Foundation of China (2019M663635).