Abstract
Aiming at the problems of poor data effectiveness, low modeling accuracy, and weak generalization in the tuning process of microwave cavity filters, a parametric model for coaxial cavity filter using kernel canonical correlation analysis (KCCA) and multioutput least squares support vector regression (MLSSVR) is proposed in this study. First, the low-dimensional tuning data is mapped to the high-dimensional feature space by kernel canonical correlation analysis, and the nonlinear feature vectors are fused by the kernel function; second, the multioutput least squares support vector regression algorithm is used for parametric modeling to solve the problems of low accuracy and poor prediction performance; third, the support vector of the parameter model is optimized by the differential evolution whale algorithm (DWA) to improve the convergence and generalization ability of the model in actual tuning. Finally, the tuning experiments of two cavity filters with different topologies are carried out. The experimental results show that the proposed method has an obvious improvement in generalization performance and prediction accuracy compared with the traditional methods.
1. Introduction
The manufacture of coaxial cavity filters usually includes two steps: design and tuning. The design is mainly based on the electromagnetic simulation software, which produces approximate theoretical errors. In addition, the difference between processing tolerance and metal coating makes it difficult to achieve the physical size of the microwave cavity filter after production, and its output response cannot match the theoretical results. Therefore, tuning has become an indispensable link. To facilitate the tuning after processing, the tuning screws are usually installed on the real object to replace the resonant and coupling rods in the cavity. By constantly changing the direction and amplitude of the screws, the waveform of the output response can meet the requirements of the performance indicators.
In view of the complex dynamic characteristics such as nonlinearity and strong coupling in the tuning process of coaxial cavity filters, how to effectively use a large number of off-line and online data generated in the tuning process to establish a relationship model that can accurately reflect the tuning law is the focus of many scholars at home and abroad. At present, the relevant modeling methods mainly include feedback neural networks, fuzzy logic, and support vector regression algorithm. The research on neural network methods is concentrated on the basis of electromagnetic simulation software, such as the model between microwave element size and parameters [1], the tuning model between tuning element and return loss [2], and the model between reflection characteristic phase and center frequency [3]. However, the parameters in these models fluctuate greatly and are easily affected by the structure of microwave components, so it is difficult to extract the return loss and reflection characteristic phases. To reduce the amount of training data and shorten the modeling time, a high-precision microwave filter parameter prediction model using the adaptive learning neural network is proposed in [4]. It is applied to the optimization design of microwave filters. Although the neural network has strong nonlinear mapping ability, it has strict requirements for data quality. For the cavity filter with insufficient field tuning data, the modeling accuracy has great limitations.
Faced with a complex tuning process, Miraftab first introduced engineers’ experience in fuzzy rules into the tuning process of cavity filters [5]. This method uses experience and data fusion to establish the relationship model between coupling and tuning elements and is applied in low-order cavity filters. The drawback of this method is that it is difficult to apply to multiple cross-coupled cavity filters, and the process of designing fuzzy rules is extremely complex. Different from fuzzy algorithms, support vector regression modeling is a novel small-sample learning method with a solid theoretical foundation [6]. It has better generalization ability than the neural network when the number of samples is small. On the basis of previous research, multicore support vector regression based on parameter self-adjustment is used to model the tuning process of cavity filters [7]. However, faced with multi-input and multioutput data from cavity filters, the above method needs to solve the complex quadratic programming problem and should consider the coupling relationship between output variables. To improve the solving efficiency of multioutput systems, some studies using multicore machine learning have been proposed [8–12], which provides an important theoretical reference for the development of this study. Nonetheless, in industrial applications, the large-scale data collection and hyperparameter optimization time used for machine learning modeling often exceeds the tuning time. From a practical operational perspective, it is difficult to achieve.
To avoid these problems, parametric models based on vector field and machine learning were proposed in [13, 14], which provides a new idea for the modeling of the tuning process. However, with the increase in the order of the coaxial cavity filter, the input-output and action space dimensions of reinforcement learning increase sharply, resulting in an exponential growth of the value function network size and an increase in the training cost. In [15, 16], a parametric modeling method based on hybrid neural networks and deep learning has been proposed successively, but the data acquisition process of cavity filter in industrial applications is cumbersome, and the time of large-scale data acquisition will even exceed the tuning time. Therefore, from a practical point of view, these methods are difficult to implement. On the other hand, although SVR has obvious advantages in parametric modeling, the parameter identification of the algorithm itself relies too much on expert experience and laboratory experiments. Therefore, intelligent biomimetic algorithms such as the genetic algorithm (GA), particle swarm optimization algorithm (PSO), and genetic whale algorithm (GWA) are generally used to calculate parameters [17–19]. However, these algorithms also have the problem of easily falling into local optima and complex parameter calculations.
Given the aforementioned research analysis, the parametric model of coaxial cavity filters will encounter three challenges. (1) The difficulty of feature data fusion under different tuning modes and the impact of data validity on model accuracy. (2) Low modeling accuracy and weak generalization ability of parameterized models for multi-input multioutput tuning processes based on multivariate, strongly coupled, and nonlinear relational data. (3) Dealing with the computational efficiency and convergence speed of optimization algorithms in the solving process. The goal of this article is to provide a parametric model for cavity filter using kernel canonical correlation analysis and multioutput least squares support vector regression.
The contribution and novelty of this work includes three aspects: (i) the feature vectors under different modes of initial and fine tuning are fused by KCCA, which solves the problems of poor data validity and large input vector dimensionality of a single feature; (ii) the output coupling and complex computation process of quadratic programming in parameter modeling are avoided by combining the least squares method with the improved support vector regression; (iii) the differential evolution whale algorithm is used for model parameter identification, which improves the optimization speed and global search ability of the model. The rest of this study is organized as follows: Section 2 describes the theoretical synthesis of the cavity filter. Section 3 presents the principal component analysis of the raw data. Sections 4 and 5 mainly introduce the establishment of tuning model and parameter optimization. Section 6 shows the simulation results and analysis. The last section is the conclusion.
2. Theoretical Synthesis of Cavity Filter
2.1. Description of the Problem
The working process of this paper mainly includes data collection and kernel canonical correlation analysis, electromechanical characteristic modeling, and model parameter identification. First, collect the input and output data pairs generated during the tuning process, where denotes the screw tuning height and denotes the corresponding output response; secondly, the coupling matrix is extracted from the scattering parameter (-parameters) as mentioned in [20], and the characteristic parameters under different modes are fused through the kernel canonical correlation analysis technology. Build a data set for electromechanical characteristic modeling; finally, the electromechanical relationship model of the cavity filter based on multioutput support vector regression is established according to the collected data sets of input and output relationships. On this basis, the differential evolution whale algorithm is used to realize the adaptive identification of model parameters. The flow chart of the parametric model for the cavity filter is shown in Figure 1.

2.2. Mechanism and Characteristics of Cavity Filter
The produced coaxial cavity filter is difficult to achieve in physical size, and its output response cannot be consistent with the theoretical results. Generally, the cavity filter is equipped with tuning screws, and the output -parameters can meet the performance requirements by changing the length of the tuning screws. The -parameters have the following nonlinear relationship with the coupling matrix [21]:where and are the input and output port coupling, respectively. denotes the diagonal matrix, denotes the identity matrix. The coupling matrix extracted from the -parameter contains different feature information, such as diagonal elements, adjacent coupling elements, and cross-coupling elements, which are interrelated in different tuning stages. To reduce the redundant information in the tuning model, this study uses the KCCA multifeature fusion method to find the key information from the multidimensional features and convert the original extracted feature vector into a new low-dimensional vector.
3. Kernel Canonical Correlation Analysis
The tuning data of the cavity filter has serious nonlinear characteristics, especially for the cavity filter with initial tuning or multiple cross coupling. In addition, there is a strong coupling between the input variables, and it is difficult to describe the internal law of tuning with a single model. Therefore, data preprocessing and correlation analysis play a crucial role in modeling accuracy and generalization ability. The KCCA is a nonlinear data analysis algorithm, which transforms the nonlinear relationship of input space into the linear relationship of feature space through the kernel function and carries out correlation analysis in the new space. , , , and . The random variables in the observation space are transformed into the kernel function of the following high-dimensional space by implicit mapping [22].where and denote the observation space and high-dimensional feature space, respectively. and denote the implicit mappings of and from the observation space to high-dimensional feature space through kernel functions, respectively. , . The essence of the kernel function is equivalent to the mapping of input data from low-dimensional space to high-dimensional space, and its mathematical expression is as follows:where and denote the constant vectors of the high-dimensional space to be solved. denotes a multiplication sign. Using equation (3), the problem of solving high-dimensional space constant vectors and is transformed into the problem of solving low-dimensional space constant vectors and [23].where and denote the transformed characteristic components. and denote the transpose and inner product, respectively. The core problem of KCCA is to solve the corresponding equations (3) and (4) when the correlation coefficient is the largest. The mathematical expression of the correlation coefficient is as follows [24]:where and denote the variance matrix and covariance matrix, respectively. The variance and covariance of and are calculated as follows:where , the following correlation coefficients can be obtained by substituting equations (5) into (6).
The problem of solving the correlation coefficient is transformed into a constrained optimization problem, and its objective function and constraint conditions are as follows:
The Lagrange multiplication is used to solve the above constrained extreme value problem, and the corresponding Lagrange equation iswhere and denote the Lagrange multiplier. Calculate the partial derivative of with respect to and and make it zero. The following equations are obtained by deriving and , respectively.where , the solution of KCCA problem is equivalent to solving the eigenvector problem corresponding to the following generalized eigenequation:where , , and and denote the eigenvectors to be solved. By solving the above equation, , , and the correlation coefficient can be calculated.
4. Parametric Modeling
The traditional multioutput regression problem is usually solved by transforming the multidimensional output into one dimensional output, which often ignores the correlation between variables, to reduce the accuracy of the model. To solve this problem and fully consider the correlation of output components, this study improves the loss function of the original algorithm and proposes a parametric modeling strategy combining KCCA and MLSSVR.
4.1. Support Vector Regression
The essence of support vector regression is to map the input space to the high-dimensional space through nonlinear mapping and perform linear regression through the estimation function in the high-dimensional space. In this study, select the training set of group sample data, in which the input variable and the output variable . This research improves the insensitive loss function in the traditional single-output function regression algorithm, using the loss function on the hypersphere instead of the loss function on the hypercube. The improved loss function is as follows [25]:where , , and denotes the width of the neutral zone, which is the tolerance for errors, denotes a nonlinear mapping function, , , denotes the number of input samples, and denotes the dimension of the output variable. The improved function takes the fitting error of each component into account, which can not only achieve the goal of overall optimization, but also weaken the noise.
4.2. Multioutput Least Squares Support Vector Regression
Different from support vector machines, LS-SVR usually rewrites the primary loss function into a quadratic loss function and changes the inequality constraints into equality constraints, which avoids the complex calculation of solving quadratic programming problems and improves the efficiency of optimization. However, the traditional LS-SVR algorithm is only applicable to the regression modeling of single-output systems. When faced with multioutput nonlinear systems, the single-output systems are usually simply combined, thus ignoring the potential correlation information between output variables. In view of this, this study improved the objective function and constraint equation and used the absolute error to describe the overall error of the sample, thus establishing a multioutput regression model. The parametric model structure is shown in Figure 2, where , , and denote the resonant cavity self-coupling, adjacent cavity coupling, and cross coupling variables, respectively. , , and denote the number of corresponding variables. The model has a simple structure and is easy to calculate. The new eigenvector in the parametric model is mapped to the high-dimensional space , and then the linear modeling is realized in the new feature space [26].where denotes the nonlinear mapping. In this study, the least squares method is introduced to change the inequality constraint in traditional support vector regression into an equality constraint, and the square term of error is used as the experience loss of the training set. Finally, the quadratic programming problem is transformed into solving linear equations, which improves the calculation speed and convergence accuracy of the algorithm. The values of and are obtained through the following optimization and constraint functions [27]:where , and denote the input and output variables. denote the output weight, denotes the bias coefficient vector, denotes the input mapping function, denotes the penalty coefficient of the overall error of the sample, denotes the -th dimension output error of the -th sample, denotes the -dimensional output error of the -th sample, and denotes the overall fitting error. The objective function and constraints are transformed into the following Lagrange function:where and denote Lagrange factor. Use equation (3) to calculate the partial derivatives of , , , and , respectively, and obtain the following expression with KKT condition:

By eliminating and . The solution of the optimization problem can be transformed into the solution of the following linear equations:where denotes the unit matrix, . When the output error of the -th dimension is greater than or equal to zero, . The values of the variables and are obtained by solving the above linear equations, and the regression function of the output of the -dimension is as follows:
The Gaussian radial basis function has good ability to handle complex nonlinear relationships between sample inputs and output. It requires fewer parameters to be determined and has high computational efficiency. Therefore, the kernel function used in this study is the Gaussian radial basis function:where denotes the kernel width parameter, which reflects the size of the training data sample space range. When its value is large, the space range is small. When the number of support vectors is large and the regression parameters have been calculated, the kernel method may suffer from the curse of dimensionality. To effectively utilize all sample information, a sparsization method based on singularity criteria is used to continuously update the sample subset. This subset effectively reduces the redundant information of input samples while maximizing the coverage of input sample information, thus significantly reducing the solution size. The main steps include (1) initialize an empty sample set , and when obtaining the first sample vector , make ; (2) when the number of samples is greater than 1, calculate the minimum distance between the new sample and the current sample set; (3) when the minimum distance is less than a preset threshold , is not added to , otherwise the prediction error is calculated; and (4) when is greater than the preset threshold , add to and update to , otherwise proceed to (2). The thresholds and are used to control the accuracy and scale of the solution. Increasing and is beneficial for reducing the size of sample set , but it will lead to a decrease in the performance of the beamformer. This article takes as 0.1 of the kernel radius and as the root mean squares value of the average steady-state mean square error, which can achieve high computational accuracy on a smaller scale. The steps to establish the parametric model of the cavity filter using MLSSVR algorithm are as follows: Step 1: Set the initial value of , , and . Based on the collected sample dataset, the MLSSVR algorithm is used to establish the regression function of the multi-input and output tuning system for the cavity filter. Step 2: Substitute the input and output matrices and relevant parameters of training samples into the subfunction. Use the established initial regression function to detect the test sample data and calculate the test error ; Step 3: For any -th dimension output, the value of is determined according to ; Step 4: According to the calculated and , the regression function of the cavity filter multi-input and output tuning system is established. The work flow chart of the improved whale algorithm is shown in Figure 3.

5. Parameters Identification
The forecast accuracy of MLSSVR is mainly affected by the kernel function parameter , the penalty factor and , where the value of the penalty factor corresponds to the empirical risk generated by the training samples. If the penalty factor is too small, the punishment for exceeding the insensitive band in the sample is too small, and the training error is too large. On the contrary, if the penalty factor is too large, the error penalty will be too large, while the model structure restriction will be reduced, resulting in the model being too complex; the kernel function parameters can reflect the distribution of sample data in high-dimensional feature space information.
5.1. Differential Evolution Whale Algorithm
DWA is a heuristic-intelligent algorithm that imitates the behavior of whale predation. Compared with the traditional optimization algorithm, it has the advantages of less control parameters, a simple implementation, and high flexibility. DWA imitates the whale’s predation method, which mainly includes three stages hunting predation, bubble net predation, and free search. The position of each individual represents the potential solution, and the global optimal solution is obtained by constantly updating the position of the whale.
Shrink Surround: Since there is no prior knowledge of the global optimal solution of the search space before solving the optimization problem, a population of individuals is randomly selected as the target for predatory activities, and other whale individuals in the population are surrounded by the optimal individuals. The position update equation is as follows [28]:where denotes the position of the -th whale at iteration . and denote the coefficient vector, and , , denotes the convergence factor of linear descent. denotes the random number in . In the process of bubble net predation, the position update between whale and prey is expressed by the following logarithmic spiral equation [29]:where denotes the distance between the search individual and the current optimal solution, denotes the spiral shape parameter, denotes a random number uniformly distributed in , and denotes the current best position vector. The algorithm can choose the predation behavior by setting the probability of the predation mechanism. The whale’s position update equation is as follows [30]:where , each whale is in the phase of gradually surrounding the current optimal solution. To ensure that all whales can search completely in the solution space, the distance between whales in the whale algorithm is used to update the position to achieve the purpose of random search. Therefore, when , the search experience will swim toward the random whale.where denotes the position of the current random individual. denotes the distance between the current search individual and the optimal solution. To increase the diversity of the population, the differential evolution is introduced in the process of population position updating, which avoids the premature phenomenon brought by the local optimal solution. The location updating formula is as follows:where and denote the optimal individual position and the random individual, respectively, denotes iterations and is a random number in . This study combines the powerful exploration capabilities of differential evolution algorithm with the whale algorithm to enhance the development space of the whale algorithm. First, the whale algorithm is used to preprocess the position of the individual population and generate an initial population. Then, the differential evolution algorithm is used to cross-select the generated population to achieve population update iteration. The algorithm process is shown in Table 1.
To prevent over- and underfitting of the model, the insensitivity coefficient is given empirically in this study. Select the following sample mean square deviation as the evaluation function of DWA:where and denote the measured and expected values of the model, respectively. The optimization goal is to make reach the minimum value. denotes the population number. The fitness function is . Parameter settings of the DWA algorithm are as follows: the evaluation number , the population size , differential variability factor , and crossover rate . Compare DWA with GA (population size , maximum generation , replication probability , crossover probability , mutation probability ), PSO (particle number , maximum particle velocity , acceleration constant ), and GWA (, , ), the parameter optimization curve is shown in Figure 4. The optimal parameters for MLSSVR prediction model using the DWA as follows: , , and .

The optimization results show that the improved whale optimization algorithm has a higher convergence accuracy and better stability than the other optimization algorithms and basic whale optimization algorithms.
6. Simulation Results and Analysis
To evaluate the accuracy of the parametric model of the coaxial cavity filter, the standard deviation (STD), the maximum absolute error (MAE), and the correlation coefficient of three different methods are used for analysis and comparison. The calculation equation is as follows:where and denote the real and predicted values of the measured data, respectively. and denote the number of samples and output variables, respectively. Substitute the tuning process relationship data into the parameter prediction model, and compare the kernel clustering back propagation neural network (KC-BPNN) and multioutput fuzzy support vector regression (MFSVR) with the KCCA-MLSSVR method proposed in this paper. The prediction results are shown in Figures 5 and 6. It can be seen from this figure that the STD and MAE of the proposed method prediction results are lower than those of the other two models; in addition, the absolute value of by the proposed model is close to 1, which is higher than the other two methods. The above results indicate that the proposed method can not only effectively reduce the impact of high-dimensional data on the model’s prediction accuracy but also improve the model’s adaptability and generalization ability. The -parameter simulation curve of the ninth-order cavity filter with three different methods is shown in Figure 7.



The purpose of cavity filter tuning is to make the -parameters of the output response meet the key performance indexes (KPI) required before production. To further verify the feasibility of the method, a ten-order cross-coupled cavity filter with complicated topology is taken as the experimental object for simulation experiments. Kpis are as follows: center frequency , passband width , the maximum return loss , maximum insertion loss . It can be seen from Figure 8 that the return loss, insertion loss, and out-of-band suppression of the proposed method can meet the KPI.

For more quantitative comparison, the insertion loss , return loss , and left and right out-of-band suppression ( or ) of -parameters calculated by different methods are shown in Tables 2 and 3. From the data given in the table, the proposed method in this study is consistent with the KPI, while the -parameters using KC-BPNN, LS-SVR, and Gaussian process regression (GPR) methods have large deviation.
7. Conclusion
The tuning process of the coaxial cavity filter has great nonlinearity, and the output characteristics vary greatly in different tuning stages. To solve this problem, a parametric model method for coaxial cavity filter fusing KCCA and MLSSVR is proposed in this study. First, the redundant eigenvectors are digested by KCCA, which not only reduces the input vector dimension of the prediction model but also accelerates the training speed of the system. Second, a parametric model using MLSSVR has higher prediction accuracy than traditional modeling methods. Finally, the model parameters are optimized by DWA, which improves the adaptive ability and convergence speed of the model. The experimental results show that the proposed method in this study has higher prediction accuracy and generalization performance in the process of cavity filter model parameter fitting.
Data Availability
The data used to support the findings of this study are available from the corresponding author upon request.
Conflicts of Interest
The authors declare that they have no conflicts of interest.
Acknowledgments
This work is supported by the National Natural Science Foundation of China under grant 62141102, Jiangxi Provincial Natural Science Foundation under grant 20202BAB202008, Doctoral Science Research Foundation of East China University of Technology under grant DHBK2019178, and the Education Department Foundation of Jiangxi Province under grant GJJ2200726.