Abstract
Aiming at the problem of the low bearing faults identification accuracy of the method based on the deep neural network under small samples and multiple working conditions, a novel bearing fault identification method combined with the coordinate delay phase space reconstruction method (CDPSR), residual network, meta-SGD algorithm, and the AdaBoost technology was proposed. The proposed method firstly calculates the high-dimensional space coordinates of bearing vibration signals using the CDPSR method and uses these coordinates to construct a training set, then learns and updates the parameters of classifier networks using the meta-SGD algorithm with the train set, iteratively trains multiple classifiers, and finally integrates those classifiers to form a strong classifier by AdaBoost technology. The 4-way and 20-shot experiments of artificial and natural bearing faults show that the proposed method can identify the fault samples and nonfault samples with 100% accuracy, and the fault location accuracy is over 90%. Compared with some state-of-the-art methods such as WDCNN and CNN-SVM, the proposed method improves the fault identification accuracy and stability to a certain extent. The proposed method has high fault identification accuracy under small samples and multiworking conditions, which makes it applicable in some practical areas of complex working conditions and difficulty obtaining bearing fault signals.
1. Introduction
From the perspective of the application, the bearing fault identification can be divided into two types, which include single work conditions and multiworking conditions. Bearing fault identification of small samples under multiworking conditions refers to the problem of predicting which fault type the test bearing samples belong to under a few fault training samples collected from complex work conditions. The problem is also called N-way and K-shot multiworking bearing fault identification problem, where the training data include N classes, each class only has K samples, and the K is no more than 20. The movement of the faulty bearing is nonlinear. In order to fit the nonlinear features included in the rolling bearing vibration signals, many studies used the neural network to learn bearing vibration signal features and got good results in bearing fault identification by establishing the mappings between the fault feature and fault. For example, literature [1] and literature [2] used convolution neural networks (CNNs) to repeatedly learn the time-frequency features of a large number of original bearing vibration signals and realized the precise classification of bearing faults. The fault identification method based on the deep neural network usually requires a large amount of training data, and it is hard to play its ability to fit nonlinear features under small samples. Literature [3] and literature [4] show that the common problem of small sample learning is that the feature extraction effect is not very high and the identification accuracy is poor. Literature [5] shows that the bearing fault identification accuracy based on rolling bearing vibration signal analysis is affected by the bearing conditions. Rotation, load, and faults are the key factors affecting the motion state of the bearing rotor. When the speed of bearing continues to increase, due to the failure occurring in bearing such as touching, the nonlinear features of the rotor will further intensify, and the movement will even evolve into chaotic motion. The bearing fault identification under multiworking conditions is more difficult than the identification in a single working condition. To improve the identification accuracy of the method based on deep neural networks for multiworking bearing faults, literature [6] and literature [7] used attention, label smoothness, and other auxiliary algorithms to optimize the training processes of the neural network. Under the limited training data condition, the identification performance of the optimized deep CNN has been improved, but the number of each class of training sample still exceeds 20. At present, the bearing fault identification of small samples under multiworking conditions is one of the research hotspots in deep neural network applications.
In recent years, some scholars have conducted research on bearing fault identification under small samples and multiworking conditions from different viewpoints. Literature [8] used CNN to extract the early bearing fault of the motor, and literature [9] used CNN to extract the features in the spectrum image of the bearing vibration signal. Different neural network structures have different capabilities for extracting bearing fault features, and the study of network structures to improve bearing vibration performance has received widespread attention. The training strategy of neural networks has also been the main direction of deep neural network research. MAML and SAMME are outstanding works in this area. Data preprocessing technology generates new training samples through geometric transformation, which was widely applied to prevent the neural network from overfitting in the process of learning small samples. Literature [10] researched data preprocessing methods based on the sparse autoencoder, and literature [11] used sliding sampling to enhance bearing vibration data. Now, fault identification of small samples based on learning strategies and data preprocessing is one of the important research directions for bearing fault diagnosis. In this paper, the fault signals collected from multiple working conditions with different speeds, different loads, and different fault degrees were used as the research objects. We first employed the coordinate delay phase space reconstruction (CDPSR) method to process bearing vibration signals, then designed a deep bearing fault identification neural network based on CNN and residual network to produce some classifiers and finally, through the Adaboost method, implemented an integrating multiply classifiers algorithm for training and integrating these bearing fault classifiers to form a stronger bearing fault classifier.
The rest of this paper is organized as follows. Section 2 introduces related technologies such as the CDPSR method, residual network, meta-SGD algorithm, and the Adaboost technology. Section 3 details our bearing fault identification method, Section 3.1 presents the processes of the data preprocessing and building new training set, Section 3.2 designs the network structure of the bearing fault classifier and its training method, and Section 3.3 develops the integrating step of multiple bearing fault classifiers; Section 4 describes the steps of conducting validation experiments using the artificial bearing fault dataset and the natural bearing fault dataset and discusses the experiment results; Finally, Section 5 concludes this paper and highlights the significance of the proposed method.
2. Related Work
2.1. Phase Space Reconstruction
The coordinate delay method [12] is a specific implementation of the phase space reconstruction technology, which can be used to calculate the phase space coordinates of the bearing vibration time-series signals. Suppose is a one-dimensional time-series signal, are delay time, and M is the embedded dimension of high-dimensional phase space. According to the coordinate delay method, the phase space coordinates of M-dimensional can be represented as equation (1), where i = 0, …, M − 1.
Solving the phase space coordinates of the time sequence requires τ and M. Calculating delay time τ has a variety of methods such as correlation coefficients and other methods based on mutual information judgment. The related function method for solving τ is simple and effective. The autocorrelation function of the time sequence is a function that changes with τ. When drops to the , the time is the best delay time τ for reconstructing phase space [13].
When the delay time τ is known, the Cao algorithm [14] can be used to solve the embedding dimension M, and it can calculate the embedding dimension with a small amount of data. Denote the i-th d-dimensional reconstruction vector as , and define , where is the reconstruction vector of the i-th (d + 1) dimension. Since , is in the d-dimensional phase space, the mean of is , and . If stops changing from a certain , then is the minimum embedding dimension to be found.
2.2. Convolution Neural Network
Convolution Neural Network [15] is made up of neurons that have learnable weights and biases; it uses the convolution layer and nonlinear activation functions to abstract the original data layer by layer as features required for specific tasks to achieve mapping of features and targets. The CNN is a sequence of layers that mainly includes the convolution layer, the pooling layer, and the fully connected layer. We can stack these layers to form a full CNN architecture. Since CNN has good nonlinear fitting capabilities, it was widely used in fields such as image feature extraction and voice feature analysis. There are many excellent CNN applications in the field of bearing fault identification [16, 17].
2.3. Residual Network
The residual refers to the gap between the observed value and the prediction value; literature [18] applied it to the neural network and proposed the concept of the residual block whose structure is shown in Figure 1.

Here, and represent the input and output of the residual blocks, represents the observed value, and the residual block can be expressed as .
If and are directly mapping functions, for example, and , the gradient of the L-layer can pass to any layer that is shallower than it [19]. Therefore, as the number of network layers increases, the residual network will not degenerate, and the ability of fitting nonlinear features becomes stronger. So far, many studies were using residual networks to extract features of business data. For example, literature [20] studied using the SELU activation function to optimize the deep residual network, used the deep residual network model to analyze information dissemination in wireless networks, and obtained a complete media information dissemination prediction of wireless networks.
Resnet is a residual network structure stacked by the residual blocks, which has performed well in image classification applications [21, 22]. Literature [23] researched the advantage of Resnet-18 by comparing Resnet-18/50, VGG-19, and Googlenet and found that Resnet-18 has the advantages of short training time and high accuracy, developed a deep learning model base on ResNet-18 to diagnose the fan blade surface damage, and got good recognition effect. Literature [24] proposed a dual attention residual network that uses a residual module from Resnet18 to detect oil spills of various shapes and scales. Resnet-18 is an implementation of the Resnet model, and it consists of five parts including Conv1_x, Conv2_x, Conv3_x, Conv4_x, and Conv5_x. Each part has a convolution and constant block. Resnet-18 was designed for the classification of thousands of pictures, its network structure is complicated, and it needs to take a long time to train. Compared to the classification of thousands of pictures, the task of bearing fault identification under dozens of working conditions is small. Therefore, this paper deleted the Conv4_x and Conv5_x network structures in the Resnet-18, modified the size of the convolution kernel, reduced the complexity of the Resnet-18, and still retained the advantages of the residual network.
2.4. Meta-SGD Algorithm
The Model-Agnostic Meta-Learning (MAML) algorithm [25] proposed by Finn et al. can be used to train the model that can be optimized by gradient descent algorithm; it is an excellent algorithm in the field of meta-learning. The difference between the MAML and other optimization algorithms such as the stochastic gradient descent (SGD) algorithm [26] is that the optimization process in MAML is divided into two steps, which first assumes that n tasks () are selected from the supporting dataset and each task is used to calculate the gradient based on the current neural network parameter to get n updated neural network parameter sets , , second calculates the update gradient on the query dataset, and determines the new neural network parameter according to .
As the α and β hyperparameters are fixed in the MAML, they cannot be adjusted with the change of the network, which makes the training process fluctuate. The Meta-SGD algorithm [27] automatically adjusts the α parameter to increase the stability, whose execution steps are as follows.
Meta-SGD can learn task-agnostic features rather than simply adapt to task-specific features [28]. Literature [29] used the learning rate of the meta-learning SGD to predict streaming time-series data online and achieved good results.
2.5. Multiclass AdaBoost Method
AdaBoost [30] is a multimodel integration method, and its final output is the weighted sum of the results of the integrated multiple classifiers. AdaBoost was originally raised by Freund to research the binary classification problem. By introducing multiclassification index loss in the front-directional model, AdaBoost also can be applied to the multiclass problem. The sample weight is constantly updated in subsequent AdaBoost training, and samples that have not been identified correctly will be set up with greater weights. This process of AdaBoost training is shown in Figure 2.

SAMME [31] is an implementation algorithm of the multiclass AdaBoost, its specific implementation steps are as shown in Algorithm1.
|
Literature [32] hybridized the AdaBoost with a linear support vector machine model and developed a diagnostic system to predict hepatitis disease; the results demonstrate that the strength of a conventional support vector machine model is improved by 6.39%. Literature [33] studied the use of the AdaBoost framework to integrate other methods and proposed a wind turbine fault feature extraction method based on the AdaBoost framework and SGMD-CS. Experiments show that AdaBoost is an effective multimodal integration framework. To realize bearing fault identification, this paper studied the use of the AdaBoost framework to integrate multiple fault identification classifier models designed based on CNN and residual networks.
3. Our Method
3.1. Data Preprocessing and Building New Feature Set
Use min-max normalization (2) to handle the training samples including N_Train samples and test samples including N_Test samples. Divide the training samples into a training set and verification set according to the ratio of 5 : 1. Select one sample separately from the training set, verification set, and test set as an input of Algorithm 2 to calculate the best value of time delay τ and phase space dimension M. Assume that the training sample has L data points. According to (1), we can get coordinates for the training set, coordinates for the verification set, and coordinates for the test set, when the sampling frequency of the bearing vibration signal is F and the phase space vector is not reused.
|
Combine the reconstructed phase space coordinates with the labels of the original signals to build new training samples. Since the phase space has the same topological properties as the original bearing vibration signal system, the regularity of the bearing time-series signal in the high-dimensional space is restored in the coordinates. Therefore, any coordinate in phase space represents the state of the original bearing vibration signal system and contains corresponding features. Compared with the features included in the original signals, the features in the phase space coordinates are more obvious and easier to be identified by the classifier.
3.2. Bearing Fault Classifier and Its Training
Our bearing fault classifier is designed based on CNN and the residual block. The classifier uses 7 network layers, of which the Conv_x network consists of convolutional layers and ReLU activation operations. The full connection layer used 100 neurons, and the output layer used 4 neurons. The first layer of our network uses a larger convolution kernel and then reduces the size of the convolution kernel. The detailed structure is shown in Figure 3.

Our bearing fault classifier takes the phase space coordinates of the bearing vibration signal as input. In the classifier, the input matrix first flows through the Conv1_x and then through Conv2_x, Conv3_x, the fully connected layer, and the output layer in sequence. In the Conv1_x module, the input matrix performed convolution operations according to (3) using 64 different convolution kernels with a size of 7 7. To avoid the deviation of the result distribution after the convolution operation, the batch normalization (BN) technique is used to standardize the convolution result, so that the convolution result obeys a normal distribution with a mean of 0 and a variance of 1. Then, the result of BN is nonlinearly transformed using the ReLU activation function (). Different from the Conv1_x module, the Conv2_x module first uses the maximum pooling technique that takes the maximum value in a fixed-size sliding window to reduce the density of data features and then uses two convolution layers with smaller kernels to perform convolution operations on the input. After the second BN operation, Conv2_x uses the sum of the result of max-pooling (as the observed value of residual block) and the result of BN as the input to the nonlinear activation function. Conv3_x also has two convolution layers, the difference is that to make the observed value of the residual block of Conv3_x own the same shape as the second convolution result of Con3_x, the observed value is processed with a convolution of size 1 1. In the fully connected layer and the output layer, the input is processed in the same way as equation (4).
In (3), represents the input matrix of the l-th layer, K is the size of the convolution kernel, and and represent the connection weight and activation parameter of the i-th convolution kernel of l-th layer (all convolution kernels of the same convolution layer shared the same activation value). In equation (4), i represents the serial number of neurons in the full connection layer, j means the position number of the input matrix, represents the weight of the i-th neurons to the j-th value of the input, and is the bias of the i-th neurons of l-th layer.
To prevent overfitting in the training process, we added the following judgment statements after step 9 of Algorithm 2 and used the modified Algorithm 2 to update the network parameters of our bearing fault classifier. If Acc_of_Train ≥ 0.9 and Acc_of_Train − Acc_of_Validation > 0.1: Number_of_ Overfitting ++; else: Number_of_ Overfitting = 0; If Number_of_ Overfitting > 5: break;
3.3. The Integrating Step of Multiple Bearing Fault Classifiers
Combining multiple weak bearing fault identification classifiers designed in Section 3.2 can generate a stronger bearing fault identification classifier using the AdaBoost algorithm; however, unlike the SAMME, our method divides the training data into a support set and a query set, calculates each sample weight of the support and query set, and updates the classification error with the sample weights. The integrating steps of our bearing fault classifiers are shown in Algorithm 3.
|
This algorithm uses the Meta-SGD learning strategy to decrease the value of the objective loss function and update the network parameters; it not only has the characteristics of the meta-learning strategy to quickly converge but also adjusts the learning rate based on the learning task. The initialization of neural network parameters has an important impact on training. In order to make the network get good initialization parameters, the algorithm initializes the parameters of the next classifier in the parameters of the previous classifier during the iterative training process.
The algorithm needs to determine 2 hyperparameters in advance. They are the learning rate β of Algorithm 2 and the number of data points of the training sample. The two parameters have a great influence on the algorithm. First, the learning rate β will affect the learning speed of meta-SGD; second, the number of data points of the training sample determines the shape of the classifier input data, and the execution effect of the coordinate delay method algorithm will be affected by this parameter. If it is too large, it will increase the interference data, and if it is too small, the high-dimensional phase space of the bearing vibration time-series signal cannot be accurately established. In this paper, we set β with 0.001, and the number of the data points is 1024.
4. Experiment
The fault identification accuracy under different working conditions is an important indicator for measuring the rolling bearing fault identification method. This paper verified the effectiveness of the proposed method by calculating the test accuracies of bearing fault identification on the artificial and natural bearing fault dataset collected from different loads, different speeds, and different fault conditions.
The experiments were conducted on the Tensorflow CPU 2.7 platform programming with python. The hardware and software environments included core i7-4790K 4.0 GHz processor, 16 G memory, and Windows Server 2018 operating system.
To eliminate the impact of accidentality, each experiment in this paper was performed 5 times independently, and the average value of 5 test accuracy was used as the experimental results.
4.1. Experiments on the CWRU Dataset
Artificial bearing fault data set CRWU [34] consists of bearing vibration time serial signals in the state of normal, internal circle fault, outer circle fault, and rolling bodies fault. The CWRU dataset is a representative data set in the field of bearing fault diagnosis. Many scholars got positive results when they used the CWRU to perform simulation experiments [35, 36].
4.1.1. Experimental Data
The bearing vibration signals used in this case were sampled from the 6205-2RSJEM SKF rolling bearings at a sampling frequency of 12 kHz. The fault types of these signals are associated with 4 different bearing damage diameters, recorded as A (0), B (0.007 inches), C (0.014 inches), and D (0.021 inches). These signals were cut into several segments. Each contained 1024 data points and was used as an experimental sample. The experiment samples are selected from the 16 kinds of work conditions described in Table 1 and are used to carry out three groups of small sample experiments. Three groups of experiments randomly selected 10 samples per class for testing, and these testing samples have the same labels and same working conditions as training samples.
4.1.2. Experiment and Result Analysis
According to Algorithm 3, three group experiments were conducted. The first group experiment was a variable power experiment. The samples of normal, inner circle fault, outer circle fault, and rolling element fault are randomly selected from 8 working conditions numbered 1, 5, 9, 13, 2, 6, 10, and 14 in Table 1 to carry out 4-way and 20-shot experiments. Similar to the first group experiment, the second group experiment was a variable fault degree experiment, in which samples were randomly selected under four operating conditions numbered 1, 2, 3, and 4. The third group of experiments was the variable power and fault degree experiments, and samples of five operating conditions of 1, 3, 4, 5, 10, and 14 were randomly selected. The sample distribution of the three experiments is shown in Table 2. Calculated the test accuracy and the test results are shown in Table 3.
In the three groups of experiments, the test accuracies of the proposed method are greater than 95% and the standard deviations of the accuracy of the three experiments are within 3 when the number of training samples of normal, inner ring fault, outer ring fault, and rolling element fault is 20 and in a variety of working conditions composed of three different factors of bearing fault degree, speed, and load. Figure 4 is the convergence process of loss value and validation accuracy under the 4-way and 20-shot experiments with variable power and fault degree.

(a)

(b)
During the training process of the first classifier, the value of the target loss function continued to decrease, the test accuracy rate continued to increase, and the value of the network parameters was continuously optimized. The test accuracy of this classifier exceeded 90%. After the first classifier, the four classifiers were fine-tuned according to the training dataset with different sample weights in the training of the second, third, fourth, and fifth four classifiers. Finally, five different classifiers were obtained, which complemented each other, their test accuracy was all above 90%, and the final test accuracy of our method reached 100%.
The above N-way and K-shot experiments show that the proposed method can filter the influence of the three factors of speed, load, and fault degree on the bearing fault identification to a certain extent whether it is applied in the constant or variable condition. The proposed method has good accuracy and stability for the fault identification of small sample bearings with different speeds, loads, and fault degrees.
4.2. Experiments on XJTU-SY Dataset
To further verify the effectiveness of the proposed method, the natural bearing fault data set, XJTU-SY [37], was used for our experiments.
4.2.1. Experimental Data
The sampling frequency is 25.6 kHz in the XJTU-SY experiment, and the XJTU-SY dataset also includes the faults raised in the outer circle, inner circle, and rolling bodies. The bearing vibration signals sampled from eight different conditions were randomly selected to form training sets and test sets. Every experiment sample has 1024 data points, and its type and its labels are shown in Table 4.
4.2.2. Experimental Results and Analysis
The 4-way and 20-shot experiment (experiment 4) is carried out by randomly selecting normal, inner fault, outer fault, and rolling element fault samples from the working condition numbered 18 in Table 4. The test results are shown in Table 5.
The results of experiment 4 show that the identification accuracy of the proposed method for natural bearing fault is still high under the small number of training samples, and the test accuracy of the 4-way and 20-shot is more than 96% when the training samples and test samples are in the same working condition. Figure 5 shows the confusion matrixes of the predicted value of the test sample in the 5 experiments.

From the confusion matrix, the predicted values of label 0 were always consistent with the actual values, and the prediction accuracy was 100%; the prediction accuracy of other labels fluctuated slightly, and the error mainly came from the wrong prediction of label 1 as a label 3. In the end, the prediction results of each classifier were excellent. The prediction accuracies of label 1 were above 90% for 4 consecutive times, the prediction accuracies of label 2 exceeded 99%, and the prediction accuracies of label 3 exceeded 90% four times. The recognition rate of 100% fault samples and nonfault samples means that the proposed method can accurately distinguish between fault samples and nonfault samples, and the fault location accuracy rate of 94.7% shows that the method also has a good ability to identify natural bearing faults under known working conditions.
Our method can be viewed as a combination of CDPSR data preprocessing, residual network, Meta-SGD, and AdaBoost, which is insensitive to changes in bearing load, rotational speed, and failure degree. To analyze each part in the proposed method, we performed the ablation experiments. First, we analyzed the contribution of the residual network designed in Section 3.2 by comparing two different types of combinations of CNN + CDPSR + Meta-SGD + AdaBoost and ResNet + CDPSR + Meta-SGD + AdaBoost, in which CNN consisted of a stack of simple convolutional layers and max-pooling layers. Second, the effect of the Meta-SGD learning strategy method is analyzed by comparing ResNet + CDPSR + Meta-SGD + AdaBoost and ResNet + CDPSR + SGD + AdaBoost, and the use of SGD was described in reference [23]. Then, by comparing Resnet + Meta-SGD + AdaBoost and ResNet + CDPSR + Meta-SGD + AdaBoost, we discussed the influence of the reconstructed bearing vibration timing signal on the method. Finally, the influence of AdaBoost on this method was analyzed by comparing ResNet + CDPSR + Meta-SGD and ResNet + CDPSR + Meta-SGD + AdaBoost. The test results of the ablation experiment are shown in Figure 6.

From the results in Figure 6, the bearing fault recognition accuracy of the classifier using the residual network was 6% higher than that of the CNN classifier, Meta-SGD had a greater improvement in bearing fault recognition than SGD, and the CDPSR method played a positive influence on the fault identification. From the variance of the test accuracy of ResNet + Meta-SGD + CDPSR and ResNet + Meta-SGD + CDPSR + AdaBoost, we can see that AdaBoost also played a positive role as a stabilizer for the proposed method. Hence, the four parts of CDPSR, residual network, Meta-SGD, and AdaBoost had made positive contributions to the proposed method. Residual network and Meta-SGD can effectively improve the test accuracy; CDPSR data preprocessing and Adaboost have significant impacts on the stability of the method.
To analyze the advantages of our method, the comparison experiment was carried out under the same conditions as experiment 4. The test accuracy of the proposed method, the WDCNN method [16], and the CNN-SVM method [17] would be compared.
The WDCNN method proposed by Zhang et al. in 2017 included a specific deep network (WDCNN network), which takes the wide kernels in the first as convolutional layers and small convolutional kernels in the preceding layers. The method slices the training samples with overlap to obtain huge amounts of data, then uses raw vibration signals as input, and trains the WDCNN network using the backpropagation algorithm detailed in reference [16].
The CNN-SVM method combines both the merits of CNN and SVM, which firstly uses the 2D representation of raw vibration signals as input, then trains the original CNN with the output layer for several epochs until the training process converges using stochastic gradient descent, and finally replaces the output layer with the SVM, which included the radial basis function (RBF) kernel. The optimum scheme used in this method was elaborated in reference [17].
The WDCNN and the CNN-SVM method are typical methods in the application of fault-bearing recognition. The fault identification results of the comparison experiment are shown in Figure 7.

In the comparison, the residual network designed for our method and the network used in the WDCNN method have the same number of network layers and the same scale of learning parameters. The average result of the five CNN-SVM experiments was 82.5%, and the average result of the WDCNN was 90.5%. The test accuracies of both the WDCNN and our method were above 90%, and our result was 96%. Compared with the test accuracy of CNN-SVM and WDCNN, the test accuracy of our method was higher. Moreover, the test results of our method were 95 to 97.5, the value range was relatively narrow, while the value range of the other two methods was relatively much wider. Therefore, the proposed method has a certain improvement in the accuracy of bearing fault identification, and its stability is better.
5. Conclusions
A novel bearing fault identification method for multiconditions and small samples was proposed to challenge the problems of lacking fault data and poor performance. To verify the effectiveness of the proposed method, the artificial and natural bearing fault signals were taken to experiment with as a case study. The result shows that the proposed method realized accurate fault signal identification under multiple working conditions and small samples, and its accuracy rate of bearing fault positioning exceeds 90%. Benefitting from the reconstruction of high-dimensional space of bearing vibration time series by coordinate delay construction method, extraction of the phase space features using the convolutional neural network, the transmission of the gradient to other layers by residual block, updation of the classifier parameters by Meta-SGD, and integration of multiple classifiers by AdaBoost method, the proposed method gets excellent bearing fault feature extraction and high fault identification ability. Finally, compared with other advanced methods, the proposed method also has certain advantages. From these cases, the proposed method is very effective.
The proposed method can accurately identify bearing faults under small samples and multiworking conditions without manually setting fault features. Therefore, the proposed method has a certain value in some application areas with complex working conditions and difficulty obtaining a large number of bearing fault samples, such as aviation bearings.
Data Availability
The data used to support the findings of this study are available from the corresponding author upon request.
Conflicts of Interest
The authors declare that they have no conflicts of interest.
Acknowledgments
This work was supported by the Growth Project of Young Scientific and Technological Talents in Guizhou for Colleges and Universities (grant no. Qian Jiao He KY2020137) and the tripartite joint fund project of Guizhou Provincial Department of Science and Technology (grant no. LH2016 7275).