Abstract

As an extension of Dempster–Shafer (D-S) theory, the evidential reasoning (ER) rule can be used as a combination strategy in ensemble learning to deeply mine classifier information through decision-making reasoning. The weight of evidence is an important parameter in the ER rule, which has a significant effect on the result of ensemble learning. However, current research results on the weight of evidence are not ideal, leveraging expert knowledge to assign weights leads to the excessive subjectivity, and using sample statistical methods to assign weights relies too heavily on the samples, so the determined weights sometimes differ greatly from the actual importance of the attributes. Therefore, to solve the problem of excessive subjectivity and objectivity of the weights of evidence, and further improve the accuracy of ensemble learning based on the ER rule, we propose a novel combination weighting method to determine the weight of evidence. The combined weights are calculated by leveraging our proposed method to combine subjective and objective weights of evidence. The regularization of these weights is studied. Then, the evidential reasoning rule is used to integrate different classifiers. Five case studies of image classification datasets have been conducted to demonstrate the effectiveness of the combination weighting method.

1. Introduction

Ensemble learning is a branch of machine learning, and using decision-making reasoning as a combination strategy helps ensemble learning to integrate better. The decision-making reasoning is the process of making a choice among many options and summarizing evidence to draw a conclusion, so it can help ensemble learning to integrate. The ER rule is a combination strategy [1], which can integrate classifiers through decision-making reasoning to obtain better results, so it is used as a decision-making strategy for ensemble learning in this paper.

In 2013, the ER rule, which considers the weight and reliability of evidence, was established by Yang and Xu [1]. The ER rule, which is an extension of the D-S theory [2, 3], clearly distinguishes the importance and reliability of evidence and constitutes a general joint probabilistic reasoning process. The counterintuitive problem encountered in Dempster’s rule is solved by assigning weight and reliability to evidence. Due to the superiority of the ER rule, it has been applied in many different applications. A fault detection method based on the data reliability and interval ER rule was proposed, which improved the accuracy of fault detection [4]. A multi-index framework was proposed based on the combination of objective quality indicators, subjective expert judgements, and patient feedback based on the ER rule and improved model robustness under uncertain conditions [5]. The above research provides a theoretical basis and technical support for the research and application of the ER rule. The ER rule has become an important element in the field of nonclassical reasoning and can deeply mine classifier information through decision-making reasoning and improve the integration model.

The weight and reliability of evidence can be as key parameters in the ER rule have been extensively studied in existing research. There are generally three ways to set the weight of evidence in the ER rule. First, the weight of evidence in the traditional ER rule is given by expert knowledge. For example, the order relation analysis method (G1) was used to quantitatively evaluate the comprehensive intelligence of autonomous vehicles [6]. The weights used in the literature [7, 8] were assigned by experts who actually assessed indicators from industrial processes. But this kind of human subjective weighting method is overpowering and prone to deviation in results. Second, some researchers have recently improved the setting of the weight of evidence. For example, Wu et al. [9] proposed popular calibration weighting methods such as generalized regression and generalized exponential tilting and verified their effectiveness. Three methods of the maximum deviation method, coefficient of variation method, and entropy weight method were combined to establish an index optimal adaptive weighting model [10]. A novel weighted majority voting ensemble (WMVE) approach was proposed to assign different weights to classifiers [11]. A comprehensive evaluation method of power quality based on the ratio of the weighted rank sum ratio method was proposed to correctly sort and classify the power quality of each bus [12]. Zhou et al. [13] carried out physical and chemical analyses and used the coefficient of variation, mean and standard deviation in a model to predict the proximate and ultimate analysis and heating value from physical compositions. These methods assigned weights in an objective way. But sometimes, heavily relying on the sample causes the weight setting to be inconsistent with the actual situation. Third, mathematical optimization was adopted to produce a dual-objective optimization model that was trained to obtain the weight of diagnostic evidence [14]. However, the mathematical optimization method needs a large number of samples, and it is difficult to obtain enough training samples in practice. Most of the proposed weighting methods are only applicable to a specific engineering background and are not universal. There is still a lack of a general method to determine the weight of evidence. Therefore, to set a more general and reasonable evidence weight, this paper proposes to use the combination weighting method to set the weight of evidence in the ER rule.

Combination weighting is a method that combines subjective weighting and objective weighting. This method can minimize the loss of information and can reflect both subjective and objective information, thereby making the determined weights more reasonable. At present, the combination weighting method has been widely used to weight the risk factors for fault modes to determine the risk priority of fault mode recognition [15]. This method was also used to determine the assessment factors of project managers, and a new method of manager performance evaluation was proposed [16]. The game theory method was proposed to combine the subjective weighting method and objective weighting method. The evaluation system of the educational informatization index was established [17], and it was verified in the Suzhou area. An et al. used the IGAHP-CRITIC combination weighting method to effectively integrate subjective and objective weight information, which effectively reduced the uncertainty and incompleteness of an equipment quality state assessment [18]. Since the setting of weights directly affects the accuracy of integrated learning using the ER rule, the complexity of objective things causes objective weighting to rely heavily on samples, and subjective weighting may lead to inaccurate weight settings due to the ambiguity of personal thought processes. Therefore, research in the literature [19] verified the rationality of the combination weighting of the maximum difference by comparing the combination weighting method of addition, multiplication and the combination weighting method of maximum difference and provided a new idea for the combination of subjective and objective weights. The weighting method of maximizing the difference is used to evaluate the operation status of the electricity market in the literature [20]. This method achieved excellent results, reflecting its practicality and innovation. The effectiveness of the combination weighting method was proven by these examples.

The application of the ER rule to ensemble learning is a new research field in the context of artificial intelligence. In ensemble learning, the generation of classifiers and the choice of combination strategies have a great impact on the quality of the results [21]. At present, a variety of methods have been applied as a combination strategy in the field of integrated learning, such as the voting method [22], average method [23], learning method [24], and D-S theory [2527]. However, these methods have difficulty effectively mining the relationship between classifiers, and the ability to integrate classifiers is limited. Based on its advantages in improving uncertainty and reasoning, the ER rule can allow integrated models to obtain better results than strategies such as voting. Additionally, the ER rule does not change the traditional probabilistic reasoning model, so it is more effective for classifiers to integrate this kind of uncertainty information. As a powerful engine and information fusion mechanism [1], the ER rule can effectively improve the effect of ensemble learning.

In summary, a new ensemble learning model is constructed based on the ER rule. The contributions of this paper are described as follows:(1)The combination weighting method is used to determine the weight of evidence in the ER rule. It can further find the effective information in the process of combining evidence, take subjective factors and objective factors of evidence into consideration, and set more general and reasonable weight of evidence for the ER rule.(2)The ER rule is an information fusion method that does not contain labels. When setting weights, whether it is necessary to regularize the combination of weights has always been a matter of debate among existing researches. Accuracy is used as an indicator to study the regularization of weights in the ER rule and our own suggestions are given in this paper.(3)An ensemble learning method is proposed based on the ER rule considering combination weighting, and it can effectively improve the effect of ensemble learning.

The rest of this paper is organized as follows: using combination weighting method to determine weights for the ensemble learning model based on the ER rule is proposed in Section 2. The combination weighting method is constructed in Section 3. The ensemble learning model based on the ER rule is constructed in Section 4. Cases are used to verify the effectiveness of the combination weighting method is presented in Section 5. The conclusion of this paper is provided in Section 6.

2. Problem Definition

In the application of ensemble learning based on the ER rule, to demonstrate that the ER rule has a better improvement effect on the ensemble learning model, it is necessary to adopt appropriate weighting methods to set the weight of evidence. The main problems to be solved in this paper are as follows:

Problem 1. construction of the combination weighting model and regularization of weight). At present, when assigning the weight of each classifier in ensemble learning, each classifier is used as a piece of evidence. Classifiers are usually generated by black-box models and are not interpretable, so it is difficult to provide accurate weight values through expert knowledge. In addition, the number of classifiers used in ensemble learning is small, and it is difficult to determine the weights in a mathematically optimized way. In the commonly used weighting method, the subjective weighting method is strongly influenced by human factors, and the objective weighting method depends a lot on samples. Therefore, subjective weighting and objective weighting are combined in this paper. We established a combination weighting model for weights:where is the combined weight of the classifier , is a combination method of subjective weights and objective weights, is the subjective weight of the classifier calculated by the subjective weighting method, is the objective weight of the classifier calculated by the objective weighting method, and is the parameter combination of the weight combination process. It is uncertain whether different combination weighting methods can make use of the advantages of subjective weights and objective weights, which needs to be verified in studies. The weight determined by the combination weighting method may exceed the range of . For weight combinations that exceed the range of , it is necessary to regularize this weight, but for weight combinations that do not exceed the range of , whether regularization is required has always been a controversial issue when using the ER rule. Assuming there are classifiers in total, the regularization process of weights can be described aswhere is the regularized result of weight and is the combined weight of the classifier .

Problem 2. (construction of the ensemble learning model based on the ER rule). In the current ensemble learning research, how to use a suitable combination strategy to combine weak learners into strong learners is the main problem. As an ensemble strategy, the ER rule can consider the relationship between classifiers, overcome the shortcomings of algorithms that only consider the numerical relationship of the classifier data, and significantly improve the effect of ensemble learning. Suppose a total of classifiers are integrated, and the process of the ensemble learning model based on the ER rule is as follows:where is the classification result of the classifier , , is the classification process of the classifier , is the dataset, and is the parameter set in the classification process. After the classification result is obtained, the ensemble result of the model can be described aswhere is the integration process of the model, is the set of classification results of classifiers, and is the parameter set in the ensemble process.
To solve the above problems, an ensemble learning model based on the ER rule considering combination weighting is proposed in this paper, as shown in Figure 1. It consists of two parts: combination weighting and the ER rule, which are introduced in detail in Section 3 and Section 4, respectively.

3. Construction of Weighting Methods

The weight of evidence is an important part of the ER rule, and the weighting method is the focus of this paper. Some representative weighting methods [28] are selected for definition in this section. The weights identified in this section will be used in Section 4.2.

3.1. Subjective Weighting

The subjective weighting method is a method of weighting based on the subjective information of the decision-maker. This method can reflect the degree of importance the decision-maker attached to different attributes and can flexibly grasp the importance of each decision-making attribute. However, its flexibility and variability can carry too much subjective arbitrariness. The subjective weighting method chosen in this paper is the analytic hierarchy process (AHP).

3.1.1. Analytic Hierarchy Process

The analytic hierarchy process is based on the experience of the decision-maker, combining quantitative analysis with qualitative analysis, judging the relative importance of each measurement target, and providing the weight of the decision plan.

The first step is to build a hierarchical structure model. The relevant factors are decomposed into several levels from the top to bottom according to different attributes, and the goals, criteria, and objects of decision-making are divided into the highest level, the middle level, and the lowest level, respectively.

The second step is that a judgement matrix is constructed. Based on the accuracy of each classifier, the classifiers are compared in pairs, a scale is given, and a matrix is formed.

The third step is hierarchical single sorting and consistency checking. Hierarchical single sorting calculates the importance weight of the index related to the level for the index of the upper level according to the judgement matrix. Then, a check is provided on whether the judgement matrix is a consistent matrix to determine the weight of the classifier.

The fourth step is that the total taxi of hierarchy occurs, and consistency is checked. After calculating the comprehensive weights of all classifiers, they are sorted to obtain the importance results.

3.2. Objective Weighting

The objective weighting method refers to the method of using the objective information of each attribute to determine the attribute weight. The decisions or evaluation results have a strong theoretical basis. However, the objective weighting method cannot address the subjective wishes of the decision-maker, which sometimes it goes against the actual importance of the attribute. The objective weighting methods used in this paper are the entropy weight method (EW), coefficient of variation method (COV), and CRITIC method.

3.2.1. Entropy Weight Method

The entropy weight method is based on the amount of information contained in the data in the evaluation index system and determines the index weights by calculating the information entropy. The calculation steps are as follows.

Suppose there are evaluation samples, classifiers, and categories. Then there will be evaluation indicators , , and is the index value of the samples ; the indicators used in this paper are all positive and the value of the evaluation index is calculated through the standardization of :

Then, according to the evaluation index value , the standardization method is used to construct the evaluation matrix , and the process is shown in the following equation:

After obtaining the standardized positive indicators, then their information entropy is computed. The calculation of the information entropy of the index in the matrix is shown in the following equation:

The normalization method is used to determine the weight of the abovementioned positive indicators. Since there are classifiers and categories, there are a total of positive indicators. The information entropy of each group of data calculated by Equation (7) is: , and the indicator weight is calculated by using Equation (8) to calculate the information entropy.where is the weight of the indicator . Therefore, the weight determined by the entropy weight method of the classifier can be expressed as , and each classifier recognizes the data of categories to obtain weight values. The weight of the index is calculated by using Equation (8). Therefore, the weight of the classifier is calculated by

3.2.2. Coefficient of Variation Method

The coefficient of variation method is based on the degree of variation between the current value of each index in the evaluation index system and the target value, and the index weight is determined by the method of calculating the degree of change of each index in the system. The calculation steps are as follows.

The steps in the entropy weight method in Section 3.2.1. Assuming there are evaluation samples, classifiers, and categories, there are evaluation indicators. The evaluation index is calculated by Equation (5).

After the evaluation index is calculated, the mean value and standard deviation of the evaluation index are also calculated, and the coefficient of variation is calculated through the mean value and standard deviation of the index. The calculation process of the coefficient of variation of the index in the matrix is as follows:where is the standard deviation of the evaluation index and is the average value of the evaluation index . Then the coefficient of variation is normalized, and the weight of each evaluation index is calculated as follows:where is the weight of the indicator . The weight determined by the coefficient of variation method for classifier can be expressed as , which can be calculated by Equation (9).

3.2.3. CRITIC Method

The CRITIC method is based on the contrast strength and conflict of each index in the evaluation index system, and it determines the index weight through the objective attributes of the data itself. The calculation steps are as follows:

The steps in the entropy weight method in Section 3.2.1 are also applied here. The evaluation index is calculated by Equation (5). Then the amount of information of the index in the matrix is calculated:where is the standard deviation of the index and is the correlation coefficient between index and index .The correlation coefficient is used to express the correlation between indicators. If the correlation between an indicator and other indicators is stronger, the conflict will be smaller, and the weight assigned to the indicator should be reduced. The weight of evaluation indicator is as follows:where is the weight of the indicator .The weight of classifier can be calculated by Equation (9).

In view of the shortcomings of the above methods, the combination weighting method can effectively solve the problem of too strong subjectivity or objectivity of a single weighting method. At the same time, no one had used the combination weighting method to determine the weight of evidence of the ER rule. Therefore, the combination weighting method is defined in Section 3.3.

3.3. Combination Weighting

Combination weighting can effectively address the subjective and objective information of attributes, so it is more reasonable for setting the weight of evidence. This paper uses three combination weighting methods, including additive synthesis, multiplicative synthesis and level difference maximization. After the combined weights are determined, they need to be regularized. The weighting process of the three combination weighting methods is shown in Figure 2.

3.3.1. Research on Additive Synthesis

The combined weights based on the additive synthesis method mainly depend on the distribution of the subjective weights and the weight coefficient of the objective weights. At present, the commonly used methods include the minimum weighting method of subjective and objective weight deviation. This paper only studies the basic weight accumulation. Assuming there are classifiers, the combination of subjective weight and objective weight of the classifier can be described as

The subjective weight and the objective weight of classifier are added to obtain the combined weight , and then it is judged whether satisfies . If the combined weight satisfies the condition, it will be retained, otherwise it will be regularized.

3.3.2. Research on Multiplicative Synthesis Method

The multiplicative synthesis method has a multiplier effect, which leads to a higher combination weight of a classifier with a higher subjective and objective weight, and a lower combination weight with a smaller weight. Assuming there are classifiers, the combination of subjective weight and objective weight of the classifier can be described as

The subjective weight and the objective weight of classifier multiplied to obtain the combined weight , and then it is judged on whether satisfies . If the combined weight satisfies the condition, it will be retained, otherwise it will be regularized.

3.3.3. Research on the Level Difference Maximization

The combination weighting method with level difference maximization is a combination weighting method with a single index as the combined unit [29]. First, the multiple subjective weights and objective weights of the classifier are computed according to different weighting methods. Then, the reasonable value space of the combined weights is determined according to the multiple weights; Finally, the maximum discrimination of the evaluation results is taken as the objective function, and the index’s reasonable value interval of the weight is used to establish an optimization model for the constraint conditions and solve the combined weight of the evaluation index. The detailed description of the method is as follows:

Suppose the scheme set is and the attribute set is . Then, is the attribute value of scheme under attribute , is the normalized result of the decision matrix , the weight vector of the attribute is , and , . Thus, the comprehensive attribute value of scheme can be calculated:

Attribute represents the total variance of all decision-making plans and other decision-making plans, and the weight vector should be selected to maximize the total variance of all the attributes for all decision-making plans, so that the deviation function can be constructed:

Thus, by solving the following single-objective optimization problem model, the weight vector of the attribute can be calculated:

In this paper, the level difference maximization is described as

After the above steps, the subjective weight and objective weight of classifier are used to obtain the combined weight by level difference maximization. Finally, it is judged whether the value range of the combined weight is satisfied. If the combined weight meets the condition, it will be retained; otherwise, it will be regularized.

4. Ensemble Learning Model Based on the ER Rule

The method of setting the weight of evidence is defined in Section 3, the ensemble learning model based on ER rule is built in this section. When using the ER rule to combine classifiers in ensemble learning, each classifier obtains its classification results for the datasets, treats the classifiers as evidence in the ER rule, and calculates the weight and reliability of each piece of evidence. Decisions are made through the ER rule, and the classification results are calculated. This process is shown in Figure 3.

4.1. Setting the Reliability of Evidence

The reliability of evidence is an inherent characteristic of evidence, reflecting the ability of the evidence to provide correct assessments or solutions to hypotheses [30]. Based on the definition of evidence reliability, when using the ER rule for ensemble learning, each classifier in the process is regarded as independent evidence, and the classification accuracy of each classifier for the dataset is the ability to correctly evaluate the sample. Therefore, through mathematical statistics, the probability that a certain classifier is correctly classified for the dataset is calculated as its reliability, which is recorded as .

4.2. Setting the Weight of Evidence

The weight of evidence is usually set by the subjective weighting method or objective weighting method, the specific steps of the weighting method are described in Section 3. This paper combines the two methods and proposes to use the combination weighting method to set weights for evidence. The process of assigning weight of evidence is shown in Figure 4. The AHP method in the subjective weighting method is combined with the EW method, COV method, and CRITIC method in the objective weighting method, and the results of the weighting are compared.

4.3. Evidential Reasoning Process

Assuming that each classifier in the ensemble process is an independent piece of evidence, there are evidence in total. The category is considered as the evaluation level and the probability of the classifier’s judgement on the sample category is considered as the belief level corresponding to the evaluation level. The reliability distribution of each piece of evidence can be expressed aswhere is the evaluation level, is the belief degree of the evaluation scheme evaluated as evaluation level under evidence , is the discernment framework including all evaluation levels, is the belief degree of indicator relative to discernment framework , namely, global ignorance, and satisfies , .

The reliability and weight of the evidence have been determined by Section 4.1 and Section 3, respectively. The weighted belief distribution of the evidence with reliability iswhere is a power set and is the mixed probability quality of indicator under level and satisfieswhere is the regularization coefficient. is the basic probability quality of indicator under level , and is an empty set.

Each classifier serves as an independent piece of evidence, and there are pieces of evidence, so the combined belief of is determined by the following equations for pieces of evidence:where is the degree of support for hypothesis of the combined result of the fusion of pieces of independent evidence, and , , and .

Supposing the utility of the evaluation level is , the expected utility of the final model is

Comparing the expected utility with the utility of the evaluation level, if , the result of the integration using the ER rule is the category .

5. Case Study

There are five types of datasets used in this study: part of a large fish dataset provided by the Kaggle platform, part of a flower recognition dataset, part of a fruit 360 dataset, Stanford’s dog dataset, and part of the dataset from the public weather data package provided by the Baidu AI Studio platform. The specific parameter information of these five datasets is shown in Table 1.

The study steps are as follows:(1)First, we use five commonly used deep learning algorithms as the classifier, i.e., DenseNet121, MobileNetV2, InceptionV3, EfficintNet, and ResNet152V2, and use 60% of the samples in the dataset as the training set and all the samples as the test set for classification prediction. Three deep learning algorithms are randomly selected for integration from the five deep learning algorithms, with the number of samples as rows, and the combination of categories and deep learning algorithms as columns, forming a matrix of rowNum×12, where rowNum is the total number of samples in different datasets.(2)Second, we calculate the accuracy of the classifier for sample prediction as the reliability of the classifier. Taking the probability matrix as the original evaluation matrix, the subjective weights and the objective weights are calculated by the AHP, EW, COV, and CRITIC methods. These weights are combined by the addition, the multiplication and the level difference maximization to obtain the combined weight of the classifier.(3)Third, we take the classifier as the evidence, the category as the evaluation level of the evidence, and the probability predicted by different classifiers as the belief distribution. The ER rule is used to integrate the classifiers, the calculated expected utility value is compared with the evaluation level, and the judgement of the image category by the ER rule is calculated.(4)Fourth, the Spearman correlation analysis is used for statistical analysis of the results to determine whether there is a significant difference in performance between the combination weighting method and other weighting methods.

In this paper, the abbreviations D, M, I, E, and R refer to five classifiers DenseNet121, MobileNetV2, InceptionV3, EfficientNet, and ResNet152V2, respectively. Five classifier combinations, DIE, DIR, IER, MIE, and MIR, are used as study objects.

5.1. Study of Datasets
5.1.1. Study Based on Large-Scale Fish Dataset

The dataset is used in this study is part of the large fish dataset on the Kaggle platform. For each picture in the dataset, five classifiers are combined with the ER rule for ensemble learning, subjective and objective weighting methods are used to calculate weights and combined, weight regularization is considered, and the final accuracy is compared. The comparison results are shown in Table 2 and Figure 5. In Table 2, “+” means using the additive synthesis method for subjective and objective weights, “(+)” means performing regularization processing on the weights after using the additive synthesis method to combine weights, “” represents the multiplicative synthesis method, “()” represents the regularization of the weights after combining the weights using the multiplicative synthesis method, and “(rg)” represents the weight by the level difference maximization method.

As shown in Table 2, the different methods used to determine weights have their own advantages and disadvantages in an ensemble. For example, when the three classifiers of DIR are combined for ensembles, the weight determined by the AHP method is the best, and the accuracy reaches 94.75%. Among the three combination methods of DIE, IER, and MIE, the weights determined by the multiplicative synthesis method are the best, and the highest accuracies of the three combination methods are 96.55%, 96.05%, and 96.28%, respectively. These five combinations are generally better than using subjective weighting and objective weighting alone to determine the weights by the level difference maximization method. In most cases, the combination weighting effect is significantly better. The bold values represent the maximum accuracy for each column of data.

As shown in Figure 5, in the MIR combination, the weights determined by the level difference maximization are generally better than the effect of using subjective weighting and objective weighting alone. After the weights determined by the additive synthesis method and the multiplicative synthesis method are regularized, all effects are better than those without regularization, and the accuracy is approximately 0.1% higher. For the other four combinations, the effect of regularization on the weights determined by the additive synthesis method is better than the effect without regularization, which is approximately 0.1%–0.3% higher. In addition, the level difference maximization method determines the weights close to the accuracy of the regularized additive synthesis method to determine the weights.

5.1.2. Study Based on Weather Dataset

The dataset used in this study is part of the weather dataset on the Kaggle platform. For the images in the dataset, five classifiers and the ER rule are combined for ensemble learning, subjective and objective weighting methods and combination weighting methods are used to calculate weights, and the ER rule ensemble accuracies of different weight determination methods are compared. The comparison result is shown in Figure 6.

As shown in Figure 6, the multiplicative synthesis method of the four combinations of DIE, DIR, IER, and MIR has the best effect in determining the weights, and the accuracy is 97.30%–98.74%. The effect after regularization is better than the effect without regularization, and the accuracy is approximately 0.03%–0.11% higher. Sometimes the multiplicative synthesis method determines that the effect of the unregularized weight is better than the effect of the regularization, but in general, the effect of regularization is still better than the effect of nonregularization.

5.1.3. Study Based on the Flower Dataset

The dataset used in this study is part of the flower dataset on the Kaggle platform, and the content is pictures of various types of flowers. Five classifiers and the ER rule are combined for ensemble learning, and the ER rule ensemble accuracy of different weight determination methods is compared. The comparison result is shown in Figure 7.

As shown in Figure 7, under these five combinations, the effect of determining the weight of the level difference maximization method is generally better than other methods. The DIE combination has the best weight determination effect under the AHP method, with an accuracy of 83.76%. In addition, the regularized weight effect under this combination is better than the unregularized effect, the accuracy is increased by approximately 0.1%. The MIE combination has the best effect under the CRITIC method, with an accuracy rate of 85.20%. The effect after weight regularization is better than the effect without regularization, and the accuracy rate is approximately 0.03%–0.3% higher. Under the combination of MIR, the accuracy of additive synthesis and multiplicative synthesis to determine the regularized weights is approximately 0.05%–0.31% higher than that of nonregularization.

5.1.4. Study Based on the Fruit Dataset

The dataset used in this study is part of the fruit 360 dataset on the Kaggle platform, and the content is pictures of different types of apples. For each picture, five classifiers and the ER rule are combined for ensemble learning, and the ER rule ensemble accuracies of different weight determination methods are compared. The comparison result is shown in Figure 8.

As shown in Figure 8, for the DIR combination, the effect of the level difference maximization method to determine the weight is generally better than other methods, and the additive synthesis method to determine the weight of the regularization effect is generally better than the effect of the nonregularization, and the accuracy is improved by approximately 0.06%–0.15%, the accuracy of the regularized weights determined by the multiplicative synthesis method is approximately 0.15%–0.36% higher than the accuracy of the unregularized. In the IER combination, the effect of weight regularization is generally better than that of nonregularization. But for the combination of DIE and MIE, there are three break points in Figure 7, because the weights determined by the additive synthesis method for the subjective and objective weights are greater than 1, which exceeds the range of weight of evidence in the ER rule, and the accuracy of the ensemble cannot be calculated. Compared with other points, the effect of regularization on the weight is generally better than the effect of nonregularization.

5.1.5. Study Based on the Dog Dataset

The dataset used in this study is part of Stanford’s dog dataset on the Kaggle platform, and the content is pictures of different kinds of dogs. For each picture, five classifiers and the ER rule are combined for ensemble learning, and the ER rule ensemble accuracies of different weight determination methods are compared. The comparison results are shown in Table 3.

As shown in Table 3, for the DIE combination, the level difference maximization method determines that the weighting effect is the best. Under the IER combination, the EW method, COV method, and EW (rg) AHP method determine the best weighting effect. In the MIE combination, the AHP method, CRITIC(+)AHP method, and CRITIC(+)AHP method have the best effect in determining the weight. In the MIR combination, the COV method has the best effect in determining the weight. However, for the four combinations of DIE, IER, MIE, and MIR, the range of weights determined when the additive synthesis method is used exceeds 1, which does not meet the range of weight of evidence in the ER rule, and the accuracy of the integration cannot be calculated, so replace it with “#“. In the DIR combination, the AHP method is used to determine the subjective weight. Since the final test coefficient is greater than 0.1, which does not meet the consistency test, the subjective weight cannot be calculated, and the combination weight cannot be calculated subsequently, so it is replaced with “&“. It can be concluded that the weights determined by the combination weighting method are not necessarily better than the weights determined by the subjective weighting method or the objective weighting method alone in all cases. The appropriate weighting method should be considered according to the actual situation. The bold values represent the maximum accuracy for each column of data.

According to the above cases, we can obtain the following conclusions:(1)First, in most cases, the combined weight determined by the level difference maximization has a better effect than the individual subjective and objective weights. The additive combination method and the multiplicative combination method also have a better effect. Compared with the individual subjective and objective weighting method, the combination weighting method has a significant improvement in the effect of the ensemble learning based on the ER rule. The accuracy of ensemble learning is improved by approximately 0.1%–0.4% through the use of combination weighting method in general. Therefore, using the combination weighting method to set the evidence weight of the ER rule has a better effect. Although the combination weighting method is effective, but additive combination method and the multiplicative combination method lack strict theoretical reasoning, so it is recommended to use the more theoretical method of the level difference maximization to combine the weights. The level difference maximization method is used as a combination weighting method for statistical analysis and comparative studies with other weighting methods in 5.2 and 5.3 Subsections.(2)Second, the effect of regularizing the weights on ensemble learning is generally better than that of nonregularization. The analysis of the level difference maximization method and the supporting evidence of the additive synthesis method and the multiplicative synthesis method show that the improvement of the ensemble learning accuracy of the regularization weight is approximately 0.01%–0.3% higher than that of the nonregularization weight. The regularization of weights can also prevent the combined weight from exceeding the range of the weight of evidence in the ER rule. In summary, the weights of evidence need to be regularized.(3)Third, the reasoning process of the ER rule is clear and the weight of evidence determined by the combination of weights is more reasonable, so the ensemble learning model based on the ER rule considering combination weighting is constructed on these basis is reasonable and universal. According to the experiments, this model can effectively improve the effect of ensemble learning.

5.2. Statistical Analysis

To show the differences in performance between the combination weighting method and the other weighting methods and verify the universality and effectiveness of the combination weighting method, we performed a statistical analysis between the different weighting methods.

The EW method, COV method, CRITIC method and combination weighting method including EW(rg)AHP, COV(rg)AHP, CRITIC(rg)AHP are tested for significance, and are denoted as “EW/EW(rg)AHP”, “COV/COV(rg)AHP”, “CRITIC/CRITIC(rg)AHP”. The AHP method and the best combination weighting method are tested for significance, and are denoted as “max(rg)”.

First, on the basis of the significance level parameter , statistical analysis is performed for each method combination, and Spearman rank correlation coefficient is calculated. Second, comparing the Spearman rank correlation coefficient () [31] with the Spearman rank correlation critical value (), if proves that the differences in performance between the combination weighting method and the other methods is statistically significant. In the DIR combination in the dog dataset, the test coefficient of the AHP method is greater than 0.1 and the weight cannot be determined, and the combination is not analysed. Therefore, the number of combinations of this dataset is , then ; the number of combinations of other datasets is , then . The comparison of significance levels is shown in Figure 9.

As shown in Figure 9, all the values are higher than in the weather dataset, flower dataset and fruit dataset; only the value of max(rg) is lower than in the fish dataset; and the values of EW/EW(rg)AHP and max(rg) are lower than in the dog dataset. It indicates differences in performance between the combination weighting method and the other methods is statistically significant.

To better judge that there are significant differences in performance between different methods, we also calculated -values. The -values are compared with the significance level, if proves the differences in performance are statistically significant. Comparison of values are shown in Table 4.

As shown in Table 4, most values are less than 0.05, only the values are greater than 0.05 in the dog dataset. Therefore, the differences in performance between the combination weighting method and other weighting methods is statistically significant.

5.3. Comparative Studies

To demonstrate the effectiveness of the combination weighting method in this paper, comparative studies are conducted in this subsection.

The combination weighting method not only considers the subjective intention of attribute decision-makers, but also contains the mathematical theoretical basis for objective weighting, so more reasonable weights of evidence can be determined. To further illustrate the universality and effectiveness of the above method, the order relation analysis method (G1) and the entropy weight method based on the rank sum ratio (RSR-EW) are selected as comparison methods [6, 12]. The level difference maximization methods are selected from the combination weighting method, including EW(rg)AHP, COV(rg)AHP, and CRITIC(rg)AHP. The accuracy of the ensemble learning model using the level difference maximization and the accuracy of the ensemble learning model using the comparative method are compared. The results of the comparison studies are shown in Table 5.

As shown in Table 5, the accuracy of the ensemble learning model using the level difference maximization method is higher than that of the ensemble learning model using the comparison method in most cases. Only the comparison method G1 is better in the fruit dataset. Therefore, it is reasonable and effective to use the combination weighting method to determine the weight of evidence, which can set a reasonable weight of evidence for the ensemble learning model based on the ER rule, thereby significantly improving the accuracy of the model.

6. Conclusion

In the field of ensemble learning based on the ER rule, problems include, a small number of classifiers in ensemble learning, difficulty in the quality judgement classifiers based on expert experience, objective weighting method that is too dependent on samples, and difficulty in setting a reasonable weight of evidence to improve the accuracy of ensemble learning. The combination weighting method is proposed to solve these problems. This method considers the subjective judgement of expert experience and the objective information of sample data, which overcomes the subjectivity or objectivity of the weight of evidence to a certain extent, making the weight of evidence more reasonable and significantly improve the recognition accuracy of the ensemble learning model. On this basis, the weight of evidence is regularized, which further improves the effect of ensemble learning model based on the ER rule.

There are three innovations in this paper. Firstly, in view of the problem that it is difficult to set a reasonable weight of evidence in traditional weighting methods, the combination weighting method is proposed to set more reasonable weights of evidence for the ER rule. Secondly, considering the lack of labels of the ER rule in the information fusion process, the weight of evidence is regularized. Lastly, based on setting the weight of evidence for the ER rule by using combination weighting method, the ensemble learning model based on the ER rule is constructed to improve the effect of integration.

Based on the above research, the main work in the future will include the following aspects: (1) In terms of the execution time of methods, the subjective weighting method is approximately 0.0003s-0.002s, the objective weighting method is approximately 0.03s-0.08s, and the combination weighting method is approximately 0.18s-0.32s, the execution time of the combination weighting method is longer. Therefore, how to reduce the execution time under the premise of ensuring the effect is the focus of future work. (2) This paper only studies the combination of several classic subjective weighting methods and objective weighting methods, and there are more combinations of combination weighting methods that can be used, such as game theory.

Data Availability

The study data in this paper come from the image classification datasets of the Kaggle platform and Baidu AI Studio platform. For the source URL of the datasets, please visit: (1) Large-scale Fish dataset: https://www.kaggle.com/crowww/a-large-scale-fish-dataset. (2) Weather dataset: https://aistudio.baidu.com/aistudio/datasetdetail/13165. (3) Flowers recognition dataset: https://www.kaggle.com/alxmamaev/flowers-recognition. (4) Fruits 360 dataset: https://www.kaggle.com/moltean/fruits. (5) Stanford’s dogs dataset: https://www.kaggle.com/jessicali9530/stanford-dogs-dataset

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Authors’ Contributions

YunYi Zhang and Cong Xu contributed equally to this work.

Acknowledgments

This work was supported in part by the Postdoctoral Science Foundation of China under Grant no. 2020M683736, the Natural Science Foundation of Heilongjiang Province of China under Grant no. LH2021F038, the Innovation Practice Project of College Students in Heilongjiang Province under Grant nos. 202010231009, 202110231024, and 202110231155, the graduate quality training and improvement project of Harbin Normal University under Grant no. 1504120015, and the graduate academic innovation project of Harbin Normal University under Grant nos. HSDSSCX2021-120 and HSDSSCX2021-29.