Abstract
Recently, bioinformatics and computational biology-enabled applications such as gene expression analysis, cellular restoration, medical image processing, protein structure examination, and medical data classification utilize fuzzy systems in offering effective solutions and decisions. The latest developments of fuzzy systems with artificial intelligence techniques enable to design the effective microarray gene expression classification models. In this aspect, this study introduces a novel feature subset selection with optimal adaptive neuro-fuzzy inference system (FSS-OANFIS) for gene expression classification. The major aim of the FSS-OANFIS model is to detect and classify the gene expression data. To accomplish this, the FSS-OANFIS model designs an improved grey wolf optimizer-based feature selection (IGWO-FS) model to derive an optimal subset of features. Besides, the OANFIS model is employed for gene classification and the parameter tuning of the ANFIS model is adjusted by the use of coyote optimization algorithm (COA). The application of IGWO-FS and COA techniques helps in accomplishing enhanced microarray gene expression classification outcomes. The experimental validation of the FSS-OANFIS model has been performed using Leukemia, Prostate, DLBCL Stanford, and Colon Cancer datasets. The proposed FSS-OANFIS model has resulted in a maximum classification accuracy of 89.47%.
1. Introduction
Microarray is an advanced technology that helps to recognize the pattern of gene expression of various genes at a time at the genomic level. It supports the researcher to investigate and analyze millions of genes in a single experiment [1]. It identifies many present diseases connected to each individual gene such as anaemia and cancer. Analysis of Gene Expression provides a method to recognize the gene that is differentially expressed [2], which is accountable to develop some diseases. Also, it shows the difference between normal and abnormal genes through a mathematical model [3, 4]. Many openly accessible datasets such as Array Express and Gene Expression Omnibus (GEO) make the task easier to identify gene patterns of rare diseases. Classification of gene expression data splits cancer samples from healthy samples that are utilized in response to treatment prediction. Due to the smaller amount of samples with a larger amount of features in the gene expression information, the standard ML method disappoints to implement better for cancer classification [5].
Recently, there has been tremendous growth in the medical field around the world. There are several computational approaches utilized in the bioinformatics field in the last few decades, for example, data mining and pattern recognition, to deal with higher-dimensional problems but still unsuccessful [6]. Thus, recently, machine learning (ML), a branch of artificial intelligence, has received considerable attention from researchers in gene expression and genomics [7]. Also, ML is a branch of data science; the main goal is to allow a model for training and learning to make decisions by itself in the future. Machine learning is widely classified into semisupervised, semi-unsupervised, supervised, and unsupervised learning [8]. For microarray data classification, the ML-based feature selection (FS) techniques such as gene selection techniques assist in selecting the essential gene [9]. Feature selection assists to preserve useful attributes. It is mainly utilized for the higher-dimensional data; simply, FS is a dimensionality reduction method. Feature selection significantly assists in the field that has relatively scarce and samples too many features, e.g., DNA Microarray and RNA sequencing [10]. This approach assists in better understanding of the feature space, preventing the scare of model overfitting, maximizing the model training time, handling the dimension, and maximizing the prediction accuracy. The results of FS are the optimum amount of features that are related to the provided class label that contributed to the prediction process.
This study introduces a novel feature subset selection with optimal adaptive neuro-fuzzy inference system (FSS-OANFIS) for gene expression classification. The FSS-OANFIS model designs an improved grey wolf optimizer-based feature selection (IGWO-FS) model to derive an optimal subset of features. Besides, the OANFIS model is employed for gene classification and the parameter tuning of the ANFIS model is adjusted by the use of coyote optimization algorithm (COA). The application of IGWO-FS and COA techniques helps in accomplishing enhanced microarray gene expression classification outcomes. For examining the enhanced outcomes of the FSS-OANFIS model, a comprehensive simulation analysis was performed on distinct datasets.
2. Related Works
In reference [11], a two-phase approach named as ML-integrated ensemble of feature selection (FS) technique is used, and then a survival study was presented. In a primary stage, it can be chosen the optimum amongst 7 ML approaches dependent upon classifier accuracy, utilizing the whole group of features (under this case miRNAs). In the secondary stage, dependent upon classifier accuracy values, the top feature in all the FS approaches is assumed for making an ensemble to offer more categorization of miRNAs. Ayyad et al. [12] presented a novel classifier approach to gene expression data. Both executions are assumed that improve the performance of KNN. An important idea is for utilizing robust neighbors in trained data with utilizing a novel weighting approach. The authors in reference [13] presented a recently developed classification named Forest DNN (fDNN) for integrating the DNN structure with a supervised forest feature detector. Utilizing this built-in feature detector, this technique is capable of learning sparse feature representation and feeding the representation to NN for mitigating the overfitting problem. Dwivedi [14] developed a structure of approaches dependent upon supervised ML with utilizing the ANN approach for gene classification.
Shukla [15] established a novel gene selection (GS) approach by integrating minimum redundancy maximum relevance (mRMR) and teaching learning-based optimization (TLBO) for accurate cancer prediction. Primarily, during the presented method, mRMR was executed for determining one of the discriminative genes in the original feature set. In SVM, mRMR was utilized as a fitness function (FF) under the presented technique for selecting relevant features that are used for estimating the prediction accuracy and classifying cancer correctly. In reference [16], a novel social network analysis-based GS method was presented. The presented approach contains 2 important objectives: relevance maximization and redundancy minimization of chosen genes. During this approach, on all iterations, a maximal community was chosen repetitively. Next amongst the present genes under this community, the suitable genes were chosen by utilizing the node centrality-based condition.
In reference [17], an ensemble DL approach was presented for reducing the dimensional features. Primarily, the reduction of dimensional with utilize of auto-encoder (AE) by utilizing several hidden layers have occurred and under the next step, a folded AE is also utilized for reducing the dimensional of identical original data. Eventually, both are combined and top feature is chosen on the fundamental of T-score value. Forestiero et al. [18] presented a multiagent technique to create a distributing approach to DNA microarray management. The group of agents, whereas all one signifying a Microarray (or chip), implement from the parallel a sequence of easy functions exploiting local data and organized virtual infrastructure was created at a global level. The word embedded method, capable of capturing the semantic context and signifying microarray with vector, was utilized for mapping the chip, thus permitting advanced agent functions.
3. The Proposed Model
In this study, a new FSS-OANFIS model has been developed for microarray gene expression data classification. The presented FSS-OANFIS model encompasses a series of processes, namely, data preprocessing, IGWO-FS-based election of features, ANFIS classification, and COA-based parameter optimization. The application of IGWO-FS and COA techniques helps in accomplishing enhanced microarray gene expression classification outcomes. Figure 1 shows the overall process of FSS-OANFIS technique.

3.1. Preprocessing
The z-score is a normalized and standardized system, which describes the count of standard deviation (SD), a raw data point, which is below or above the population mean [18]. It ideally lies in the range of and . It standardizes the dataset to the aforementioned scale to change data with distinct scales to default scale. Thus, reflecting that several SD a point is below/above the mean as follows, but refers to the value of particular instance, signifies the mean, and depicts the SD:
3.2. Steps Involved in IGWO-FS Technique
Once the input data is preprocessed, the next stage is to choose an optimal subset of features. The GWO algorithm is naturally inspired by the behavior and social leadership of the grey wolves [19]. The population of wolves can be classified into alpha, beta, delta, and omega for establishing the social hierarchy of wolves. The fittest solution is called alpha , whereas beta and delta represent the 2nd and 3rd most efficient options, respectively. Omega represents semblance of a hopeful solution. The arithmetical expression of readapting position is shown as follows:where denotes the location of the alpha, represent the location of the beta, indicates the location of the delta, and , and and , and signifies random vector, that is, the location of the existing solution, and shows the amount of iterations. It can be expressed in the following equation:where is represented in equation (3), indicates the iteration number, denote the vector coefficient, and and represent the praise and grey wolf locations. The vectors are calculated in the following equation:
and denote vectors with arbitrary numbers within and parameter is linearly reduced from 2 to all over the iteration. Usually, the alpha is responsible for the chase. To change the position with the optimal searching agent position, the first three optimal solutions attained up until now compel another searching agent. Then, the wolves position can be upgraded as follows:
The variable governs the balance between exploration and exploitation. Here, the variable can be updated linearly in all the iterations, which ranges from 2 to , with being the iteration number and be the overall iterations allowed for the optimization:
The wolf's location reflects attribute set selection and the solution space can be made by each probable attribute selection. The fitness function of the IGWO-FS technique can be utilized for determining whether an attribute subset would be selected or not:
represents the selected attribute length subset and denotes the classification quality of attribute set in relation to decision . The overall amount of quality indicates the letter and , are two respective values for the attribute subset length and classification quality, repectively. Both have dissimilar implications for the attribute reduction task. The set and attribute subset length are less important when compared to the quality classification. The higher ensure that the ideal location is a rough set reduction as a minimum. The fitness function evaluates the quality of location. After defining the fitness level, important feature is taken as well as removing the unwanted feature.
The performance of the GWO algorithm can be improved by the design of IGWO algorithm with the inclusion of adaptive ‐hill climbing (AβHC). It is a recently presented local search-based technique, that is, basically, a modified version of ‐hill climbing [20]. The study has established that AβHC gives optimum performance than several other famous local search techniques. For boosting the techniques exploitation capability and the quality of last solutions, has been combined with the fundamental GTO for support searching the neighborhoods of optimum solution under this study. The explanation of has been demonstrated mathematically as follows:
In order to provide an existing solution , is iteratively created an improved solution on the fundamental of 2 control operators: operator and ‐operator. The ‐operator primarily transfers to a novel neighborhood solution.
that is demonstrated in equations (9) and (10) as follows:where refers the arbitrary number between the interval of 0 and 1 represents the value of decision variable from the dimensional, stands for the existing iteration, denotes the maximal count of iterations, signifies the bandwidth distance amongst the existing solution and their neighbor, refers to the spatial dimensionality, and the parameter is a constant.
3.3. Optimal ANFIS-Based Classification
At the final stage, the OANFIS model has been employed for the detection and classification of gene expression data into multiple classes. A network with 2 inputs, and and one output, is considered. The ANFIS is a fuzzy Sugeno method. For presenting the ANFIS structure, 2 fuzzy if-then rules dependent upon a first‐order Sugeno method are assumed as follows [21]:(i)Rule 1: is and is , then (ii)Rule 2: if is and is , then
Where and are inputs, and imply fuzzy sets, is the output of fuzzy system, and , , and represent the design parameters that are defined in the training procedure. The ANFIS structure for implementing these 2 rules, whereas a circle represents the set node and a square refers the adaptive node. The ANFIS infrastructure has 5 layers. Figure 2 showss the framework of ANFIS.

Layer 1. All nodes from layer1 are adaptation nodes. The resultant of layer 1 is are fuzzy membership grade of the inputs that are provided as follows:where and refer the inputs to node , refers the linguistic label, and and are some fuzzy membership functions. Generally, is chosen aswhere , , and are the parameters of membership bell‐shaped function.
Layer 2. The node of this layer is labeled , signifying that it can be implemented as an easy multiplier The resultant of this layer is demonstrated as follows:
Layer 3. It comprises set nodes which compute the ratio of firing strength of the rules as follows:
Layer 4. During this layer, the adaptive node is used. The resultants of this layer are calculated by the following equation:
signifies the normalized firing strength in layer 3.
Layer 5. The node executes the summation of every incoming signal. Therefore, an entire result of the model is provided as follows:
It could be realized that there are 2 adaptive layers under this ANFIS structure such as the 1st layer and 4th layer. During the 1st layer, there are 3 modifiable parameters that are connected to the input membership function. These parameters are usually named as premise parameters. During the 4th layer, there are also 3 modifiable parameters relating to the first‐order polynomial. This parameter is supposed the consequent parameter.
For tuning the ANFIS parameters, the COA is applied to it. The COA is a mathematical model that depends on smart diversity [22]. Chasing, driving, attacking, and blocking are archived by four distinct kinds of chimps that are attained by chasers, drivers, attackers, and obstacles. These hunting steps are accomplished in two phases such as exploration and exploitation stages. The exploration phase involves chasing, driving, and blocking the prey. The exploitation phase should attack the prey, and the chasing and driving are characterized as follows:where denotes the vector of prey location, indicates the vector of chimp location, denotes the amount of present iterations, , , and represent coefficient vectors and are calculated as follows:where declined nonlinearly from 2.5 to , and denote the random value within [0, 1], and represents the chaotic vector. The dynamic coefficient is selected for distinct slopes and curves; therefore, chimps employ distinct capabilities for searching the prey. Chimps upgrade the position according to the other chimps, and the arithmetical expression can be given in the following equation:
4. Experimental Validation
In this section, the experimental validation of the FSS-OANFIS model has been performed using four benchmark datasets [23–26]. The details of the dataset are given in Table 1. The results are investigated and the outcomes are assessed in terms of different measures. For experimental validation, a 10-fold cross-validation process is utilized.
4.1. Results Analysis of Proposed Model
Figure 3 illustrates a set of confusion matrices offered by the FSS-OANFIS model on test datasets. The figure reported that the FSS-OANFIS model has properly recognized the class labels on all datasets. For instance, on the Leukemia dataset, the FSS-OANFIS model has identified 8 samples in class 0 and 2 samples in class 1. In addition, on the Prostate dataset, the FSS-OANFIS system has identified 16 samples in class 0 and 9 samples in class 1. Also, on the DLBCL Stanford dataset, the FSS-OANFIS approach has identified 2 samples in class 0 and 9 samples in class 1. Besides, on the Colon Cancer dataset, the FSS-OANFIS algorithm has identified 12 samples in class 0 and 5 samples in class 1.

(a)

(b)

(c)

(d)
Table 2 provides the overall classification outcomes of the FSS-OANFIS model on the test datasets. The experimental outcomes pointed out that the FSS-OANFIS model has offered effective outcomes on all the datasets applied. For instance, with the Leukemia dataset, the FSS-OANFIS model has resulted in average , , , , and of 83.33%, 75%, 75%, 77.78%, and 80.08%, respectively. Following this, with the Prostate dataset, the FSS-OANFIS methodology has resulted in average , , , , and of 80.65%, 80%, 80%, 79.61%, and 81.37%, respectively. Meanwhile, with the DLBCL Stanford dataset, the FSS-OANFIS algorithm has resulted in average , , , , and of 73.33%, 66.67%, 66.67%, 65.91%, and 70.47%, respectively. Eventually, with the Colon Cancer dataset, the FSS-OANFIS technique has resulted in average , , , , and of 89.47%, 87.82%, 87.82%, 87.82%, and 87.82%, respectively.
Figure 4 shows the precision-recall curves offered by the FSS-OANFIS model on the test Leukemia dataset. The figure indicated that the FSS-OANFIS model has depicted effective precision-recall values on the classification of two classes, namely, class 0 and class 1 on the test Leukemia dataset.

Next, Figure 5 shows the precision-recall curves offered by the FSS-OANFIS model on the test Prostate dataset. The figure revealed that the FSS-OANFIS technique has depicted effective precision-recall values on the classification of two classes, namely, class 0 and class 1 on the test Prostate dataset.

Similarly, Figure 6 shows the precision-recall curves offered by the FSS-OANFIS system on the test DLBCL Stanford dataset. The figure exposed that the FSS-OANFIS model has depicted effective precision-recall values on the classification of two classes, namely, class 0 and class 1 on the test DLBL Stanford dataset.

Figure 7 shows the precision-recall curves offered by the FSS-OANFIS method on the test Colon Cancer dataset. The figure indicated that the FSS-OANFIS approach has depicted effective precision-recall values on the classification of two classes, namely, class 0 and class 1 on the test Colon Cancer dataset.

A brief ROC investigation of the FSS-OANFIS model on the distinct four datasets is described in Figure 8. The results indicated that the FSS-OANFIS technique has exhibited its ability in categorizing two different classes such as class 0 and 1 on the test four datasets.

(a)

(b)

(c)

(d)
Figure 9 shows the accuracy and loss graph analysis of the ODBN-IDS technique under four datasets. The results show that the accuracy value tends to increase and the loss value tends to decrease with an increase in epoch count. It is also observed that the training loss is low and validation accuracy is high under four datasets.

(a)

(b)

(c)

(d)

(e)

(f)

(g)

(h)
4.2. Discussion
Finally, a detailed comparative study of the FSS-OANFIS model with recent methods on distinct datasets is shown in Table 3 and Figure 10 [27]. The experimental results indicated that the FSS-OANFIS model has shown effectual outcomes under all datasets. For instance, with the Leukemia dataset, the DE and AHSA-GS models have depicted lower performance over the other methods. Though the PSO algorithm has resulted in slightly reasonable performance with , , , and of 80.59%, 74.95%, 73.96%, and 68.07%, the FSS-OANFIS model has resulted in higher , , , and of 83.33%, 75%, 75%, and 80.08%, respectively.

(a)

(b)

(c)

(d)
At the same time, with the Prostate dataset, the DE and AHSA-GS models have depicted lower performance over the other methods. Likewise, the PSO algorithm has resulted in somewhat reasonable performance with , , , and of 68.78%, 63.63%, 70.15%, and 66.01% and the FSS-OANFIS methodology has resulted in superior , , , and of 80.65%, 80%, 80%, and 81.37%, respectively. In addition, with the DLBCL Stanford dataset, the DE and AHSA-GS techniques have showcased lesser performance over the other methods. Though the PSO algorithm has resulted in slightly reasonable performance with , , , and of 72.80%, 60.82%, 60.17%, and 61.74%, the FSS-OANFIS approach has resulted in higher , , , and of 733.33%, 66.67%, 66.67%, and 70.47%, respectively.
Along with that, with the Colon Cancer dataset, the DE and AHSA-GS models have portrayed lower performance over the other methods. But, the PSO approach has resulted in slightly reasonable performance with , , , and of 59%, 43.02%, 58.34%, and 36.66%, and the FSS-OANFIS system has resulted in superior , , , and of 89.47%, 87.80%, 87.82%, and 87.82%, respectively. After examining the results and discussion, it is apparent that the FSS-OANFIS model has accomplished maximum performance in the microarray gene expression classification process.
5. Conclusion
In this study, a new FSS-OANFIS model has been developed for microarray gene expression data classification. The presented FSS-OANFIS model encompasses a series of processes, namely, data pre-processing, IGWO-FS-based election of features, ANFIS classification, and COA-based parameter optimization. The application of IGWO-FS and COA techniques helps in accomplishing enhanced microarray gene expression classification outcomes. For examining the enhanced outcomes of the FSS-OANFIS model, a wide range of simulations were performed on distinct datasets. The experimental results indicated that the FSS-OANFIS model has resulted in enhanced performance over the recent approaches. In future, the feature reduction and clustering approaches can be integrated to enhance gene expression classification outcomes.
Data Availability
Data are available and can be provided upon direct request to the corresponding author.
Ethical Approval
This article does not contain any studies with human participants performed by any of the authors.
Consent
Not applicable.
Conflicts of Interest
The authors declare that they have no conflicts of interest.
Acknowledgments
The authors extend their appreciation to the Deanship of Scientific Research at King Khalid University for funding this work under grant number RGP 2/180/43, Princess Nourah bint Abdulrahman Universitysupporting project number PNURSP2022R151, Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia. The authors would like to thank the Deanship of Scientific Research at Umm Al-Qura University for supporting this work under grant code: 22UQU4340237DSR15.