Literature DB >> 35607459

Feature Subset Selection with Optimal Adaptive Neuro-Fuzzy Systems for Bioinformatics Gene Expression Classification.

Anwer Mustafa Hilal1, Areej A Malibari2, Marwa Obayya3, Jaber S Alzahrani4, Mohammad Alamgeer5, Abdullah Mohamed6, Abdelwahed Motwakel1, Ishfaq Yaseen1, Manar Ahmed Hamza1, Abu Sarwar Zamani1.   

Abstract

Recently, bioinformatics and computational biology-enabled applications such as gene expression analysis, cellular restoration, medical image processing, protein structure examination, and medical data classification utilize fuzzy systems in offering effective solutions and decisions. The latest developments of fuzzy systems with artificial intelligence techniques enable to design the effective microarray gene expression classification models. In this aspect, this study introduces a novel feature subset selection with optimal adaptive neuro-fuzzy inference system (FSS-OANFIS) for gene expression classification. The major aim of the FSS-OANFIS model is to detect and classify the gene expression data. To accomplish this, the FSS-OANFIS model designs an improved grey wolf optimizer-based feature selection (IGWO-FS) model to derive an optimal subset of features. Besides, the OANFIS model is employed for gene classification and the parameter tuning of the ANFIS model is adjusted by the use of coyote optimization algorithm (COA). The application of IGWO-FS and COA techniques helps in accomplishing enhanced microarray gene expression classification outcomes. The experimental validation of the FSS-OANFIS model has been performed using Leukemia, Prostate, DLBCL Stanford, and Colon Cancer datasets. The proposed FSS-OANFIS model has resulted in a maximum classification accuracy of 89.47%.
Copyright © 2022 Anwer Mustafa Hilal et al.

Entities:  

Mesh:

Substances:

Year:  2022        PMID: 35607459      PMCID: PMC9124108          DOI: 10.1155/2022/1698137

Source DB:  PubMed          Journal:  Comput Intell Neurosci


1. Introduction

Microarray is an advanced technology that helps to recognize the pattern of gene expression of various genes at a time at the genomic level. It supports the researcher to investigate and analyze millions of genes in a single experiment [1]. It identifies many present diseases connected to each individual gene such as anaemia and cancer. Analysis of Gene Expression provides a method to recognize the gene that is differentially expressed [2], which is accountable to develop some diseases. Also, it shows the difference between normal and abnormal genes through a mathematical model [3, 4]. Many openly accessible datasets such as Array Express and Gene Expression Omnibus (GEO) make the task easier to identify gene patterns of rare diseases. Classification of gene expression data splits cancer samples from healthy samples that are utilized in response to treatment prediction. Due to the smaller amount of samples with a larger amount of features in the gene expression information, the standard ML method disappoints to implement better for cancer classification [5]. Recently, there has been tremendous growth in the medical field around the world. There are several computational approaches utilized in the bioinformatics field in the last few decades, for example, data mining and pattern recognition, to deal with higher-dimensional problems but still unsuccessful [6]. Thus, recently, machine learning (ML), a branch of artificial intelligence, has received considerable attention from researchers in gene expression and genomics [7]. Also, ML is a branch of data science; the main goal is to allow a model for training and learning to make decisions by itself in the future. Machine learning is widely classified into semisupervised, semi-unsupervised, supervised, and unsupervised learning [8]. For microarray data classification, the ML-based feature selection (FS) techniques such as gene selection techniques assist in selecting the essential gene [9]. Feature selection assists to preserve useful attributes. It is mainly utilized for the higher-dimensional data; simply, FS is a dimensionality reduction method. Feature selection significantly assists in the field that has relatively scarce and samples too many features, e.g., DNA Microarray and RNA sequencing [10]. This approach assists in better understanding of the feature space, preventing the scare of model overfitting, maximizing the model training time, handling the dimension, and maximizing the prediction accuracy. The results of FS are the optimum amount of features that are related to the provided class label that contributed to the prediction process. This study introduces a novel feature subset selection with optimal adaptive neuro-fuzzy inference system (FSS-OANFIS) for gene expression classification. The FSS-OANFIS model designs an improved grey wolf optimizer-based feature selection (IGWO-FS) model to derive an optimal subset of features. Besides, the OANFIS model is employed for gene classification and the parameter tuning of the ANFIS model is adjusted by the use of coyote optimization algorithm (COA). The application of IGWO-FS and COA techniques helps in accomplishing enhanced microarray gene expression classification outcomes. For examining the enhanced outcomes of the FSS-OANFIS model, a comprehensive simulation analysis was performed on distinct datasets.

2. Related Works

In reference [11], a two-phase approach named as ML-integrated ensemble of feature selection (FS) technique is used, and then a survival study was presented. In a primary stage, it can be chosen the optimum amongst 7 ML approaches dependent upon classifier accuracy, utilizing the whole group of features (under this case miRNAs). In the secondary stage, dependent upon classifier accuracy values, the top feature in all the FS approaches is assumed for making an ensemble to offer more categorization of miRNAs. Ayyad et al. [12] presented a novel classifier approach to gene expression data. Both executions are assumed that improve the performance of KNN. An important idea is for utilizing robust neighbors in trained data with utilizing a novel weighting approach. The authors in reference [13] presented a recently developed classification named Forest DNN (fDNN) for integrating the DNN structure with a supervised forest feature detector. Utilizing this built-in feature detector, this technique is capable of learning sparse feature representation and feeding the representation to NN for mitigating the overfitting problem. Dwivedi [14] developed a structure of approaches dependent upon supervised ML with utilizing the ANN approach for gene classification. Shukla [15] established a novel gene selection (GS) approach by integrating minimum redundancy maximum relevance (mRMR) and teaching learning-based optimization (TLBO) for accurate cancer prediction. Primarily, during the presented method, mRMR was executed for determining one of the discriminative genes in the original feature set. In SVM, mRMR was utilized as a fitness function (FF) under the presented technique for selecting relevant features that are used for estimating the prediction accuracy and classifying cancer correctly. In reference [16], a novel social network analysis-based GS method was presented. The presented approach contains 2 important objectives: relevance maximization and redundancy minimization of chosen genes. During this approach, on all iterations, a maximal community was chosen repetitively. Next amongst the present genes under this community, the suitable genes were chosen by utilizing the node centrality-based condition. In reference [17], an ensemble DL approach was presented for reducing the dimensional features. Primarily, the reduction of dimensional with utilize of auto-encoder (AE) by utilizing several hidden layers have occurred and under the next step, a folded AE is also utilized for reducing the dimensional of identical original data. Eventually, both are combined and top feature is chosen on the fundamental of T-score value. Forestiero et al. [18] presented a multiagent technique to create a distributing approach to DNA microarray management. The group of agents, whereas all one signifying a Microarray (or chip), implement from the parallel a sequence of easy functions exploiting local data and organized virtual infrastructure was created at a global level. The word embedded method, capable of capturing the semantic context and signifying microarray with vector, was utilized for mapping the chip, thus permitting advanced agent functions.

3. The Proposed Model

In this study, a new FSS-OANFIS model has been developed for microarray gene expression data classification. The presented FSS-OANFIS model encompasses a series of processes, namely, data preprocessing, IGWO-FS-based election of features, ANFIS classification, and COA-based parameter optimization. The application of IGWO-FS and COA techniques helps in accomplishing enhanced microarray gene expression classification outcomes. Figure 1 shows the overall process of FSS-OANFIS technique.
Figure 1

Overall process of FSS-OANFIS technique.

3.1. Preprocessing

The z-score is a normalized and standardized system, which describes the count of standard deviation (SD), a raw data point, which is below or above the population mean [18]. It ideally lies in the range of −3 and +3. It standardizes the dataset to the aforementioned scale to change data with distinct scales to default scale. Thus, reflecting that several SD a point is below/above the mean as follows, but x refers to the value of particular instance, μ signifies the mean, and σ depicts the SD:

3.2. Steps Involved in IGWO-FS Technique

Once the input data is preprocessed, the next stage is to choose an optimal subset of features. The GWO algorithm is naturally inspired by the behavior and social leadership of the grey wolves [19]. The population of wolves can be classified into alpha, beta, delta, and omega for establishing the social hierarchy of wolves. The fittest solution is called alpha (α), whereas beta (δ) and delta (δ) represent the 2nd and 3rd most efficient options, respectively. Omega (ω) represents semblance of a hopeful solution. The arithmetical expression of readapting position 0 is shown as follows:where denotes the location of the alpha, represent the location of the beta, indicates the location of the delta, and , and and , and signifies random vector, that is, the location of the existing solution, and shows the amount of iterations. It can be expressed in the following equation:where is represented in equation (3), u indicates the iteration number, denote the vector coefficient, and and represent the praise and grey wolf locations. The vectors are calculated in the following equation: s 1 and s2 denote vectors with arbitrary numbers within [0,1] and parameter is linearly reduced from 2 to 0 all over the iteration. Usually, the alpha is responsible for the chase. To change the position with the optimal searching agent position, the first three optimal solutions attained up until now compel another searching agent. Then, the wolves position can be upgraded as follows: The variable b governs the balance between exploration and exploitation. Here, the variable can be updated linearly in all the iterations, which ranges from 2 to 0, with u being the iteration number and m be the overall iterations allowed for the optimization: The wolf's location reflects attribute set selection and the solution space can be made by each probable attribute selection. The fitness function of the IGWO-FS technique can be utilized for determining whether an attribute subset would be selected or not: |S| represents the selected attribute length subset and γ(E) denotes the classification quality of attribute set S in relation to decision E. The overall amount of quality indicates the letter |D|∝∈0,1 and β=1 − ∝, are two respective values for the attribute subset length and classification quality, repectively. Both have dissimilar implications for the attribute reduction task. The set ∝=0.9, β=0.1 and attribute subset length are less important when compared to the quality classification. The higher ensure that the ideal location is a rough set reduction as a minimum. The fitness function evaluates the quality of location. After defining the fitness level, important feature is taken as well as removing the unwanted feature. The performance of the GWO algorithm can be improved by the design of IGWO algorithm with the inclusion of adaptive β‐hill climbing (AβHC). It is a recently presented local search-based technique, that is, basically, a modified version of β‐hill climbing (βHC) [20]. The study has established that AβHC gives optimum performance than several other famous local search techniques. For boosting the techniques exploitation capability and the quality of last solutions, AβHC has been combined with the fundamental GTO for support searching the neighborhoods of optimum solution under this study. The explanation of AβHC has been demonstrated mathematically as follows: In order to provide an existing solution X=(x, x,…, x), AβHC is iteratively created an improved solution X″=(x″, x″,…, x″) on the fundamental of 2 control operators: 𝒩− operator and β‐operator. The N‐operator primarily transfers X to a novel neighborhood solution. X ′=(x′, x′,…, x′) that is demonstrated in equations (9) and (10) as follows:where U(O, 1) refers the arbitrary number between the interval of 0 and 1 , x represents the value of decision variable from the j dimensional, t stands for the existing iteration, Maxiter denotes the maximal count of iterations, N signifies the bandwidth distance amongst the existing solution and their neighbor, D refers to the spatial dimensionality, and the parameter K is a constant.

3.3. Optimal ANFIS-Based Classification

At the final stage, the OANFIS model has been employed for the detection and classification of gene expression data into multiple classes. A network with 2 inputs, x and y and one output, f is considered. The ANFIS is a fuzzy Sugeno method. For presenting the ANFIS structure, 2 fuzzy if-then rules dependent upon a first‐order Sugeno method are assumed as follows [21]: Rule 1: ifx is A1 and y is B1, then f1=p1x+q1y+r1 Rule 2: if x is A2 and y is B2, then f2=p2x+q2y+r2 Where x and y are inputs, A1 and B imply fuzzy sets, f is the output of fuzzy system, and p, q, and r represent the design parameters that are defined in the training procedure. The ANFIS structure for implementing these 2 rules, whereas a circle represents the set node and a square refers the adaptive node. The ANFIS infrastructure has 5 layers. Figure 2 showss the framework of ANFIS.
Figure 2

Structure of ANFIS.

Layer 1. All nodes from layer1 are adaptation nodes. The resultant of layer 1 is are fuzzy membership grade of the inputs that are provided as follows:where x and y refer the inputs to node i, A refers the linguistic label, and ∝(x) and μ(x) are some fuzzy membership functions. Generally, ∝(x) is chosen aswhere a, b, and c are the parameters of membership bell‐shaped function. Layer 2. The node of this layer is labeled M, signifying that it can be implemented as an easy multiplier The resultant of this layer is demonstrated as follows: Layer 3. It comprises set nodes which compute the ratio of firing strength of the rules as follows: Layer 4. During this layer, the adaptive node is used. The resultants of this layer are calculated by the following equation: signifies the normalized firing strength in layer 3. Layer 5. The node executes the summation of every incoming signal. Therefore, an entire result of the model is provided as follows: It could be realized that there are 2 adaptive layers under this ANFIS structure such as the 1st layer and 4th layer. During the 1st layer, there are 3 modifiable parameters {a, bc} that are connected to the input membership function. These parameters are usually named as premise parameters. During the 4th layer, there are also 3 modifiable parameters {pqr} relating to the first‐order polynomial. This parameter is supposed the consequent parameter. For tuning the ANFIS parameters, the COA is applied to it. The COA is a mathematical model that depends on smart diversity [22]. Chasing, driving, attacking, and blocking are archived by four distinct kinds of chimps that are attained by chasers, drivers, attackers, and obstacles. These hunting steps are accomplished in two phases such as exploration and exploitation stages. The exploration phase involves chasing, driving, and blocking the prey. The exploitation phase should attack the prey, and the chasing and driving are characterized as follows:where Xprey denotes the vector of prey location, xchimp indicates the vector of chimp location, t denotes the amount of present iterations, a, c, and m represent coefficient vectors and are calculated as follows:where f declined nonlinearly from 2.5 to 0, r1 and r2 denote the random value within [0, 1], and m represents the chaotic vector. The dynamic coefficient f is selected for distinct slopes and curves; therefore, chimps employ distinct capabilities for searching the prey. Chimps upgrade the position according to the other chimps, and the arithmetical expression can be given in the following equation:

4. Experimental Validation

In this section, the experimental validation of the FSS-OANFIS model has been performed using four benchmark datasets [23-26]. The details of the dataset are given in Table 1. The results are investigated and the outcomes are assessed in terms of different measures. For experimental validation, a 10-fold cross-validation process is utilized.
Table 1

Dataset details.

DatasetLeukemiaProstateDLBCL StanfordColon Cancer
No. of genes71291260040262000
Class 027522440
Class 111502322
Total no. of samples381024762

4.1. Results Analysis of Proposed Model

Figure 3 illustrates a set of confusion matrices offered by the FSS-OANFIS model on test datasets. The figure reported that the FSS-OANFIS model has properly recognized the class labels on all datasets. For instance, on the Leukemia dataset, the FSS-OANFIS model has identified 8 samples in class 0 and 2 samples in class 1. In addition, on the Prostate dataset, the FSS-OANFIS system has identified 16 samples in class 0 and 9 samples in class 1. Also, on the DLBCL Stanford dataset, the FSS-OANFIS approach has identified 2 samples in class 0 and 9 samples in class 1. Besides, on the Colon Cancer dataset, the FSS-OANFIS algorithm has identified 12 samples in class 0 and 5 samples in class 1.
Figure 3

Confusion matrix of FSS-OANFIS technique under four datasets.

Table 2 provides the overall classification outcomes of the FSS-OANFIS model on the test datasets. The experimental outcomes pointed out that the FSS-OANFIS model has offered effective outcomes on all the datasets applied. For instance, with the Leukemia dataset, the FSS-OANFIS model has resulted in average accu, reca, spec, Fscore, and Gmeasure of 83.33%, 75%, 75%, 77.78%, and 80.08%, respectively. Following this, with the Prostate dataset, the FSS-OANFIS methodology has resulted in average accu, reca, spec, Fscore, and Gmeasure of 80.65%, 80%, 80%, 79.61%, and 81.37%, respectively. Meanwhile, with the DLBCL Stanford dataset, the FSS-OANFIS algorithm has resulted in average accu, reca, spec, Fscore, and Gmeasure of 73.33%, 66.67%, 66.67%, 65.91%, and 70.47%, respectively. Eventually, with the Colon Cancer dataset, the FSS-OANFIS technique has resulted in average accu, reca, spec, Fscore, and Gmeasure of 89.47%, 87.82%, 87.82%, 87.82%, and 87.82%, respectively.
Table 2

Result analysis of FSS-OANFIS technique with different measures and datasets.

Class labelsAccuracyRecallSpecificityF-scoreG-measure
Leukemia dataset

Class 083.33100.0050.0088.8989.44

Class 183.3350.00100.0066.6770.71

Average83.3375.0075.0077.7880.08

Prostate dataset

Class 080.65100.0060.0084.2185.28

Class 180.6560.00100.0075.0077.46

Average80.6580.0080.0079.6181.37

Stanford dataset

Class 073.3333.33100.0050.0057.74

Class 173.33100.0033.3381.8283.21

Average73.3366.6766.6765.9170.47

Colon dataset

Class 089.4792.3183.3392.3192.31

Class 189.4783.3392.3183.3383.33

Average89.4787.8287.8287.8287.82
Figure 4 shows the precision-recall curves offered by the FSS-OANFIS model on the test Leukemia dataset. The figure indicated that the FSS-OANFIS model has depicted effective precision-recall values on the classification of two classes, namely, class 0 and class 1 on the test Leukemia dataset.
Figure 4

Precision-recall analysis of FSS-OANFIS technique under the Leukemia dataset.

Next, Figure 5 shows the precision-recall curves offered by the FSS-OANFIS model on the test Prostate dataset. The figure revealed that the FSS-OANFIS technique has depicted effective precision-recall values on the classification of two classes, namely, class 0 and class 1 on the test Prostate dataset.
Figure 5

Precision-recall analysis of FSS-OANFIS technique under the Prostate dataset.

Similarly, Figure 6 shows the precision-recall curves offered by the FSS-OANFIS system on the test DLBCL Stanford dataset. The figure exposed that the FSS-OANFIS model has depicted effective precision-recall values on the classification of two classes, namely, class 0 and class 1 on the test DLBL Stanford dataset.
Figure 6

Precision-recall analysis of FSS-OANFIS technique under the DLBCL Stanford dataset.

Figure 7 shows the precision-recall curves offered by the FSS-OANFIS method on the test Colon Cancer dataset. The figure indicated that the FSS-OANFIS approach has depicted effective precision-recall values on the classification of two classes, namely, class 0 and class 1 on the test Colon Cancer dataset.
Figure 7

Precision-recall analysis of FSS-OANFIS technique under the Colon Cancer dataset.

A brief ROC investigation of the FSS-OANFIS model on the distinct four datasets is described in Figure 8. The results indicated that the FSS-OANFIS technique has exhibited its ability in categorizing two different classes such as class 0 and 1 on the test four datasets.
Figure 8

ROC analysis of FSS-OANFIS technique under different datasets.

Figure 9 shows the accuracy and loss graph analysis of the ODBN-IDS technique under four datasets. The results show that the accuracy value tends to increase and the loss value tends to decrease with an increase in epoch count. It is also observed that the training loss is low and validation accuracy is high under four datasets.
Figure 9

Accuracy and loss analysis of FSS-OANFIS technique under various datasets.

4.2. Discussion

Finally, a detailed comparative study of the FSS-OANFIS model with recent methods on distinct datasets is shown in Table 3 and Figure 10 [27]. The experimental results indicated that the FSS-OANFIS model has shown effectual outcomes under all datasets. For instance, with the Leukemia dataset, the DE and AHSA-GS models have depicted lower performance over the other methods. Though the PSO algorithm has resulted in slightly reasonable performance with accu, sens, spec, and Gmeasure of 80.59%, 74.95%, 73.96%, and 68.07%, the FSS-OANFIS model has resulted in higher accu, sens, spec, and Gmeasure of 83.33%, 75%, 75%, and 80.08%, respectively.
Table 3

Comparative analysis of FSS-OANFIS technique with existing approaches.

MethodsAccuracySensitivitySpecificityG-measure
Leukemia dataset

FSS-OANFIS83.3375.0075.0080.08

AHSA-GS75.4969.6674.8145.94

PSO algorithm80.5974.9573.9668.07

DE algorithm68.6763.0163.6264.80

Prostate dataset

FSS-OANFIS80.6580.0080.0081.37

AHSA-GS71.1953.8279.7979.84

PSO algorithm68.7863.6370.1566.01

DE algorithm62.7760.3763.2267.94

Stanford dataset

FSS-OANFIS73.3366.6766.6770.47

AHSA-GS71.2762.6465.1568.82

PSO algorithm72.8060.8260.1761.74

DE algorithm66.9363.8059.1661.48

Colon dataset

FSS-OANFIS89.4787.8287.8287.82

AHSA-GS61.0248.0764.0443.62

PSO algorithm59.0043.0258.3436.66

DE algorithm50.3833.6338.7658.07
Figure 10

Comparative analysis of FSS-OANFIS technique with existing approaches.

At the same time, with the Prostate dataset, the DE and AHSA-GS models have depicted lower performance over the other methods. Likewise, the PSO algorithm has resulted in somewhat reasonable performance with accu, sens, spec, and Gmeasure of 68.78%, 63.63%, 70.15%, and 66.01% and the FSS-OANFIS methodology has resulted in superior accu, sens, spec, and Gmeasure of 80.65%, 80%, 80%, and 81.37%, respectively. In addition, with the DLBCL Stanford dataset, the DE and AHSA-GS techniques have showcased lesser performance over the other methods. Though the PSO algorithm has resulted in slightly reasonable performance with accu, sens, spec, and Gmeasure of 72.80%, 60.82%, 60.17%, and 61.74%, the FSS-OANFIS approach has resulted in higher accu, sens, spec, and Gmeasure of 733.33%, 66.67%, 66.67%, and 70.47%, respectively. Along with that, with the Colon Cancer dataset, the DE and AHSA-GS models have portrayed lower performance over the other methods. But, the PSO approach has resulted in slightly reasonable performance with accu, sens, spec, and Gmeasure of 59%, 43.02%, 58.34%, and 36.66%, and the FSS-OANFIS system has resulted in superior accu, sens, spec, and Gmeasure of 89.47%, 87.80%, 87.82%, and 87.82%, respectively. After examining the results and discussion, it is apparent that the FSS-OANFIS model has accomplished maximum performance in the microarray gene expression classification process.

5. Conclusion

In this study, a new FSS-OANFIS model has been developed for microarray gene expression data classification. The presented FSS-OANFIS model encompasses a series of processes, namely, data pre-processing, IGWO-FS-based election of features, ANFIS classification, and COA-based parameter optimization. The application of IGWO-FS and COA techniques helps in accomplishing enhanced microarray gene expression classification outcomes. For examining the enhanced outcomes of the FSS-OANFIS model, a wide range of simulations were performed on distinct datasets. The experimental results indicated that the FSS-OANFIS model has resulted in enhanced performance over the recent approaches. In future, the feature reduction and clustering approaches can be integrated to enhance gene expression classification outcomes.
  19 in total

1.  Gene expression cancer classification using modified K-Nearest Neighbors technique.

Authors:  Sarah M Ayyad; Ahmed I Saleh; Labib M Labib
Journal:  Biosystems       Date:  2019-01-03       Impact factor: 1.973

2.  Feature specific quantile normalization enables cross-platform classification of molecular subtypes using gene expression data.

Authors:  Jennifer M Franks; Guoshuai Cai; Michael L Whitfield
Journal:  Bioinformatics       Date:  2018-06-01       Impact factor: 6.937

3.  Broad patterns of gene expression revealed by clustering analysis of tumor and normal colon tissues probed by oligonucleotide arrays.

Authors:  U Alon; N Barkai; D A Notterman; K Gish; S Ybarra; D Mack; A J Levine
Journal:  Proc Natl Acad Sci U S A       Date:  1999-06-08       Impact factor: 11.205

4.  Machine learning integrated ensemble of feature selection methods followed by survival analysis for predicting breast cancer subtype specific miRNA biomarkers.

Authors:  Jnanendra Prasad Sarkar; Indrajit Saha; Anasua Sarkar; Ujjwal Maulik
Journal:  Comput Biol Med       Date:  2021-01-28       Impact factor: 4.589

5.  Molecular classification of cancer: class discovery and class prediction by gene expression monitoring.

Authors:  T R Golub; D K Slonim; P Tamayo; C Huard; M Gaasenbeek; J P Mesirov; H Coller; M L Loh; J R Downing; M A Caligiuri; C D Bloomfield; E S Lander
Journal:  Science       Date:  1999-10-15       Impact factor: 47.728

6.  Gene selection for microarray data classification via multi-objective graph theoretic-based method.

Authors:  Mehrdad Rostami; Saman Forouzandeh; Kamal Berahmand; Mina Soltani; Meisam Shahsavari; Mourad Oussalah
Journal:  Artif Intell Med       Date:  2021-12-03       Impact factor: 5.326

7.  Feature selection methods on gene expression microarray data for cancer classification: A systematic review.

Authors:  Esra'a Alhenawi; Rizik Al-Sayyed; Amjad Hudaib; Seyedali Mirjalili
Journal:  Comput Biol Med       Date:  2021-11-23       Impact factor: 4.589

Review 8.  Gene Expression-Assisted Cancer Prediction Techniques.

Authors:  Tanima Thakur; Isha Batra; Monica Luthra; Shanmuganathan Vimal; Gaurav Dhiman; Arun Malik; Mohammad Shabaz
Journal:  J Healthc Eng       Date:  2021-08-19       Impact factor: 2.682

9.  Machine learning integration of scleroderma histology and gene expression identifies fibroblast polarisation as a hallmark of clinical severity and improvement.

Authors:  Dana E Orange; Jessica K Gordon; Kimberly Showalter; Robert Spiera; Cynthia Magro; Phaedra Agius; Viktor Martyanov; Jennifer M Franks; Roshan Sharma; Heather Geiger; Tammara A Wood; Yaxia Zhang; Caryn R Hale; Jackie Finik; Michael L Whitfield
Journal:  Ann Rheum Dis       Date:  2020-10-07       Impact factor: 19.103

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.