Skip to main content

An effective heuristic for developing hybrid feature selection in high dimensional and low sample size datasets

Abstract

Background

High-dimensional datasets with low sample sizes (HDLSS) are pivotal in the fields of biology and bioinformatics. One of core objective of HDLSS is to select most informative features and discarding redundant or irrelevant features. This is particularly crucial in bioinformatics, where accurate feature (gene) selection can lead to breakthroughs in drug development and provide insights into disease diagnostics. Despite its importance, identifying optimal features is still a significant challenge in HDLSS.

Results

To address this challenge, we propose an effective feature selection method that combines gradual permutation filtering with a heuristic tribrid search strategy, specifically tailored for HDLSS contexts. The proposed method considers inter-feature interactions and leverages feature rankings during the search process. In addition, a new performance metric for the HDLSS that evaluates both the number and quality of selected features is suggested. Through the comparison of the benchmark dataset with existing methods, the proposed method reduced the average number of selected features from 37.8 to 5.5 and improved the performance of the prediction model, based on the selected features, from 0.855 to 0.927.

Conclusions

The proposed method effectively selects a small number of important features and achieves high prediction performance.

Peer Review reports

Background

In the current data era, information is recorded intricately across various dimensions. Consequently, the dimensionality of the data increases rapidly. High-dimensional datasets refer to data where the number of features (p) is considerably larger than the sample size (n). High-dimensional datasets have driven significant scientific discoveries in fields such as biology and bioinformatics by revealing previously unknown complex patterns. However, analyzing and utilizing high dimensional and low sample size (HDLSS) datasets remain a challenge for researchers [1,2,3,4,5,6].

Microarray data are prime examples of HDLSS datasets. With the capacity to capture the expression of tens of thousands of genes, the core aim is to identify genetic variations in specific cancers or other genetically-related diseases [3]. However, a notable proportion of these genes either lacks direct relevance to the disease or exhibits substantial overlap in feature information, leading to redundancy [7]. Furthermore, owing to the patient-centric nature of these data, sample size was inherently limited. These aspects amplify risks such as overfitting and the curse of dimensionality [3, 4]. Consequently, developing cancer classification models and analyzing HDLSS datasets such as microarrays, pose intricate challenges for computer science researchers [7].

A fundamental approach for addressing these challenges in machine learning is dimensionality reduction via feature selection. The primary objective of feature selection is to filter out irrelevant and redundant features, thereby honing the most pertinent features relevant to the subject of interest. This enhances the clarity, generalizability, and predictive accuracy of classification models, concurrently reducing computational demands and boosting efficiency [1]. In bioinformatics, specifically, feature selection plays a critical role in uncovering potential avenues for drug development and offering insights into disease diagnostics and causation [1,2,3, 8].

Identifying the optimal features is an NP-hard problem [3, 6, 8, 9]. Numerous feature-selection methods have been proposed to address this challenge. These methods are typically classified into three categories based on their interactions with the classification model: filter, wrapper, and embedded methods [3, 4]. The filter method takes a model-independent stance, assigns rankings to features based on metrics such as statistical or information-theoretic properties, and subsequently selects those that exceed a predetermined threshold. In contrast, the wrapper method incorporates a model-dependent approach, by selecting feature subsets based on the performance of a specific model. The embedded method differs from the wrapper method by integrating feature selection directly into the model-training process.

Recent advancements in feature selection include methods based on graphs and deep learning frameworks. These methods aim to capture the intricate relationships among features. In the graph-based approach, features are visualized as nodes and their interrelationships as edges. This method determines the importance of features by analyzing the structural properties of a graph [6, 10,11,12]. Deep learning techniques for feature selection utilize neural networks with features as inputs [6,7,8,9, 12,13,14]. The significance of each feature is gauged by the magnitude of the gradients during network training.

For HDLSS data, several studies on feature selection have adopted a hybrid approach involving two stages, filtering and searching, as illustrated on the upper side of Fig. 1 [7]. This approach seeks to harness the computational efficiency of filter methods and the high performance of wrapper methods. The filter method can be viewed as a preprocessing step that significantly reduces the search scope by eliminating unnecessary features. By contrast, the wrapper method refines the search within this narrow scope, aiming to identify the most optimal features more precisely [3, 7].

Fig. 1
figure 1

General process of feature selection for HDLSS datasets and improved points with the proposed method

From the collective insights of previous studies on feature selection, we can draw the following conclusions:

  1. 1.

    Understanding intricate relationships, such as the interactions between features, is pivotal. Previous methods often overlooked this aspect. For example, filter methods assume that each feature is independent and does not interact with any other feature.

  2. 2.

    Finding the optimal features for HDLSS is realistically difficult. An alternative is to find a better solution among several suboptimal features. Therefore, avoiding local optima is essential, as wrapper methods are particularly prone to this issue.

  3. 3.

    These two objectivess can be achieved by balancing the diversity and focus in the feature search space, introducing randomness during the search process, and employing a hybrid method that leverages both filter and wrapper techniques.

In this study, we propose a new metaheuristic method [15,16,17,18,19] to improve upon previous feature selection approaches for HDLSS datasets, aiming to achieve near-optimal performance with minimal features. By integrating gradual permutation filtering with a diverse search strategy, we designed our approach based on core principles. The key highlights of our method are as follows:

  1. 1.

    Filtering: Leveraging permutation importance, filtering accounts for feature interactions and establishes a correct threshold.

  2. 2.

    Search: Through “consolation matches,” the nesting effect can be overcome, wherein a once-selected (or excluded) feature remains unaltered. Additionally, by varying “first-choice features,” the search range is expanded to approach a more global optimum.

  3. 3.

    Runtime optimization: Using ranked features, the method efficiently reduces the search space.

  4. 4.

    Assessing the fitness of the selected features: We crafted a unique performance metric for the HDLSS, providing a holistic perspective on feature quantity and quality.

The details are outlined in the next section.

Methods

Overall procedure

The proposed method also follows the general procedure of feature selection for HDLSS datasets. As illustrated on the lower side of Fig. 1 and elaborated in Algorithm 1, the proposed method comprises two primary phases: filtering and searching.

  1. 1.

    Gradual Permutation Filtering (GPF): This phase ingests all the HDLSS data features. It ranks features based on their permutation importance and subsequently eliminates irrelevant features (Algorithm 2).

  2. 2.

    Heuristic Tribrid Search (HTS): By leveraging the features ranked by the GPF, HTS employs a heuristic blend of forward search, consolation match, and backward elimination. This approach aims to identify the near-optimal feature set by considering both the number of features and classification performance using the features (see Algorithms 3–6).

The subsequent sections will delve into the specifics of each phase.

Algorithm 1
figure a

Overall procedure

Gradual permutation filtering

The GPF stage prioritizes features based on their permutation importance and eliminates unimportant features. This method chooses features with an importance value greater than zero because values below this threshold often indicate redundancy or suggest potential noise.

This method offers a refined approach for evaluating feature importance by adopting a gradual process of eliminating noise and recalculating the importance, thereby minimizing the potential biases associated with single-step elimination. Considering that the permutation importance is evaluated within the evaluation group (features), the approach emphasizes filtering out irrelevant features and reevaluating the importance of enhancing purity, which is defined solely based on performance-impacting features. Notably, a gradual feature filtering method, as opposed to filtering all at once, provides a more precise selection of features with importance on the verge of 0.

To ensure robust feature selection, the GPF measures the permutation importance of each feature multiple times; i.e., 50 times in our case. In Algorithm 2, the constant variable M refers to the number of permutation test trials used for feature evaluation. Because permutation involves randomness, performing the test only once does not allow for an accurate assessment of feature importance. Based on our experiments, we determined that setting M to approximately 50 trials is appropriate. From these measurements, the features that exceed the importance of zero for a specific threshold number were selected. For the selected features, the permutation importance was recalculated by applying a progressively higher threshold each time. This iterative process, detailed in Algorithm 2, refines the ranking of significant features. The final ranking of the features was determined by averaging the last measured importance values of the features selected until the end.

Algorithm 2
figure b

Gradual_permutation_filtering (FILTERED, CL, RATIO)

Heuristic tribrid search

While the GPF is designed to filter out unimportant features, HTS focuses on identifying the most informative features from a pool refined by the GPF. HTS operates in three distinct stages, forward search, consolation match, and backward elimination, as outlined in Algorithm 3.

We slightly modified the original forward search technique. The original technique began with an empty feature set and determined its first feature. In contrast, the proposed method uses “first-choice feature” that is chosen from ranked feature list of GPF. This procedure incrementally adds to the feature set, guided by the performance increments detailed in Algorithm 4. The performance was measured using a classification metric based on the selected features.

When the performance increase caused by the selected features stops during the forward search, the process shifts to “consolation match.” This stage aims to enhance the performance by swapping a single feature between selected and unselected feature pools, as depicted in Algorithm 5. This strategy offers an escape route from potential local optima. Notably, the duration of this stage depends on the volume of features initially filtered by the GPF. Techniques for expediting this stage are explored in subsequent sections.

If the consolation match yields improved performance, the forward search resumes further enhancements. However, if no such gains are evident, HTS transitions to a backward elimination phase. This final stage discards any remaining unimportant features, thereby ensuring the relevance of the end feature set.

Importantly, at each stage of the HTS (Algorithms 4–6), the performance is gauged using a novel metric introduced in this study. This metric considers both the classification performance of the model and feature count. An elaboration of this metric is presented in the following section.

Algorithm 3
figure c

Heuristic_tribird_search (FILTERED, CL, i, N)

Algorithm 4
figure d

Forward_search (FILTERED, CL, SELECTED, max_fit, N)

Algorithm 5
figure e

Consolation_match (FILTERED, CL, SELECTED, max_fit, N)

Log comprehensive metric for HDLSS datasets

Balancing the number of selected features with the classification performance is imperative when working with HDLSS datasets. Therefore, this study proposes log-comprehensive metric (LCM). This metric, which is a refined iteration of the conventional fitness function, was meticulously designed for the HDLSS datasets, as elaborated in Algorithm 6.

Algorithm 6
figure f

Log_comprehensive_metric (CL, SELECTED, N)

The widely used fitness function is described by Eq. (1). It constitutes a weighted sum of the model error and the ratio of the selected features to the total. The aim is to minimize the following function:

$$fitness=\theta *Error+\left(1-\theta \right)*\left(\frac{Number\, of\, selected\, features}{Total\, number\, of\, features}\right), where\, 0\le \theta \le 1$$
(1)

By contrast, the LCM, expressed in Eq. (2), fuses the model’s classification success with the log reduction rate (LRR) using weighted factors. The objective is to maximize.

$$LCM=\theta *Performance+\left(1-\theta \right)*LRR, where 0\le \theta \le 1$$
(2)

The parameter performance in Eq. (2) denotes the evaluation metric for the model used in the HTS stage, which is a variation of the wrapper method. In this study, AUC is used as the performance metric; however, it is not restricted to AUC. Other metrics, such as accuracy or sensitivity—both ranging between 0 and 1—can also be applied. Furthermore, the model for evaluating performance is not limited to XGBoost, offering flexibility in the choice of algorithms. In Algorithm 6, performance values are computed as the average metric obtained from fivefold cross-validation.

LRR is further explained by Eq. (3). Its value fluctuates between 0 and 1 and approaches 1 as the number of selected features decreases.

$$LRR=1-{\text{log}}_{a}b=1-\text{log}b/\text{log}a,$$
$$where:0\le LRR\le 1 ,$$
$$a=Total\, number\, of\, features, b= Number\, of\, selected\, features$$
(3)

In contrast to the reduction rate (RR), which is formulated as RR = 1—(Number of selected features / Total number of features), the advantage of LRR lies in its enhanced sensitivity, especially when the number of selected features is small.

Table 1 demonstrates the effectiveness of the LCM measure. While the Performance (AUC) values are hypothetical, the LRR values are actual calculations based on the number of features, and the LCM values are derived from the Performance (AUC) and LRR. By focusing solely on the AUC performance, one might opt for the set with 50 features that presents the highest AUC. However, when evaluating based on the LCM, the most optimal feature set is that with only 10 features. This LCM-based selection results in a minor AUC performance drop of 0.002, but with the benefit of reducing the selected feature count by 40.

Table 1 Comparison of two measures, LRR and LCM

Experimental procedure

Benchmark datasets

To evaluate the previous and proposed feature selection methods, we utilized 24 publicly available datasets designed for binary classification, as presented in Table 2. Of these, No. 1 to 17 are cancer-related microarray datasets. The features within these microarray datasets represent gene expression profiles, which are often referred to as probes. These datasets were acquired from a widely-used R package, “datamicroarray” [20]. We also included additional datasets (No. 18 to 24) obtained from public databases [21, 22] to demonstrate that the proposed method is effective for non-microarray or non- HDLSS data. No. 1 to 18 are HDLSS datasets, while No. 19 to 24 are not.

Table 2 List of benchmark datasets

Compared feature selection methods and experimental environment

To assess the efficacy of the proposed method, we compared it with various established feature selection methods, as presented in Table 3.

Table 3 List of benchmark feature selection methods

Among these, the first four adopted a filter-based approach. For each technique, feature rankings were computed and the top 100 features were selected. In the case of GRACES, a fixed number of features were produced. Hence, to maintain parity with the number of features selected by the other methods, we set this count to 100 and used the resulting rankings. The rankings for both INF and GRACES were determined after parameter optimization using fivefold cross validation.

Direct comparisons between any filter method and the proposed approach are challenging because filter methods do not explicitly indicate a subset of features that yield optimal performance. Consequently, it is imperative to identify the optimal feature subsets for benchmarking methods. For this purpose, we employed two distinct strategies: ‘simple sequential search’ and ‘forward search.’ In simple sequential search, the performance of the classification model was evaluated at each step while increasing the number of features from 1 to 100 based on the ranking of each feature, and the subset of features with the highest performance was selected. Forward search is similar to simple sequential search, but it continues selecting good features until no further improvement in performance is observed.

To evaluate the quality of the selected features, we applied the eXtreme Gradient Boosting (XGB) classifier, equipped with specific parameters, nrounds = 5, objective = ”multi:softmax”, while retaining default configurations for the remaining parameters. Owing to its established potency and computational efficiency, XGB is an appropriate choice for evaluating feature selection methods.

For the performance evaluation, we incorporated two metrics: Area Under the Curve (AUC) and LCM. AUC is a useful measure for comparing the performance of prediction models, especially when the classes of the dataset are imbalanced. As explained previously, LCM is suitable for evaluating both the classification performance and the number of selected features in the HDLSS dataset. For our analysis, a θ value of 0.8 was adopted as the reference point. The choice of θ = 0.8 was informed by a performance comparison between the proposed method and existing methods across 24 datasets. Our method consistently outperformed Simple Sequential Search and Forward Search for θ values ranging from 0.7 to 0.9, with the optimal balance observed at 0.8. Figure 2 illustrates the overall process of the comparative experiments.

Fig. 2
figure 2

Overall procedure of the comparative experiment

To orchestrate the efficacious experiments, we created two computational environments, R and Python, as shown in Table 4. Although the rankings for GRACE and INF were procured within a Python framework, the search and evaluation phases for all the techniques were uniformly executed in the R environment. The associated experimental code is available at https://bitldku.github.io/home/sw/heuristic_search.html.

Table 4 Experimental environment

Results

Number of selected features through GPF and HTS

The number of selected features is an important criterion for feature selection in the HDLSS datasets. The proposed method comprised two procedures for feature selection: GPF and HTS. The GPF filters useless features, and the HTS chooses informative features from GPF results. Table 5 summarizes the results of the GPF and HTS analyses. Details are provided in Supplementary Table S1. During the GPF process, approximately 98.2% of the total features were eliminated from the 18 HDLSS datasets. Conversely, in the six non-HDLSS datasets, the average feature reduction rate was 61.2%, indicating considerable variability among the datasets. Notably, datasets with fewer than 100 features were not filtered. In the HTS phase, an additional 97.6% of the previously reduced features in the HDLSS dataset were excluded. The non-HDLSS datasets also exhibited a significant elimination rate, with an average of 98.3%. In conclusion, after the integrated application of both the GPF and HTS methodologies, the HDLSS datasets consistently selected 12 or fewer features, representing an average of merely 0.07% of the initial feature set.

Table 5 Average number of features determined by proposed GPF and HTS procedures

Performance comparison based on simple sequential search

In this section, we compare the classification performance the previous and proposed methods using the selected features. To determine the performance of the filter method, a simple sequential search was performed as described in the “Compared feature selection methods” section. AUC and LCM were used as performance criteria. A summary of the comparison results is presented in Table 6 and Fig. 3. Details are provided in Supplementary Table S2 and S3.

Table 6 Comparison of the number of selected features, AUC, and LCM between the proposed method and previous works using simple sequential search
Fig. 3
figure 3

Comparison result of AUC, LCM, and No of features in Table 6

When compared using the AUC performance metric, the proposed method exhibited equivalent or superior performance in 16 out of 24 datasets, while maintaining the lowest performance variability among all methods, as detailed in Supplementary Table S2 and S3. The AUC of our method exceeded that of comparative approaches, with margins ranging from 2.2 to 13.1%. However, the number of selected features was markedly reduced, averaging only 5.5 features, which is between 1/4 and 1/10 that of the other methods selected.

Utilizing the LCM performance metric with a θ value of 0.8, our method displayed enhanced performance in 20 out of 24 datasets, representing 83.3%. This indicates that when considering both the performance and influence of the number of selected features, the proposed method excels, highlighting its ability to achieve comparable or better results with fewer features.

Among the traditional methods, permutation exhibits the best performance, followed by F-statistics, mRMR, GRACES, and INF. Notably, GRACES and INF exhibited a tendency to select a greater number of features than the other methods. Furthermore, INF was unable to produce results for datasets exceeding 20,000 features, and GRACES also encountered difficulties in generating results for specific datasets.

Performance comparison based on forward search

In this section, we compare the classification performances of previous methods and the proposed method. To determine the performance of the filter method, a forward and a simple sequential search was used as described in the “Compared feature selection methods” section. A summary of the results is presented in Table 7 and Fig. 4, and the complete experimental results are detailed in Supplementary Table S4 and S5.

Table 7 Comparison of the number of selected features, AUC, and LCM between the proposed method and previous works using forward search
Fig. 4
figure 4

Comparison result of AUC, LCM, and No of features in Table 7

Compared to the results in previous section, most filter methods showed an improvement in the AUC and significantly reduced the number of selected features. This trend suggests that the top 100 features chosen using traditional filtering methods exhibit high redundancy. A forward search helped remove many overlapping features. In specific datasets subh as Alon, both the feature count and AUC decreased, which might indicate a fall into local optima.

Using the AUC metric, the proposed method matched or exceeded the performance of the other methods on 20 of the 24 datasets (83.3%). It also exhibited the lowest variability in performance. The AUC of the proposed method was between 2.2 and 11.1% higher than those of the other methods. Although the comparative methods selected fewer features on average than the sequence results, the proposed method consistently selected the least number of features. Furthermore, when using the LCM with θ at 0.8, the proposed method outperformed other methods in 21 of the 24 datasets (87.5%).

Execution time of the proposed method

Supplementary Table S6 lists the average runtimes of our method over 30 iterations for the HDLSS datasets. The Arcene dataset required approximately 12 min, the longest duration, whereas 16 out of the 18 datasets completed the feature selection within 5 min. Notably, the HTS phase constituted 77% of the total duration, with 94% spiking in some cases. The execution times for both GPF and HTS increased proportionally with instance size. Conversely, apart from SMK_CAN_187, the GPF duration remained relatively consistent irrespective of the feature count, whereas HTS showed a pronounced rise as the features increased. In essence, the duration of the HTS phase significantly shapes the total runtime, with the efficiency of the GPF in transmitting a reduced feature set being pivotal to the HTS timeframe. In conclusion, the average execution time of the proposed method is 169 s (2.8 min), which is reasonable.

Discussion and conclusions

The essence of feature selection lies in effectively finding a subset that is close to the global optimum, where the relevance to the class is high and the redundancy between features is low. We proposed two methods, GPF for filtering and HTS for searching and verified their superior performance compared to those of other methods. Particularly, for microarray datasets with HDLSS properties, we selected a minimal subset of core features within a reasonable time. The advantages of the proposed method are summarized as follows:

Incorporating feature interactions for relevance measurement in the filtering stage: Oh [26] experimentally and theoretically proved that the importance of a feature can be decomposed into its intrinsic predictive power (feature power) and the effect arising when combined with other features (interaction). The importance of a feature not only lies in its intrinsic power but also varies depending on the surrounding features. Therefore, the relevance between the subset and class can be evaluated more accurately by considering feature interactions. As we progressively filtered out features with low importance in the filtering stage and then measured their importance within the refined subset, we were able to accurately measure the interactions based on this subset. Although the permutation importance method we used does not explicitly distinguish between feature power and interaction, it is clear that the interaction effect is reflected within the importance.

Elimination of redundancy among features during the search stage: While basic permutation importance can account for interactions between features, it does not capture redundancy. Thus, there is a possibility that highly redundant features exist within the filtered subset. By contrast, the proposed HTS can exclude redundancy among the features. To incorporate a new feature into the existing subset in the HTS process, the performance should be improved. If a perfectly redundant feature exists, no performance improvement observed, which leads to its exclusion. Therefore, HTS can play a complementary role in removing redundant features present in GPF.

Clear threshold criteria in the filtering stage: Traditional filter methods do not consider the best candidate features for the next selection stage. Therefore, the users randomly select the best n features. In other methods, the candidate features produced depend on threshold. Therefore, users must tune these parameters to obtain better candidate features. By contrast, the proposed method produces effective candidate features without parameter settings. For our filtering process, we utilized a fixed threshold value based on theoretical evidence. A permutation importance of zero indicates that the feature has no effect on the performance improvement [27]. If a feature degrades the performance or causes a negative interaction, its importance becomes less than 0. Hence, setting the threshold to zero allows for a clear and intuitive filtering of features that contribute to performance enhancement.

Improvement of computational efficiency using feature ranking: In the filtering stage, existing methods utilize feature evaluation values exclusively for the removal of weak features, whereas our proposed method harnesses these values to optimize subsequent computations. Specifically, SFFS [28] thoroughly explores the whole feature space, encompassing both selected and unselected features, to identify pairs that can overcome the nesting effect. However, by leveraging the feature-ranking information already obtained during the filtering stage, our approach constrained the search space. This not only enhances the computational efficiency during the search phase but also avoids performance degradation.

Expended search space to get better features: Beyond “consolation match”, we incorporated a semi-random approach to pinpoint the global optimum. At the start of the search, we considered multiple candidate features with high importance to alter the first-choice feature and selected the feature subset with the best performance. From our observations, the optimal feature set frequently emerged when the first-choice feature was not ranked as the top feature.

Selection of a sufficiently good features for microarray datasets: To understand or diagnose complex diseases, it is essential to not only achieve high predictive (classification) performance but also identify pivotal biomarkers (features) [27]. For example, a predictive accuracy of 0.82 with 20 features is more meaningful than an accuracy of 0.85 with 120 features. Although the proposed method does not always result in a better predictive performance than the other methods, the derived features are very compact with a decent prediction performance. Therefore, it would be helpful to identify novel genes associated with diseases.

Usefulness of selected features by the proposed method: Microarray-related papers typically present a list of important genes derived through simple statistical analysis, without experimental validation. Therefore, it is difficult to demonstrate that the features (genes) identified by the proposed method are experimentally useful. In the case of chiaretti dataset, the proposed method selected five genes—ITM2A, IGF2R, FADS1, CDK14, and PON2—which are related to lymphoblastic leukemia. The original paper [29] identified 31 important genes, among which only ITM2A overlaps with the proposed method’s results; the other four genes are newly discovered. Through a literature analysis, we confirmed that these four genes are also associated with lymphoblastic leukemia, as shown in Table 8. These findings indirectly demonstrate that the proposed method produces biologically meaningful results.

Table 8 Studies demonstrating the association of selected genes with lymphoblastic leukemia

In our proposed method, during the consolation match, we specifically operated on the latter half of the selected features and the former half of the non-selected features. Unexpectedly, when we expanded this search range during our experiments, a decline in the performance was observed. This phenomenon raises questions that require further exploration, particularly concerning feature interactions. This is a topic for further research.

Availability of data and materials

No datasets were generated or analysed during the current study.

References

  1. Ang JC, Mirzal A, Haron H, Hamed HNA. Supervised, unsupervised, and semi-supervised feature selection: a review on gene selection. IEEE/ACM Trans Comput Biol Bioinform. 2016;13(5):971–89. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/TCBB.2015.2478454.

    Article  PubMed  Google Scholar 

  2. Lu H, Chen J, Yan K, Jin Q, Xue Y, Gao Z. A hybrid feature selection algorithm for gene expression data classification. Neurocomputing. 2017;256:56–62. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.neucom.2016.07.080.

    Article  Google Scholar 

  3. Almugren N, Alshamlan H. A survey on hybrid feature selection methods in microarray gene expression data for cancer classification. IEEE Access. 2019;7:78533–48. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/ACCESS.2019.2922987.

    Article  Google Scholar 

  4. Bommert A, Sun X, Bischl B, Rahnenführer J, Lang M. Benchmark for filter methods for feature selection in high-dimensional classification data. Comput Stat Data Anal. 2020;143:106839. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.csda.2019.106839.

    Article  Google Scholar 

  5. Manikandan G, Abirami S. Feature selection is important: state-of-the-art methods and application domains of feature selection on high-dimensional data. In: Kumar R, Paiva S, editors. Applications in Ubiquitous Computing. Cham: Springer; 2021. p. 177–96.

    Chapter  Google Scholar 

  6. Chen C, Weiss ST, Liu YY. Graph convolutional network-based feature selection for high-dimensional and low-sample size data. Bioinformatics. 2023. https://doiorg.publicaciones.saludcastillayleon.es/10.1093/bioinformatics/btad135.

    Article  PubMed  PubMed Central  Google Scholar 

  7. Alhenawi E, Al-Sayyed R, Hudaib A, Mirjalili S. Feature selection methods on gene expression microarray data for cancer classification: a systematic review. Comput Biol Med. 2022;140:105051. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.compbiomed.2021.105051.

    Article  CAS  PubMed  Google Scholar 

  8. Liu B, Wei Y, Zhang Y, Yang Q. Deep neural networks for high dimension, low sample size data. In: Proceedings of the 26th International Joint Conference on Artificial Intelligence. Melbourne: AAAI, 2017; 2287–2293.

  9. Li K, Wang F, Yang L, Liu R. Deep feature screening: feature selection for ultra high-dimensional data via deep neural networks. Neurocomputing. 2023;538:126186. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.neucom.2023.03.047.

    Article  Google Scholar 

  10. Roffo G, Melzi S, Castellani U, Vinciarelli A, Cristani M. Infinite feature selection: a graph-based feature filtering approach. IEEE Trans Pattern Anal Mach Intell. 2020;43(12):4396–410. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/TPAMI.2020.3002843.

    Article  Google Scholar 

  11. Rostami M, Forouzandeh S, Berahmand K, Soltani M, Shahsavari M, Oussalah M. Gene selection for microarray data classification via multi-objective graph theoretic-based method. Artif Intell in Med. 2022;123:102228. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.artmed.2021.102228.

    Article  Google Scholar 

  12. Pati SK, Banerjee A, Manna S. Gene selection of microarray data using heatmap analysis and graph neural network. Appl Soft Comput. 2023;135:110034. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.asoc.2023.110034.

    Article  Google Scholar 

  13. Li Y, Chen CY, Wasserman WW. Deep feature selection: theory and application to identify enhancers and promoters. J Comput Biol. 2016;23(5):322–36. https://doiorg.publicaciones.saludcastillayleon.es/10.1089/cmb.2015.0189.

    Article  CAS  PubMed  Google Scholar 

  14. Chowdhury S, Dong X, Li X. Recurrent neural network based feature selection for high dimensional and low sample size micro-array data. In: Proceedings of 2019 IEEE International Conference on Big Data. Piscataway: IEEE, 2019; 4823–4828.

  15. Agrawal P, Abutarboush HF, Ganesh T, Mohamed AW. Metaheuristic algorithms on feature selection: a survey of one decade of research (2009–2019). IEEE Access. 2021;9:26766–91. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/ACCESS.2021.3056407.

    Article  Google Scholar 

  16. Dokeroglu T, Deniz A, Kiziloz HE. A comprehensive survey on recent metaheuristics for feature selection. Neurocomputing. 2022;494:269–96. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.neucom.2022.04.083.

    Article  Google Scholar 

  17. Ferri FJ, Pudil P, Hatef M, Kittler J. Comparative study of techniques for large-scale feature selection. Mach Intell Pattern Recognit. 1994;16:403–13. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/B978-0-444-81892-8.50040-7.

    Article  Google Scholar 

  18. Wang L, Wang Y, Chang Q. Feature selection methods for big data bioinformatics: a survey from the search perspective. Methods. 2016;111:21–31. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.ymeth.2016.08.014.

    Article  CAS  PubMed  Google Scholar 

  19. Hsu HH, Hsieh CW, Lu MD. Hybrid feature selection by combining filters and wrappers. Expert Syst Appl. 2011;38(7):8144–50. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.eswa.2010.12.156.

    Article  Google Scholar 

  20. Ramey J. Datamicroarray. Available at https://github.com/ramhiser/datamicroarray (accessed 19 Jan. 2024).

  21. Arizona State University (ASU). Feature selection datasets. Available at https://jundongl.github.io/scikit-feature/datasets.html (accessed 19 Jan. 2024).

  22. OpenML. A worldwide machine learning lab. Available at https://www.openml.org/ (accessed 19 May 2024).

  23. Song Q, Jiang H, Liu J. Feature selection based on FDA and F-score for multi-class classification. Expert Syst Appl. 2017;81:22–7.

    Article  Google Scholar 

  24. Berrendero JR, Cuevas A, Torrecilla JL. The mRMR variable selection method: a comparative study for functional data. J Stat Comput Simul. 2016;86(5):891–907.

    Article  Google Scholar 

  25. Altmann A, Toloşi L, Sander O, Lengauer T. Permutation importance: a corrected feature importance measure. Bioinformatics. 2010;26(10):1340–7.

    Article  CAS  PubMed  Google Scholar 

  26. Oh S. Predictive case-based feature importance and interaction. Inform Sci. 2022;593:155–76. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.ins.2022.02.003.

    Article  Google Scholar 

  27. Mi X, Zou B, Zou F, Hu J. Permutation-based identification of important biomarkers for complex diseases via machine learning models. Nature Commun. 2021;12(1):3008. https://doiorg.publicaciones.saludcastillayleon.es/10.1038/s41467-021-22756-2.

    Article  Google Scholar 

  28. Pudil P, Novovičová J, Kittler J. Floating search methods in feature selection. Pattern Recognit Lett. 1994;15(11):1119–25. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/0167-8655(94)90127-9.

    Article  Google Scholar 

  29. Chiaretti S, Zini G, Bassan R. Diagnosis and subclassification of acute lymphoblastic leukemia. Mediterr J Hematol Infect Dis. 2014;6(1):e2014073.

    Article  PubMed  PubMed Central  Google Scholar 

  30. Coelho-Silva JL, Machado-Neto JA, Fernandes JC, de Lima ASG, Scheucher PS, Rego EM, Traina F. IGF signaling predicts outcomes and is a promising target therapy for acute myeloid leukemia. Blood. 2017;130:3966.

    Google Scholar 

  31. Heravi G, Jang H, Wang X, Long Z, Peng Z, Kim S, Liu W. Fatty acid desaturase 1 (FADS1) is a cancer marker for patient survival and a potential novel target for precision cancer treatment. Front Oncol. 2022;12:942798.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  32. Chadeau-Hyam M, Vermeulen RCH, Hebels DGAJ, Castagné R, Campanella G, Portengen L, et al. Prediagnostic transcriptomic markers of Chronic lymphocytic leukemia reveal perturbations 10 years before diagnosis. Ann Oncol. 2014;25(5):1065–72.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  33. Pan L, Hong C, Chan LN, Xiao G, Malvi P, Robinson ME, et al. PON2 subverts metabolic gatekeeper functions in B cells to promote leukemogenesis. Proc Natl Acad Sci. 2021;118(7):e2016553118.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

Download references

Acknowledgements

None.

Funding

This work was supported by the Institute for Information & Communications Technology Planning & Evaluation (IITP) grand funded by the Ministry of Science, ICT (MSIT), Korea (No. RS-2023–00222191, Development of data fabric technology to support logical data integration and compound analysis of distributed data).

Author information

Authors and Affiliations

Authors

Contributions

Shin and Oh conceptualized the research. Shin contributed to development of the model. Shin collected and annotated data. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Sejong Oh.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interest

The authors declare no competing interests.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Shin, H., Oh, S. An effective heuristic for developing hybrid feature selection in high dimensional and low sample size datasets. BMC Bioinformatics 25, 390 (2024). https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12859-024-06017-9

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12859-024-06017-9

Keywords