Abstract

Motivation

Omics features, often measured by high-throughput technologies, combined with clinical features, significantly impact the understanding of many complex human diseases. Integrating key omics biomarkers with clinical risk factors is essential for elucidating disease mechanisms, advancing early diagnosis, and enhancing precision medicine. However, the high dimensionality and intricate associations between disease outcomes and omics profiles present substantial analytical challenges.

Results

We propose a high-dimensional feature importance test (HiFIT) framework to address these challenges. Specifically, we develop an ensemble data-driven biomarker identification tool, Hybrid Feature Screening (HFS), to construct a candidate feature set for downstream machine learning models. The pre-screened candidate features from HFS are further refined using a computationally efficient permutation-based feature importance test employing machine learning methods to flexibly model the potential complex associations between disease outcomes and molecular biomarkers. Through extensive numerical simulation studies and practical applications to microbiome-associated weight changes following bariatric surgery, as well as the examination of gene-expression-associated kidney pan-cancer survival data, we demonstrate HiFIT’s superior performance in both outcome prediction and feature importance identification.

Availability and implementation

An R package implementing the HiFIT algorithm is available on GitHub (https://github.com/BZou-lab/HiFIT).

Introduction

High-dimensional omics data, such as genomics, proteomics, and other types of biomedical data generated from high-throughput technologies, have revolutionized clinical research and personalized medicine by providing detailed molecular profiles of individuals (Meyer et al. 2013, Ibrahim et al. 2016). Omics data offer complementary patient information in addition to low-dimensional baseline demographic and clinical features. This information helps healthcare professionals gaining a deep understanding of the genetic and molecular mechanisms underlying complex human diseases, enabling improved early disease diagnoses and effective personalized treatment strategies tailored to individual patients or subpopulations (Issa et al. 2014). However, accurately predicting disease outcomes remains highly challenging due to complex disease mechanisms, including nonlinear impacts of molecular biomarkers and clinical features on disease outcomes, as well as interactive effects among these risk factors. Consequently, conventional parametric methods, such as multiple linear or logistic regression, prove ineffective in constructing powerful predictive models and identifying clinically meaningful molecular biomarkers.

Machine learning algorithms, particularly deep neural networks (DNNs, LeCun et al. 2015), support vector machines (SVMs, Cortes and Vapnik 1995), random forests (RFs, Breiman 2001), and gradient boosting machines such as XGBoost (Chen and Guestrin 2016) have demonstrated potential in robustly handling the intricate associations between molecular biomarkers, biological, and clinical features, and disease outcomes (Zhu et al. 2020, Poirion et al. 2021). Despite powerful and flexible, these models are complicated by the high-dimensionality of multi-omics data, commonly referred to as the “curse of dimensionality” (Mirza et al. 2019, Reel et al. 2021). Predictive models with high-dimensional input features are susceptible to overfitting, which often results in high training accuracy but poor testing performance. As a result, directly incorporating omics features alongside conventional clinical variables may not enhance disease outcome predictions, even with robustness techniques like pruning, dropout (Hinton et al. 2012), and bootstrap bagging and scoring (Mi et al. 2019). In addition, interpreting machine learning models trained on high-dimensional multi-omics data presents another significant challenge. Though there are existing methods to identify important biomarkers using machine learning methods, such as the Permutation Feature Importance Test (PermFIT, Mi et al. 2021), Shapley Additive Explanations (SHAP, Lundberg and Lee 2017), and Knock-off Randomized Testing (Candès et al. 2018), all struggle with the curse of dimensionality. They are limited in their ability to differentiate true important features from noise features when the data dimension is high.

To address the dimensionality issue for omics data integration, researchers extensively utilized shrinkage linear or generalized linear models. These include methods such as the least absolute shrinkage and selection operator (Lasso, Tibshirani 1996) and elastic nets (Zou and Hastie 2005). While powerful, these shrinkage algorithms suffer from performance degradation as the number of nuisance input features increases. Consequently, feature pre-screening becomes necessary. One such approach is Sure Independence Screening (SIS) (Fan and Lv 2008), which selects features with strong marginal effects from high-dimensional data before applying more refined analyses like Lasso (Fan and Lv 2008, Fan and Song 2010). Evaluating the marginal association between each input feature and a disease outcome can be done using metrics such as mutual information (MI), Spearman correlation (SPC), maximal information coefficient (MIC, Reshef et al. 2011), or Kendall rank correlation coefficient (Kendall’s tau). For capturing nonlinear associations, researchers may consider methods like the Hilbert-Schmidt independence criterion (HSIC Gretton et al. 2007) and kernelized partial correlation (KPC, Huang et al. 2022). For multi-omics applications, many existing differential expression (DE) or differential abundance (DA) analysis tools (Love et al. 2014, Finak et al. 2015) can be naturally incorporated into the SIS framework.

Despite significant advances, several challenges persist. First, the marginal screening approaches mentioned earlier are constrained by individual criteria for measuring marginal dependency, and their performance is often data-dependent. Parametric methods, constrained by assumptions about data-generating models—such as Gaussian or log-Gaussian—may not be suitable for all multi-omics data. For instance, popular parametric DE methods for genomics data may not generalize well to metatranscriptomics or microbiome data (Cho et al. 2023). While non-parametric methods relax these distributional assumptions, they can be less effective at capturing simple relationships, such as linear correlations, and no single screening method consistently outperforms others in all cases (Zhang 2019). This, along with the challenges of validating marginal dependency methods in real-world applications (Rapaport et al. 2013, Wu et al. 2015, Cho et al. 2023), makes selecting an appropriate dependency criterion for multi-omics datasets difficult for researchers. Second, integrating omics data with low-dimensional clinical features—such as treatment, medication, or disease history—requires an analysis of interactive effects. However, classical methods like SIS, DE, and DA are designed to measure marginal associations rather than conditional effects, limiting their applicability in these contexts. Third, determining cutoffs for significant omics features can be ambiguous, especially for non-parametric methods. Although non-parametric approaches excel at uncovering novel, nonlinear associations between multi-omics and disease phenotypes, the absence of robust statistical testing and feature validation can lead to significant false discoveries.

To address these challenges, we propose an efficient Hybrid Feature Selection (HFS) framework that combines multiple dependency metrics. HFS identifies important biomarkers by assembling metrics, minimizing the risk of missing important features by relying only on one specific dependency measure. To capture interactions between omics and clinical features, HFS can be easily extended to assess partial associations between disease outcomes and omics traits, conditioned on clinical variables. Additionally, we introduce a novel data-driven method that uses the isolation forest algorithm (Liu et al. 2008) to determine the optimal cutoff for dependency statistics, enabling principled identification of important features to boost feature pre-screening performance. While HFS significantly filters out many nuisance biomarkers, it inevitably selects some noise features due to its marginal screening nature. Therefore, a further refinement process is necessary to fine-tune the HFS list and determine the most relevant features for outcome prediction. Furthermore, it is critically important to evaluate the impact of each individual pre-selected biomarker on disease outcomes by adjusting for the confounding effects of other biomarkers under complex association settings. This can not only deeper our understanding of disease mechanisms but also informs better clinical decisions. To achieve this objective, we leverage PermFIT (Mi et al. 2021), a computationally efficient framework to evaluate each individual pre-screened feature’s impact on disease outcome with a rigorous statistical inference under potential complex associations for machine learning models. Combining PermFIT with HFS, we alleviate the curse of dimensionality, allowing more effective detection of complicated impacts such as nonlinear interactions among omics features on disease outcomes. We consolidate the entire process into a comprehensive high-dimensional feature importance test (HiFIT) framework. This framework encompasses feature pre-screening, refinement, and final predictive modeling, achieving robust, scalable, and interpretable disease outcome predictions in high-dimensional settings as delineated in the following section.

Materials and methods

The proposed HiFIT framework comprises two main components: feature pre-screening by HFS and machine learning-based feature importance testing using the PermFIT algorithm (Mi et al. 2021). HFS pre-screens high-dimensional multi-omic features by evaluating their complex marginal association with the outcome, addressing the curse of dimensionality. Following pre-screening, the selected high-dimensional features, together with low-dimensional clinical variables, are incorporated into machine learning models—such as DNN, RF, XGBoost, and SVM—to develop initial predictive models. PermFIT then assesses the impact of each HFS-selected feature on the outcome under intricate associations. This helps to further filter out unimportant features, control the false discoveries, and boost prediction performance.

HFS: hybrid feature screening

Let y=(y1,,yn)T be the clinical outcome of interest across n samples, and X={xij} be an n×p matrix, where xij is the jth feature of sample i, and p is the number of input features. Further, define xi·=(xi1,,xip)T and x·j=(x1j,,xnj)T, and assume that among the p input features, there exists an important feature subset S with |S|<p such that E[yi|xi·]=E[yi|{xij,jS}]. To ensure the selection power of important features, we follow the sparsity assumption of SIS such that the true number of important features |S|p. We also follow SIS to refer the important features as important features while the other features are referred to as noise features. In contrast to SIS, HFS screens important features whose effects on the outcome are linear or more complex than linear by combining (i) parametric utility metrics, such as statistics in generalized linear models commonly used in DE analysis (Finak et al. 2015), and (ii) non-parametric utility metrics like kernel-based correlations. In this article, we set the utility metrics to (i) the adjusted R-squared for polynomial regression and (ii) the kernel partial correlation (KPC, Huang et al. 2022) coefficient for HFS. To determine a cutoff for screened features, HFS leverages the isolation forest algorithm (Liu et al. 2008), which assigns an anomaly score to each feature, with higher scores indicating stronger associations with the outcome.

To compute the parametric utility function, we fit the following linear or generalized linear model independently for each feature (say jth feature),

(1)

with an appropriate link function g corresponding to the outcome type, for example, g(x)=x for continuous outcome, and g(x)=logit(x) for binary outcome, and a pre-specified order M polynomial. The adjusted R-square or McFadden’s pseudo R-square (McFadden 1974), denoted as ρj1, is computed as follows:

where y¯ is the sample mean of the outcome, and y^i is the predicted outcome of sample i; LLj is the maximal log-likelihood of model (1), and LL0 is the maximal log-likelihood of the null model with an intercept only. In our analysis, in line with many parametric DE/DA methods, we search for the linear and quadratic associations by setting M to 2.

For more general marginal non-linear association detection, a natural extension of model (1) involves increasing M to capture associations in higher orders, potentially extending to infinite moments. Alternatively, we adopt the KPC coefficient of each feature within the Reproducing Kernel Hilbert Space (RKHS) framework (Huang et al. 2022) to approximate the infinite extension of model (1), where the n×n sample kernel matrices are defined as:

with κ being a kernel function. The centered kernel matrices are denoted as Kj˜=HKjH and Ky˜=HKyH where H=I1n11T. Here, I denotes an n by n identity matrix, and 1 denotes a 1-vector of size n. The empirical KPC between y and the jth feature is:

(2)

where Oj=Kj˜(Kj˜+nδI)1 with δ being a positive constant. Notably, if a linear kernel κ(a,b)=aTb is used, the KPC coefficient in Equation (2) approximates the adjusted R-square of model (1) when M=1. Higher-order variations of model (1) can be approximated using KPC with the corresponding polynomial kernel. In our analysis, we employ an infinite-dimensional kernel, or the radial basis function kernel κ(a,b)=exp{||ab||22/2} to flexibly detect complex omics-phenotype associations.

Both utility functions have their limitations. First, model (1) with lower orders may fail to capture complex higher-order associations, while KPC is less effective at detecting linear associations. Second, model (1) allows for the use of a likelihood ratio test to derive a p-value, providing a clear criterion for significance. In contrast, the value of the utility function (2) is more difficult to interpret, making it challenging to determine an appropriate cutoff for significance. To address the first challenge, HFS combines the two utility functions of the jth feature into a correlation vector, ρj=(ρj1,ρj2)T to generate a biomarker list highly correlated with the disease outcomes, referred to as the HFS list. To address the second challenge, and recalling the sparsity assumption in SIS, we model the distribution of ρjs through a mixture model. In this model, we assume that the vector corresponding to the noise features forms one cluster with a distribution of g0, while the vector associated with the remaining ones forms another cluster with a distribution of g1. More specifically, we have

where π is the proportion of the noise features. The cumulative distribution function of g0, G0(x), is assumed to be no smaller than G1(x), the cumulative distribution function of g1. The features classified to the second cluster are considered anomaly features which are expected to overlap largely with set S as long as the marginal utilities of the important features are not too small. We utilize the isolation forest (Liu et al. 2008), a bootstrapping ensemble of isolation trees to estimate the anomaly probability of each feature as follows: for B bootstrap replicates, the anomaly score of feature j is given by

where ψ is the bootstrapped sample size regularized by c(ψ), and h(b)(·) represents the path length from the root to the leaf node that the sample belongs to, adjusted by the average path length of the bth isolation tree. For a given cutoff τ, the selected feature set is given by

(3)

To integrate multi-omics features with low-dimensional clinical data, it is important to model their interactive effects. For example, specific omics features may interact with treatments, influencing disease outcomes and enabling personalized medicine. To capture these interactions, we can replace the default utility functions in HFS with their conditional counterparts that measure if certain omics traits are correlated with the disease outcomes given the level of clinical features. This adjustment will allow the HFS framework to identify a greater number of features that interact with clinical variables. Further details on this modification are available in the Supplementary Section S1.

Machine learning based feature importance test

To provide quality checking and formal statistical testing for features in the HFS list, HiFIT employs the PermFIT procedure, a permutation-based feature importance score test originally developed for seemingly black-box machine learning models (Mi et al. 2021). Suppose y=f(x)+ϵ, and f^(·) is the estimated function of f(·) from a machine learning model. While HFS ensures that the biomarkers x are marginally correlated with the disease outcome y—such as genes that are differentially express—the f(·) can model the conditional effect of each biomarker given other elements in x. Thus, HiFIT can refine the HFS lists by accounting for more confounding structures within x. In the remaining part, we outline the PermFIT procedure to derive feature-wise p-values. Notably, we use some notation shortcuts and assume that the input feature vector now contains only the features in the HFS list S^. Consequently, p now refers to the size of S^, which is much smaller than the original number of high-dimensional input features. In this article, we focus on four popular machine learning algorithms, although the framework is not limited to them: (i) SVM—implemented via R package “e1071” using Radial kernels, (ii) RF—implemented via R package “randomForest”, (iii) XGBoost—implemented via R package “xgboost”, and (iv) ensemble DNN—implemented via R package “deepTL” (Mi et al. 2019).

For the jth feature in S^, we define its importance score Λj as Λj=Exi,·,xi(j)[{f(xi.)f^(xi(j))}2{f(xi.)f^(xi·)}2], where xi(j)=(xi,1,,xi,jpermuted,,xi,p)T with xi,jpermuted being the ith element of the permuted vector of x·j. The importance score Λj equals zero when the contribution of the jth feature to y is null, and that, the stronger the impact of the jth feature on the outcome, the larger Λj is. PermFIT then empirically estimates Λj by Λ^j=1ni=1n[{yif^(xi(j))}2{yif^(xi·)}2]. To avoid potential overfitting of f for data with finite samples, PermFIT employs a data splitting strategy in which it divides the data into training and validation sets. It uses the training set for generating f^(·) and the validation set for evaluating the distribution of Λj. That is, let f^T(·) denote the estimate of f(·) from a training set, and ΩV={yi,xi·}i=1nV be the validation set, we obtain the feature importance score estimate Λ^j, and its associated variance as Λ^j=1nVi=1nVΛ^ij and Var^[Λ^j]=1nVi=1nV[Λ^ijΛ^j]2, respectively, with Λ^ij=1nVi=1nV[{yif^T(xi(j))}2{yif^T(xi·)}2]. The statistics for feature importance test of Xj can be constructed as λj=Λ^jVar^[Λ^j]N(0,1) under the null. With the proposed test, S^ is further refined to a final feature list, denoted as S^final and used by HiFIT to build the final predictive model. The importance score test for the binary outcome can be constructed in a similar manner, and the details can be found in Mi et al. (2021).

Adaptive selection of the HFS cutoff parameter

In this section, we propose a data-driven approach that heuristically and computationally efficiently searches for an optimal τ from a set of candidates (ordered from the smallest to the largest) (τ0,τ1,,τR), with {S^1,,S^R} being the corresponding selected feature sets. These sets consist of features with their anomaly scores falling in the intervals of the candidate cutoffs. That is,

For the rth feature sets, we estimate its set importance score ΛSr=ExS,x(Sr)[f(xS)f(x(Sr))]2, analogous to the way that PermFIT defines feature importance score, where xS denotes the features with the HFS anomaly score larger than τ0, and x(Sr) is a rearranged xS with features in Sr replaced by random permutations. For computational efficiency, instead of testing the importance of each feature, we estimate the set importance scores here. The smallest τ among all candidates, for which the feature sets have a P-value smaller than 0.1 (p(Sr)<0.1), will be selected by HiFIT as the final cutoff. In this article, we set the candidate list as (0.5,0.55,0.6,0.65,0.7) for both simulation studies and real data analysis.

Results

To evaluate the performance of HiFIT in eliminating nuisance features and predicting outcomes, we conducted comprehensive simulation studies under various data complexity scenarios with varying numbers of features and data generation schemes. For comparison, we included the gold standard approach, which uses only important features as input, alongside Lasso and other machine learning algorithms. Additionally, HiFIT was applied to two real-world datasets: the weight loss data after bariatric surgery (BS) (Heinberg et al. 2020, Fouladi et al. 2021), and RNA sequence data from three kidney cancer studies in The Cancer Genome Atlas (TCGA, Cancer Genome Atlas Research Network et al. 2013). Given the unknown biological ground truth of the associations between clinical and omics features and disease phenotypes, we evaluated the performance of each comparison method by focusing on the prediction accuracy of the final HiFIT model in real data applications.

Simulation studies

We examine the performance of the proposed methods under the following simulation scenarios: (a) Linear case: where the effects of the important features are linear and additive; (b) Non-linear case: where the effects of the important features are non-linear and non-additive. Specifically, we simulate y as follows:

for cases (a) and (b) respectively, where x=(x1,,xp)T is a p-dimensional random variable drawn from a multivariate normal distribution, multinomial distribution, or Poisson log-normal distribution with block correlation, the error term ϵ follows a standard normal distribution, that is, N(0,1), and the vector β=(β1,,β10) is drawn from a uniform distribution U(1,1.5) at the beginning of the simulation and remains unchanged throughout subsequent Monte Carlo simulations. For both cases, we set the total number of important features to 10, with the remaining p10 features being nuisance variables. The total number of samples is set to 500. To evaluate the impact of input feature dimensions, we vary p across the values {500, 1000, 10 000}. We split the simulated data into training and testing sets in a 9:1 ratio. For each scenario, the total number of simulation is set to 100. Since the performance of HFS and HiFIT is similar across different distributions of x, we primarily present the results with x drawn from multivariate normal distributions and leave remaining results in the Supplementary Figs S1–S6.

We begin by comparing the performance of HFS with other feature pre-screening methods, including Lasso, Pearson Correlation (PC), Spearman Correlation (SPC), MIC, and HSIC. Figure 1 illustrates the ranks of correlation scores for important features relative to all features. Since Lasso does not directly produce a correlation score, we rank the penalty parameter associated with the important features. Specifically, we consider the largest penalty parameter at which the coefficient of a feature becomes non-zero. Under linear settings, when p=500 or 1000, HFS performs comparably to parametric models like Lasso and PC in ranking important features. As the dimensionality p increases to 10 000, HFS still outperforms non-parametric methods such as HSIC and MIC. Notably, HSIC is significantly influenced by noisy variables and ranks half of the important features with smaller linear effects much lower than HFS. The advantage of HFS becomes even more prominent in non-linear scenarios. Parametric methods like Lasso and PC become ineffective and fail to identify important features, as shown in the second panel of Fig. 1. On the other hand, while SPC and MIC relax the parametric assumptions of Lasso and PC, they only rank half of the important features at the top. Specifically, SPC struggles to detect monotonic associations, while MIC tends to capture spurious correlations between nuisance features and outcomes. Furthermore, their performance deteriorates as the dimensionality increases to 10,000. In contrast, HFS not only identifies the largest number of important features but also remains robust in high-dimensional data scenarios. In summary, HFS effectively combines the advantages of both parametric and non-parametric methods for feature screening by leveraging different marginal association dependencies. Other methods are either constrained by their parametric assumptions or limited in their ability to handle high-dimensional data.

Average rank of important features selected by pre-screening methods. The x-axis denotes the number of selected important features, and the corresponding value of the y-axis represents the average rank of this feature over 100 repetitions. The curves are generated by locally estimated scatterplot smoothing.
Figure 1.

Average rank of important features selected by pre-screening methods. The x-axis denotes the number of selected important features, and the corresponding value of the y-axis represents the average rank of this feature over 100 repetitions. The curves are generated by locally estimated scatterplot smoothing.

In practice, the cutoff for the HFS score is determined using a data-driven approach (see section Materials and methods). Features with HFS scores higher than the cutoff are retained for downstream analysis. Since many comparison methods do not provide a data-driven cutoff, we select the same number of features as HFS across all comparison methods to ensure a fair comparison. We evaluate the quality of the selected feature list based on recall and precision. Recall is defined as the ratio of selected important features to the total number of important features, while precision is defined as the ratio of selected important features to the total number of selected features. Figure 2a illustrates the quality of feature sets. HSIC and MIC tend to overlook important features with smaller effects under linear settings. Meanwhile, Lasso, PC, and SPC struggle to detect log-quadratic or interaction terms. In comparison, HFS consistently identifies more important features across all simulation scenarios. However, as the dimensionality increases, the precision of all pre-screening methods decreases. To address this issue, HiFIT refines the pre-screening feature set obtained from HFS using machine learning algorithms. Figure 2b compares (i) HFS feature sets with the cutoff parameter determined by XGB, RF, SVM, and DNN (denoted as S-XGB, S-RF, S-SVM, and S-DNN); (ii) HiFIT feature sets obtained by applying PermFIT to the corresponding HFS feature lists, retaining features with p-values smaller than 0.1 (denoted as HF-XGB, HF-RF, HF-SVM, and HF-DNN).

High-dimensional feature pre-screening and selection results. (a) Performance of feature pre-screening methods. (b) Feature selection results of HiFIT models. Recall and precision are averaged over 100 simulations.
Figure 2.

High-dimensional feature pre-screening and selection results. (a) Performance of feature pre-screening methods. (b) Feature selection results of HiFIT models. Recall and precision are averaged over 100 simulations.

HiFIT feature sets retain most of the important features identified by HFS. Specifically, the recall of HF-XGB, HF-RF, and HF-DNN match those of the corresponding HFS feature sets, and the recall of HF-SVM is comparable to S-SVM. Furthermore, HiFIT improves the precision of HFS. For instance, when p=500 or 1000, the precision of HF-XGB, HF-SVM, and HF-DNN is controlled at 0.9. As p increases to 10 000, HF models improve the precision of the corresponding HFS sets from 0.5 to more than 0.8, and further reductions are possible with p-value adjustments.

Next, we compare the importance scores and p-values from HiFIT with those from PermFIT. Figure 3 provides detailed feature importance scores for the following scenarios: (i) HiFIT models: HF-DNN, HF-SVM, HF-XGB, and HF-RF; and (ii) PermFIT models: PermFIT-DNN, PermFIT-SVM, PermFIT-SVM, and PermFIT-XGB (applied to the same models without pre-screening). We observe that all HiFIT models successfully identify important features, consistently estimating high importance scores for these features across various simulation scenarios. HiFIT assigns low importance scores to nuisance features, demonstrating its ability to control type-I errors reasonably in high-dimensional settings, regardless of association complexity. In contrast, PermFIT models struggle with nonlinear effects. PermFIT-DNN and PermFIT-SVM identify only one nonlinear important feature out of ten on average. While PermFIT-RF and PermFIT-XGB successfully identify all nonlinear features, their importance scores for important features are substantially lower than those from HF-XGB and HF-RF, highlighting HiFIT’s superior efficiency. For instance, the importance score of the interaction term X10 from PermFIT-RF is lower than that from HF-RF, indicating that PermFIT-RF is more likely to overlook this feature. As data dimensionality increases, PermFIT’s computation cost becomes overwhelming when p=10000, whereas HiFIT remains scalable to high-dimensional data.

HiFIT feature interpretation results. Average feature importance scores for 10 important features (denoted as X1,…,X10) and the feature set of nuisance features (denoted as null) over 100 repetitions. Importance scores of features not selected by HFS are set to zero.
Figure 3.

HiFIT feature interpretation results. Average feature importance scores for 10 important features (denoted as X1,,X10) and the feature set of nuisance features (denoted as null) over 100 repetitions. Importance scores of features not selected by HFS are set to zero.

Finally, we assess the impact of HFS and HiFIT on prediction accuracy for various machine learning models. Figure 4 summarizes the predicted average mean squared error (MSE) and Pearson correlation coefficient (PCC) for the following scenarios: (i) Taking the full features as input: Denote the corresponding models as Lasso, DNN, XGB, RF, and SVM; (ii) Taking the HFS pre-screened features as input: Denote the corresponding models as S-Lasso, S-DNN, S-XGB, S-RF, and S-SVM; (iii) Taking the HiFIT refined features (with P.1) as input: Denote the corresponding models as HF-DNN, HF-XGB, HF-RF, and HF-SVM. Under linear settings, and for scenario i), all machine learning models perform worse than Lasso, and their prediction errors increase with the data dimensionality. When P=10000, RF, DNN, and SVM all fail to converge. After HFS, S-SVM and S-DNN achieve performance comparable to Lasso across dimensions, with similar PCC and MSE. S-RF and S-XGBoost perform slightly worse than Lasso but close to their optimal predictions (using only true important features, see Supplementary Section S2.1). Similarly, under the nonlinear cases, without pre-screening, all four machine learning methods struggle to make reliable predictions (see Fig. 4). Trained with the only 10 important features, SVM and DNN outperform XGBoost and RF, highlighting their better ability in capturing complex feature-outcome relationships when a set of right features are provided. However, in high-dimensional settings with many nuisance features, their performance becomes inferior to RF and XGBoost, and they even become computationally infeasible for P=10000. In contrast, HFS significantly improves performance of these models, especially for SVM and DNN. S-DNN achieves the lowest prediction error and highest PCC across dimensions, surpassing other machine learning algorithms. Even at P=10000, S-DNN maintains a high PCC of 0.77. Overall, HFS enhances machine learning models by reducing MSE (by 30%) and increasing PCC (by 20%) for P=500 and 1000. The models remain stable as feature dimension increases, with PCC consistently above 0.7.

Average MSE and PCC for methods in comparison. Lasso, XGB, RF, SVM, and DNN: specific models with all features; S-Lasso, S-XGB, S-RF, S-SVM, S-DNN: specific models with HFS pre-screening; HF-XGB, HF-RF, HF-SVM, HF-DNN: specific models with HiFIT feature selection. Simulation in each scenario is repeated 100 times.
Figure 4.

Average MSE and PCC for methods in comparison. Lasso, XGB, RF, SVM, and DNN: specific models with all features; S-Lasso, S-XGB, S-RF, S-SVM, S-DNN: specific models with HFS pre-screening; HF-XGB, HF-RF, HF-SVM, HF-DNN: specific models with HiFIT feature selection. Simulation in each scenario is repeated 100 times.

In addition to the simulation above, we conducted numerical studies to evaluate the performance of HiFIT for scenarios with non-Gaussian predictors, mixture of linear and nonlinear effects, and binary outcomes (Supplementary Sections S2.2–S2.4). Similar conclusions were observed, and detailed results are presented in the supplementary materials. Beyond the simulation studies, we applied the proposed HiFIT framework to two practical studies, i.e., post BS weight loss and kidney pan-cancer (KIPAN) study, as described below. Similar to the simulation, we randomly split the each real dataset into training and testing sets in a 9:1 ratio and replicate 100 times.

Weight loss after bariatric surgery study

In the first real application, we applied the HiFIT to a weight loss study for BS. One of the primary objectives of this study was to use the baseline microbiome profiles along with other demographic and clinical features to predict postoperative weight loss and identify associated important features. The weight loss microbiome cohort consists of 144 participants undergoing BS with 50% of them having Roux-en-Y Gastric Bypass (RYGB) and the other 50% having Sleeve Gastrectomy (SG) (Heinberg et al. 2020, Fouladi et al. 2021). The body mass index (BMI) and fecal material were collected from individuals at 1, 6, 12, 18, and 24 months post-surgery. The microbial profiles of the BS study were characterized through shotgun Whole Genome Sequencing across multiple time points of 135 subjects (n=135), resulting in total 430 measurements of BMI change with 1533 microbial genera. There are also four demographic features of participants, for example, age, race, height, and sex. In this study, we aim to predict patients’ BMI change from the day of surgery to their last recorded measurement, using their surgery type, time since surgery, demographic features, and the gut microbiome abundance collected from their first visit. Since the effect of the surgery type and demographic features on weight change is of particular interest (Heinberg et al. 2020), we retain all these features and only perform pre-screening on the microbiome features with HFS. Figure 5 presents the model performance, demonstrating that HFS substantially enhances the prediction accuracy of Lasso, SVM, RF, and DNN. HiFIT further improves the performance of RF and DNN. Specifically, HF-DNN achieves the highest prediction accuracy among all methods, with the smallest predicted MSE and the largest PCC.

Model performance for the real data analysis. The first column presents the MSE and PCC for the weight loss microbiome study. The second column presents the accuracy and AUC for TCGA kidney pan cancer cohort. All metrics are averaged on separate testing sets consisting of 10% observations with random repeats for 100 times.
Figure 5.

Model performance for the real data analysis. The first column presents the MSE and PCC for the weight loss microbiome study. The second column presents the accuracy and AUC for TCGA kidney pan cancer cohort. All metrics are averaged on separate testing sets consisting of 10% observations with random repeats for 100 times.

Figure 6a presents the feature-wise P-values from the four machine learning models in HiFIT. Some demographic features, including time, age, and race, have significant effects on weight loss after BS. The finding on age is consistent with clinical findings (Contreras et al. 2013). HiFIT provides further insights into microbial effects on weight loss. As shown in Fig. 6a, all four machine learning models identify important microbiota related to weight loss. This implies that gut microbial abundance offers a distinct source of information on weight loss beyond patients’ demographic features, aligning with previous studies (Zhang et al. 2009, Ferrocino et al. 2015). Specifically, all HiFIT models identify the beneficial microbe Hyphobacterium (Ferrocino et al. 2015) as an important predictor for post-BS weight loss, and three HiFIT models highlight Panacibacter as an important genus. Figure 7 further illustrates the nonlinear relationship between BMI reduction and microbial abundance. Unlike the approximately linear effect of age on weight loss, most microbial effects exhibit quadratic or even non-polynomial patterns, underscoring the need for nonlinear feature screening and flexible modeling using machine learning methods included in HiFIT. Moreover, the effects of the two aforementioned genera, Hyphobacterium and Panacibacter, resemble the patterns observed in several other significant microbial genera, including Roseiflexus, Tsuneonella, and Cyclobacterium. Both low and high abundances of these genera imply decreased diversity and richness of gut bacteria, which can be further associated with less weight loss outcomes (Zhang et al. 2009). While higher abundances of most identified genera tend to correlate with less significant weight loss after BS, increased levels of Schaalia, Saccharomonospora, and Shimwellia are associated with more weight loss. Lautropia and Aureimonas exhibit more complex effects that require further investigation. Although the BS surgery type is not identified as a significant biomarker contributing to the post-surgery weight reduction, HFS and HiFIT with conditional utility functions determined that RYGB and SG interact with patient age and microbial levels. More details are available in Supplementary Section S3.1.

Negative  log 10  P-values for biomarkers from real datasets. (a) Feature importance for the weight loss data. (b) Feature importance for the TCGA data.
Figure 6.

Negative log10  P-values for biomarkers from real datasets. (a) Feature importance for the weight loss data. (b) Feature importance for the TCGA data.

Association between BMI reduction and microbiome abundance. The x-axis represents the logarithmic abundance of microbial genera identified by HiFIT in Fig. 6a. The y-axis represents the reduction in BMI following bariatric surgery. We also include the association between age and BMI reduction as a reference. The smoothing curves and 95% confidence bands are obtained by locally polynomial regression. The color of the curves represents the number of HiFIT models in Fig. 6a identifying the feature as significant.
Figure 7.

Association between BMI reduction and microbiome abundance. The x-axis represents the logarithmic abundance of microbial genera identified by HiFIT in Fig. 6a. The y-axis represents the reduction in BMI following bariatric surgery. We also include the association between age and BMI reduction as a reference. The smoothing curves and 95% confidence bands are obtained by locally polynomial regression. The color of the curves represents the number of HiFIT models in Fig. 6a identifying the feature as significant.

TCGA kidney cancer data from the KIPAN cohort

We further applied the HiFIT to analyze binary outcomes using TCGA data. Though TCGA has a large collection of publicly available clinical and omics data (Cancer Genome Atlas Research Network et al. 2013), we focus on the KIPAN (n=941) in our analysis. We aim to predict the patients’ survival status using normalized counts of RNA sequence data from Illumina HiSeq platform at gene level. For simplicity, we categorize patients into two groups: long-term survival ( 5 years of survival) and short-term survival (<5 years of survival). Out of the total 941 samples in this cohort, 193 participants achieved long-term survival, while 242 participants achieved short-term survival. The remaining samples were lost to follow-up and removed from the analysis. Additionally, expression profiles from 20 189 genes and five clinical features—age, tumor stage, histological type (cancer type), tumor purity, gender, and race—are available for these patients. Due to the large number of features and the limited sample size in the cohort, RF, SVM, and DNN fail to converge when using the full set of features, making reliable predictions impossible. Instead, we implement these algorithms using the top 1000 genes with largest variance as training input. Figure 5 presents the comparison of prediction accuracy and area under the curve (AUC) using HFS/HiFIT selected genes. HFS improves the performance of all four machine learning models and Lasso. Specifically, S-RF yields the highest prediction accuracy and AUC among all methods. Although HiFIT does not further improve the performance of machine learning algorithms, the prediction accuracy and AUC of HF-RF, HF-SVM and HF-DNN are still comparable to S-RF, S-SVM, and S-DNN, outperforming models using all top 1000 over-dispersed genes as input. Despite the minor improvement in prediction accuracy, HiFIT offers feature importance evaluation with solid statistical inference, offering in-depth understanding of disease mechanisms.

The feature-wise P-value for kidney cancer-associated features is summarized in Fig. 6b. We next summarize important features that are identified by at least three models from HF-DNN, HF-RF, HF-SVM, and HF-XGB. First, age and tumor stage are two important demographic features. Second, we uncover the following important genes: KIF18A, PABPC1, and SPIC, which are consistent with previous cancer studies. Specifically, KIF18A is required for chromosomally unstable tumor cells for proliferation (Marquis et al. 2021) and exhibits association with pan cancer survival across multiple cohorts (Liu et al. 2023). PABPC1 was shown to promote cell proliferation and metastasis in pancreatic cancer (Yao et al. 2023). The upregulation of SNHG4 is associated with lymph node involvement, distant metastasis, and reduced overall survival for renal cell carcinoma patients (Wu et al. 2020). Lastly, SPIC has been observed to drive cancer progression in mice (Ranzani et al. 2013). More details of the features can be found in Supplementary Section S3.2.

Discussion

High-dimensional omics profiles, when combined with low-dimensional biological and clinical features, play a significant role in influencing the onset and severity of many complex human diseases. To gain fundamental insights into disease mechanisms and improve early diagnosis and precision medicine, it is essential to integrate critical molecular biomarkers with clinical risk factors. However, the high-dimensional nature of omics data, along with the intricate associations between disease outcomes and these profiles, poses substantial analytical challenges. To address these challenges, HiFIT first screens for important features using HFS, followed by refining this feature list through machine learning models.

Our research demonstrates the superior performance of HiFIT through extensive numerical simulations and two real-world data applications. In simulation studies, by combining metrics, HFS exhibits a “minimax” property, showing superior power in detecting both linear and nonlinear associations. In contrast, using a single nonlinear dependency measure may overlook biomarkers that exhibit simple correlations with disease outcomes. Although HiFIT was not specifically designed for DE or DA analysis, its success in boosting the recall and precision for nonlinear biomarker detection suggests it could have promising future applications in improving DE/DA analysis. Our real data applications include analysis of multi-omics—microbiome and genome—and their associations with various disease phenotypes. Although the biological ground truth of important genera or gene lists is unknown, the observed boost in predictive accuracy suggests that HFS and HiFIT selected highly predictive biomarkers in both studies. Remarkably, real data results highlight the practical challenges of verifying model assumptions for high-dimensional data. In the weight loss study, both simple linear and complex nonlinear effects are present (Fig. 7), while in the TCGA cohort, associations with the binary outcomes are difficult to verify through visualization. Applying parametric models for such data can easily result in model mis-specifications.

A potential direction for future research is to relax the marginal dependency assumptions when handling high-dimensional omics data. Although HFS and HiFIT demonstrated strong performance in our simulation models where feature effects are mostly additive, the marginal screening approach inherent in HFS may result in selecting highly correlated features, potentially overlooking more complex yet biologically relevant biomarkers.

In summary, the proposed HiFIT framework bridges the gap between high-dimensional omics profiles and low-dimensional biological and clinical features. By effectively handling high-dimensional data and capturing complex associations between molecular biomarkers, biological and clinical features, and disease outcomes, HiFIT facilitates robust feature importance identification. This leads to more accurate outcome predictions in high-dimensional settings, enabling scalable, interpretable, and robust disease outcome predictions. As such, HiFIT provides a valuable framework for improved disease management and personalized treatment strategies.

Acknowledgments

We thank the editor, associate editor, and three anonymous reviewers for their insightful comments and constructive suggestions, which have greatly improved the article.

Supplementary data

Supplementary data are available at Bioinformatics online.

Conflict of interest: None declared.

Funding

This work was partially supported by National Institutes of Health R56 (1R56LM013784) and R01 (R01LM014407 and 1R01HL173044) grants.

Data availability

The metagenomic sequences used in this study can be found at the National Center for Biotechnology information Sequence Read Archive (https://www.ncbi.nlm.nih.gov/sra) under BiopProject PRJNA668357 and PRJNA668472. The TCGA dataset is publicly available at the LinkedOmics website (http://linkedomics.org), where the KIPAN study (i.e. KIRC, KICH, and KIRP studies) is used in our analysis.

Code availability

HiFIT is implemented in our R package “HiFIT” which is available along with source code for replicating the simulation studies and real data applications on GitHub (https://github.com/BZou-lab/HiFIT).

References

Breiman
L.
 
Random forests
.
Mach Learn
 
2001
;
45
:
5
32
.

Cancer Genome Atlas Research Network
 et al.  
The cancer genome atlas pan-cancer analysis project
.
Nat Genet
 
2013
;
45
:
1113
20
.

Candès
E
,
Fan
Y
,
Janson
L
 et al.  
Panning for gold: ‘Model-X’ knockoffs for high dimensional controlled variable selection
.
J R Stat Soc Ser B Stat Methodol
 
2018
;
80
:
551
77
.

Chen
T
,
Guestrin
C.
Xgboost: a scalable tree boosting system. In: Proceedings of the 22nd ACM sigkdd International Conference on Knowledge Discovery and Data Mining, New York, NY, USA: Association for Computing Machinery, p.
785
794
,
2016
.

Cho
H
,
Qu
Y
,
Liu
C
 et al.  
Comprehensive evaluation of methods for differential expression analysis of metatranscriptomics data
.
Briefings in Bioinformatics
 
2023
;
24
:
bbad279
.

Contreras
JE
,
Santander
C
,
Court
I
 et al.  
Correlation between age and weight loss after bariatric surgery
.
Obes Surg
 
2013
;
23
:
1286
9
.

Cortes
C
,
Vapnik
V.
 
Support-vector networks
.
Mach Learn
 
1995
;
20
:
273
97
.

Fan
J
,
Lv
J.
 
Sure independence screening for ultrahigh dimensional feature space
.
J R Stat Soc Ser B Stat Method
 
2008
;
70
:
849
911
.

Fan
J
,
Song
R.
 
Sure independence screening in generalized linear models with NP-dimensionality
.
The Annals of Statistics
 
2010
;
38
:
3567
604
.

Ferrocino
I
,
Di Cagno
R
,
De Angelis
M
 et al.  
Human gut microbiome before and after bariatric surgery in obese patients with and without type 2 diabetes
.
Obesity Surgery
 
2015
;
25
:
677
87
.

Finak
G
,
McDavid
A
,
Yajima
M
 et al.  
Mast: a flexible statistical framework for assessing transcriptional changes and characterizing heterogeneity in single-cell RNA sequencing data
.
Genome Biol
 
2015
;
16
:
278
13
.

Fouladi
F
,
Carroll
IM
,
Sharpton
TJ
 et al.  
A microbial signature following bariatric surgery is robustly consistent across multiple cohorts
.
Gut Microbes
 
2021
;
13
:
1930872
.

Gretton
A
,
Fukumizu
K
,
Teo
C
 et al.  
A kernel statistical test of independence
.
Adv Neural Inf Process Syst
 
2007
;
20
:
585
92
.

Heinberg
LJ
,
Bond
DS
,
Carroll
I
 et al.  
Identifying mechanisms that predict weight trajectory after bariatric surgery: rationale and design of the biobehavioral trial
.
Surg Obes Relat Dis
 
2020
;
16
:
1816
26
.

Hinton
GE
,
Srivastava
N
,
Krizhevsky
A
 et al. Improving neural networks by preventing co-adaptation of feature detectors.
2012
.

Huang
Z
,
Deb
N
,
Sen
B.
 
Kernel partial correlation coefficient—a measure of conditional dependence
.
Journal of Machine Learning Research
 
2022
;
23
:
1
58
.

Ibrahim
R
,
Pasic
M
,
Yousef
GM.
 
Omics for personalized medicine: defining the current we swim in
.
Expert Rev Mol Diagn
 
2016
;
16
:
719
22
.

Issa
NT
,
Byers
SW
,
Dakshanamurthy
S.
 
Big data: the next frontier for innovation in therapeutics and healthcare
.
Expert Rev Clin Pharmacol
 
2014
;
7
:
293
8
.

LeCun
Y
,
Bengio
Y
,
Hinton
G.
 
Deep learning
.
Nature
 
2015
;
521
:
436
44
.

Liu
FT
,
Ting
KM
,
Zhou
Z-H.
Isolation forest. In: 2008 Eighth IEEE International Conference on Data Mining, p.
413
422
,
2008
.

Liu
T
,
Yang
K
,
Chen
J
 et al.  
Comprehensive pan-cancer analysis of kif18a as a marker for prognosis and immunity
.
Biomolecules
 
2023
;
13
:
326
.

Love
MI
,
Huber
W
,
Anders
S.
 
Moderated estimation of fold change and dispersion for rna-seq data with deseq2
.
Genome Biol
 
2014
;
15
:
550
21
.

Lundberg
SM
,
Lee
S-I.
 
A unified approach to interpreting model predictions
.
Adv Neural Inf Process Syst
 
2017
;
30
:
4768
77
.

Marquis
C
,
Fonseca
CL
,
Queen
KA
 et al.  
Chromosomally unstable tumor cells specifically require kif18a for proliferation
.
Nat Commun
 
2021
;
12
:
1213
.

McFadden
D.
 
The measurement of urban travel demand
.
J Pub Econ
 
1974
;
3
:
303
28
.

Meyer
UA
,
Zanger
UM
,
Schwab
M.
 
Omics and drug response
.
Annu Rev Pharmacol Toxicol
 
2013
;
53
:
475
502
.

Mi
X
,
Zou
B
,
Zou
F
 et al.  
Permutation-based identification of important biomarkers for complex diseases via machine learning models
.
Nat Commun
 
2021
;
12
:
3008
.

Mi
X
,
Zou
F
,
Zhu
R.
 
Bagging and deep learning in optimal individualized treatment rules
.
Biometrics
 
2019
;
75
:
674
84
.

Mirza
B
,
Wang
W
,
Wang
J
 et al.  
Machine learning and integrative analysis of biomedical big data
.
Genes (Basel)
 
2019
;
10
:
87
.

Poirion
OB
,
Jing
Z
,
Chaudhary
K
 et al.  
Deepprog: an ensemble of deep-learning and machine-learning models for prognosis prediction using multi-omics data
.
Genome Med
 
2021
;
13
:
112
.

Ranzani
M
,
Annunziato
S
,
Adams
DJ
 et al.  
Cancer gene discovery: exploiting insertional mutagenesis
.
Mol Cancer Res
 
2013
;
11
:
1141
58
.

Rapaport
F
,
Khanin
R
,
Liang
Y
 et al.  
Comprehensive evaluation of differential gene expression analysis methods for rna-seq data
.
Genome Biol
 
2013
;
14
:
R95
13
.

Reel
PS
,
Reel
S
,
Pearson
E
 et al.  
Using machine learning approaches for multi-omics data analysis: a review
.
Biotechnol Adv
 
2021
;
49
:
107739
.

Reshef
DN
,
Reshef
YA
,
Finucane
HK
 et al.  
Detecting novel associations in large data sets
.
Science
 
2011
;
334
:
1518
24
.

Tibshirani
R.
 
Regression shrinkage and selection via the lasso
.
J R Stat Soc Ser B Stat Methodol
 
1996
;
58
:
267
88
.

Wu
H
,
Wang
C
,
Wu
Z.
 
Proper: comprehensive power evaluation for differential expression using rna-seq
.
Bioinformatics
 
2015
;
31
:
233
41
.

Wu
J
,
Liu
T
,
Sun
L
 et al.  
Long noncoding RNA snhg4 promotes renal cell carcinoma tumorigenesis and invasion by acting as cerna to sponge mir-204-5p and upregulate runx2
.
Cancer Cell Int
 
2020
;
20
:
514
.

Yao
W
,
Yao
Y
,
He
W
 et al.  
Pabpc1 promotes cell proliferation and metastasis in pancreatic adenocarcinoma by regulating col12a1 expression
.
Immun Inflamm Dis
 
2023
;
11
:
e919
.

Zhang
H
,
DiBaise
JK
,
Zuccolo
A
 et al.  
The importance of the microbiome in bariatric surgery: a systematic review
.
Obes Surg
 
2009
;
19
:
2365
70
.

Zhang
K.
 
Bet on independence
.
J Am Stat Assoc
 
2019
;
114
:
1620
37
.

Zhu
W
,
Xie
L
,
Han
J
 et al.  
The application of deep learning in cancer prognosis prediction
.
Cancers (Basel)
 
2020
;
12
:
603
.

Zou
H
,
Hastie
T.
 
Regularization and variable selection via the elastic net
.
J R Stat Soc Ser B Stat Method
 
2005
;
67
:
301
20
.

This is an Open Access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted reuse, distribution, and reproduction in any medium, provided the original work is properly cited.
Associate Editor: Inanc Birol
Inanc Birol
Associate Editor
Search for other works by this author on:

Supplementary data