 Methodology
 Open Access
 Published:
Sparse generalized linear model with L _{0} approximation for feature selection and prediction with big omics data
BioData Mining volume 10, Article number: 39 (2017)
Abstract
Background
Feature selection and prediction are the most important tasks for big data mining. The common strategies for feature selection in big data mining are L _{1}, SCAD and MC+. However, none of the existing algorithms optimizes L _{0}, which penalizes the number of nonzero features directly.
Results
In this paper, we develop a novel sparse generalized linear model (GLM) with L _{0} approximation for feature selection and prediction with big omics data. The proposed approach approximate the L _{0} optimization directly. Even though the original L _{0} problem is nonconvex, the problem is approximated by sequential convex optimizations with the proposed algorithm. The proposed method is easy to implement with only several lines of code. Novel adaptive ridge algorithms (L _{0}ADRIDGE) for L _{0} penalized GLM with ultra high dimensional big data are developed. The proposed approach outperforms the other cutting edge regularization methods including SCAD and MC+ in simulations. When it is applied to integrated analysis of mRNA, microRNA, and methylation data from TCGA ovarian cancer, multilevel gene signatures associated with suboptimal debulking are identified simultaneously. The biological significance and potential clinical importance of those genes are further explored.
Conclusions
The developed Software L _{0}ADRIDGE in MATLAB is available at https://github.com/liuzqx/L0adridge.
Background
Integrating multilevel molecular and clinical data to design preventive, diagnostic, and therapeutic solutions that are individually tailored to each patient’s requirements is the ultimate goal of precision medicine. However, the huge number of features makes it neither practical nor feasible to predict clinical outcomes with all omics features directly. Thus, selecting a small subset of informative features (biomarkers) to conduct association studies and clinical predictions has become an important step toward effective big data mining. Statistical tests or univariate correlation analysis for feature selection ignore the interacting relationship among genes. To evaluate the predictive power of the features, one appealing approach for feature selection is L _{0} regularized sparse modeling, which penalizes the number of nonzero features directly. L _{0} is known as the most essential sparsity measure and has nice theoretical properties. However, it is computational impossible to perform an exhaustive search when analyzing omics data sets with millions of features. L _{0} penalized optimization is known to be NPhard in general (Lin et al. 2010).
One common strategy for feature selection is to replace the nonconvex L _{0} with the L _{1} norm. L _{1} is a convex relaxation and loose approximation of L _{0}. Although L _{1} penalized sparse models [1] can be solved efficiently, the estimators with L _{1} are penalized too much and asymptotically biased. In addition, L _{1} inclines to select more spurious features than necessary, and may not always choose the true model consistently [2]. Theoretically, L _{1} never outperforms L _{0} by a constant [3]. Depending on the location of true optimum, L _{1} may perform much worse than L _{0} [4, 5]. As a result, the convex relaxation techniques have been shown to be suboptimal in many cases [6]. More recent approaches aimed to reduce bias and overcome discontinuity include the nonconvex SCAD [7] and MC+ [8]. However, none of the existing algorithms directly approximate the L _{0} optimization problem. Either SCAD or MC+ has been rarely used for feature selection in big data analytics because of their computational intensity with multiple tuning parameters. On the other hand, recent research works including ours show that sparse regression models with L _{0} penalty (local solution) outperforms L _{1} (global solution) by a substantial margin [5, 9–11].
Debulking cytoreductive surgery is a standard treatment for ovarian cancer. The goal of debulking is to remove as much visible cancer as possible. However, if tumor nodules have invaded vital organs, surgeons may not be able to remove them without compromising the patient’s life. Leaving tumor nodules larger than 1 cm is defined as suboptimal debulking (cytoreduction). It has been shown that suboptimal debulking is associated with reduced chemosensitivity and poor survival in ovarian cancer. Biomarkers derived from multiomics data may help physicians decide which patients should undergo surgery and which should be treated with chemotherapy first [12–14]. Identifying biomarkers from multiomics data has been an exciting but challenging task. Sparse modeling is one of the important approaches for simultaneous phenotype prediction and biomarker identification. In this paper, we propose a L _{0} penalized generalized linear regression (GLM) for feature selection and prediction. Adaptive ridge algorithm (L _{0}ADRIDGE) is developed to approximate L _{0} penalized GLM with sequential convex optimization and is efficient in handling ultra highdimensional omics data. The proposed method outperforms other cuttingedge convex and nonconvex penalties including L _{1}, SCAD and MC+ with simulations. When applied to the important suboptimal debulking prediction problem in ovarian cancer, the proposed approach identifies multilevel molecular signatures through mining methylation, microRNA and mRNA expression data jointly from TCGA. The identified molecular signatures are further evaluated using public databases.
Materials and methods
Given an input X _{ N×P }, where N≪P, and output Y, we have a generalized linear model with canonical link in the following form:
where G is a canonical link function. Different link functions lead to different models. For instance, a logit link function leads to logistic regression, while an exponential link function leads to Poisson regression.
L _{0} penalized GLM
The distribution of Y in GLM is assumed to be from the exponential families with the following probability (density) function:
where ϕ is a dispersion parameter, and different functions A(∗), B(∗) and C(∗) are for different distributions Y [15]. The corresponding mean and variance are:
where V(μ)=B ^{′′}(θ). Let Y=[Y _{1},…,Y _{ N }]^{t}, X=[x _{1},x _{2},…,x _{ N }]^{t}, and μ=[μ _{1},…,μ _{ N }]^{t}, so \(\mu _{i} = G({\theta }_{i}) = G\left (\mathbf {x}_{i}^{t} {\beta }\right)\) and \({\theta }_{i} = \mathbf {x}_{i}^{t} {\beta } \). The loglikelihood of Y is
Dropping the constants A(ϕ), and C(Y _{ i },ϕ), we have the simplified log likelihood as follows:
Hence, L _{0} penalized error function to minimize is
where \( {\beta } _{0} = \sum _{j=1}^{P} I({\beta }_{j} \neq 0)\) is the number of nonzero elements in β, μ _{ i }=G(θ _{ i }) and \({\theta }_{i} = \mathbf {x}_{i}^{t} {\beta } \). If we define \(\frac {0}{0} = 0\), then \( {\beta } _{0} = \sum _{j} I({\beta }_{j} \neq 0) = \sum _{j} \frac { {\beta }_{j}^{2}}{ {\beta }_{j}^{2}}\). Equation (2) is equivalent to
which is equivalent to the following system:
Given η and \({\theta }_{i} = \mathbf {x}_{i}^{t} {\beta } \), the derivative of E w.r.t. β is
where ⊘ indicates elementwise division. The Hessian matrix is
Let
where V _{ i }=V(μ _{ i })=B ^{″}(θ _{ i })=G ^{′}(θ _{ i }), i=1,…,N, and let \(\tilde {Y} = [Y_{1} B'({\theta }_{1}), \ldots, Y_{N} B'({\theta }_{N})]^{t} = [Y_{1} \mu _{1}, \ldots, Y_{N} \mu _{N}]^{t}\), we have
The NewtonRaphson iteration for β is
Let \(Z = VX {\beta }^{old}+ \tilde {Y}\), we have
Different link functions will lead to different regression models as shown as in Table 1.
Other GLMs such as negative binomial, gamma, and inverse Gaussian can be implemented accordingly with a different V(μ). When dealing with big data problems with N≪P,where N is the number of samples and P is the number of parameters, the inverse of a P×P matrix is timeconsuming and computational challenging. We proposed an efficient algorithm to calculate the inverse of a much smaller N×N matrix as follows (Liu et al. 2015):
So that when N≪P, we have a much efficient estimation:
The adaptive ridge algorithm (L _{0}ADRIDGEA) is implemented in MATLAB are as follows:
The algorithm is easy to implement and very efficient for either small sample size and large dimension or large sample size and small dimension big data problem. The regularized parameter λ can be determined either by crossvalidation or by AIC and BIC with λ=2 and λ= log(N), respectively. We further discuss that the proposed method is a L _{0} approximation and converges to L _{0} when the number of iterations m→∞.
Algorithm justification: Given a highdimensional big feature matrix X _{ N×P }(N≪P) and a threshold γ for the coefficient estimates, L _{0} rejects all the coefficient estimates below γ to 0 and keeps the large coefficients unchanged. This is the same as defining a binary vector s=[…,1,0,…,1]^{t}, with the value of 0 or 1 for each feature, where s _{ j }=1 if the coefficient estimate for that feature is above the threshold γ, and 0 otherwise. Let S=diag(s) be a matrix with s on its diagonal, we have the selected feature matrix X _{ S }=XS. We can build the standard models with the matrix X _{ S }, if we know s in advance. For instance, we can estimate the coefficients of a GLM with L _{2} regulation given X _{ S } and Y with
where \(Z = VX {\beta }^{old}+ \tilde {Y}\), \(\tilde {Y} = [Y_{1} \mu _{1}, \ldots, Y_{N} \mu _{N}]^{t}\), and \(X_{S}^{t}VX_{S} = SX^{t}VXS = SX^{t}VX\) because of the special structure of matrix S. It is guaranteed that the estimate is 0 for feature j with s _{ j }=0. However, in reality we do not know s. Estimating both s and θ is an NPhard problem, since we need to solve a mixedinteger optimization problem. Comparing Eq. (7) with Eq. (5), β ^{new}=(DX ^{t} VX+λI)^{−1} DX ^{t} Z, it is clear that S is replaced by D and a binary s _{ j } is approximated by a continuous \(\eta _{j}^{2}\) in proposed algorithm. Therefore, the proposed method is a L _{0} approximation.
Recall the iterative system in Eq. (3), note that each feature is penalized by a different penalty, which is inversely proportional to the squared magnitude of that parameter estimator η _{ j }. i.e.,
Smaller β _{ j } will lead to larger λ _{ j }. A tiny β _{ j }, will become smaller and λ _{ j } will be getting larger in each iteration of L _{0}ADRIDGE algorithm. β _{ j }→0, and λ _{ j }→∞. On the other hand, a larger β _{ j } will lead a finite λ _{ j }, and nonzero β _{ j }, when the number of iteration goes to ∞. The solution of L _{0}ADRIDGE will converge to that of Eq. (7), because the effect of nonzero η _{ j } will be canceled out in Eq. (5). Note that our proposed methods will find a sparse solution with a large number of iterations and small ε, even though the solution of L _{2} regularized modeling is not sparse. Small parameters (β _{ j }s) become smaller at each iteration and will eventually go to zero (below the machine ε). We can also set a parameter to 0 if it is below predefined ε=1e−6 to speed up the convergence of the algorithm.
Results
Simulations
Poisson Regression: Our first simulation was used to evaluate the performance of our method for high dimensional Poisson regression. The data was generated from Poisson distribution with different sample sizes (N) and dimensions (P). However, only features 1, 5, 10 and the constant term are used to generate the Poisson counts with [β _{0},β _{1},β _{5},β _{10}]=[1,0.5,0.5,0.4]. The count Y is generated with Y=Poisson(μ), where mean μ= exp(β X). The proposed method is compared with the glmnet ([16] and SparseReg package [17, 18]. glmnet and SparseReg implemented the elastic net, SCAD, and MC+ penalties with an efficient path algorithm. We compare the performance of our approach with L _{1} (glmnet), SCAD and MC+ using the popular BIC (λ= log(N)) criteria. Our L _{0}ADRIDGE is compared to the glmnet for L _{1} and SparseReg for both SCAD and MC+. The results of different methods are presented in Table 2.
Table 2 shows that our L _{0}ADRIDGE consistently achieved the best performance with BIC and different sample sizes and dimensions. With BIC, although MC+ has the lowest square root of mean squared error (rMSE), and fits the data better, L _{0}ADRIDGE achieves the least absolute bias \(\hat { {\beta }}  {\beta } \), highest percentage of identified true model (PTM), and lowest false discovery rate (FDR) under different simulation settings. The average number of selected features (ANSF) with L _{0}ADRIDGE is also closest to the true number 4. Particularly, L _{0}ADRIDGE found 100% true model with the lowest average absolute bias (0.086) under the dimension of P=10,000 and sample size of N=500, indicating that the proposed approach is efficient under extrahigh dimensional setting. Another interesting finding is that the square root of mean squared errors and absolute biases with L _{0}ADRIDGE did not vary much across different simulation setting, indicating the robustness of the proposed approach. Moreover, L _{0}ADRIDGE with BIC is slightly faster than different routines implemented in glmnet and SparseReg in computational time. Finally, BIC apparently is not a good model selection criteria for L _{1}, SCAD and MC+. More features are selected than necessary. A larger λ is needed for selecting the correct model. We reported the results with a larger λ on Additional file 1: Table S1, and demonstrated that both SCAD and MC+ can achieve a much smaller FDR, but a larger absolute bias and rMSE.
Logistic regression: The logistic regression data was generated with the coefficients of [β _{1},β _{5},β _{10}]=[0.5,0.5,−0.4], respectively, and the remaining coefficients were set to zero. The score z=X β+ε, where ε is the random noise with the signal to noise ratio of 4. Then, the probability y is generated from the logistic function y=1/(1+e ^{−z}). Note that y is the true probability instead of binary (1/0) in this simulation. Unlike the previous example, the optimal values of λ in this simulation were selected with the standard 5fold crossvalidation. We divided the λ from λ _{min}=1e−4, to λ _{max} into 100 equal intervals in logscale, then chose the optimal λ with the smallest test error. The simulation was also repeated 100 times. The computational results were reported in Table 3. The values in the parenthesis are the positive/negative standard deviation.
Table 3 shows that L _{0}ADRIDGE outperforms L _{1}, SCAD and MC+ with a substantial margin under the 5fold crossvalidation. Crossvalidation is a standard tool for parameter selection in machine learning. L _{0}ADRIDGE achieved the smallest test square root of mean squared error, least absolute biases, the lowest FDR, and highest percentages of identified true models The average number of selected features are 3.33 and 3.41 for the dimensions of 100 and 1000, respectively, which are the closest to the true number of features 3. In contrary, L _{1}, SCAD and MC+ selected unnecessary features. L _{1} on average identified 17.1 and 50.92 features, and SCAD selected 18.35 and 73.03 features on average for the dimensions of 100 and 1000, respectively, while MC+ performed slightly better, choosing 10.41 and 24.8 features for the dimensions of 100 and 1000, respectively. More impressively, out of 100 simulations, L _{0}ADRIDGE identified the true model 81 and 80 times with different dimensions, while L _{1} and SCAD could not find the true model once, and MC+ only identified the true model 2 times for the dimension of 100, indicating the super performance of L _{0}ADRIDGE under crossvalidation. Finally, L _{0}ADRIDGE is robust. The test square root of mean squared error and other performance measures did not vary much when the dimension increased from 100 to 1000. It is worth noting that our proposed method performs well with the popular statistical model selection criteria such as BIC and crossvalidation. Other popular methods such as L _{1}, SCAD, and MC+ select more features than necessary with such criteria. Therefore, many popular packages including the commercial MATLAB usually choose a larger λ one standard deviation above the minimum test error with crossvalidation, which is arbitrary and leads to larger bias. To overcome such bias in parameter estimation, some packages reestimate the parameters with the selected features and standard GLM model. Unlike these methods, our proposed method performed much better without any postprocessing. Finally, the algorithm is very robust with different initialization. With N=100, P=1000 and 100 times of different randomized initialization, we achieved the trMSE of 0.437(±.003), average absolute bias of 0.0763(±.07), ANSF of 3.39(±1.154), PTM of 85% and FDR of 6.9%, which is quite similar to the results with a fixed initialization.
TCGA ovarian cancer data
The Cancer Genomic Atlas (TCCA) has generated a large amount of next generation sequencing and other omics data for ovarian adenocarcinoma (OC). In this study, we conducted integrated analysis of RNAseq, miRNA expression, promoter methylation, and debulking status data from 367 OC patients. There are 342 microRNAs, 13,911 mRNA expression (in FPKM), and 21,985 promoter methylation values available. We first normalized different omics data and screened the debulking associated microRNA, mRNAs, and methylation promoters with the Pvalues of less than 0.01 with the training data only. Based on the central dogma of biology, suboptimal debulking is associated with microRNA expression, gene expression, and DNA methylation; gene expression is a function of microRNA expression and DNA methylation; and microRNA expression is regulated by DNA methylation. L _{0} Logistic regression was used for suboptimal debulking prediction, while L _{0} penalized Poisson regression was used for gene expression and microRNA expression prediction with FPKM. FPKM, representing fragments per kilobase of exon per million fragments mapped, measures the normalized read counts for RNAseq. Threefold cross validation was used for gene selection and validation. We reported the gene signatures with the best predicted area under the ROC curves (AUCs). Molecular signatures that are directly or indirectly associated with suboptimal debulking are shown in Fig. 1.
Figure 1 indicates that there are 16 gene signatures including 7 mRNAs and 9 epigenetic markers directly associated with debulking status. Even though there is no microRNA directly associated with debulking, eight microRNA signatures are indirectly associated with debulking through their association with mRNA signatures. Moreover, there are additional 18 epigenetic markers indirectly associated with debulking. The 7 mRNAs directly associated with debulking are EIF3D, PPP1R7, ADA, HSD17B1, SRBD1, ZNF621, and BARX1, where EIF3D, PPP1R7, BARX1 and ZNF621 have positive correlations and the other 3 genes have negative correlations with suboptimal debulking. Among the 7 mRNAs, ADA (Adenosine Deaminase) is a wellstudied gene in ovarian neoplasms. ADA levels were found to be significantly higher in patients with ovarian cancers as compared with benign ovarian tumors [19]. ADA has been regarded as a potential biomarker for diagnosis and an agent for the treatment of ovarian cancer [20]. Other mRNAs such as BARX1, EIF3D, PPP1R7, and HSD17B1 are also known to be associated with different cancers or other diseases. At the microRNA level, there are 8 microRNAs indirectly associated with debulking including mir183, let7b, mir91, mir377, mir202, mir758, mir375, and mir30c2. While let7b, mir30c2, and mir377 are positively correlated with suboptimal debulking through mRNAs ADA and BARX1 indirectly, the other 5 microRNAs have indirectly negative correlations with suboptimal debulking. Seven of eight microRNAs except for mir758 are known to be associated with ovarian cancer. Particularly, let7b is known to be an unfavorable prognostic biomarker and predict of molecular and clinical subclasses in highgrade serous ovarian carcinoma, and it may also be useful for discriminating between controls and patients with serous ovarian cancer [21, 22]. Mir183 is known to be associated with multiple cancers. It regulates target oncogene (Tiam1), and reduce the migration, invasion and viability of ovarian cancer cells [23]. Finally, at the DNA level, nine epigenetically modified genes directly associated with debulking are SSX1, TBR1, ZNF621, ORC3L, COL22A1, SPEF2, SSU72, EEF1D, and ZNF621, where EEF1D, SSU72, and ORC3L are positively associated with suboptimal debulking, while 6 other epigenetic genes are negatively correlated with suboptimal debulking. In addition, 18 other epigenetic genes indirectly associated with debulking may also have biological implications. Finally, integration of multiomic data increases the prediction power substantially. Besides analyzing three types of omics data together, we performed the same threefold cross validation for gene expression, methylation, and microRNA expression separately. The AUC curves are in Fig. 2.
Figure 2 shows that the best predicted AUC over 100 simulations for integrated data is 0.88, while the best predictive AUCs for gene expression, methylation, and microRNA over 100 simulations are 0.81, 0.84, and 0.76, respectively. The AUC with integrated data achieved the highest AUC, indicating the importance of multiomics data mining. Genes selected with mRNA, microRNA, and methylations separately are reported in the supplementary document. In addition, we also compare the selected features and the same number of top genes identified with statistical test. The results are reported on Additional file 1: Table S2, and demonstrate that although individual genes are more statistically significant, combination of a panel of genes with standard logistic regression has less predictive power and test AUC (0.79).
Conclusions
Biomarkers from multiomics data may predict disease status and help physicians to make clinical decisions. L _{0} based GLM, which directly penalizes the number of nonzero parameters, has nice theoretical properties and leads to essential sparsity for biomarker discovery. Optimizing the L _{0} regularization is a crucial, but difficult problem. We have developed an adaptive ridge algorithm (L _{0}ADRIDGE) for approximating L _{0} penalized GLM. The algorithm is easy to implement and efficient for problems with either an ultrahigh dimension and small sample size, or a lowdimension and large sample size. It outperforms the other cutting edge regularization methods including L _{1}, SCAD and MC+ through simulations. When applied to the integration of multilevel omics data from TCGA and the prediction of suboptimal debulking from ovarian cancer, it can identify a panel of gene signatures achieving the best prediction power. We also demonstrate that prediction power of a model with multiomics data increases substantially, when comparing with a model with one omics data, indicating the importance of big data mining.
Abbreviations
 AUC:

The area under an ROC curve
 GLM:

Generalized linear model
 L _{0}ADRIDGE:

L _{0} adaptive ridge algorithms
 MC+:

Minimax concave penalty. SCAD: Smoothly clipped absolute deviation
 TCGA:

The cancer genome atlas
References
 1
Tibshirani R. Regression shrinkage and selection via the lasso. J Roy Stat Soc B. 1996; 58:267–88.
 2
Zou H. The adaptive lasso and its oracle properties. J Am Stat Assoc. 2006; 101:1418–29.
 3
Lin D, Foster D, Ungar L. A risk ratio comparison of l0 and l1 penalized regressions. Tech. rep.,University of Pennsylvania; 2010.
 4
Kakade S, Shamir O, Sridharan K, Tewari A. Learning exponential families in high dimensions: strong convexity and sparsity. JMLR. 2013; 9:381–8.
 5
Bahmani S, Raj B, Boufounos P. Greedy sparsityconstrained optimization. J Mach Learn Res. 2013; 14(3):807–41.
 6
Zhang T. Multistage convex relaxation for feature selection. Bernoulli. 2012; 19:2153–779.
 7
Fan J, Li R. Variable selection via nonconcave penalized likelihood and its oracle properties. J Am Stat Assoc. 2001; 96:1348–61.
 8
Zhang C. Nearly unbiased variable selection under minimax concave penalty. Ann Stat. 2010; 38:894–942.
 9
Liu Z, Lin S, Deng N, McGovern D, Piantadosi S. Sparse inverse covariance estimation with L0 Penalty for Network Construction with Omics Data. J Comput Biol. 2016; 23(3):192–202.
 10
Liu Z, Li G. Efficient Regularized Regression with L _{0} Penalty for Variable Selection and Network Construction. Comput Math Methods Med. 2016; 2016:3456153.
 11
Bahmani S, Boufounos P, Raj B. Learning ModelBased Sparsity via Projected Gradient Descent. IEEE Trans Info Theory. 2016; 62(4):2092–9.
 12
Riester M, Wei W, Waldron L, Culhane A, Trippa L, Oliva E, Kim S, Michor F, Huttenhower C, Parmigiani G, Birrer M. Risk prediction for latestage ovarian cancer by metaanalysis of 1525 patient samples. J Natl Cancer Inst. 2014;106(5):Apr 3.
 13
Tucker S, Gharpure K, Herbrich S, Unruh A, Nick A, Crane E, Coleman R, Guenthoer J, Dalton H, Wu S, Rupaimoole R, LopezBerestein G, Ozpolat B, Ivan C, Hu W, Baggerly K, Sood A. Molecular biomarkers of residual disease after surgical debulking of highgrade serous ovarian cancer. Clin Cancer Res. 2014; 20(12):3280–8.
 14
Liu Z, Beach J, Agadjanian H, Jia D, Aspuria P, Karlan B, Orsulic S. Suboptimal cytoreduction in ovarian carcinoma is associated with molecular pathways characteristic of increased stromal activation. Gynecol Oncol. 2015; 139(3):394–400.
 15
Wood S. Generalized Additive Models: An Introduction with R. New York: Chapman & Hall/CRC; 2006.
 16
Friedman J, Hastie T, Tibshirani R. Regularization Paths for Generalized Linear Models via Coordinate Descent. J Stat Softw. 2011; 33(1):1–22.
 17
Zhou H, Lange K.A path algorithm for constrained estimation. J Comput Graph Stat. 2013; 22(2):261–83.
 18
Zhou H, Wu Y. A generic path algorithm for regularized statistical estimation. J Am Stat Assoc. 2014; 109(506):686–99.
 19
Urunsak I, Gulec U, Paydas S, Seydaoglu G, Guzel A, Vardar M. Adenosine deaminase activity in patients with ovarian neoplasms. Arch Gynecol Obstet. 2012; 286(1):155–9.
 20
Shirali S, Aghaei M, Shabani M, Fathi M, Sohrabi M, Moeinifard M. Adenosine induces cell cycle arrest and apoptosis via cyclinD1/Cdk4 and Bcl2/Bax pathways in human ovarian cancer cell line OVCAR3. Tumour Biol. 2013; 34(2):1085–95.
 21
Tang Z, Ow G, Thiery J, Ivshina A, Kuznetsov V. Metaanalysis of transcriptome reveals let7b as an unfavorable prognostic biomarker and predicts molecular and clinical subclasses in highgrade serous ovarian carcinoma. Int J Cancer. 2014; 134(2):306–18.
 22
Chung Y, Bae H, Song J, Lee J, Lee N, Kim T, Lee K. Detection of microRNA as novel biomarkers of epithelial ovarian cancer from the serum of ovarian cancer patients. Int J Gynecol Cancer. 2013; 23(4):673–9.
 23
Li J, Liang S, Jin H, Xu C, Ma D, Lu X. Tiam1, negatively regulated by miR22, miR183 and miR31, is involved in migration, invasion and viability of ovarian cancer cells. Oncol Rep. 2012; 27(6):1835–42.
Acknowledgements
This work was partially supported by the DMS1343506 grant from the National Science Foundation (ZL), and P01 CA09891210 and UL1 TR000188101 grants from NIH. The funders had no role in the preparation of the article.
Funding
DMS1343506 of NSF (ZL), P01 CA09891210 and UL1 TR000188101.
Availability of data and materials
L _{0} ADRIDGE in MATLAB is available at https://github.com/liuzqx/L0adridge.
Author information
Affiliations
Contributions
ZL conceptualized and designed method, developed the software, and wrote the manuscript. SF, MDP helped in method design and manuscript writing and revised the manuscript critically. All authors read and approved the final manuscript.
Corresponding author
Correspondence to Zhenqiu Liu.
Ethics declarations
Ethics approval and consent to participate
Not Applicable.
Consent for publication
Not Applicable.
Competing interests
The authors declare that they have no competing interests.
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Additional file
Additional file 1
Table S1. Performance of different GLM methods for Poisson regression over 100 simulations, where values in the parenthesis are the standard deviations, and ANSF: Average number of selected features; rMSE: Average square root of mean squared error; \(\hat {\beta }  \beta  = {\sum \nolimits }_{i}\hat {\beta }  \beta _{i}\): average absolute bias when comparing true and estimated parameters. PMS: Performance Measures. PTM: Percentage of true models. FDR: False discovery rates. The L0ADRIDGE is compared to the best performance chosen from λ=0.9λ _{max} and λ=0.5λ _{ max } with both for SCAD and MC+. Table S2. The comparison of performance of the our sparse modeling approach and the top genes selected with Student’s ttest. The results demonstrate that although each gene is more statistically significant with statistical test, the combination of the panel of genes has less predictive power and test AUC with standard logistic regression and threefold cross valida tion, indicating the collinearity among theses genes. (PDF 86 kb)
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
About this article
Cite this article
Liu, Z., Sun, F. & McGovern, D.P. Sparse generalized linear model with L _{0} approximation for feature selection and prediction with big omics data. BioData Mining 10, 39 (2017). https://doi.org/10.1186/s130400170159z
Received:
Accepted:
Published:
Keywords
 Sparse modeling
 L _{0} penalty
 Big data mining
 Multiomics data
 GLM
 Classification
 Suboptimal debulking