PLoS ONE
Home Efficient sparse estimation on interval-censored data with approximated L0 norm: Application to child mortality
Efficient sparse estimation on interval-censored data with approximated L0 norm: Application to child mortality
Efficient sparse estimation on interval-censored data with approximated L0 norm: Application to child mortality

Competing Interests: The authors have declared that no competing interests exist.

Article Type: Research Article Article History
Abstract

A novel penalty for the proportional hazards model under the interval-censored failure time data structure is discussed, with which the subject of variable selection is rarely studied. The penalty comes from an idea to approximate some information criterion, e.g., the BIC or AIC, and the core process is to smooth the ℓ0 norm. Compared with usual regularization methods, the proposed approach is free of heavily time-consuming hyperparameter tuning. The efficiency is further improved by fitting the model and selecting variables in one step. To achieve this, sieve likelihood is introduced, which simultaneously estimates the coefficients and baseline cumulative hazards function. Furthermore, it is shown that the three desired properties for penalties, i.e., continuity, sparsity, and unbiasedness, are all guaranteed. Numerical results show that the proposed sparse estimation method is of great accuracy and efficiency. Finally, the method is used on data of Nigerian children and the key factors that have effects on child mortality are found.

Chen,Zhao,and Shi: Efficient sparse estimation on interval-censored data with approximated L0 norm: Application to child mortality

Introduction

Interval-censored failure time data, which means the failure time of interest is only known to belong to a period instead of observed directly, is commonly seen in many fields, such as demography, medicine, and ecology. The statistical analysis of this special data structure has attracted much attention since first being addressed by Finkelstein (1986 [1]), and many researchers have developed significant works related to model estimations (Huang 1996 [2]; Zhang et al. 2005 [3], Zeng, Cai, and Shen 2006 [4], Lin and Wang 2010 [5]). Sun (2006 [6]) made a thorough review of research on interval-censored failure time data. Compared with right-censored data, interval-censored data can be more challenging when modeled in two ways. First, interval-censored data can be more complicated, such that sometimes it is a mixture of interval censoring and right censoring. Right censoring can be considered a special form of interval censoring with the right bound extending to infinity. Second, when implementing the proportional hazards model (Cox 1972 [7]) on right-censored data, one can use partial likelihood and does not have to estimate the baseline hazards function simultaneously with the parameters of interest.

Several methods exist that deal with interval-censored failure time data. Tong, Chen, and Sun (2008 [8]) and Goggins and Finkelstein (2000 [9]) developed approaches for interval censoring, but under strict independent assumptions. These approaches are restricted to several models; for example, the former is only applicable to the additive hazards model (Lin and Ying 1994 [10]). In this paper, a sieve maximum likelihood method with Bernstein polynomials is proposed as a general way that can be applied to many semi-parametric survival models. More information is presented below.

Although basic theories on interval-censored data are well established, studies on variable selection under this data structure are very limited. To the best of our knowledge, penalized partial likelihood has been effectively used since it is intuitive to add a penalization term to the likelihood function. Tibshirani (1997 [11]) and Fan and Li(2002 [12]) successfully applied LASSO and SCAD penalties to the proportional hazards model right after they proposed them. Zou (2006 [13]) developed adaptive LASSO (ALASSO) and Zhang and Lu (2007 [14]) used it with partial likelihood. However, penalized partial likelihood is bounded to the analysis of right-censored data. For variable selection on interval-censored failure time data, piecewise constant functions are occasionally used (Wu and Cook 2015 [15]) to represent the baseline hazards model, incorporated with several penalties and the EM algorithm is applied to optimize the likelihood function. Wang et al. (2019 [16]) introduced a Bayesian adaptive lasso penalty for the additive hazards model with Case I interval-censored data, also known as current status data, in which the subjects are only visited once and one only knows whether it has failed at the exact observation time. Zhao (2020 [17]) developed a broken adaptive ridge (BAR) penalized procedure and, with iterations, some parameters finally shrank to zero. The simulation studies show satisfying results, whereas it is still found to be computationally costly due to the heavy optimization procedures and that there are two parameters to tune.

The object of interest in the present paper is another type of covariate selection technique, i.e., best subset selection (BSS). A typical BSS method is to list all the variable subsets, model with each one of them, and use some information criterion, such as AIC (Akaike 1974 [18]) and BIC (Schwarz 1978 [19]) to judge every subset. For a dataset with n0 uncensored samples and p dimensions of parameters, a criterion has the form as:

, where k(kp) represents the number of selected parameters and l(⋅) represents the log-likelihood function. φ0 is fixed as 2 or log(n0) when a AIC or BIC is applied and it makes the criterion free of tuning, which can be a heavily time-consuming process when a common penalty such as LASSO, SCAD, or MCP (Zhang 2010 [20]) is used. The most significant problem of this method is that in the criterion a ℓ0 norm is involved, the discrete nature of which makes the method a NP-hard problem. Although stepwise regressions are available to help with the optimization, BSS is still infeasible when it comes to a moderately large p. Su et al. successfully developed an approximated form of information criterion as a penalty term [minimum information criterion (MIC)] for general linear models (GLM, 2018 [21]) and the Cox model with right-censored data (2016 [22]), which extends the BSS method to a large variable dimension.

In this study, a form of approximated information criterion is proposed under the interval-censored data structure. The result provides several major contributions to the current literature. First, an approximated BSS method is introduced into the analysis of interval-censored data, with which the variable selection approaches are rarely studied. Second, great efficiency is achieved by conducting the estimation of both the coefficients and baseline hazards function with free-tuning covariate selection simultaneously.

The rest of this paper is organized as follows. In the first section, the notation and detailed estimation procedure of interval-censored data are given. In the next section, the approximation of BSS is presented. In the third section, the simulation results of the proposed method are shown along with several other commonly seen penalties. The application section contains a survey example and the last section concludes this research and addresses a short discussion.

Notation, assumptions, and models

Interval censoring

Consider a failure time study that involves n independent subjects and for an ith subject there is a p-dimensional vector Zi(1 ≤ in) that may affect its failure time Ti. According to the proportional hazards model, the cumulative hazard function Λ(t) is given by

, where Λ0(t) denotes the baseline cumulative hazard function and β = (β1, β2, …, βp)′ the regression coefficients. The corresponding survival function is S(t|Zi)=P(Tit|Zi)=exp(-Λ(t|Zi))=exp(Λ0(t)exp(Ziβ)). Under the interval-censored data structure, observations will be recorded as O = {Oi = {Li, Ri, Zi}, i = 1, …, n}, with (Li, Ri] denoting the interval in which the failure of the ith subject belongs. Then, the likelihood function can be given as

In practice, if one does not observe the failure of some samples during the entire experiment, these samples will be considered right-censored. For right-censored subject, Ri = ∞ and thus S(Ri|Zi) = 0.

To estimate ξ = (β, Λ0), the traditional approach is to maximize the log-likelihood function ln(β, Λ0), usually by finding the zeros of the derivatives. The main difficulty is to estimate finite- and infinite-dimensional parameters at the same time. This problem is discussed in the next section.

Regularized sieve maximum likelihood estimation

To deal with the infinite-parameter estimation problem mentioned above, a sieve method (Huang and Rossini 1997 [23] and Zhou, Hu, and Sun [24]) is developed for our study. Now, consider a parameter space

, where B={βRp,|β|M} with M a positive constant and F representing the function set that contains all bounded, continuous, non-negative, and non-decreasing functions. One common way to model λ0(⋅) is using splines (Wood, Pya, and Säfken 2017 [25] and Wang et al. 2019 [26]), but here, with the given restricted shape of the baseline cumulative hazards function, use of the Bernstein basis polynomials (Wang and Ghosh 2012 [27]) is preferred. Hence, a sieve parameter space
is defined, where
on the domain of observed data, recorded as [u, v]. Here, Mn is a positive constant and bj(x, u, v, m) is defined as
where m decides the number of terms in the Bernstein basis polynomials and ω˜=(ω˜0,ω˜1,,ω˜m) denotes the coefficient vector of the terms. Note that for the non-decreasing and non-negative requirements of Λ0n, one needs ω˜ following the inequality 0ω˜0ω˜1ω˜m. This constraint can be ensured by reparameterization, introducing a novel vector ω = (ω0, ω1, …, ωm) and let ω˜l=k=0lexp(ωk).

By focusing on sieve space Ξn, the complex estimations of both infinite- and finite-dimensional parameters are converted into a much simpler estimation problem that contains only finite-dimensional parameters (β, ω). Thus, given the argument matrix Z = (Z1, Z2, …, Zn), our likelihood function has the form

To estimate parameters and select covariates simultaneously, minimizing the sieve log-likelihood function with a penalty term is considered:

It is intuitive to replace pen(β) with various developed penalties. For LASSO, let pen(β)=i=1p|βi|; for SCAD, let pen(β)=0|βj|min{1,(aλx)+/(aλλ)}dx with a usually fixed at 3.7; for MCP, let pen(β)=0βj(1xγλ)+dx on [0, + ∞]. However, the aforementioned penalties all need a time-consuming tuning process for hyperparameter φ. The approximate information criterion is introduced as a penalty term in the following section, which frees us from tuning the parameter φ and greatly reduces computing time.

Approximated information criterion

Approximation of information criterion

A novel sparse estimation method for interval-censored data is sought from the idea of approximation. The BSS method with certain information criteria does not need the parameter-tuning process, but it is infeasible for a large p. In this part, a smooth approximation of information criteria is developed that can be further used in the way of regularization methods.

The core task is to approximate the ℓ0 norm in the information criteria. The ℓ0 norm can be defined by indicator functions

The essential job is to approximate the indicator functions, for which one introduces a function η(x) that satisfies (1) η(0) = 0, (2) η(x) = η(−x), (3) lim|x|→∞ η(x) = 1, and (4) η(⋅); the latter is a smooth function and non-decreasing on R+. Clearly, η(⋅) has captured the key features of the indicator I(x ≠ 0).

One natural thought is to adopt sigmoid functions, which are commonly used as a smooth output unit in binary classification. The classic choice is the logistic activation sigmoid σ(x)=11+e-x, and by making some minor changes on the independent variable and the intercept one can successfully develop our solution as follows:

, where θ > 0 and γ control the shape of the function. Fig 1a and 1b plot η(⋅) with γ = 1 and γ = 2, respectively. θ varies from 1 to 100. It can be seen from Fig 1b that, in general, the functions η(⋅) with both choices on γ are good approximations of I(β ≠ 0), and it seems that, with a larger θ, η(⋅) will be more like the target indicator function. Nevertheless, comparing Fig 1a and 1b, we avoid directly setting γ = 1 because there is a cusp at x = 0, although it appears to give a better performance on sparsity. This concern restricts one to set γ = 2. To achieve balance smoothness and sparsity together on one penalty, our motivation is to use the reparameterization procedure.

η(⋅) with various values of θ.
Fig 1

η(⋅) with various values of θ.

Clearly shown is that η(⋅) makes a promising approximation of ℓ0 norm. Note that in both plots the parameter θ varies from 1 to 100 and the curves become sharper with larger θ. (a)γ = 1. (b)γ = 2.

Reparameterization

By introducing η(⋅), the smoothness problem of the ℓ0 norm is preliminarily solved by setting γ = 2. Fan and Li (2001 [28]) proposed three properties that a good penalty should possess: unbiasedness, sparsity, and continuity. Unbiasedness and continuity are apparently ensured by the definition. The sparsity needs to be enforced since we have chose γ = 2. For this purpose, the following reparameterization procedures are considered.

Set a vector ϕ=(ϕ1,ϕ2,,ϕp)Rp and relate ϕ to β by βi = ϕi η(ϕi). Define a matrix

and then the reparameterization can be written as β = H ϕ. In this way, (3) is rewritten as follows:
where tr(H) denotes the trace of H. By reparameterizing β with ϕ, two goals can be achieved simultaneously: one is to keep the smoothness of the regularization problem, and the other is to obtain good performance on sparsity. These two aspects will be explained in the next section with figures.

    Smoothness. In (4), the second term of the right-hand side ϕ0 tr(H) is smoothed by the definition of η(⋅) when γ = 2, so the problem remaining here is to check the smoothness of the first term, which is essentially decided by H ϕ. β = H ϕ is composed of formula βi = ϕiη(ϕi), i = 1, …, p, and it is commonly known that the product of two smooth functions is also smooth. The relationship of β = H ϕ and ϕ is illustrated in Fig 2, in which a desired one-to-one mapping can be seen.

    Sparsity. In the preceding section, the reason for choosing γ = 2 was explained, although γ = 1 is favorable in sparsity, which is shown in Fig 1a and 1b. Here, after reparameterization, the relationship between H, which determines the penalty, and β, the true coefficients, is explored. η(ϕ) and β are plotted in Fig 3, which demonstrates that the scale of penalty really assembles the situation when one sets γ = 1, in which the regularization will penalize the likelihood function with a considerably small coefficient allocated to wrong covariates.

Plot of βi and ϕi.
Fig 2

Plot of βi and ϕi.

This plot shows desired mutually one-to-one mapping, which is strong evidence supporting reparameterizations. With γ varying from 1 to 100, the function becomes closer to f(ϕ) = ϕ.

Plot of βi and η(ϕi).
Fig 3

Plot of βi and η(ϕi).

This plot displays relationship between true coefficients and penalty term.

One returns to the three properties that Fan and Li recommended, i.e., continuity, sparsity, and unbiasedness, after reparameterizing β with ϕ. Unbiasedness is ensured by the definition of η(⋅). Continuity lies naturally in maintaining smoothness. The goal of sparsity is attained by reparameterization, as explained in this section. Accordingly, the penalty developed herein fulfills all requirements.

Noted that after reparameterization, the penalty term becomes non-convex on β. This implies the penalized log-likelihood function (4) can have multiple local optima and the initial value of optimization process will effect the result. Thus, we use simulated annealing (Belisle 1992 [29]) and BFGS quasi-Newton algorithm (see, e.g., Jorge Nocedal 1980 [30]) to obtain the final result. The simulated annealing is robust and seeks the global optimum, and BFGS algorithm is very fast and assures the result converges to a critical point. Both methods are built-in in Matlab with function simulannealbnd and fminunc. To further scrutinize our method, the numerical results are presented in the next section.

Simulation study

An extensive simulation study is conducted in this section to assess the performance of the proposed sparse estimation method with the approximated information criterion (appIC) on finite interval-censored samples. In this study, a p-dimensional covariate set Zp × n = (Zi, Z2, …, Zp) is considered, and Zi is generated from multivariate normal distribution, with mean zero and covariance matrix Σ. Σ is defined by

Here we set three scenarios: n = 100, p = 10, n = 300, p = 10 and n = 300, p = 30. For the true coefficient vector, we set the first and last two components at b0 (b0 = 0.5 or 1), with other components zero. The baseline cumulative hazards function takes the forms Λ0(t) = t and Λ0(t) = log(t + 1), respectively.

When constructing interval-censored data structure, M visiting points are set in the interval [u, v] with a uniform gap (vu)/M. To simulate the real situation in which some samples will be difficult to reach, every point is allocated a 50% chance to actually observe a certain sample. In this simulation, u is fixed at 0 and v is set as 3, thus when b0 = 0.5, there are approximately 25% and 35% right-censored portions for Λ0(t) = t and Λ0(t) = log(t + 1), and when b0 = 1, the portions are approximately 30% and 40%. Meanwhile, the number pf observation points M are set at 10 and 20, and the latter is supposed to bring more information.

The hyperparameters of the proposed method are assigned as follows: (1) we set Bernstein polynomials parameter m = 3 because we find it sufficient to characterize the baseline cumulative hazards function; (2) we set γ = 2 for smoothness as described in the previous sections; (3) we fix φ at log(n0) (n0 denotes the number of samples that are not right-censored) as BIC; (4) we assign θ with n0 and the robustness of the estimate with different value of θ is shown in S1 File. The results are presented in Tables 15. In Tables 14, the performance is measured in three ways: mean, bias and standard deviation (SD), which indicate the accuracy and stability of our estimates. In Table 5, the performance under all scenarios are assessed with four measurements: average selected size (Size), the average true positive size (TP), the average false positive size (TP) and the median and standard deviation of (β^-β)E(ZZ)(β^-β). It can be seen that the bias and standard deviation both greatly reduce when the sample size increases from 100 to 300. When p jumps to 30, appIC method remains feasible and generate good results. And the estimating accuracy and stability generally improve with a more frequent visiting pattern, which is indicated by a large M. Besides, it is shown that the selection correctness of the appIC method increases significantly with greater signals (b0 = 1), compared to b0 = 0.5. Meanwhile, the baseline cumulative hazards function is well modeled by Bernstein polynomials and some of the results are displayed in Fig 4. It is obvious that when n increases, the polynomials fit the hazards function better.

Table 1
Simulation result of appIC sparse estimation with b0 = 0.5 and baseline cumulative hazards function Λ0(t) = t.
MeanBiasSDMeanBiasSD
M = 10M = 20
n = 100, p = 10
β1 = 0.50.5370.0370.2430.5370.0370.233
β2 = 0.50.494-0.0060.2700.486-0.0140.249
β9 = 0.50.5190.0190.2670.5090.0090.250
β10 = 0.50.5010.0010.2510.5090.0090.215
n = 300, p = 10
β1 = 0.50.5100.0100.0920.5130.0130.084
β2 = 0.50.5090.0090.0980.5090.0090.095
β9 = 0.50.499-0.0010.1200.5000.0000.100
β10 = 0.50.5100.0100.1000.5070.0070.088
n = 300, p = 30
β1 = 0.50.5240.0240.1010.5180.0180.093
β2 = 0.50.5090.0090.1120.5020.0020.115
β9 = 0.50.5060.0060.1210.5060.0060.100
β10 = 0.50.5120.0120.1050.5070.0070.096
Table 2
Simulation result of appIC sparse estimation with b0 = 0.5 and baseline cumulative hazards function Λ0(t) = log(t + 1).
MeanBiasSDMeanBiasSD
M = 10M = 20
n = 100, p = 10
β1 = 0.50.5440.0440.2680.5590.0590.227
β2 = 0.50.5770.0770.2580.480-0.0200.265
β9 = 0.50.5830.0830.2440.5090.0090.258
β10 = 0.50.5590.0590.2300.5260.0260.221
n = 300, p = 10
β1 = 0.50.5180.0180.0930.5190.0190.091
β2 = 0.50.5130.0130.1110.5150.0150.108
β9 = 0.50.5040.0040.1190.5060.0060.109
β10 = 0.50.5200.0200.1020.5140.0140.095
n = 300, p = 30
β1 = 0.50.5280.0280.1040.5300.0300.094
β2 = 0.50.5140.0140.1080.5100.0100.108
β9 = 0.50.5180.0180.1150.5180.0180.108
β10 = 0.50.5170.0170.1100.5150.0150.104
Table 3
Simulation result of appIC sparse estimation with b0 = 1 and baseline cumulative hazards function Λ0(t) = t.
MeanBiasSDMeanBiasSD
M = 10M = 20
n = 100, p = 10
β1 = 11.1520.1520.2761.1320.1320.251
β2 = 11.0840.0840.3491.0860.0860.307
β9 = 11.1070.1070.3431.1100.1100.276
β10 = 11.1180.1180.2821.1050.1050.253
n = 300, p = 10
β1 = 11.0390.0390.1291.0360.0360.119
β2 = 11.0270.0270.1401.0210.0210.124
β9 = 11.0210.0210.1341.0200.0200.115
β10 = 11.0350.0350.1361.0280.0280.122
n = 300, p = 30
β1 = 11.0650.0650.1381.0480.0480.119
β2 = 11.0460.0460.1431.0450.0450.125
β9 = 11.0460.0460.1331.0410.0410.121
β10 = 11.0500.0500.1451.0370.0370.127
Table 4
Simulation result of appIC sparse estimation with b0 = 1 and baseline cumulative hazards function Λ0(t) = log(t + 1).
MeanBiasSDMeanBiasSD
M = 10M = 20
n = 100, p = 10
β1 = 11.1650.1650.3151.1380.1380.277
β2 = 11.1150.1150.3601.1120.1120.303
β9 = 11.1550.1550.3551.1430.1430.314
β10 = 11.1150.1150.3221.1060.1060.289
n = 300, p = 10
β1 = 11.0510.0510.1331.0480.0480.121
β2 = 11.0430.0430.1481.0360.0360.128
β9 = 11.0330.0330.1381.0380.0380.132
β10 = 11.0480.0480.1461.0400.0400.131
n = 300, p = 30
β1 = 11.0680.0680.1411.0710.0710.132
β2 = 11.0560.0560.1581.0600.0600.139
β9 = 11.0600.0600.1581.0570.0570.135
β10 = 11.0590.0590.1601.0580.0580.141
Table 5
Comparison of different scenarios with the appIC model.
SizeTPFPMMSE(SD)SizeTPFPMMSE(SD)
Λ0(t) = tΛ0(t) = ln(t + 1)
b0 = 0.5
M = 10
n = 100, p = 103.8433.4800.3630.240(0.188)3.9333.5370.3970.256(0.186)
n = 300, p = 104.1773.9670.2100.040(0.043)4.2633.9800.2830.047(0.047)
n = 300, p = 304.6173.9570.6600.061(0.061)4.8333.9770.8570.071(0.061)
M = 20
n = 100, p = 103.8673.5670.3000.198(0.162)3.9503.5700.3800.218(0.184)
n = 300, p = 104.2333.9830.2500.035(0.034)4.2103.9830.2270.042(0.042)
n = 300, p = 304.4603.9570.5030.047(0.047)4.7173.9770.7400.062(0.064)
b0 = 1
M = 10
n = 100, p = 104.3573.9400.4170.449(0.492)4.3733.9600.4130.535(0.764)
n = 300, p = 104.2674.0000.2670.077(0.063)4.3534.0000.3530.087(0.068)
n = 300, p = 304.9434.0000.9430.089(0.081)5.1034.0001.1030.184(0.203)
M = 20
n = 100, p = 104.2833.9670.3170.441(0.586)4.3433.9770.3670.540(0.915)
n = 300, p = 104.2104.0000.2100.071(0.073)4.3074.0000.3070.090(0.080)
n = 300, p = 304.6534.0000.6530.100(0.095)4.9334.0000.9330.145(0.134)
The estimated baseline cumulative functions.
Fig 4

The estimated baseline cumulative functions.

True and simulated baseline cumulative hazards functions shown in black and yellow, respectively. Simulated cumulative hazards function are generated with b0 = 0.5, M = 10 and p = 10. Note here (a)n = 100,Λ0(t) = t. (b)n = 100,Λ0(t) = log(t + 1). (c)n = 300,Λ0(t) = t. (d)n = 300,Λ0(t) = log(t + 1).

To comprehensively assess the performance of appIC, its estimation results were compared with other commonly used approaches to variable selection: LASSO, SCAD, MCP, and BAR under the most serious conditions above, that is M = 10 and b0 = 0.5. The parameters of these penalties are tuned with 5-fold cross validation with the largest log-likelihood value on the validation sets. The estimations with true covariates (oracle) and full models (without selecting the covariates) are presented alongside in Tables 6 and 7. The measurements are described in the previous paragraph. It is obvious that the appIC sparse estimation performs well in both selection correctness and estimation accuracy. The variable size that it has chosen is especially close to the true size, 4, with relatively low FP. That is to say, the proposed method is very unlikely to include an irrelevant variable. Meanwhile, the TP of the appIC method rises to a satisfactory level when n increases from 100 to 300, performing close to SCAD and MCP. Besides, the square error of the proposed method is low with both n = 100 and n = 300. Furthermore, it is worth mentioning that our method is far faster than common penalties owing to the free-tuning on the hyperparameter. The CPU time of different methods under various conditions can be found in S1 File.

Table 6
Estimation results with different methods.
SizeTPFPMMSE(SD)
Λ0(t) = t
n = 100, p = 10
Full10.0004.0006.0000.694(0.814)
Oracle4.0004.0000.0000.092(0.112)
BAR4.1833.8130.3700.140(0.152)
Lasso5.3463.9631.3830.199(0.116)
Alasso4.3843.7070.6770.211(0.172)
SCAD4.3703.7800.5900.354(0.168)
MCP4.2003.6130.5870.208(0.196)
appIC3.8433.4800.3630.240(0.188)
n = 300, p = 10
Full10.0004.0006.0000.095(0.096)
Oracle4.0004.0000.0000.028(0.030)
BAR4.1764.0000.1760.035(0.029)
Lasso5.3304.0001.3300.081(0.043)
Alasso4.4063.9930.4130.061(0.062)
SCAD4.4603.9830.4770.063(0.051)
MCP4.3933.9930.4000.031(0.041)
appIC4.1773.9670.2100.040(0.043)
n = 300, p = 30
Full10.0004.0006.0000.256(0.211)
Oracle4.0004.0000.0000.029(0.031)
BAR4.3174.0000.3170.041(0.032)
Lasso5.7834.0001.7830.187(0.071)
Alasso4.7563.9430.8130.141(0.125)
SCAD4.5263.9830.5430.191(0.081)
MCP4.5403.9630.5770.033(0.056)
appIC4.6173.9570.6600.061(0.061)

Note that Λ0(t) = t.

Table 7
Estimation results with different methods.
SizeTPFPMMSE(SD)
Λ0(t) = log(t + 1)
n = 100, p = 10
Full10.0006.0004.0000.854(1.140)
Oracle4.0004.0000.0000.090(0.149)
BAR4.2363.7530.4830.181(0.192)
Lasso5.4763.9431.5330.204(0.115)
Alasso4.7503.7271.0230.218(0.151)
SCAD4.4163.6660.7500.366(0.168)
MCP4.1463.5230.6230.252(0.290)
appIC3.9333.5370.3970.256(0.186)
n = 300, p = 10
Full10.0006.0004.0000.104(0.083)
Oracle4.0004.0000.0000.033(0.037)
BAR4.2034.0000.2030.033(0.036)
Lasso5.2534.0001.2530.070(0.045)
Alasso4.4643.9970.4670.062(0.063)
SCAD4.4003.9970.4030.081(0.056)
MCP4.2803.9930.2870.038(0.285)
appIC4.2633.9800.2830.047(0.047)
n = 300, p = 30
Full10.0006.0004.0000.304(0.322)
Oracle4.0004.0000.0000.028(0.038)
BAR4.1913.9940.1970.038(0.039)
Lasso6.3334.0002.3330.131(0.056)
Alasso4.6333.9200.7130.172(0.162)
SCAD4.5503.9870.5630.166(0.085)
MCP4.6103.9930.6170.040(0.057)
appIC4.8333.9770.8570.071(0.061)

Note that Λ0(t) = log(t + 1).

Application

In this section, the focus is on Nigerian child mortality data from the Demographic and Health Surveys (DHS) Program (https://www.dhsprogram.com). The dataset records women’s information from various aspects, including their children. The survey was very detailed; nevertheless, due to some practical restrictions the survival time of the children are only recorded in months or years, which makes it an interval-censored data structure. Meanwhile, it is found that the sample child mortality rate is over 20%, significantly higher than the global average, and our goal is to identify the key factors that affects children’s survival status.

A total of 24 potential factors are listed in Table 8. After excluding the samples that hold null value in either one of the variables, 8, 671 valid child samples are found, out of which 6, 830 are right censored. Note that variables 1-14 in Table 8 are dummy variables and are assigned 1 when the corresponding statements are true. Variables 15-24 are standardized so that the significance of all the factors can be compared. Most variables are concerned with the mother, and four variables marked with asterisks in Table 8 are child specific.

Table 8
Variables that possibly affect child mortality with six chosen by proposed sparse estimation method.
Variable numberFactorCoefficient
V1De facto place of residence-city-
V2De facto place of residence-countryside-
V3Has electricity-0.110
V4Has telephone-1.011
V5Presence of soap/ash/other cleansing agent in household-
V6Knowledge of ovulatory cycle-
V7Ever use of any modern contraception methods-0.394
V8Visited health facilities last 12m-
V9Smokes nothing-
V10Sex of household head-
V11When child is seriously ill, probably can not decide whether to get medical aid0.201
V12Child is twin*1.185
V13Sex of child*-
V14Proper body mass index-
V15Education in single years-
V16Number of household members-0.107
V17Number of children 5 and under-
V18Number of eligible women in HH-
V19Total children ever born0.241
V20Age of respondent at 1st birth-
V21Age at first marriage-
V22Ideal number of children (grp)-
V23Preceding birth interval*-0.440
V24Mother age of birth*-

Note that four variables marked with asterisks are child specific and variables 1-14 are dummy variables, assigned as 1 when corresponding statements are true.

To apply the appIC regression procedure here, m = 3, γ = 2, θ = n0 = 1841, and φ = log(n0) (BIC) are set. Meanwhile, the observation interval on the children is set as [0, 144]; that is to say, if the child lives to 144 months, or 12 years old, he or she is recorded as right censored. The results are shown in Table 8. According to the present research, eight variables have effects on child mortality. Having telephone, longer preceding birth interval of the mother, the usage of modern contraceptive methods, having electricity and more household members can reduce child mortality hazards, in the order of effectiveness. Meanwhile, the hazards increase with the mother having had more children and the child being twin. If the family can not decide whether to get medical aid when the child is seriously ill, the mortality hazards also increase. The baseline cumulative hazards function and baseline survival function are presented in Fig 5.

Estimated baseline cumulative hazards function and baseline survival function of children in survey.
Fig 5

Estimated baseline cumulative hazards function and baseline survival function of children in survey.

Note that t is measured in months. We see that the hazards function curve and survival function curve become flat as children grow up, which is consistent with reality.

Conclusion

In this paper, an approximated information criterion for the proportional hazards model under the interval-censored failure time data structure is discussed. The common penalties usually need a time-consuming hyperparameter tuning process, which is not necessary if one uses BSS with some well-known information criteria, such as BIC or AIC. The modified logistic sigmoid function is used herein to emulate the ℓ0 norm and accordingly convert the BIC as a penalized likelihood function that can be implemented in the way of regularizations. This method literally builds a bridge between BSS and the regularizations, with a special and novel strength in efficiency since it simulates the baseline hazards function, estimates coefficients of covariates, and chooses variables simultaneously, without tuning hyperparameters for the penalty term. The numerical results, including an application to child mortality, show that this method possesses great potential to facilitate mainstream sparse estimation for interval-censored data, with which the subject of variable selection is rarely studied.

There exist some interesting directions of planned future work. First, in this paper only the situation that the censoring time is independent of the failure time is considered, which sometimes may not conform with practice. Many studies discussing informative censoring exist and one can explore the proposed methods under that circumstance. The second direction is to change the survival model. Herein, only the proportional hazards model is applied, but several other superb semi-parameter survival models exist, e.g., the additive hazards model. One can compare the estimation accuracy or efficiency and show the reason. Third, in practice, some datasets with a very large covariate dimension are seen, a typical one of which is genetic data. Study on this problem is absolutely meaningful and clearly more research is needed in this direction.

Acknowledgements

Thanks are due to the Demographic and Health Surveys (DHS) Program for the application data.

References

DMFinkelstein. A proportional hazards model for interval-censored failure time data. Biometrics. 1986;42(4):845854. 10.2307/2530698

JHuang. Efficient estimation for the proportional hazards model with interval censoring. Ann Stat. 1996;24(2):540568. 10.1214/aos/1032894452

ZZhang, LSun, XZhao, JSun. Regression analysis of interval-censored failure time data with linear transformation models. Can J Stat-Rev Can Stat. 2005;33(1):6170. 10.1002/cjs.5540330105

DZeng, JCai, YShen. Semiparametric additive risks model for interval-censored data. Stat Sin. 2006; p. 287302.

XLin, LWang. A semiparametric probit model for case 2 interval-censored failure time data. Stat Med. 2010;29(9):972981. 10.1002/sim.3832

JSun. The statistical analysis of interval-censored failure time data. Springer; 2006.

DRCox. Regression models and life-tables. J R Stat Soc Ser B-Stat Methodol. 1972;34(2):187202.

XTong, MHChen, JSun. Regression Analysis of Multivariate Interval-Censored Failure Time Data with Application to Tumorigenicity Experiments. Biom J. 2008;50(3):364374. 10.1002/bimj.200710418

WBGoggins, DMFinkelstein. A proportional hazards model for multivariate interval-censored failure time data. Biometrics. 2000;56(3):940943. 10.1111/j.0006-341X.2000.00940.x

10 

DYLin, ZYing. Semiparametric Analysis of the Additive Risk Model. Biometrika. 1994;81(1):6171. 10.1093/biomet/81.1.61

11 

RTibshirani. The lasso method for variable selection in the Cox model. Stat Med. 1997;16(4):385395. 10.1002/(SICI)1097-0258(19970228)16:4<385::AID-SIM380>3.0.CO;2-3

12 

JFan, RLi. Variable selection for Cox’s proportional hazards model and frailty model. Ann Stat. 2002; p. 7499.

13 

HZou. The adaptive lasso and its oracle properties. J Am Stat Assoc. 2006;101(476):14181429. 10.1198/016214506000000735

14 

HHZhang, WLu. Adaptive Lasso for Cox’s proportional hazards model. Biometrika. 2007;94(3):691703. 10.1093/biomet/asm037

15 

YWu, RJCook. Penalized regression for interval-censored times of disease progression: Selection of HLA markers in psoriatic arthritis. Biometrics. 2015;71(3):782791. 10.1111/biom.12302

16 

CWang, QLi, XSong, XDong. Bayesian adaptive lasso for additive hazard regression with current status data. Stat Med. 2019;38(20):37033718. 10.1002/sim.8137

17 

HZhao, QWu, GLi, JSun. Simultaneous estimation and variable selection for interval-censored data with broken adaptive ridge regression. J Am Stat Assoc. 2020;115(529):204216. 10.1080/01621459.2018.1537922

18 

HAkaike. A new look at the statistical model identification. IEEE Trans Autom Control. 1974;19(6):716723. 10.1109/TAC.1974.1100705

19 

GSchwarz. Estimating the Dimension of a Model. Ann Stat. 1978;6(2):461464. 10.1214/aos/1176344136

20 

CHZhang, et al. Nearly unbiased variable selection under minimax concave penalty. Ann Stat. 2010;38(2):894942. 10.1214/09-AOS729

21 

XSu, JFan, RALevine, MENunn, CTsai. Sparse estimation of generalized linear models (GLM) via approximated information criteria. Stat Sin. 2018;28(3):15611581. 10.5705/ss.202016.0353

22 

XSu, CSWijayasinghe, JFan, YZhang. Sparse estimation of Cox proportional hazards models via approximated information criteria. Biometrics. 2016;72(3):751759. 10.1111/biom.12484

23 

JHJRossini. Sieve Estimation for the Proportional-Odds Failure-Time Regression Model with Interval Censoring. J Am Stat Assoc. 1997;92(439):960967. 10.1080/01621459.1997.10474050

24 

QZhou, THu, JSun. A Sieve Semiparametric Maximum Likelihood Approach for Regression Analysis of Bivariate Interval-Censored Failure Time Data. J Am Stat Assoc. 2016;112(518):664672. 10.1080/01621459.2016.1158113

25 

SNWood, NPya, BSfken. Smoothing parameter and model selection for general smooth models. J Am Stat Assoc. 2017;111(516):15481563. 10.1080/01621459.2016.1180986

26 

CWang, QLi, XSong, XDong. Bayesian adaptive lasso for additive hazard regression with current status data. Stat Med. 2019;38(20). 10.1002/sim.8137

27 

JWang, SKGhosh. Shape restricted nonparametric regression with Bernstein polynomials. Comput Stat Data Anal. 2012. 10.1016/j.csda.2012.02.018

28 

JFan, RLi. Variable selection via nonconcave penalized likelihood and its oracle properties. J Am Stat Assoc. 2001;96(456):13481360. 10.1198/016214501753382273

29 

CJPBelisle. Convergence Theorems for a Class of Simulated Annealing Algorithms. J Appl Probab. 1992;29:885895. 10.1017/S002190020004376X

30 

JNocedal. Updating quasi-Newton matrices with limited storage. Math. Comp. 1980;35:773782. 10.1090/S0025-5718-1980-0572855-7