首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
In some applications of survival analysis with covariates, the commonly used semiparametric assumptions (e.g., proportional hazards) may turn out to be stringent and unrealistic, particularly when there is scientific background to believe that survival curves under different covariate combinations will cross during the study period. We present a new nonparametric regression model for the conditional hazard rate using a suitable sieve of Bernstein polynomials. The proposed nonparametric methodology has three key features: (i) the smooth estimator of the conditional hazard rate is shown to be a unique solution of a strictly convex optimization problem for a wide range of applications; making it computationally attractive, (ii) the model is shown to encompass a proportional hazards structure, and (iii) large sample properties including consistency and convergence rates are established under a set of mild regularity conditions. Empirical results based on several simulated data scenarios indicate that the proposed model has reasonably robust performance compared to other semiparametric models particularly when such semiparametric modeling assumptions are violated. The proposed method is further illustrated on the gastric cancer data and the Veterans Administration lung cancer data.  相似文献   

2.
A general technique is proposed for efficient computation of the nonparametric maximum likelihood estimate (NPMLE) of a survival function. The main idea is to include a new support interval that has the largest gradient value between inclusively every two neighbouring support intervals in the support set at each iteration. It is thus able to expand the support set exponentially fast during the initial stage of computation and tends to produce the same support set of the NPMLE afterward. The use of the proposed technique needs to be combined with an algorithm that can effectively find and remove redundant support intervals, for example, the constrained Newton method, the iterative convex minorant algorithm and the subspace-based Newton method. Numerical studies show that the dimension-reducing technique works very well, especially for purely interval-censored data, where a significant computational improvement via dimension reduction is possible. Strengths and weaknesses of various algorithms are also discussed and demonstrated.  相似文献   

3.
The Cox model with frailties has been popular for regression analysis of clustered event time data under right censoring. However, due to the lack of reliable computation algorithms, the frailty Cox model has been rarely applied to clustered current status data, where the clustered event times are subject to a special type of interval censoring such that we only observe for each event time whether it exceeds an examination (censoring) time or not. Motivated by the cataract dataset from a cross-sectional study, where bivariate current status data were observed for the occurrence of cataracts in the right and left eyes of each study subject, we develop a very efficient and stable computation algorithm for nonparametric maximum likelihood estimation of gamma-frailty Cox models with clustered current status data. The algorithm proposed is based on a set of self-consistency equations and the contraction principle. A convenient profile-likelihood approach is proposed for variance estimation. Simulation and real data analysis exhibit the nice performance of our proposal.  相似文献   

4.
A competing risks model based on Lomax distributions is considered under progressive Type-II censoring. Maximum likelihood estimates for the distribution parameters are established. Moreover, the expected Fisher information matrix is computed and optimal Fisher information based censoring plans are discussed. In particular, it turns out that the optimal censoring scheme depends on the particular parametrization of the Lomax distributions.  相似文献   

5.
By executing different fingerprint-image matching algorithms on large data sets, it reveals that the match and non-match similarity scores have no specific underlying distribution function. Thus, it requires a nonparametric analysis for fingerprint-image matching algorithms on large data sets without any assumption about such irregularly discrete distribution functions. A precise receiver operating characteristic (ROC) curve based on the true accept rate (TAR) of the match similarity scores and the false accept rate (FAR) of the non-match similarity scores can be constructed. The area under such an ROC curve computed using the trapezoidal rule is equivalent to the Mann-Whitney statistic directly formed from the match and non-match similarity scores. Thereafter, the Z statistic formulated using the areas under ROC curves along with their variances and the correlation coefficient is applied to test the significance of the difference between two ROC curves. Four examples from the extensive testing of commercial fingerprint systems at the National Institute of Standards and Technology are provided. The nonparametric approach presented in this article can also be employed in the analysis of other large biometric data sets.  相似文献   

6.
Recently, a nonparametric marginal structural model (NPMSM) approach to Causal Inference has been proposed [Neugebauer, R., van der Laan, M., 2006. Nonparametric causal effects based on marginal structural models. J. Statist. Plann. Inference (in press), 〈www http://www.sciencedirect.com/science/journal/03783758〉.] as an appealing practical alternative to the original parametric MSM (PMSM) approach introduced by Robins [Robins, J., 1998a. Marginal structural models. In: 1997 Proceedings of the American Statistical Association, American Statistical Association, Alexandria, VA, pp. 1-10]. The new MSM-based causal inference methodology generalizes the concept of causal effects: the proposed nonparametric causal effects are interpreted as summary measures of the causal effects defined with PMSMs. In addition, causal inference with NPMSM does not rely on the assumed correct specification of a parametric MSM but instead defines causal effects based on a user-specified working causal model which can be willingly misspecified. The NPMSM approach was developed for studies with point treatment data or with longitudinal data where the outcome is not time-dependent (typically collected at the end of data collection). In this paper, we generalize this approach to longitudinal studies where the outcome is time-dependent, i.e. collected throughout the span of the studies, and address the subsequent estimation inconsistency which could easily arise from a hasty generalization of the algorithm for maximum likelihood estimation. More generally, we provide an overview of the multiple causal effect representations which have been developed based on MSMs in longitudinal studies.  相似文献   

7.
The problems arising when there are outliers in a data set that follow a circular distribution are considered. A robust estimation of the unknown parameters is obtained using the methods of weighted likelihood and minimum disparity, each of which is defined for a general parametric family of circular data. The class of power divergence and the related residual adjustment function is investigated in order to improve the performance of the two methods which are studied for the Von Mises (circular normal) and the Wrapped Normal distributions. The techniques are illustrated via two examples based on a real data set and a Monte Carlo study, which also enables the discussion of various computational aspects.  相似文献   

8.
The performance of model based bootstrap methods for constructing point-wise confidence intervals around the survival function with interval censored data is investigated. It is shown that bootstrapping from the nonparametric maximum likelihood estimator of the survival function is inconsistent for the current status model. A model based smoothed bootstrap procedure is proposed and proved to be consistent. In fact, a general framework for proving the consistency of any model based bootstrap scheme in the current status model is established. In addition, simulation studies are conducted to illustrate the (in)-consistency of different bootstrap methods in mixed case interval censoring. The conclusions in the interval censoring model would extend more generally to estimators in regression models that exhibit non-standard rates of convergence.  相似文献   

9.
The Lindley distribution applied to competing risks lifetime data   总被引:1,自引:0,他引:1  
Competing risks data usually arises in studies in which the death or failure of an individual or an item may be classified into one of k ≥ 2 mutually exclusive causes. In this paper a simple competing risks distribution is proposed as a possible alternative to the Exponential or Weibull distributions usually considered in lifetime data analysis. We consider the case when the competing risks have a Lindley distribution. Also, we assume that the competing events are uncorrelated and that each subject can experience only one type of event at any particular time.  相似文献   

10.
In reliability analysis, accelerated life-testing allows for gradual increment of stress levels on test units during an experiment. In a special class of accelerated life tests known as step-stress tests, the stress levels increase discretely at pre-fixed time points, and this allows the experimenter to obtain information on the parameters of the lifetime distributions more quickly than under normal operating conditions. Moreover, when a test unit fails, there are often more than one fatal cause for the failure, such as mechanical or electrical. In this article, we consider the simple step-stress model under time constraint when the lifetime distributions of the different risk factors are independently exponentially distributed. Under this setup, we derive the maximum likelihood estimators (MLEs) of the unknown mean parameters of the different causes under the assumption of a cumulative exposure model. Since it is found that the MLEs do not exist when there is no failure by any particular risk factor within the specified time frame, the exact sampling distributions of the MLEs are derived through the use of conditional moment generating functions. Using these exact distributions as well as the asymptotic distributions, the parametric bootstrap method, and the Bayesian posterior distribution, we discuss the construction of confidence intervals and credible intervals for the parameters. Their performance is assessed through Monte Carlo simulations and finally, we illustrate the methods of inference discussed here with an example.  相似文献   

11.
王建宏 《控制与决策》2014,29(2):358-362
针对仿射结构形式在丢失数据下的条件极大似然辨识问题, 首先引入交换矩阵将原随机矢量分解成观测和丢失部分; 然后确定出观测数据在丢失数据下的条件均值和条件方差, 以此建立条件似然函数; 进而从理论上给出了条件极大似然函数关于未知参数矢量、未知白噪声方差值和丢失数据的求导公式, 并从工程上给出一种可分离的优化算法; 最后通过仿真算例验证了该辨识方法的有效性.  相似文献   

12.
For multivariable equation-error systems with an autoregressive moving average noise, this paper applies the decomposition technique to transform a multivariable model into several identification sub-models based on the number of the system outputs, and derives a data filtering and maximum likelihood-based recursive least-squares algorithm to reduce the computation complexity and improve the parameter estimation accuracy. A multivariable recursive generalised extended least-squares method and a filtering-based recursive extended least-squares method are presented to show the effectiveness of the proposed algorithm. The simulation results indicate that the proposed method is effective and can produce more accurate parameter estimates than the compared methods.  相似文献   

13.
Population models are used to describe the dynamics of different subjects belonging to a population and play an important role in drug pharmacokinetics. A nonparametric identification scheme is proposed in which both the average impulse response of the population and the individual ones are modelled as Gaussian stochastic processes. Assuming that the average curve is an integrated Wiener process, it is shown that its estimate is a cubic spline. An empirical Bayes algorithm for estimating both the average and the individual curves is worked out. The model is tested on simulated data sets as well as on xenobiotics pharmacokinetic data.  相似文献   

14.
This paper extends to multi-input multi-output (MIMO) systems the results on frequency response estimation for normalized coprime factors (NCF) under a stochastic framework from closed-loop frequency domain experimental data. Under the condition that the covariance matrices are available for the external disturbances and measurement errors, an analytic solution has been obtained for the maximum likelihood estimate (MLE), on the basis of a linear fractional transformation (LFT) representation for all the plant possible normalized right coprime factors (NRCF). It is proved that the estimate can be expressed by a linear combination of a normalized random matrix with all its columns having independent complex normal distributions. Some methods are suggested to reduce the estimation bias when high-quality experimental data can be obtained. A numerical example is included that confirms the theoretical results.  相似文献   

15.
Semiparametric methods for longitudinal data with dependence within subjects have recently received considerable attention. Existing approaches that focus on modeling the mean structure require a correct specification of the covariance structure as misspecified covariance structures may lead to inefficient or biased mean parameter estimates. Besides, computation and estimation problems arise when the repeated measurements are taken at irregular and possibly subject-specific time points, the dimension of the covariance matrix is large, and the positive definiteness of the covariance matrix is required. In this article, we propose a profile kernel approach based on semiparametric partially linear regression models for the mean and model covariance structures simultaneously, motivated by the modified Cholesky decomposition. We also study the large-sample properties of the parameter estimates. The proposed method is evaluated through simulation and applied to a real dataset. Both theoretical and empirical results indicate that properly taking into account the within-subject correlation among the responses using our method can substantially improve efficiency.  相似文献   

16.
Trace contaminants in water, including metals and organics, often are measured at sufficiently low concentrations to be reported only as values below the instrument detection limit. Interpretation of these “less thans” is complicated when multiple detection limits occur. Statistical methods for multiply censored, or multiple-detection limit, datasets have been developed for medical and industrial statistics, and can be employed to estimate summary statistics or model the distributions of trace-level environmental data.We describe S-language-based software tools that perform robust linear regression on order statistics (ROS). The ROS method has been evaluated as one of the most reliable procedures for developing summary statistics of multiply censored data. It is applicable to any dataset that has 0 to 80% of its values censored. These tools are a part of a software library, or add-on package, for the R environment for statistical computing. This library can be used to generate ROS models and associated summary statistics, plot modeled distributions, and predict exceedance probabilities of water-quality standards.  相似文献   

17.
One of the features inherent in nested Archimedean copulas, also called hierarchical Archimedean copulas, is their rooted tree structure. A nonparametric, rank-based method to estimate this structure is presented. The idea is to represent the target structure as a set of trivariate structures, each of which can be estimated individually with ease. Indeed, for any three variables there are only four possible rooted tree structures and, based on a sample, a choice can be made by performing comparisons between the three bivariate margins of the empirical distribution of the three variables. The set of estimated trivariate structures can then be used to build an estimate of the target structure. The advantage of this estimation method is that it does not require any parametric assumptions concerning the generator functions at the nodes of the tree.  相似文献   

18.
Hazard function plays an important role in reliability and survival analysis. In some real life applications, abrupt changes in the hazard function may be observed and it is of interest to detect the location and the size of the change. Hazard models with a change-point are considered when the observations are subject to random left truncation and right censoring. For a piecewise constant hazard function with a single change-point, two estimation methods based on the maximum likelihood ideas are considered. The first method assumes parametric families of distributions for the censoring and truncation variables, whereas the second one is based on conditional likelihood approaches. A simulation study is carried out to illustrate the performances of the proposed estimators. The results indicate that the fully parametric method performs better especially for estimating the size of the change, however the difference between the two methods vanish as the sample size increases. It is also observed that the full likelihood approach is not robust to model misspecification.  相似文献   

19.
When considering competing risks survival data, the cause specific hazard functions are often modelled by the proportional hazards Cox regression model. First, we present how to estimate the parameters in this model when some of the covariates are allowed to have exactly the same effect on several causes of failure. In many cases, the focus is not on the parameter estimates, but rather on the probability of observing a failure from a specific cause for individuals with specified covariate values. These probabilities, the cumulative incidences, are not simple functions of the parameters and they are, so far, not provided by the standard statistical software packages. We present two SAS macros: a SAS macro named CumInc for estimation of the cumulative incidences and a SAS macro named CumIncV for estimation of the cumulative incidences and the variances of the estimated cumulative incidences. The use of the macros is demonstrated through an example.  相似文献   

20.
Multivariate methods often rely on a sample covariance matrix. The conventional estimators of a covariance matrix require complete data vectors on all subjects—an assumption that can frequently not be met. For example, in many fields of life sciences that are utilizing modern measuring technology, such as mass spectrometry, left-censored values caused by denoising the data are a commonplace phenomena. Left-censored values are low-level concentrations that are considered too imprecise to be reported as a single number but known to exist somewhere between zero and the laboratory’s lower limit of detection. Maximum likelihood-based covariance matrix estimators that allow the presence of the left-censored values without substituting them with a constant or ignoring them completely are considered. The presented estimators efficiently use all the information available and thus, based on simulation studies, produce the least biased estimates compared to often used competing estimators. As the genuine maximum likelihood estimate can be solved fast only in low dimensions, it is suggested to estimate the covariance matrix element-wise and then adjust the resulting covariance matrix to achieve positive semi-definiteness. It is shown that the new approach succeeds in decreasing the computation times substantially and still produces accurate estimates. Finally, as an example, a left-censored data set of toxic chemicals is explored.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号