首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
This article concerns the problem of estimating a continuous distribution in a diseased or nondiseased population when only group-based test results on the disease status are available. The problem is challenging in that individual disease statuses are not observed and testing results are often subject to misclassification, with further complication that the misclassification may be differential as the group size and the number of the diseased individuals in the group vary. We propose a method to construct nonparametric estimation of the distribution and obtain its asymptotic properties. The performance of the distribution estimator is evaluated under various design considerations concerning group sizes and classification errors. The method is exemplified with data from the National Health and Nutrition Examination Survey study to estimate the distribution and diagnostic accuracy of C-reactive protein in blood samples in predicting chlamydia incidence.  相似文献   

2.
When the sample size is not large or when the underlying disease is rare, to assure collection of an appropriate number of cases and to control the relative error of estimation, one may employ inverse sampling, in which one continues sampling subjects until one obtains exactly the desired number of cases. This paper focuses discussion on interval estimation of the simple difference between two proportions under independent inverse sampling. This paper develops three asymptotic interval estimators on the basis of the maximum likelihood estimator (MLE), the uniformly minimum variance unbiased estimator (UMVUE), and the asymptotic likelihood ratio test (ALRT). To compare the performance of these three estimators, this paper calculates the coverage probability and the expected length of the resulting confidence intervals on the basis of the exact distribution. This paper finds that when the underlying proportions of cases in both two comparison populations are small or moderate (≤0.20), all three asymptotic interval estimators developed here perform reasonably well even for the pre-determined number of cases as small as 5. When the pre-determined number of cases is moderate or large (≥50), all three estimators are essentially equivalent in all the situations considered here. Because application of the two interval estimators derived from the MLE and the UMVUE does not involve any numerical iterative procedure needed in the ALRT, for simplicity we may use these two estimators without losing efficiency.  相似文献   

3.
Qu A  Li R 《Biometrics》2006,62(2):379-391
Nonparametric smoothing methods are used to model longitudinal data, but the challenge remains to incorporate correlation into nonparametric estimation procedures. In this article, we propose an efficient estimation procedure for varying-coefficient models for longitudinal data. The proposed procedure can easily take into account correlation within subjects and deal directly with both continuous and discrete response longitudinal data under the framework of generalized linear models. The proposed approach yields a more efficient estimator than the generalized estimation equation approach when the working correlation is misspecified. For varying-coefficient models, it is often of interest to test whether coefficient functions are time varying or time invariant. We propose a unified and efficient nonparametric hypothesis testing procedure, and further demonstrate that the resulting test statistics have an asymptotic chi-squared distribution. In addition, the goodness-of-fit test is applied to test whether the model assumption is satisfied. The corresponding test is also useful for choosing basis functions and the number of knots for regression spline models in conjunction with the model selection criterion. We evaluate the finite sample performance of the proposed procedures with Monte Carlo simulation studies. The proposed methodology is illustrated by the analysis of an acquired immune deficiency syndrome (AIDS) data set.  相似文献   

4.
Leveraging information in aggregate data from external sources to improve estimation efficiency and prediction accuracy with smaller scale studies has drawn a great deal of attention in recent years. Yet, conventional methods often either ignore uncertainty in the external information or fail to account for the heterogeneity between internal and external studies. This article proposes an empirical likelihood-based framework to improve the estimation of the semiparametric transformation models by incorporating information about the t-year subgroup survival probability from external sources. The proposed estimation procedure incorporates an additional likelihood component to account for uncertainty in the external information and employs a density ratio model to characterize population heterogeneity. We establish the consistency and asymptotic normality of the proposed estimator and show that it is more efficient than the conventional pseudopartial likelihood estimator without combining information. Simulation studies show that the proposed estimator yields little bias and outperforms the conventional approach even in the presence of information uncertainty and heterogeneity. The proposed methodologies are illustrated with an analysis of a pancreatic cancer study.  相似文献   

5.
We explore a hierarchical generalized latent factor model for discrete and bounded response variables and in particular, binomial responses. Specifically, we develop a novel two-step estimation procedure and the corresponding statistical inference that is computationally efficient and scalable for the high dimension in terms of both the number of subjects and the number of features per subject. We also establish the validity of the estimation procedure, particularly the asymptotic properties of the estimated effect size and the latent structure, as well as the estimated number of latent factors. The results are corroborated by a simulation study and for illustration, the proposed methodology is applied to analyze a dataset in a gene–environment association study.  相似文献   

6.
Quantitative determination of protein using the binding of Coomassie Brilliant Blue G-250 was investigated with respect to interference with the density gradient material metrizamide, and compared with the corresponding interference using the Lowry method. The background absorption obtained with metrizamide in the absence of protein was less than 10% of that obtained with the Lowry method. In the presence of 0–4% metrizamide, parallel standard curves were obtained with 0–67 μg of protein in the samples. The curves overlapped in the range 0–40 μg of protein when metrizamide was included in the blanks. With up to 2% final concentration of metrizamide in the assay, the curves overlapped at all protein concentrations tested (0–67 μg). Correction for metrizamide interference is thus a simple procedure and a precise estimation of the metrizamide concentration is less critical than when the Lowry assay is used. The method is well suited for quantitation of protein in samples collected from metrizamide grandients.  相似文献   

7.
流域生物信息流是流域生态学研究中的重要内容,是流域生态系统中的物质输移和能量输移过程的信息标记,是用eDNA技术调查评估河流水体中物种组成空间特征的基础。估算流域生物信息流是流域生态系统过程研究和eDNA技术调查评估河流水体中物种组成空间特征的关键。在有限的调查采样中,平行样的数量如何影响流域生物信息流的估算,尚待解答。基于随机抽样调查的基本原理,提出假设--采样数量不影响流域生物信息流估算结果的准确度,但会影响其精密度,然后通过问题简化转化和模拟计算,对该假设进行了检验。模拟计算结果显示,随着样点生物信息检出度(平行样数量)的增大,流域生物信息流估算结果会从偏小逐渐靠近流域生物信息流实际值,同时其99.9%置信区间也逐渐集中于流域生物信息流实际值。即样点生物信息检出度(平行样数量)对流域生物信息流估算的准确度和精密度均有影响。在实际调查研究过程中,建议先在所研究区域对平行样数量和样点生物信息检出度的关系进行预评估,然后基于流域生物信息流估算可信度目标在正式实施方案中经济有效地设置平行样,基于多平行样调查结果估算流域生物信息流,再根据各样点生物信息检出状况对流域生物信息流估算结果进行后验评估。  相似文献   

8.
In a bioassay, under certain experimental circumstances, information on concentration (dose rate) and time to response for some subjects can be combined in a single analysis. An underlying logistic random variable is assumed and the resulting mixed- (continuous-quantal) response model is analyzed by likelihood methods. The estimation procedure for the mean and the variance is described, and expressions for asymptotic variances are obtained. A comparison of results from the mixed model and from the standard quantal-response model shows that there is a substantial reduction in the variance of the estimators for the mixed model. On the basis of the table of asymptotic variances, some design implications are discussed. An example from insect pheromone research is used to illustrate the main ideas.  相似文献   

9.
Summary Many well‐known methods are available for estimating the number of species in a forest community. However, most existing methods result in considerable negative bias in applications, where field surveys typically represent only a small fraction of sampled communities. This article develops a new method based on sampling with replacement to estimate species richness via the generalized jackknife procedure. The proposed estimator yields small bias and reasonably accurate interval estimation even with small samples. The performance of the proposed estimator is compared with several typical estimators via simulation study using two complete census datasets from Panama and Malaysia.  相似文献   

10.
In preparing intracellular microbial samples for one- or two-dimensional electrophoresis, trichloroacetic acid (TCA) precipitation is frequently used to remove interfering compounds. Solubilization of TCA precipitate typically requires the addition of a number of chaotropes or detergents, in a multistep process, that requires hours to carry out. In this study, a simple, rapid, one-step method to solubilize TCA precipitated proteins is presented. Precipitated proteins are pretreated with 0.2 M NaOH for less than 5 min, followed by addition of standard sample solubilization buffer (SSSB). When compared to solubilization with SSSB alone, NaOH pretreatment of TCA-precipitated intracellular protein from Aspergillus oryzae and Escherichia coli shows an approximate 5-fold increase in soluble protein. In addition, two-dimensional gel electrophoresis on resolubilized proteins shows an equivalent number of proteins in samples with and without NaOH pretreatment.  相似文献   

11.
Statistical analysis of longitudinal data often involves modeling treatment effects on clinically relevant longitudinal biomarkers since an initial event (the time origin). In some studies including preventive HIV vaccine efficacy trials, some participants have biomarkers measured starting at the time origin, whereas others have biomarkers measured starting later with the time origin unknown. The semiparametric additive time-varying coefficient model is investigated where the effects of some covariates vary nonparametrically with time while the effects of others remain constant. Weighted profile least squares estimators coupled with kernel smoothing are developed. The method uses the expectation maximization approach to deal with the censored time origin. The Kaplan–Meier estimator and other failure time regression models such as the Cox model can be utilized to estimate the distribution and the conditional distribution of left censored event time related to the censored time origin. Asymptotic properties of the parametric and nonparametric estimators and consistent asymptotic variance estimators are derived. A two-stage estimation procedure for choosing weight is proposed to improve estimation efficiency. Numerical simulations are conducted to examine finite sample properties of the proposed estimators. The simulation results show that the theory and methods work well. The efficiency gain of the two-stage estimation procedure depends on the distribution of the longitudinal error processes. The method is applied to analyze data from the Merck 023/HVTN 502 Step HIV vaccine study.  相似文献   

12.
Bootstrap confidence intervals for adaptive cluster sampling   总被引:2,自引:0,他引:2  
Consider a collection of spatially clustered objects where the clusters are geographically rare. Of interest is estimation of the total number of objects on the site from a sample of plots of equal size. Under these spatial conditions, adaptive cluster sampling of plots is generally useful in improving efficiency in estimation over simple random sampling without replacement (SRSWOR). In adaptive cluster sampling, when a sampled plot meets some predefined condition, neighboring plots are added to the sample. When populations are rare and clustered, the usual unbiased estimators based on small samples are often highly skewed and discrete in distribution. Thus, confidence intervals based on asymptotic normal theory may not be appropriate. We investigated several nonparametric bootstrap methods for constructing confidence intervals under adaptive cluster sampling. To perform bootstrapping, we transformed the initial sample in order to include the information from the adaptive portion of the sample yet maintain a fixed sample size. In general, coverages of bootstrap percentile methods were closer to nominal coverage than the normal approximation.  相似文献   

13.
The objective of this article is to propose an algorithm for the on-line estimation of the specific growth rate in a batch or a fed-batch fermentation process. The algorithm shows the practical procedure for the estimation method utilizing the macroscopic balance and the extended Kalman filter. A number of studies of the on line estimation have been presented. However, there are few studies discussing about the selection of the observed variables and for the tuning of some parameters of the extended Kalman filter, such as covariance matrix and initial values of the state.The beginning of this article is devoted to explain the selection of the observed variable. This information is very important in terms of the practical know-how for using technique. It is discovered that the condition number is a practically useful and valid criterion for number is a practically useful and valid criterion for choosing the variable to be observed.Next, when the extended Kalman filter in applied to the online estimation of the specific growth rate, which is directly unmeasurable, criteria for judging the validity of the estimated value from the observed data are proposed. Based on the proposed criterial, the system equation of the specific growth rate is selected and initial value of the state variable and covariance matrix of the system noises are adjusted. From many experiments, it is certified that the specific growth rate in the batch or fed -batch fermentation can be estimated accurately by means of the algorithm proposed here. In these experiments, that is, when the cell concentration is measured directly, the extended Kalman filter using the convariance matrix with a constant element can estimate more accurately values of the specific growth rate than the adaptive extended Kalman filter does.  相似文献   

14.
The relative risk (RR) is one of the most frequently used indices to measure the strength of association between a disease and a risk factor in etiological studies or the efficacy of an experimental treatment in clinical trials. In this paper, we concentrate attention on interval estimation of RR for sparse data, in which we have only a few patients per stratum, but a moderate or large number of strata. We consider five asymptotic interval estimators for RR, including a weighted least-squares (WLS) interval estimator with an ad hoc adjustment procedure for sparse data, an interval estimator proposed elsewhere for rare events, an interval estimator based on the Mantel-Haenszel (MH) estimator with a logarithmic transformation, an interval estimator calculated from a quadratic equation, and an interval estimator derived from the ratio estimator with a logarithmic transformation. On the basis of Monte Carlo simulations, we evaluate and compare the performance of these five interval estimators in a variety of situations. We note that, except for the cases in which the underlying common RR across strata is around 1, using the WLS interval estimator with the adjustment procedure for sparse data can be misleading. We note further that using the interval estimator suggested elsewhere for rare events tends to be conservative and hence leads to loss of efficiency. We find that the other three interval estimators can consistently perform well even when the mean number of patients for a given treatment is approximately 3 patients per stratum and the number of strata is as small as 20. Finally, we use a mortality data set comparing two chemotherapy treatments in patients with multiple myeloma to illustrate the use of the estimators discussed in this paper.  相似文献   

15.
The Coomassie brilliant blue assay for the determination of protein has been extended to rapidly and conveniently measure the protein concentration of cells growing in culture in a 96-well microtiter format. Modifications of the standard assay include sodium hydroxide to solubilize the cells and ovalbumin, instead of bovine serum albumin, as a protein standard. The procedure allows a large number of small samples to be assayed simultaneously. Two examples of its use, enzyme-specific activity and drug resistance, are shown. An assay for acetylcholinesterase activity in the same culture plate is demonstrated. G418, an inhibitor of cell protein synthesis, is frequently used to select for cells transfected with the neomycin resistance gene. The required concentration of G418 can be easily determined with this protein assay.  相似文献   

16.
Clustered interval-censored failure time data occur when the failure times of interest are clustered into small groups and known only to lie in certain intervals. A number of methods have been proposed for regression analysis of clustered failure time data, but most of them apply only to clustered right-censored data. In this paper, a sieve estimation procedure is proposed for fitting a Cox frailty model to clustered interval-censored failure time data. In particular, a two-step algorithm for parameter estimation is developed and the asymptotic properties of the resulting sieve maximum likelihood estimators are established. The finite sample properties of the proposed estimators are investigated through a simulation study and the method is illustrated by the data arising from a lymphatic filariasis study.  相似文献   

17.
Problems of establishing equivalence or noninferiority between two medical diagnostic procedures involve comparisons of the response rates between correlated proportions. When the sample size is small, the asymptotic tests may not be reliable. This article proposes an unconditional exact test procedure to assess equivalence or noninferiority. Two statistics, a sample-based test statistic and a restricted maximum likelihood estimation (RMLE)-based test statistic, to define the rejection region of the exact test are considered. We show the p-value of the proposed unconditional exact tests can be attained at the boundary point of the null hypothesis. Assessment of equivalence is often based on a comparison of the confidence limits with the equivalence limits. We also derive the unconditional exact confidence intervals on the difference of the two proportion means for the two test statistics. A typical data set of comparing two diagnostic procedures is analyzed using the proposed unconditional exact and asymptotic methods. The p-value from the unconditional exact tests is generally larger than the p-value from the asymptotic tests. In other words, an exact confidence interval is generally wider than the confidence interval obtained from an asymptotic test.  相似文献   

18.
Procedures for comparing samples with multiple endpoints   总被引:18,自引:0,他引:18  
P C O'Brien 《Biometrics》1984,40(4):1079-1087
Five procedures are considered for the comparison of two or more multivariate samples. These procedures include a newly proposed nonparametric rank-sum test and a generalized least squares test. Also considered are the following tests: ordinary least squares, Hotelling's T2, and a Bonferroni per-experiment error-rate approach. Applications are envisaged in which each variable represents a qualitatively different measure of response to treatment. The null hypothesis of no treatment difference is tested with power directed towards alternatives in which at least one treatment is uniformly better than the others. In all simulations the nonparametric procedure provided relatively good power and accurate control over the size of the test, and is recommended for general use. Alternatively, the generalized least squares procedure may also be useful with normally distributed data in moderate or large samples. A convenient expression for this procedure is obtained and its asymptotic relative efficiency with respect to the ordinary least squares test is evaluated.  相似文献   

19.
Wang L  Li R 《Biometrics》2009,65(2):564-571
Summary .  Shrinkage-type variable selection procedures have recently seen increasing applications in biomedical research. However, their performance can be adversely influenced by outliers in either the response or the covariate space. This article proposes a weighted Wilcoxon-type smoothly clipped absolute deviation (WW-SCAD) method, which deals with robust variable selection and robust estimation simultaneously. The new procedure can be conveniently implemented with the statistical software R . We establish that the WW-SCAD correctly identifies the set of zero coefficients with probability approaching one and estimates the nonzero coefficients with the rate   n −1/2  . Moreover, with appropriately chosen weights the WW-SCAD is robust with respect to outliers in both the x and y directions. The important special case with constant weights yields an oracle-type estimator with high efficiency in the presence of heavier-tailed random errors. The robustness of the WW-SCAD is partly justified by its asymptotic performance under local shrinking contamination. We propose a Bayesian information criterion type tuning parameter selector for the WW-SCAD. The performance of the WW-SCAD is demonstrated via simulations and by an application to a study that investigates the effects of personal characteristics and dietary factors on plasma beta-carotene level.  相似文献   

20.
Cheung YK 《Biometrics》2005,61(2):524-531
When comparing follow-up measurements from two independent populations, missing records may arise due to censoring by events whose occurrence is associated with baseline covariates. In these situations, inferences based only on the completely followed observations may be biased if the follow-up measurements and the covariates are correlated. This article describes exact inference for a class of modified U-statistics under covariate-dependent dropouts. The method involves weighing each permutation according to the retention probabilities, and thus requires estimation of the missing data mechanism. The proposed procedure is nonparametric in that no distributional assumption is necessary for the outcome variables and the missingness patterns. Monte Carlo approximation by the Gibbs sampler is proposed, and is shown to be fast and accurate via simulation. The method is illustrated in two small data sets for which asymptotic inferential procedures may not be appropriate.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号