首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
Modelling digit preference in fecundability studies.   总被引:4,自引:0,他引:4  
M S Ridout  B J Morgan 《Biometrics》1991,47(4):1423-1433
Retrospective studies of fecundability, in which women are asked how many cycles they required to become pregnant, are often affected by problems of digit preference. A probability model for such digit preference is proposed in which misreporting favours 6 or 12 (and possibly also 3) cycles. It is assumed that in the absence of misreporting the number of cycles follows a beta-geometric distribution. The model is applied to two data sets, with clear-cut results: The inclusion of additional parameters to model the misreporting can lead to substantial improvements in fit, but causes little change to the estimated parameters of the underlying beta-geometric distribution. In some cases misreporting parameters may be regarded as nuisance parameters, while in others they may be of interest. We have found estimates of these parameters to vary between different categories of women in an interpretable manner. The models may also be used to estimate the percentage of couples in any study that misreport their conception waiting time.  相似文献   

2.
We consider longitudinal studies in which the outcome observed over time is binary and the covariates of interest are categorical. With no missing responses or covariates, one specifies a multinomial model for the responses given the covariates and uses maximum likelihood to estimate the parameters. Unfortunately, incomplete data in the responses and covariates are a common occurrence in longitudinal studies. Here we assume the missing data are missing at random (Rubin, 1976, Biometrika 63, 581-592). Since all of the missing data (responses and covariates) are categorical, a useful technique for obtaining maximum likelihood parameter estimates is the EM algorithm by the method of weights proposed in Ibrahim (1990, Journal of the American Statistical Association 85, 765-769). In using the EM algorithm with missing responses and covariates, one specifies the joint distribution of the responses and covariates. Here we consider the parameters of the covariate distribution as a nuisance. In data sets where the percentage of missing data is high, the estimates of the nuisance parameters can lead to highly unstable estimates of the parameters of interest. We propose a conditional model for the covariate distribution that has several modeling advantages for the EM algorithm and provides a reduction in the number of nuisance parameters, thus providing more stable estimates in finite samples.  相似文献   

3.
Many late-phase clinical trials recruit subjects at multiple study sites. This introduces a hierarchical structure into the data that can result in a power-loss compared to a more homogeneous single-center trial. Building on a recently proposed approach to sample size determination, we suggest a sample size recalculation procedure for multicenter trials with continuous endpoints. The procedure estimates nuisance parameters at interim from noncomparative data and recalculates the sample size required based on these estimates. In contrast to other sample size calculation methods for multicenter trials, our approach assumes a mixed effects model and does not rely on balanced data within centers. It is therefore advantageous, especially for sample size recalculation at interim. We illustrate the proposed methodology by a study evaluating a diabetes management system. Monte Carlo simulations are carried out to evaluate operation characteristics of the sample size recalculation procedure using comparative as well as noncomparative data, assessing their dependence on parameters such as between-center heterogeneity, residual variance of observations, treatment effect size and number of centers. We compare two different estimators for between-center heterogeneity, an unadjusted and a bias-adjusted estimator, both based on quadratic forms. The type 1 error probability as well as statistical power are close to their nominal levels for all parameter combinations considered in our simulation study for the proposed unadjusted estimator, whereas the adjusted estimator exhibits some type 1 error rate inflation. Overall, the sample size recalculation procedure can be recommended to mitigate risks arising from misspecified nuisance parameters at the planning stage.  相似文献   

4.
We are interested in the estimation of average treatment effects based on right-censored data of an observational study. We focus on causal inference of differences between t-year absolute event risks in a situation with competing risks. We derive doubly robust estimation equations and implement estimators for the nuisance parameters based on working regression models for the outcome, censoring, and treatment distribution conditional on auxiliary baseline covariates. We use the functional delta method to show that these estimators are regular asymptotically linear estimators and estimate their variances based on estimates of their influence functions. In empirical studies, we assess the robustness of the estimators and the coverage of confidence intervals. The methods are further illustrated using data from a Danish registry study.  相似文献   

5.
Sample size calculations in the planning of clinical trials depend on good estimates of the model parameters involved. When the estimates of these parameters have a high degree of uncertainty attached to them, it is advantageous to reestimate the sample size after an internal pilot study. For non-inferiority trials with binary outcome we compare the performance of Type I error rate and power between fixed-size designs and designs with sample size reestimation. The latter design shows itself to be effective in correcting sample size and power of the tests when misspecification of nuisance parameters occurs with the former design.  相似文献   

6.
Outcome-dependent sampling (ODS) schemes can be a cost effective way to enhance study efficiency. The case-control design has been widely used in epidemiologic studies. However, when the outcome is measured on a continuous scale, dichotomizing the outcome could lead to a loss of efficiency. Recent epidemiologic studies have used ODS sampling schemes where, in addition to an overall random sample, there are also a number of supplemental samples that are collected based on a continuous outcome variable. We consider a semiparametric empirical likelihood inference procedure in which the underlying distribution of covariates is treated as a nuisance parameter and is left unspecified. The proposed estimator has asymptotic normality properties. The likelihood ratio statistic using the semiparametric empirical likelihood function has Wilks-type properties in that, under the null, it follows a chi-square distribution asymptotically and is independent of the nuisance parameters. Our simulation results indicate that, for data obtained using an ODS design, the semiparametric empirical likelihood estimator is more efficient than conditional likelihood and probability weighted pseudolikelihood estimators and that ODS designs (along with the proposed estimator) can produce more efficient estimates than simple random sample designs of the same size. We apply the proposed method to analyze a data set from the Collaborative Perinatal Project (CPP), an ongoing environmental epidemiologic study, to assess the relationship between maternal polychlorinated biphenyl (PCB) level and children's IQ test performance.  相似文献   

7.
A Gibbs sampling approach to linkage analysis.   总被引:9,自引:0,他引:9  
We present a Monte Carlo approach to estimation of the recombination fraction theta and the profile likelihood for a dichotomous trait and a single marker gene with 2 alleles. The method is an application of a technique known as 'Gibbs sampling', in which random samples of each of the unknowns (here genotypes, theta and nuisance parameters, including the allele frequencies and the penetrances) are drawn from their posterior distributions, given the data and the current values of all the other unknowns. Upon convergence, the resulting samples derive from the marginal distribution of all the unknowns, given only the data, so that the uncertainty in the specification of the nuisance parameters is reflected in the variance of the posterior distribution of theta. Prior knowledge about the distribution of theta and the nuisance parameters can be incorporated using a Bayesian approach, but adoption of a flat prior for theta and point priors for the nuisance parameters would correspond to the standard likelihood approach. The method is easy to program, runs quickly on a microcomputer, and could be generalized to multiple alleles, multipoint linkage, continuous phenotypes and more complex models of disease etiology. The basic approach is illustrated by application to data on cholesterol levels and an a low-density lipoprotein receptor gene in a single large pedigree.  相似文献   

8.
Schweder T 《Biometrics》2003,59(4):974-983
Maximum likelihood estimates of abundance are obtained from repeated photographic surveys of a closed stratified population with naturally marked and unmarked individuals. Capture intensities are assumed log-linear in stratum, year, and season. In the chosen model, an approximate confidence distribution for total abundance of bowhead whales, with an accompanying likelihood reduced of nuisance parameters, is found from a parametric bootstrap experiment. The confidence distribution depends on the assumed study protocol. A confidence distribution that is exact (except for the effect of discreteness) is found by conditioning in the unstratified case without unmarked individuals.  相似文献   

9.
J. Tufto  S. Engen    K. Hindar 《Genetics》1996,144(4):1911-1921
A new maximum likelihood method to simultaneously estimate the parameters of any migration pattern from gene frequencies in stochastic equilibrium is developed, based on a model of multivariate genetic drift in a subdivided population. Motivated by simulations of this process in the simplified case of two subpopulations, problems related to the nuisance parameter q, the equilibrium gene frequency, are eliminated by conditioning on the observed mean gene frequency. The covariance matrix of this conditional distribution is calculated by constructing an abstract process that mimics the behavior of the original process in the subspace of interest. The approximation holds as long as there is limited differentiation between subpopulations. The bias and variance of estimates of long-range and short-range migration in a finite stepping stone model are evaluated by fitting the model to simulated data with known values of the parameters. Possible ecological extensions of the model are discussed.  相似文献   

10.
Stratified data arise in several settings, such as longitudinal studies or multicenter clinical trials. Between-strata heterogeneity is usually addressed by random effects models, but an alternative approach is given by fixed effects models, which treat the incidental nuisance parameters as fixed unknown quantities. This approach presents several advantages, like computational simplicity and robustness to confounding by strata. However, maximum likelihood estimates of the parameter of interest are typically affected by incidental parameter bias. A remedy to this is given by the elimination of stratum-specific parameters by exact or approximate conditioning. The latter solution is afforded by the modified profile likelihood, which is the method applied in this paper. The aim is to demonstrate how the theory of modified profile likelihoods provides convenient solutions to various inferential problems in this setting. Specific procedures are available for different kinds of response variables, and they are useful both for inferential purposes and as a diagnostic method for validating random effects models. Some examples with real data illustrate these points.  相似文献   

11.
Kang SH  Shin D 《Human heredity》2004,58(1):10-17
Many scientific problems can be formulated in terms of a statistical model indexed by parameters, only some of which are of scientific interest and the other parameters, called nuisance parameters, are not of interest in themselves. For testing the Hardy-Weinberg law, a relation among genotype and allele probabilities is of interest and allele probabilities are of no interest and now nuisance parameters. In this paper we investigate how the size (the maximum of the type I error rate over the nuisance parameter space) of the chi-square test for the Hardy-Weinberg law is affected by the nuisance parameters. Whether the size is well controlled or not under the nominal level has been frequently investigated as basic components of statistical tests. The size represents the type I error rate at the worst case. We prove that the size is always greater than the nominal level as the sample size increases. Extensive computations show that the size of the chi-squared test (worst type I error rate over the nuisance parameter space) deviates more upwardly from the nominal level as the sample size gets larger. The value at which the maximum of the type I error rate was found moves closer to the edges of the the nuisance parameter space with increasing sample size. An exact test is recommended as an alternative when the type I error is inflated.  相似文献   

12.
Modelling paired survival data with covariates   总被引:4,自引:0,他引:4  
The objective of this paper is to consider the parametric analysis of paired censored survival data when additional covariate information is available, as in the Diabetic Retinopathy Study, which assessed the effectiveness of laser photocoagulation in delaying loss of visual acuity. Our first approach is to extend the fully parametric model of Clayton (1978, Biometrika 65, 141-151) to incorporate covariate information. Our second approach is to obtain parameter estimates from an independence working model together with robust variance estimates. The approaches are compared in terms of efficiency and computational considerations. A fundamental consideration in choosing a strategy for the analysis of paired survival data is whether the correlation within a pair is a nuisance parameter or a parameter of intrinsic scientific interest. The approaches are illustrated with the Diabetic Retinopathy Study.  相似文献   

13.
Collinearity among metrics of habitat loss and habitat fragmentation is typically treated as a nuisance in landscape ecology, and it is the norm to use statistical approaches that remove collinear information prior to estimating model parameters. However, collinearity may arise from causal relationships among landscape metrics and may therefore signal the occurrence of indirect effects (where one model predictor influences the response variable by driving changes in another influential predictor). Here we suggest that, far from being merely a statistical nuisance, collinearity may be crucial for accurately quantifying the effects of habitat loss versus habitat fragmentation. We use simulation modelling to create datasets of collinear landscape metrics in which collinearity arose from causal relationships, then test the ability of two statistical approaches to estimate the effects of these metrics on a simulated response variable: 1) multiple regression, which statistically removes collinearity, and was identified in a recent study as the best approach for estimating the effects of collinear landscape metrics (although this study did not account for any indirect effects implied by collinearity among metrics); and 2) path analysis, which accounts for the causal basis of collinearity. In agreement with this previous study, we found that multiple regression gave unbiased estimates of direct effects (effects not mediated by other model predictors). However, it gave biased estimates of total (direct + indirect) effects when indirect effects occurred. In contrast, path analysis reliably identified the causal basis of collinearity and gave unbiased estimates of direct, indirect, and total effects. We suggest that effective research on the impacts of habitat loss versus fragmentation will often require tools that can empirically test whether collinear landscape metrics are causally related, and if so, account for the indirect effects that these causal relationships imply. Path analysis, but not multiple regression, provides such a tool.  相似文献   

14.
Analyses of biomedical studies often necessitate modeling longitudinal causal effects. The current focus on personalized medicine and effect heterogeneity makes this task even more challenging. Toward this end, structural nested mean models (SNMMs) are fundamental tools for studying heterogeneous treatment effects in longitudinal studies. However, when outcomes are binary, current methods for estimating multiplicative and additive SNMM parameters suffer from variation dependence between the causal parameters and the noncausal nuisance parameters. This leads to a series of difficulties in interpretation, estimation, and computation. These difficulties have hindered the uptake of SNMMs in biomedical practice, where binary outcomes are very common. We solve the variation dependence problem for the binary multiplicative SNMM via a reparameterization of the noncausal nuisance parameters. Our novel nuisance parameters are variation independent of the causal parameters, and hence allow for coherent modeling of heterogeneous effects from longitudinal studies with binary outcomes. Our parameterization also provides a key building block for flexible doubly robust estimation of the causal parameters. Along the way, we prove that an additive SNMM with binary outcomes does not admit a variation independent parameterization, thereby justifying the restriction to multiplicative SNMMs.  相似文献   

15.
The epidemiologic concept of the adjusted attributable risk is a useful approach to quantitatively describe the importance of risk factors on the population level. It measures the proportional reduction in disease probability when a risk factor is eliminated from the population, accounting for effects of confounding and effect-modification by nuisance variables. The computation of asymptotic variance estimates for estimates of the adjusted attributable risk is often done by applying the delta method. Investigations on the delta method have shown, however, that the delta method generally tends to underestimate the standard error, leading to biased confidence intervals. We compare confidence intervals for the adjusted attributable risk derived by applying computer intensive methods like the bootstrap or jackknife to confidence intervals based on asymptotic variance estimates using an extensive Monte Carlo simulation and within a real data example from a cohort study in cardiovascular disease epidemiology. Our results show that confidence intervals based on bootstrap and jackknife methods outperform intervals based on asymptotic theory. Best variants of computer intensive confidence intervals are indicated for different situations.  相似文献   

16.
MOTIVATION: The desire to compare molecular phylogenies has stimulated the design of numerous tests. Most of these tests are formulated in a frequentist framework, and it is not known how they compare with Bayes procedures. I propose here two new Bayes tests that either compare pairs of trees (Bayes hypothesis test, BHT), or test each tree against an average of the trees included in the analysis (Bayes significance test, BST). RESULTS: The algorithm, based on a standard Metropolis-Hastings sampler, integrates nuisance parameters out and estimates the probability of the data under each topology. These quantities are used to estimate Bayes factors for composite vs. composite hypotheses. Based on two data sets, the BHT and BST are shown to construct similar confidence sets to the bootstrap and the Shimodaira Hasegawa test, respectively. This suggests that the known difference among previous tests is mainly due to the null hypothesis considered.  相似文献   

17.
Liao JG 《Biometrics》1999,55(1):268-272
This paper introduces a hierarchical Bayesian model for combining multiple 2 x 2 tables that allows the flexibility of different odds ratio estimates for different tables and at the same time allows the tables to borrow information from each other. The proposed model, however, is different from a full Bayesian model in that the nuisance parameters are eliminated by conditioning instead of integration. The motivation is a more robust model and a faster and more stable Gibbs algorithm. We work out a Gibbs scheme using the adaptive rejection sampling for log concave density and an algorithm for the mean and variance of the noncentral hypergeometric distribution. The model is applied to a multicenter ulcer clinical trial.  相似文献   

18.
Maximizing the homogeneity lod is known to be an appropriate procedure for estimating parameters of the trait model in an approximately 'ascertainment assumption free' (AAF) manner. We have investigated whether this same property also holds for the heterogeneity lod (HLOD). We show that, when the genetic models at linked and unlinked loci differ, HLODs are not AAF, and maximizing the HLOD yields parameter estimates that are for all practical purposes meaningless; indeed, the admixture parameter alpha does not even measure the proportion of linked families within the sample, as is commonly supposed. In spite of this, our results confirm a large body of evidence supporting the use of HLODs as robust tools for linkage detection, and suggest further that maximizing the HLOD over both alpha and parameters of the trait model can improve accuracy in estimation of the recombination fraction theta;. These findings have important implications for the optimal handling of nuisance parameters in linkage analysis, particularly when evaluating the evidence for or against linkage based on multiple independent heterogeneous sets of data.  相似文献   

19.
Tai JJ  Hsiao CK 《Human heredity》2001,51(4):192-198
In human genetic analysis, data are collected through the so-called 'ascertainment procedure'. Statistically this sampling scheme can be thought of as a multistage sampling method. At the first stage, one or several probands are ascertained. At the subsequent stages, a sequential sampling scheme is applied. Sampling in such a way is virtually a nonrandom procedure, which, in most cases, causes biased estimation which may be intractable. This paper focuses on the underlying causes of the intractability problem of ascertained genetic data. Three types of parameters, i.e. target, design and nuisance parameters, are defined as the essences to formulate the true likelihood of a set of data. These parameters are also classified into explicit or implicit parameters depending on whether they can be expressed explicity in the likelihood function. For ascertained genetic data, a sequential scheme is regarded as an implicit design parameter, and a true pedigree structure as an implicit nuisance parameter. The intractability problem is attributed to loss of information of any implicit parameter in likelihood formulation. Several approaches to build a likelihood for estimation of the segregation ratio when only an observed pedigree structure is available are proposed.  相似文献   

20.
Many facets of neuromuscular activation patterns and control can be assessed via electromyography and are important for understanding the control of locomotion. After spinal cord injury, muscle activation patterns can affect locomotor recovery. We present a novel application of reversible jump Markov chain Monte Carlo simulation to estimate activation patterns from electromyographic data. We assume the data to be a zero-mean, heteroscedastic process. The variance is explicitly modeled using a step function. The number and location of points of discontinuity, or change-points, in the step function, the inter-change-point variances, and the overall mean are jointly modeled along with the mean and variance from baseline data. The number of change-points is considered a nuisance parameter and is integrated out of the posterior distribution. Whereas current methods of detecting activation patterns are deterministic or provide only point estimates, ours provides distributional estimates of muscle activation. These estimates, in turn, are used to estimate physiologically relevant quantities such as muscle coactivity, total integrated energy, and average burst duration and to draw valid statistical inferences about these quantities.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号