共查询到20条相似文献,搜索用时 0 毫秒
1.
We consider sample size calculations for testing differences in means between two samples and allowing for different variances in the two groups. Typically, the power functions depend on the sample size and a set of parameters assumed known, and the sample size needed to obtain a prespecified power is calculated. Here, we account for two sources of variability: we allow the sample size in the power function to be a stochastic variable, and we consider estimating the parameters from preliminary data. An example of the first source of variability is nonadherence (noncompliance). We assume that the proportion of subjects who will adhere to their treatment regimen is not known before the study, but that the proportion is a stochastic variable with a known distribution. Under this assumption, we develop simple closed form sample size calculations based on asymptotic normality. The second source of variability is in parameter estimates that are estimated from prior data. For example, we account for variability in estimating the variance of the normal response from existing data which are assumed to have the same variance as the study for which we are calculating the sample size. We show that we can account for the variability of the variance estimate by simply using a slightly larger nominal power in the usual sample size calculation, which we call the calibrated power. We show that the calculation of the calibrated power depends only on the sample size of the existing data, and we give a table of calibrated power by sample size. Further, we consider the calculation of the sample size in the rarer situation where we account for the variability in estimating the standardized effect size from some existing data. This latter situation, as well as several of the previous ones, is motivated by sample size calculations for a Phase II trial of a malaria vaccine candidate. 相似文献
2.
We prove that the generalized Poisson distribution GP(theta, eta) (eta > or = 0) is a mixture of Poisson distributions; this is a new property for a distribution which is the topic of the book by Consul (1989). Because we find that the fits to count data of the generalized Poisson and negative binomial distributions are often similar, to understand their differences, we compare the probability mass functions and skewnesses of the generalized Poisson and negative binomial distributions with the first two moments fixed. They have slight differences in many situations, but their zero-inflated distributions, with masses at zero, means and variances fixed, can differ more. These probabilistic comparisons are helpful in selecting a better fitting distribution for modelling count data with long right tails. Through a real example of count data with large zero fraction, we illustrate how the generalized Poisson and negative binomial distributions as well as their zero-inflated distributions can be discriminated. 相似文献
3.
This article investigates the performance, in a frequentist sense, of Bayesian confidence intervals (CIs) for the difference of proportions, relative risk, and odds ratio in 2 x 2 contingency tables. We consider beta priors, logit-normal priors, and related correlated priors for the two binomial parameters. The goal was to analyze whether certain settings for prior parameters tend to provide good coverage performance regardless of the true association parameter values. For the relative risk and odds ratio, we recommend tail intervals over highest posterior density (HPD) intervals, for invariance reasons. To protect against potentially very poor coverage probabilities when the effect is large, it is best to use a diffuse prior, and we recommend the Jeffreys prior. Otherwise, with relatively small samples, Bayesian CIs using more informative (even uniform) priors tend to have poorer performance than the frequentist CIs based on inverting score tests, which perform uniformly quite well for these parameters. 相似文献
4.
Animal experiments in chemical carcinogenesis are often analysed by tests for equal proportions in a 2 × 2 table. Since hereby nothing is said about the power of the tests, we propose to describe the outcome of such an experiment by an upper confidence limit of an adjusted difference between the two response rates. 相似文献
5.
Frank Krummenauer 《Biometrical journal. Biometrische Zeitschrift》1998,40(7):823-832
Power investigations, for example, in statistical procedures for the assessment of agreement among multiple raters often require the simultaneous simulation of several dependent binomial or Poisson distributions to appropriately model the stochastical dependencies between the raters' results. Regarding the rather large dimensions of the random vectors to be generated and the even larger number of interactions to be introduced into the simulation scenarios to determine all necessary information on their distributions' dependence stucture, one needs efficient and fast algorithms for the simulation of multivariate Poisson and binomial distributions. Therefore two equivalent models for the multivariate Poisson distribution are combined to obtain an algorithm for the quick implementation of its multivariate dependence structure. Simulation of the multivariate Poisson distribution then becomes feasible by first generating and then convoluting independent univariate Poisson variates with appropriate expectations. The latter can be computed via linear recursion formulae. Similar means for simulation are also considered for the binomial setting. In this scenario it turns out, however, that exact computation of the probability function is even easier to perform; therefore corresponding linear recursion formulae for the point probabilities of multivariate binomial distributions are presented, which only require information about the index parameter and the (simultaneous) success probabilities, that is the multivariate dependence structure among the binomial marginals. 相似文献
6.
7.
James D. Stamey Dean M. Young Tom L. Bratcher 《Biometrical journal. Biometrische Zeitschrift》2004,46(5):572-578
We develop three Bayesian predictive probability functions based on data in the form of a double sample. One Bayesian predictive probability function is for predicting the true unobservable count of interest in a future sample for a Poisson model with data subject to misclassification and two Bayesian predictive probability functions for predicting the number of misclassified counts in a current observable fallible count for an event of interest. We formulate a Gibbs sampler to calculate prediction intervals for these three unobservable random variables and apply our new predictive models to calculate prediction intervals for a real‐data example. (© 2004 WILEY‐VCH Verlag GmbH & Co. KGaA, Weinheim) 相似文献
8.
A stability study is usually conducted to ensure that a drug product can meet the approved specifications prior to its expiration dating period (shelf life). Several approaches for determination of drug shelf life assuming random batches have been proposed. In this paper, we examine sampling distributions of the estimated shelf lives proposed by Shao and Chow (1994, Biometrics 50, 753-763). An application to some stability data from the pharmaceutical industry is presented. 相似文献
9.
J. Campione-Piccardo 《Biometrical journal. Biometrische Zeitschrift》1983,25(8):807-815
The use of an approximation to the median of the Poisson distribution to represent each occurrence of mutations in a growing clone permits the prediction of the number of mutants per clone without the limitations imposed by more heuristic expressions. Its application to the evaluation of mutation rates yields results comparable to those obtained by fluctuation analysis. 相似文献
10.
Bo Xu 《Developmental biology》2010,348(1):3-11
The categorical data set is an important data class in experimental biology and contains data separable into several mutually exclusive categories. Unlike measurement of a continuous variable, categorical data cannot be analyzed with methods such as the Student's t-test. Thus, these data require a different method of analysis to aid in interpretation. In this article, we will review issues related to categorical data, such as how to plot them in a graph, how to integrate results from different experiments, how to calculate the error bar/region, and how to perform significance tests. In addition, we illustrate analysis of categorical data using experimental results from developmental biology and virology studies. 相似文献
11.
The paper provides a comprehensive review of methodology for setting confidence intervals for the parameter of a Poisson distribution. The results are illustrated by a numerical example. 相似文献
12.
A simple extension of the Poisson process results in binomially distributed counts of events in a time interval. A further extension generalises this to probability distributions under‐ or over‐dispersed relative to the binomial distribution. Substantial levels of under‐dispersion are possible with this modelling, but only modest levels of over‐dispersion – up to Poisson‐like variation. Although simple analytical expressions for the moments of these probability distributions are not available, approximate expressions for the mean and variance are derived, and used to re‐parameterise the models. The modelling is applied in the analysis of two published data sets, one showing under‐dispersion and the other over‐dispersion. More appropriate assessment of the precision of estimated parameters and reliable model checking diagnostics follow from this more general modelling of these data sets. 相似文献
13.
Directly standardized rates continue to be an integral tool for presenting rates for diseases that are highly dependent on age, such as cancer. Statistically, these rates are modeled as a weighted sum of Poisson random variables. This is a difficult statistical problem, because there are k observed Poisson variables and k unknown means. The gamma confidence interval has been shown through simulations to have at least nominal coverage in all simulated scenarios, but it can be overly conservative. Previous modifications to that method have closer to nominal coverage on average, but they do not achieve the nominal coverage bound in all situations. Further, those modifications are not central intervals, and the upper coverage error rate can be substantially more than half the nominal error. Here we apply a mid‐p modification to the gamma confidence interval. Typical mid‐p methods forsake guaranteed coverage to get coverage that is sometimes higher and sometimes lower than the nominal coverage rate, depending on the values of the parameters. The mid‐p gamma interval does not have guaranteed coverage in all situations; however, in the (not rare) situations where the gamma method is overly conservative, the mid‐p gamma interval often has at least nominal coverage. The mid‐p gamma interval is especially appropriate when one wants a central interval, since simulations show that in many situations both the upper and lower coverage error rates are on average less than or equal to half the nominal error rate. 相似文献
14.
15.
16.
We propose a new Poisson method to estimate the variance for prevalence estimates obtained by the counting method described by Gail et al. (1999, Biometrics 55, 1137-1144) and to construct a confidence interval for the prevalence. We evaluate both the Poisson procedure and the procedure based on the bootstrap proposed by Gail et al. in simulated samples generated by resampling real data. These studies show that both variance estimators usually perform well and yield coverages of confidence intervals at nominal levels. When the number of disease survivors is very small, however, confidence intervals based on the Poisson method have supranominal coverage, whereas those based on the procedure of Gail et al. tend to have below-nominal coverage. For these reasons, we recommend the Poisson method, which also reduces the computational burden considerably. 相似文献
17.
Approaches like multiple interval mapping using a multiple-QTL model for simultaneously mapping QTL can aid the identification of multiple QTL, improve the precision of estimating QTL positions and effects, and are able to identify patterns and individual elements of QTL epistasis. Because of the statistical problems in analytically deriving the standard errors and the distributional form of the estimates and because the use of resampling techniques is not feasible for several linked QTL, there is the need to perform large-scale simulation studies in order to evaluate the accuracy of multiple interval mapping for linked QTL and to assess confidence intervals based on the standard statistical theory. From our simulation study it can be concluded that in comparison with a monogenetic background a reliable and accurate estimation of QTL positions and QTL effects of multiple QTL in a linkage group requires much more information from the data. The reduction of the marker interval size from 10 cM to 5 cM led to a higher power in QTL detection and to a remarkable improvement of the QTL position as well as the QTL effect estimates. This is different from the findings for (single) interval mapping. The empirical standard deviations of the genetic effect estimates were generally large and they were the largest for the epistatic effects. These of the dominance effects were larger than those of the additive effects. The asymptotic standard deviation of the position estimates was not a good criterion for the accuracy of the position estimates and confidence intervals based on the standard statistical theory had a clearly smaller empirical coverage probability as compared to the nominal probability. Furthermore the asymptotic standard deviation of the additive, dominance and epistatic effects did not reflect the empirical standard deviations of the estimates very well, when the relative QTL variance was smaller/equal to 0.5. The implications of the above findings are discussed. 相似文献
18.
Given a sample, the problem of choosing one of a variety of theoretical distributions for data analysis has been a major concern to both theoretical and applied statisticians. On the basis of a graph drawn using a normal characterization, a method is given to detect departure from normality. Special mathematical study is presented for the proposed graphs. Tukey's g and h family of distributions are used as an alternate to normality. 相似文献
19.
20.
通过两点分布和几何分布模型,给出了物种总数N的点估计,并讨论了它的性质.根据F分布和二项分布、几何分布的关系,本文得到了N的区间估计和检验统计量. 相似文献