首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 0 毫秒
1.
2.
We propose a new procedure for constructing inferences about a measure of interobserver agreement in studies involving a binary outcome and multiple raters. The proposed procedure, based on a chi-square goodness-of-fit test as applied to the correlated binomial model (Bahadur, 1961, in Studies in Item Analysis and Prediction, 158-176), is an extension of the goodness-of-fit procedure developed by Donner and Eliasziw (1992, Statistics in Medicine 11, 1511-1519) for the case of two raters. The new procedure is shown to provide confidence-interval coverage levels that are close to nominal over a wide range of parameter combinations. The procedure also provides a sample-size formula that may be used to determine the required number of subjects and raters for such studies.  相似文献   

3.
    
Aim To explore the impacts of imperfect reference data on the accuracy of species distribution model predictions. The main focus is on impacts of the quality of reference data (labelling accuracy) and, to a lesser degree, data quantity (sample size) on species presence–absence modelling. Innovation The paper challenges the common assumption that some popular measures of model accuracy and model predictions are prevalence independent. It highlights how imperfect reference data may impact on a study and the actions that may be taken to address problems. Main conclusions The theoretical independence of prevalence of popular accuracy measures, such as sensitivity, specificity, true skills statistics (TSS) and area under the receiver operating characteristic curve (AUC), is unlikely to occur in practice due to reference data error; all of these measures of accuracy, together with estimates of species occurrence, showed prevalence dependency arising through the use of a non‐gold‐standard reference. The number of cases used also had implications for the ability of a study to meet its objectives. Means to reduce the negative effects of imperfect reference data in study design and interpretation are suggested.  相似文献   

4.
A new approach that extends the classical Clopper‐Pearson procedure is proposed for the estimation of the (1–α)% confidence interval of a proportion with over‐dispersion. Over‐dispersion occurs when a proportion of interest shows more variation (variance inflation) than predicted by the binomial distribution. There are two steps in the approach. The first step consists of the estimation of the variance inflation factor. In the second step, an extended Clopper‐Pearson procedure is applied to calculate the confidence interval after the effective sample size is obtained by adjusting with the estimated variance inflation factor. The performance of the extended Clopper‐Pearson procedure is evaluated via a Monte Carlo study under the setup motivated from head lice studies. It is demonstrated that the 95% confidence intervals constructed from the new approach generally have the closest coverage rate to target (95%) when compared with those constructed from competing procedures.  相似文献   

5.
    
To construct a confidence interval of effect size in paired studies, we propose four approximate methods--Wald method, variance-stabilizing transformation method, and signed and modified signed log-likelihood ratio methods. We compare these methods using simulation to determine those that have good performance in terms of coverage probability. In particular, simulations show that the modified signed log-likelihood ratio method produces a confidence interval with a nearly exact coverage probability and highly accurate and symmetric error probabilities even for very small samples. We apply the methods to data from an iron deficiency anemia study.  相似文献   

6.
7.
Scientists often need to test hypotheses and construct corresponding confidence intervals. In designing a study to test a particular null hypothesis, traditional methods lead to a sample size large enough to provide sufficient statistical power. In contrast, traditional methods based on constructing a confidence interval lead to a sample size likely to control the width of the interval. With either approach, a sample size so large as to waste resources or introduce ethical concerns is undesirable. This work was motivated by the concern that existing sample size methods often make it difficult for scientists to achieve their actual goals. We focus on situations which involve a fixed, unknown scalar parameter representing the true state of nature. The width of the confidence interval is defined as the difference between the (random) upper and lower bounds. An event width is said to occur if the observed confidence interval width is less than a fixed constant chosen a priori. An event validity is said to occur if the parameter of interest is contained between the observed upper and lower confidence interval bounds. An event rejection is said to occur if the confidence interval excludes the null value of the parameter. In our opinion, scientists often implicitly seek to have all three occur: width, validity, and rejection. New results illustrate that neglecting rejection or width (and less so validity) often provides a sample size with a low probability of the simultaneous occurrence of all three events. We recommend considering all three events simultaneously when choosing a criterion for determining a sample size. We provide new theoretical results for any scalar (mean) parameter in a general linear model with Gaussian errors and fixed predictors. Convenient computational forms are included, as well as numerical examples to illustrate our methods.  相似文献   

8.
    
The effective population size Ne is an important parameter in population genetics and conservation biology. In recent years, there has been great interest in the use of molecular markers to estimate Ne. Although the point estimates from molecular markers in general suffer from a low reliability, the use of single nucleotide polymorphism (SNP) markers over a wide range of genome is expected to remarkably improve the reliability. In this study, expressions were derived for interval estimates of Ne from one published method, the heterozygote‐excess method, when it is applied to SNP markers. The conditional variance theory is applied to the derivation of a confidence interval for Ne under random union of gametes, monogamy and polygyny. Stochastic simulation shows that the obtained confidence interval is slightly conservative, but fairly useful for practical applications. The result is illustrated with real data on SNP markers in a pig strain.  相似文献   

9.
Importance sampling and the nested bootstrap   总被引:2,自引:0,他引:2  
HINKLEY  D. V.; SHI  S. 《Biometrika》1989,76(3):435-446
  相似文献   

10.
Precision of the estimate of the population mean using ranked set sample (RSS) relative to using simple random sample (SRS), with the same number of quantified units, depends upon the population and success in ranking. In practice, even ranking a sample of moderate size and observing the ith ranked unit (other than the extremes) is a difficult task. Therefore, in this paper we introduce a variety of extreme ranked set sample (ERSSs) to estimate the population mean. ERSSs is more practical than the ordinary ranked set sampling, since in case of even sample size we need to identify successfully only the first and/or the last ordered unit or in case of odd sample size the median unit. We show that ERSSs gives an unbiased estimate of the population mean in case of symmetric populations and it is more efficient than SRS, using the same number of quantified units. Example using real data is given. Also, parametric examples are given.  相似文献   

11.
Estimation of heritability from varietal trials data   总被引:2,自引:0,他引:2  
We present the estimation of heritabilities of an observed trait in situations where evaluation of several pure breeding lines is performed in a trial at a single location and in trials from several locations. For the single location situation, we evaluate exact confidence intervals, the probability of invalid estimates, and the percentage points of the distribution of heritability. Simulations were performed to numerically verify the results. Additionally, approximations to the bias and standard error of the estimate were obtained and are presented along with their simulated values and coefficients of skewness and kurtosis. For trials in several locations, explicit expressions for exact values of confidence limits are not available. Further, one would require knowledge of one more parameter, represented by the ratio of genotype x environment (G x E) interaction variance to error variance, in addition to the number of genotypes, replication and true heritability value. Approximations were made for bias and the standard error of estimates of heritability. The evaluation of the distribution of heritability and its moments was recognized as a problem of the linear function of an independent chi-square. The methods have been illustrated by data from experiments on grain and straw yield of 64 barley genotypes evaluated at three locations.  相似文献   

12.
When comparing two treatments, we often use the simple difference between the probabilities of response to measure the efficacy of one treatment over the other. When the measurement of outcome is unreliable or the cost of obtaining additional subjects is high relative to that of additional measurements from the obtained subjects, we may often consider taking more than one measurement per subject to increase the precision of an interval estimator. This paper focuses discussion on interval estimation of simple difference when we take repeated measurements per subject. This paper develops four asymptotic interval estimators of simple difference for any finite number of measurements per subject. This paper further applies Monte Carlo simulation to evaluate the finite‐sample performance of these estimators in a variety of situations. Finally, this paper includes a discussion on sample size determination on the basis of both the average length and the probability of controlling the length of the resulting interval estimate proposed elsewhere.  相似文献   

13.
14.
We show that the symmetric “confidence intervals” of WESTLAKE (1972, 1976), widely used and referred to in bioequivalence studies, are not confidence intervals in any accepted sense. Nevertheless, meaningful symmetric intervals can be constructed in the context of Bayesian or fiducial inference.  相似文献   

15.
16.
Approaches like multiple interval mapping using a multiple-QTL model for simultaneously mapping QTL can aid the identification of multiple QTL, improve the precision of estimating QTL positions and effects, and are able to identify patterns and individual elements of QTL epistasis. Because of the statistical problems in analytically deriving the standard errors and the distributional form of the estimates and because the use of resampling techniques is not feasible for several linked QTL, there is the need to perform large-scale simulation studies in order to evaluate the accuracy of multiple interval mapping for linked QTL and to assess confidence intervals based on the standard statistical theory. From our simulation study it can be concluded that in comparison with a monogenetic background a reliable and accurate estimation of QTL positions and QTL effects of multiple QTL in a linkage group requires much more information from the data. The reduction of the marker interval size from 10 cM to 5 cM led to a higher power in QTL detection and to a remarkable improvement of the QTL position as well as the QTL effect estimates. This is different from the findings for (single) interval mapping. The empirical standard deviations of the genetic effect estimates were generally large and they were the largest for the epistatic effects. These of the dominance effects were larger than those of the additive effects. The asymptotic standard deviation of the position estimates was not a good criterion for the accuracy of the position estimates and confidence intervals based on the standard statistical theory had a clearly smaller empirical coverage probability as compared to the nominal probability. Furthermore the asymptotic standard deviation of the additive, dominance and epistatic effects did not reflect the empirical standard deviations of the estimates very well, when the relative QTL variance was smaller/equal to 0.5. The implications of the above findings are discussed.  相似文献   

17.
18.
    
In clinical trials, sample size reestimation is a useful strategy for mitigating the risk of uncertainty in design assumptions and ensuring sufficient power for the final analysis. In particular, sample size reestimation based on unblinded interim effect size can often lead to sample size increase, and statistical adjustment is usually needed for the final analysis to ensure that type I error rate is appropriately controlled. In current literature, sample size reestimation and corresponding type I error control are discussed in the context of maintaining the original randomization ratio across treatment groups, which we refer to as “proportional increase.” In practice, not all studies are designed based on an optimal randomization ratio due to practical reasons. In such cases, when sample size is to be increased, it is more efficient to allocate the additional subjects such that the randomization ratio is brought closer to an optimal ratio. In this research, we propose an adaptive randomization ratio change when sample size increase is warranted. We refer to this strategy as “nonproportional increase,” as the number of subjects increased in each treatment group is no longer proportional to the original randomization ratio. The proposed method boosts power not only through the increase of the sample size, but also via efficient allocation of the additional subjects. The control of type I error rate is shown analytically. Simulations are performed to illustrate the theoretical results.  相似文献   

19.
20.
    
Liu A  Schisterman EF  Wu C 《Biometrics》2006,62(4):1190-1196
We introduce sequential testing procedures for the planning and analysis of reliability studies to assess an exposure's measurement error. The designs allow repeated evaluation of reliability of the measurements and stop testing if early evidence shows the measurement error is within the level of tolerance. Methods are developed and critical values tabulated for a number of two-stage designs. The methods are exemplified using an example evaluating the reliability of biomarkers associated with oxidative stress.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号