首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 0 毫秒
1.
Two common goals when choosing a method for performing all pairwise comparisons of J independent groups are controlling experiment wise Type I error and maximizing power. Typically groups are compared in terms of their means, but it has been known for over 30 years that the power of these methods becomes highly unsatisfactory under slight departures from normality toward heavy-tailed distributions. An approach to this problem, well-known in the statistical literature, is to replace the sample mean with a measure of location having a standard error that is relatively unaffected by heavy tails and outliers. One possibility is to use the trimmed mean. This paper describes three such multiple comparison procedures and compares them to two methods for comparing means.  相似文献   

2.
DUNNETT (1955) developed a procedure simultaneously comparing k treatments to one control with an exact overall type I error of α when all sampling distributions are normal. Sometimes it is desirable to compare k treatments to m≧2 controls, in particular to two controls. For instance, several new therapies (e.g., pain relievers) could be compared to two standard therapies (e.g., Aspirin and Tylenol). Alternatively, a standard therapy could be very expensive, difficult to apply and/or have bad side effects, making it useful to compare each new therapy to both standard therapy and no therapy (Placebo). Dunnett's method is expanded here to give comparisons of mean values for k treatments to mean values for m≧2 controls at an exact overall type I error of α when all sampling distributions are normal. Tabled values needed to make exact simultaneous comparisons at α = .05 are given for m = 2. An application is made to an example from the literature.  相似文献   

3.
When a trial involves an invasive laboratory test procedure or requires patients to make a commitment to follow a restrictive test schedule, we can often lose a great proportion of our sampled patients due to refusal of participation into our study. Therefore, incorporating the possible loss of patients into sample size calculation is certainly important in the planning stage of a study. In this paper, we have generalized the sample size calculation procedure for intraclass correlation by accounting for the random loss of patients in the beginning of a trial. We have demonstrated that the simple ad hoc procedure, that raises the estimated sample size in the absence of loss of patients by the factor 1/po, where po is the retention probability for a randomly selected patient, is adequate when po is large (=0.80). When po is small (i.e., a high refusal rate), however, use of this simple ad hoc procedure tends to underestimate the required sample size. Furthermore, we have found that if the individual retention probability varied substantially among patients, then the magnitude of the above underestimation could even be critical and therefore, the application of the simple direct adjustment procedure in this situation should be avoided.  相似文献   

4.
  总被引:1,自引:0,他引:1  
Horn M  Vollandt R  Dunnett CW 《Biometrics》2000,56(3):879-881
Laska and Meisner (1989, Biometrics 45, 1139-1151) dealt with the problem of testing whether an identified treatment belonging to a set of k + 1 treatments is better than each of the other k treatments. They calculated sample size tables for k = 2 when using multiple t-tests or Wilcoxon-Mann-Whitney tests, both under normality assumptions. In this paper, we provide sample size formulas as well as tables for sample size determination for k > or = 2 when t-tests under normality or Wilcoxon-Mann-Whitney tests under general distribution assumptions are used.  相似文献   

5.
An approximate and practical solution is proposed for the Behrens-Fisher problem. This solution is compared to the solutions considered by Mehta and Srinivasan (1970) and Welch's (1937) approximate t-test in terms of the stability of the size and magnitude of the power. It is shown that the stability of the size of the new test is better than that of Welch's t when at least one of the sample sizes is small. When the sample sizes are moderately large or large the sizes and powers of all the recommended tests are almost the same.  相似文献   

6.
    
Sexual dimorphism in primate species expresses the effects of phylogeny, life history, behavior, and ontogeny. The causes and implications of sexual dimorphism have been studied in several different primates using a variety of morphological databases such as body weight, canine length, and coat color and ornamentation. In addition to these different patterns of dimorphism, the degree to which a species is dimorphic results from a variety of possible causes. In this study we test the general hypothesis that a species highly dimorphic for one size-based index of dimorphism will be equally dimorphic (relative to other species) for other size-based indices. Specifically, the degree and pattern of sexual dimorphism in Cebus and several other New World monkey species is measured using craniometric data as a substitute for the troublesome range of variation in body weight estimates. In general, the rank ordering of species for dimorphism ratios differs considerably across neural vs. non-neural functional domains of the cranium. The relative degree of sexual dimorphism in different functional regions of the cranium is affected by the independent action of natural selection on those regions. Regions of the cranium upon which natural selection is presumed to have acted within a species show greater degrees of dimorphism than do the same regions in closely related taxa. Within Cebus, C. apella is consistently more dimorphic than other Cebus species for facial measurements, but not for neural or body weight measurements. The pattern in C. apella indicates no single best measurement of the degree of dimorphism in a species; rather, the relative degree of dimorphism applies only to the region being measured and may be enhanced by other selective pressures on morphology. Am J Phys Anthropol 107:243–256, 1998. © 1998 Wiley-Liss, Inc.  相似文献   

7.
    
The intraclass version of kappa coefficient has been commonly applied as a measure of agreement for two ratings per subject with binary outcome in reliability studies. We present an efficient statistic for testing the strength of kappa agreement using likelihood scores, and derive asymptotic power and sample size formula. Exact evaluation shows that the score test is generally conservative and more powerful than a method based on a chi‐square goodness‐of‐fit statistic (Donner and Eliasziw , 1992, Statistics in Medicine 11 , 1511–1519). In particular, when the research question is one directional, the one‐sided score test is substantially more powerful and the reduction in sample size is appreciable.  相似文献   

8.
9.
The use of variance components and multivariate linear models in genetics applications has a long history that dates back to (at least) Fisher's seminal 1918 paper “The correlation between relatives on the supposition of Mendelian inheritance” [Phil. Trans. 52: 399–433]. Although extensions and elaborations of Fisher's insights have been offered in recent times, relatively few studies exist which examine the theoretical and operational properties variance component models possess in complicated genetic analysis settings. In this paper variance component models, as well as some of their properties (e.g., power, efficiency, and sample size considerations) are discussed in the context of each of the following genetic analysis settings: 1. the detection of general polygenic additive and dominance effects; 2. the detection of genetic effects in the presence of environmental effects (and vice versa); 3. the detection of pleiotropic gene action; 4. aspects of the detection of genotype by environment interaction; and 5. sequential tests for general hypotheses framed in the context of settings 1 through 4. Exposition of the proposed methods and results are facilitated through a special emphasis placed on pedigree covariance structure modeling.  相似文献   

10.
A robust test (to be referred to as M* test) is proposed for testing equality of several group means without assuming normality and equality of variances. This test statistic is obtained by combining Tiku's MML robust procedure with the James statistic. Monte Carlo simulation studies indicate that the M* test is more powerful than the Welch test, the James test, and the tests based on Huber's M-estimators over a wide range of nonnormal universes. It is also more powerful than the Brown and Forsythe test under most of nonnormal distributions and has substantially the same power as the Brown and Forsythe test under normal distribution. Comparing with Tan-Tabatabai test, M* is almost as powerful as Tan-Tabatabai test.  相似文献   

11.
12.
Summary As the nonparametric generalization of the one‐way analysis of variance model, the Kruskal–Wallis test applies when the goal is to test the difference between multiple samples and the underlying population distributions are nonnormal or unknown. Although the Kruskal–Wallis test has been widely used for data analysis, power and sample size methods for this test have been investigated to a much lesser extent. This article proposes new power and sample size calculation methods for the Kruskal–Wallis test based on the pilot study in either a completely nonparametric model or a semiparametric location model. No assumption is made on the shape of the underlying population distributions. Simulation results show that, in terms of sample size calculation for the Kruskal–Wallis test, the proposed methods are more reliable and preferable to some more traditional methods. A mouse peritoneal cavity study is used to demonstrate the application of the methods.  相似文献   

13.
    
Nam JM 《Biometrics》1999,55(1):289-293
Power and sample-size formulas for testing the homogeneity of relative risks using the score method are presented. The homogeneity score test (Gart, 1985, Biometrika 72, 673-677) is formally equivalent to the Pearson chi-square test, although they look different. Results of this paper may be useful in assessing the validity of the model of a common relative risk before combining several 2 x 2 tables or in designing a prospective study for detecting heterogeneity of relative risks.  相似文献   

14.
    
The most commonly used method in evolutionary biology for combining information across multiple tests of the same null hypothesis is Fisher's combined probability test. This note shows that an alternative method called the weighted Z-test has more power and more precision than does Fisher's test. Furthermore, in contrast to some statements in the literature, the weighted Z-method is superior to the unweighted Z-transform approach. The results in this note show that, when combining P-values from multiple tests of the same hypothesis, the weighted Z-method should be preferred.  相似文献   

15.
We present a survey of sample size formulas derived in other papers for pairwise comparisons of k treatments and for comparisons of k treatments with a control. We consider the calculation of sample sizes with preassigned per‐pair, any‐pair and all‐pairs power for tests that control either the comparisonwise or the experimentwise type I error rate. A comparison exhibits interesting similarities between the parametric, nonparametric and binomial case.  相似文献   

16.
    
There has been growing interest, when comparing an experimental treatment with an active control with respect to a binary outcome, in allowing the non-inferiority margin to depend on the unknown success rate in the control group. It does not seem universally recognized, however, that the statistical test should appropriately adjust for the uncertainty surrounding the non-inferiority margin. In this paper, we inspect a naive procedure that treats an \"observed margin\" as if it were fixed a priori, and explain why it might not be valid. We then derive a class of tests based on the delta method, including the Wald test and the score test, for a smooth margin. An alternative derivation is given for the asymptotic distribution of the likelihood ratio statistic, again for a smooth margin. We discuss the asymptotic behavior of these tests when applied to a piecewise smooth margin. A simple condition on the margin function is given which allows the likelihood ratio test to carry over to a piecewise smooth margin using the same critical value as for a smooth margin. Simulation experiments are conducted, under a smooth margin and a piecewise linear margin, to evaluate the finite-sample performance of the asymptotic tests studied.  相似文献   

17.
    
Clinical trials are often planned with high uncertainty about the variance of the primary outcome variable. A poor estimate of the variance, however, may lead to an over‐ or underpowered study. In the internal pilot study design, the sample variance is calculated at an interim step and the sample size can be adjusted if necessary. The available recalculation procedures use the data of those patients for sample size recalculation that have already completed the study. In this article, we consider a variance estimator that takes into account both the data at the endpoint and at an intermediate point of the treatment phase. We derive asymptotic properties of this estimator and the relating sample size recalculation procedure. In a simulation study, the performance of the proposed approach is evaluated and compared with the procedure that uses only long‐term data. Simulation results demonstrate that the sample size resulting from the proposed procedure shows in general a smaller variability. At the same time, the Type I error rate is not inflated and the achieved power is close to the desired value.  相似文献   

18.
One of multiple testing problems in drug finding experiments is the comparison of several treatments with one control. In this paper we discuss a particular situation of such an experiment, i.e., a microarray setting, where the many-to-one comparisons need to be addressed for thousands of genes simultaneously. For a gene-specific analysis, Dunnett's single step procedure is considered within gene tests, while the FDR controlling procedures such as Significance Analysis of Microarrays (SAM) and Benjamini and Hochberg (BH) False Discovery Rate (FDR) adjustment are applied to control the error rate across genes. The method is applied to a microarray experiment with four treatment groups (three microarrays in each group) and 16,998 genes. Simulation studies are conducted to investigate the performance of the SAM method and the BH-FDR procedure with regard to controlling the FDR, and to investigate the effect of small-variance genes on the FDR in the SAM procedure.  相似文献   

19.
The test statistics used until now in the CFA have been developed under the assumption of the overall hypothesis of total independence. Therefore, the multiple test procedures based on these statistics are really only different tests of the overall hypothesis. If one likes to test a special cell hypothesis, one should only assume that this hypothesis is true and not the whole overall hypothesis. Such cell tests can then be used as elements of a multiple test procedure. In this paper it is shown that the usual test procedures can be very anticonservative (except of the two-dimensional, and, for some procedures, the three-dimensional case), and corrected test procedures are developed. Furthermore, for the construction of multiple tests controlling the multiple level, modifications of Holm's (1979) procedure are proposed which lead to sharper results than his general procedure and can also be performed very easily.  相似文献   

20.
Cope's rule is the tendency for body size to increase over time along a lineage. A set of 65 phylogenetically independent comparisons, between earlier and later genera, show that Cope's rule applied in dinosaurs: later genera were on average about 25% longer than the related earlier genera to which they were compared. The tendency for size to increase was not restricted to a particular clade within the group, nor to a particular time within its history. Small lineages were more likely to increase in size, and large lineages more likely to decrease: this pattern may indicate an intermediate optimum body size, but can also be explained as an artefact of data error. The rate of size increase estimated from the phylogenetic comparisons is significantly higher than the rate seen across the fauna as a whole. This difference could indicate that within-lineage selection for larger size was opposed by clade selection favouring smaller size, but data limitations mean that alternative explanations (which we discuss) cannot be excluded. We discuss ways of unlocking the full potential usefulness of phylogenies for studying the dynamics of evolutionary trends.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号