首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
Data with varying age at disease onset arise frequently in studies of mapping disease associated genes. Naively combining affected subjects with different ages at onset may result in a much reduced power in detecting the disease genes. In this paper we present a weighted score test statistic to detect the linkage between marker and latent disease loci using affected sibpairs, where the weight is used for assigning differential contribution due to the varying age at onset of each affected sibpair to the test statistic. We show that the weighted test has a correct type I error rate asymptotically. For an illustrative purpose, we analyze a data set from the 12th Genetic Analysis Workshop. The result shows that the weighted tests appear to be able to pinpoint the location of latent disease genes better than the mean IBD test with equal weight with respect to the age at onset. To avoid the potential power loss due to the improper weight, we propose to use a combined test statistic, taking the maximum of two tests, one that is weighted by the age-dependent penetrance function and the other that may be invariant to the age. We conduct an analytical study, comparing the combined test with weighted and equal weight with respect to age test. It shows that the combined test retains the most power of the better one of the two tests being combined.  相似文献   

2.
The two‐sided Simes test is known to control the type I error rate with bivariate normal test statistics. For one‐sided hypotheses, control of the type I error rate requires that the correlation between the bivariate normal test statistics is non‐negative. In this article, we introduce a trimmed version of the one‐sided weighted Simes test for two hypotheses which rejects if (i) the one‐sided weighted Simes test rejects and (ii) both p‐values are below one minus the respective weighted Bonferroni adjusted level. We show that the trimmed version controls the type I error rate at nominal significance level α if (i) the common distribution of test statistics is point symmetric and (ii) the two‐sided weighted Simes test at level 2α controls the level. These assumptions apply, for instance, to bivariate normal test statistics with arbitrary correlation. In a simulation study, we compare the power of the trimmed weighted Simes test with the power of the weighted Bonferroni test and the untrimmed weighted Simes test. An additional result of this article ensures type I error rate control of the usual weighted Simes test under a weak version of the positive regression dependence condition for the case of two hypotheses. This condition is shown to apply to the two‐sided p‐values of one‐ or two‐sample t‐tests for bivariate normal endpoints with arbitrary correlation and to the corresponding one‐sided p‐values if the correlation is non‐negative. The Simes test for such types of bivariate t‐tests has not been considered before. According to our main result, the trimmed version of the weighted Simes test then also applies to the one‐sided bivariate t‐test with arbitrary correlation.  相似文献   

3.
This paper investigates homogeneity test of rate ratios in stratified matched-pair studies on the basis of asymptotic and bootstrap-resampling methods. Based on the efficient score approach, we develop a simple and computationally tractable score test statistic. Several other homogeneity test statistics are also proposed on the basis of the weighted least-squares estimate and logarithmic transformation. Sample size formulae are derived to guarantee a pre-specified power for the proposed tests at the pre-given significance level. Empirical results confirm that (i) the modified score statistic based on the bootstrap-resampling method performs better in the sense that its empirical type I error rate is much closer to the pre-specified nominal level than those of other tests and its power is greater than those of other tests, and is hence recommended, whilst the statistics based on the weighted least-squares estimate and logarithmic transformation are slightly conservative under some of the considered settings; (ii) the derived sample size formulae are rather accurate in the sense that their empirical powers obtained from the estimated sample sizes are very close to the pre-specified nominal powers. A real example is used to illustrate the proposed methodologies.  相似文献   

4.
Chen JJ  Lin KK  Huque M  Arani RB 《Biometrics》2000,56(2):586-592
A typical animal carcinogenicity experiment routinely analyzes approximately 10-30 tumor sites. Comparisons of tumor responses between dosed and control groups and dose-related trend tests are often evaluated for each individual tumor site/type separately. p-Value adjustment approaches have been proposed for controlling the overall Type I error rate or familywise error rate (FWE). However, these adjustments often result in reducing the power to detect a dose effect. This paper proposes using weighted adjustments by assuming that each tumor can be classified as either class A or class B based on prior considerations. The tumors in class A, which are considered as more critical endpoints, are given less adjustment. Two weighted methods of adjustments are presented, the weighted p adjustment and weighted alpha adjustment. A Monte Carlo simulation shows that both weighted adjustments control the FWE well. Furthermore, the power increases if a treatment-dependent tumor is analyzed as in class A tumors and the power decreases if it is analyzed as in class B tumors. A data set from a National Toxicology Program (NTP) 2-year animal carcinogenicity experiment with 13 tumor types/sites observed in male mice was analyzed using the proposed methods. The modified poly-3 test was used to test for increased carcinogenicity since it has been adopted by the NTP as a standard test for a dose-related trend. The unweighted adjustment analysis concluded that there was no statistically significant dose-related trend. Using the Food and Drug Administration classification scheme for the weighted adjustment analyses, two rare tumors (with background rates of 1% or less) were analyzed as class A tumors and 11 common tumors (with background rates higher than 1%) as class B. Both weighted analyses showed a significant dose-related trend for one rare tumor.  相似文献   

5.
Burman CF  Sonesson C 《Biometrics》2006,62(3):664-669
Flexible designs allow large modifications of a design during an experiment. In particular, the sample size can be modified in response to interim data or external information. A standard flexible methodology combines such design modifications with a weighted test, which guarantees the type I error level. However, this inference violates basic inference principles. In an example with independent N(mu, 1) observations, the test rejects the null hypothesis of mu < or = 0 while the average of the observations is negative. We conclude that flexible design in its most general form with the corresponding weighted test is not valid. Several possible modifications of the flexible design methodology are discussed with a focus on alternative hypothesis tests.  相似文献   

6.
OBJECTIVES: The association of a candidate gene with disease can be evaluated by a case-control study in which the genotype distribution is compared for diseased cases and unaffected controls. Usually, the data are analyzed with Armitage's test using the asymptotic null distribution of the test statistic. Since this test does not generally guarantee a type I error rate less than or equal to the significance level alpha, tests based on exact null distributions have been investigated. METHODS: An algorithm to generate the exact null distribution for both Armitage's test statistic and a recently proposed modification of the Baumgartner-Weiss-Schindler statistic is presented. I have compared the tests in a simulation study. RESULTS: The asymptotic Armitage test is slightly anticonservative whereas the exact tests control the type I error rate. The exact Armitage test is very conservative, but the exact test based on the modification of the Baumgartner-Weiss-Schindler statistic has a type I error rate close to alpha. The exact Armitage test is the least powerful test; the difference in power between the other two tests is often small and the comparison does not show a clear winner. CONCLUSION: Simulation results indicate that an exact test based on the modification of the Baumgartner-Weiss-Schindler statistic is preferable for the analysis of case-control studies of genetic markers.  相似文献   

7.
The Cochran-Armitage test has commonly been used for a trend test in binomial proportions. The quasi-likelihood method provides a simple approach to model extra-binomial proportions. Two versions of the score and Wald tests using different parameterizations for the extra-binomial variance were investigated: one in terms of intercluster correlation, and another in terms of variance. The Monte Carlo simulation was used to evaluate the performance of the each version of the score test and the Wald test, and the Cochran-Armitage test. The simulation shows that the Cochran-Armitage test has the proper size only for the binomial sample data, and the test is no longer valid when applied to the extra-binomial data. The Wald test is more likely to exceed the nominal level than the score test under either intercluster correlation model or variance model. Both score tests performed very well even with the binomial data; the tests control the type I error and in the meantime maintain the power of detecting the dose effects. Based on the design considered in this paper, the two scores test are comparable. The score test based on the intercluster correlations model seems better controlling the Type I error but appears less powerful than that based on the variance model. An example from a developmental toxicity experiment is given.  相似文献   

8.
Jonckheere's test is a frequently used nonparametric trend test for the evaluation of preclinical studies and clinical dose-finding trials. In this paper, a modification of Jonckheere's test is proposed. If the exact permutation distribution is used for inference, the modified test can fill out the level of the type I error in a much more complete way and is substantially more powerful than the common Jonckheere test. If the asymptotic normality is used for inference, the modified test is slightly more powerful. In addition, a maximum test is investigated which is more robust concerning an a priori unknown dose-response shape. The robustness is advantageous, especially in a closed testing procedure. The different tests are applied to two example data sets.  相似文献   

9.
McNemar's test is used to assess the difference between two different procedures (treatments) using independent matched-pair data. For matched-pair data collected in clusters, the tests proposed by Durkalski et al. and Obuchowski are popular and commonly used in practice since these tests do not require distributional assumptions or assumptions on the structure of the within-cluster correlation of the data. Motivated by these tests, this note proposes a modified Obuchowski test and illustrates comparisons of the proposed test with the extant methods. An extensive Monte Carlo simulation study suggests that the proposed test performs well with respect to the nominal size, and has higher power; Obuchowski's test is most conservative, and the performance of the Durkalski's test varies between the modified Obuchowski test and the original Obuchowski's test. These results form the basis for our recommendation that (i) for equal cluster size, the modified Obuchowski test is always preferred; (ii) for varying cluster size Durkalski's test can be used for a small number of clusters (e.g. K < 50), whereas for a large number of clusters (e.g. K ≥ 50) the modified Obuchowski test is preferred. Finally, to illustrate practical application of the competing tests, two real collections of clustered matched-pair data are analyzed.  相似文献   

10.
Observed variations in rates of taxonomic diversification have been attributed to a range of factors including biological innovations, ecosystem restructuring, and environmental changes. Before inferring causality of any particular factor, however, it is critical to demonstrate that the observed variation in diversity is significantly greater than that expected from natural stochastic processes. Relative tests that assess whether observed asymmetry in species richness between sister taxa in monophyletic pairs is greater than would be expected under a symmetric model have been used widely in studies of rate heterogeneity and are particularly useful for groups in which paleontological data are problematic. Although one such test introduced by Slowinski and Guyer a decade ago has been applied to a wide range of clades and evolutionary questions, the statistical behavior of the test has not been examined extensively, particularly when used with Fisher's procedure for combining probabilities to analyze data from multiple independent taxon pairs. Here, certain pragmatic difficulties with the Slowinski-Guyer test are described, further details of the development of a recently introduced likelihood-based relative rates test are presented, and standard simulation procedures are used to assess the behavior of the two tests in a range of situations to determine: (1) the accuracy of the tests' nominal Type I error rate; (2) the statistical power of the tests; (3) the sensitivity of the tests to inclusion of taxon pairs with few species; (4) the behavior of the tests with datasets comprised of few taxon pairs; and (5) the sensitivity of the tests to certain violations of the null model assumptions. Our results indicate that in most biologically plausible scenarios, the likelihood-based test has superior statistical properties in terms of both Type I error rate and power, and we found no scenario in which the Slowinski-Guyer test was distinctly superior, although the degree of the discrepancy varies among the different scenarios. The Slowinski-Guyer test tends to be much more conservative (i.e., very disinclined to reject the null hypothesis) in datasets with many small pairs. In most situations, the performance of both the likelihood-based test and particularly the Slowinski-Guyer test improve when pairs with few species are excluded from the computation, although this is balanced against a decline in the tests' power and accuracy as fewer pairs are included in the dataset. The performance of both tests is quite poor when they are applied to datasets in which the taxon sizes do not conform to the distribution implied by the usual null model. Thus, results of analyses of taxonomic rate heterogeneity using the Slowinski-Guyer test can be misleading because the test's ability to reject the null hypothesis (equal rates) when true is often inaccurate and its ability to reject the null hypothesis when the alternative (unequal rates) is true is poor, particularly when small taxon pairs are included. Although not always perfect, the likelihood-based test provides a more accurate and powerful alternative as a relative rates test.  相似文献   

11.
In this paper we give the generalization of the score tests covering the case of ties and we give examples where the expressions in matrix form are completely specified for the weighted tests and the score tests for the case of r groups. It is worth mentioning that although the score tests are not generally included in the commercial software, these tests should be used if it can be assumed that the censoring mechanism is equal in the r groups or if there is no censoring (Lawless , 1982). We establish the equivalence between “numerators” of these families of tests. As result of this equivalence we define four new tests that complete the classification of score and weighted tests. The Kruskal‐Wallis test (1952) appears as a particular case of the score tests for the case of non‐censoring. A simulation study has been done in order to compare the performance of the tests described in this paper. An example is included to make the understanding of the paper easier.  相似文献   

12.
Much forensic inference based upon DNA evidence is made assuming that the Hardy-Weinberg equilibrium (HWE) is valid for the genetic loci being used. Several statistical tests to detect and measure deviation from HWE have been devised, each having advantages and limitations. The limitations become more obvious when testing for deviation within multiallelic DNA loci is attempted. Here we present an exact test for HWE in the biallelic case, based on the ratio of weighted likelihoods under the null and alternative hypotheses, the Bayes factor. This test does not depend on asymptotic results and minimizes a linear combination of type I and type II errors. By ordering the sample space using the Bayes factor, we also define a significance (evidence) index, P value, using the weighted likelihood under the null hypothesis. We compare it to the conditional exact test for the case of sample size n = 10. Using the idea under the method of chi(2) partition, the test is used sequentially to test equilibrium in the multiple allele case and then applied to two short tandem repeat loci, using a real Caucasian data bank, showing its usefulness.  相似文献   

13.
The performance of diagnostic tests is often evaluated by estimating their sensitivity and specificity with respect to a traditionally accepted standard test regarded as a “gold standard” in making the diagnosis. Correlated samples of binary data arise in many fields of application. The fundamental unit for analysis is occasionally the site rather than the subject in site-specific studies. Statistical methods that take into account the within-subject corelation should be employed to estimate the sensitivity and the specificity of diagnostic tests since site-specific results within a subject can be highly correlated. I introduce several statistical methods for the estimation of the sensitivity and the specificity of sitespecific diagnostic tests. I apply these techniques to the data from a study involving an enzymatic diagnostic test to motivate and illustrate the estimation of the sensitivity and the specificity of periodontal diagnostic tests. I present results from a simulation study for the estimation of diagnostic sensitivity when the data are correlated within subjects. Through a simulation study, I compare the performance of the binomial estimator pCBE, the ratio estimator pCBE, the weighted estimator pCWE, the intracluster correlation estimator pCIC, and the generalized estimating equation (GEE) estimator PCGEE in terms of biases, observed variances, mean squared errors (MSE), relative efficiencies of their variances and 95 per cent coverage proportions. I recommend using PCBE when σ == 0. I recommend use of the weighted estimator PCWE when σ = 0.6. When σ == 0.2 or σ == 0.4, and the number of subjects is at least 30, PCGEE performs well.  相似文献   

14.
Two recently developed methods for the analysis of rare variants include the sequence kernel association test (SKAT) and the kernel-based adaptive cluster test (KBAC). While SKAT represents a type of variance component score test, and KBAC computes a weighted integral representing the difference in risk between variants, they appear to be developed using different initial principles. In this note, we show in fact that the KBAC can be modified to yield a test statistic with operating characteristics more similar to SKAT. Such a development relies on U- and V-statistic theory from mathematical statistics. Some simulation studies are used to evaluate the new proposed tests.  相似文献   

15.
When applying the Cochran‐Armitage (CA) trend test for an association between a candidate allele and a disease in a case‐control study, a set of scores must be assigned to the genotypes. Sasieni (1997, Biometrics 53 , 1253–1261) suggested scores for the recessive, additive, and dominant models but did not examine their statistical properties. Using the criteria of minimizing the required sample size of the CA trend test to achieve prespecified type I and type II errors, we show that the scores given by Sasieni (1997) are optimal for the recessive and dominant models and locally optimal for the additive one. Moreover, the additive scores are shown to be locally optimal for the multiplicative model. The tests are applied to a real dataset.  相似文献   

16.
Yan Li  Barry I. Graubard 《Biometrics》2009,65(4):1096-1104
Summary For studies on population genetics, the use of representative random samples of the target population can avoid ascertainment bias. Genetic variation data from over a hundred genes were collected in a U.S. nationally representative sample in the Third National Health and Nutrition Examination Survey (NHANES III). Surveys such as the NHANES have complex stratified multistage cluster sample designs with sample weighting that can inflate variances and alter the expectations of test statistics. Thus, classical statistical tests of Hardy–Weinberg equilibrium (HWE) and homogeneity of HW disequilibrium (HHWD) for simple random samples are not suitable for data from complex samples. We propose using Wald tests for HWE and generalized score tests for HHWD that have been modified for complex samples. Monte Carlo simulation studies are used to investigate the finite sample properties of the proposed tests. Rao–Scott corrections applied to the tests were found to improve their type I error properties. Our methods are applied to the NHANES III genetic data for three loci involved in metabolizing lead in the body.  相似文献   

17.
In the analysis of gene expression by microarrays there are usually few subjects, but high-dimensional data. By means of techniques, such as the theory of spherical tests or with suitable permutation tests, it is possible to sort the endpoints or to give weights to them according to specific criteria determined by the data while controlling the multiple type I error rate. The procedures developed so far are based on a sequential analysis of weighted p-values (corresponding to the endpoints), including the most extreme situation of weighting leading to a complete order of p-values. When the data for the endpoints have approximately equal variances, these procedures show good power properties. In this paper, we consider an alternative procedure, which is based on completely sorting the endpoints, but smoothed in the sense that some perturbations in the sequence of the p-values are allowed. The procedure is relatively easy to perform, but has high power under the same restrictions as for the weight-based procedures.  相似文献   

18.
A statistical model for doubled haploids and backcrosses based on the interval-mapping methodology has been used to carry out power studies to investigate the effects of different experimental designs, heritabilities of the quantitative trait, and types of gene action, using two test statistics, the F of Fisher-Snedecor and the LOD score. The doubled haploid experimental design is more powerful than backcrosses while keeping actual type I errors similar to nominal ones. For the doubled haploid design, individual QTLs, showing heritabilities as low as 5% were detected in about 90% of the cases using only 250 individuals. The power to detect a given QTL is related to its contribution to the heritability of the trait. For a given nominal type I error, tests using F values are more powerful than with LOD scores. It seems that more conservative levels should be used for the LOD score in order to increase the power and obtain type I errors similar to nominal ones.  相似文献   

19.
Rotnitzky A  Li L  Li X 《Biometrika》2010,97(4):997-1001
Standardized means, commonly used in observational studies in epidemiology to adjust for potential confounders, are equal to inverse probability weighted means with inverse weights equal to the empirical propensity scores. More refined standardization corresponds with empirical propensity scores computed under more flexible models. Unnecessary standardization induces efficiency loss. However, according to the theory of inverse probability weighted estimation, propensity scores estimated under more flexible models induce improvement in the precision of inverse probability weighted means. This apparent contradiction is clarified by explicitly stating the assumptions under which the improvement in precision is attained.  相似文献   

20.
In ecological field surveys, observations are gathered at different spatial locations. The purpose may be to relate biological response variables (e.g., species abundances) to explanatory environmental variables (e.g., soil characteristics). In the absence of prior knowledge, ecologists have been taught to rely on systematic or random sampling designs. If there is prior knowledge about the spatial patterning of the explanatory variables, obtained from either previous surveys or a pilot study, can we use this information to optimize the sampling design in order to maximize our ability to detect the relationships between the response and explanatory variables?
The specific questions addressed in this paper are: a) What is the effect (type I error) of spatial autocorrelation on the statistical tests commonly used by ecologists to analyse field survey data? b) Can we eliminate, or at least minimize, the effect of spatial autocorrelation by the design of the survey? Are there designs that provide greater power for surveys, at least under certain circumstances? c) Can we eliminate or control for the effect of spatial autocorrelation during the analysis? To answer the last question, we compared regular regression analysis to a modified t‐test developed by Dutilleul for correlation coefficients in the presence of spatial autocorrelation.
Replicated surfaces (typically, 1000 of them) were simulated using different spatial parameters, and these surfaces were subjected to different sampling designs and methods of statistical analysis. The simulated surfaces may represent, for example, vegetation response to underlying environmental variation. This allowed us 1) to measure the frequency of type I error (the failure to reject the null hypothesis when in fact there is no effect of the environment on the response variable) and 2) to estimate the power of the different combinations of sampling designs and methods of statistical analysis (power is measured by the rate of rejection of the null hypothesis when an effect of the environment on the response variable has been created).
Our results indicate that: 1) Spatial autocorrelation in both the response and environmental variables affects the classical tests of significance of correlation or regression coefficients. Spatial autocorrelation in only one of the two variables does not affect the test of significance. 2) A broad‐scale spatial structure present in data has the same effect on the tests as spatial autocorrelation. When such a structure is present in one of the variables and autocorrelation is found in the other, or in both, the tests of significance have inflated rates of type I error. 3) Dutilleul's modified t‐test for the correlation coefficient, corrected for spatial autocorrelation, effectively corrects for spatial autocorrelation in the data. It also effectively corrects for the presence of deterministic structures, with or without spatial autocorrelation.
The presence of a broad‐scale deterministic structure may, in some cases, reduce the power of the modified t‐test.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号