首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
2.
A parametric approach fits particular classes of parametric models to the data, uses the model parameter estimates as summaries and tests for differences between groups by comparing fits with and without the assumption of common parameter values across groups. The paper discusses how a parametric approach can be implemented in the specific context of a single‐factor replicated spatial experiment and uses simulations to show when the parametric approach can be efficient or potentially misleading. An analysis of the spatial distribution of pyramidal neurons in human patients is also shown.  相似文献   

3.
This paper presents new biostatistical methods for the analysis of microbiome data based on a fully parametric approach using all the data. The Dirichlet-multinomial distribution allows the analyst to calculate power and sample sizes for experimental design, perform tests of hypotheses (e.g., compare microbiomes across groups), and to estimate parameters describing microbiome properties. The use of a fully parametric model for these data has the benefit over alternative non-parametric approaches such as bootstrapping and permutation testing, in that this model is able to retain more information contained in the data. This paper details the statistical approaches for several tests of hypothesis and power/sample size calculations, and applies them for illustration to taxonomic abundance distribution and rank abundance distribution data using HMP Jumpstart data on 24 subjects for saliva, subgingival, and supragingival samples. Software for running these analyses is available.  相似文献   

4.
Concern about the sustainability of intercontinental‐scale migration systems grows apace with global change. Traditional organism‐centred approaches to this problem have provided insights at the population level, but not at the systems level. We are sceptical that an accumulation of data from a species‐by‐species approach will yield an understanding of these systems in the near term. As an alternative, we advocate a new research programme that grows from an explicitly system‐based framework that leverages existing Earth observation infrastructure to make inferences directly at the macrosystem level. We illustrate how this approach can be used to generate and test system‐level predictions, using NEXRAD radar data as an example. We urge organismal ecologists to recognize that some of the most urgent migration questions are at the macrosystem scale and that tackling these questions requires an interdisciplinary approach if we are to make progress at a pace that exceeds that of climate change.  相似文献   

5.
PARCAT is a computer program which implements alternative tests for average partial association in three-way contingency tables within the framework of the product multiple hypergeometric probability model. Primary attention is directed at the relationship between two of the variables, controlling for the effects of a covariable. This approach is essentially a multivariate extension of the Cochran/Mantel-Haenszel test to sets of (s x r) tables. A set of scores such as uniform, ridits, or probits can be assigned to categories which are ordinally scaled. In particular, if ridit scores with midranks assigned for ties are utilized, this procedure is equivalent to a partial Kruskal-Wallis test when one variable is ordinally scaled, and is equivalent to a partial Spearman rank correlation test when both variables are ordinally scaled.  相似文献   

6.
In clinical and epidemiological studies information on the primary outcome of interest, that is, the disease status, is usually collected at a limited number of follow‐up visits. The disease status can often only be retrieved retrospectively in individuals who are alive at follow‐up, but will be missing for those who died before. Right‐censoring the death cases at the last visit (ad‐hoc analysis) yields biased hazard ratio estimates of a potential risk factor, and the bias can be substantial and occur in either direction. In this work, we investigate three different approaches that use the same likelihood contributions derived from an illness‐death multistate model in order to more adequately estimate the hazard ratio by including the death cases into the analysis: a parametric approach, a penalized likelihood approach, and an imputation‐based approach. We investigate to which extent these approaches allow for an unbiased regression analysis by evaluating their performance in simulation studies and on a real data example. In doing so, we use the full cohort with complete illness‐death data as reference and artificially induce missing information due to death by setting discrete follow‐up visits. Compared to an ad‐hoc analysis, all considered approaches provide less biased or even unbiased results, depending on the situation studied. In the real data example, the parametric approach is seen to be too restrictive, whereas the imputation‐based approach could almost reconstruct the original event history information.  相似文献   

7.
For nonnormal data we suggest a test of location based on a broader family of distributions than normality. Such a test will in a sense fall between the standard parametric and non parametric tests. We see that the Wald tests based on this family of distributions have some advantages over the score tests and that they perform well in comparison to standard parametric and nonparametric tests in a variety of situations. We also consider when and how to apply such tests in practice.  相似文献   

8.
9.
Evaluating statistical trends in high‐dimensional phenotypes poses challenges for comparative biologists, because the high‐dimensionality of the trait data relative to the number of species can prohibit parametric tests from being computed. Recently, two comparative methods were proposed to circumvent this difficulty. One obtains phylogenetic independent contrasts for all variables, and statistically evaluates the linear model by permuting the phylogenetically independent contrasts (PICs) of the response data. The other uses a distance‐based approach to obtain coefficients for generalized least squares models (D‐PGLS), and subsequently permutes the original data to evaluate the model effects. Here, we show that permuting PICs is not equivalent to permuting the data prior to the analyses as in D‐PGLS. We further explain why PICs are not the correct exchangeable units under the null hypothesis, and demonstrate that this misspecification of permutable units leads to inflated type I error rates of statistical tests. We then show that simply shuffling the original data and recalculating the independent contrasts with each iteration yields significance levels that correspond to those found using D‐PGLS. Thus, while summary statistics from methods based on PICs and PGLS are the same, permuting PICs can lead to strikingly different inferential outcomes with respect to statistical and biological inferences.  相似文献   

10.
Although linear rank statistics for the two‐sample problem are distribution free tests, their power depends on the distribution of the data. In the planning phase of an experiment, researchers are often uncertain about the shape of this distribution and so the choice of test statistic for the analysis and the determination of the required sample size are based on vague information. Adaptive designs with interim analysis can potentially overcome both problems. And in particular, adaptive tests based on a selector statistic are a solution to the first. We investigate whether adaptive tests can be usefully implemented in flexible two‐stage designs to gain power. In a simulation study, we compare several methods for choosing a test statistic for the second stage of an adaptive design based on interim data with the procedure that applies adaptive tests in both stages. We find that the latter is a sensible approach that leads to the best results in most situations considered here. The different methods are illustrated using a clinical trial example.  相似文献   

11.
Species’ responses at the genetic level are key to understanding the long‐term consequences of anthropogenic global change. Herbaria document such responses, and, with contemporary sampling, provide high‐resolution time‐series of plant evolutionary change. Characterizing genetic diversity is straightforward for model species with small genomes and a reference sequence. For nonmodel species—with small or large genomes—diversity is traditionally assessed using restriction‐enzyme‐based sequencing. However, age‐related DNA damage and fragmentation preclude the use of this approach for ancient herbarium DNA. Here, we combine reduced‐representation sequencing and hybridization‐capture to overcome this challenge and efficiently compare contemporary and historical specimens. Specifically, we describe how homemade DNA baits can be produced from reduced‐representation libraries of fresh samples, and used to efficiently enrich historical libraries for the same fraction of the genome to produce compatible sets of sequence data from both types of material. Applying this approach to both Arabidopsis thaliana and the nonmodel plant Cardamine bulbifera, we discovered polymorphisms de novo in an unbiased, reference‐free manner. We show that the recovered genetic variation recapitulates known genetic diversity in A. thaliana, and recovers geographical origin in both species and over time, independent of bait diversity. Hence, our method enables fast, cost‐efficient, large‐scale integration of contemporary and historical specimens for assessment of genome‐wide genetic trends over time, independent of genome size and presence of a reference genome.  相似文献   

12.
Numerous statistical methods have been developed for analyzing high‐dimensional data. These methods often focus on variable selection approaches but are limited for the purpose of testing with high‐dimensional data. They are often required to have explicit‐likelihood functions. In this article, we propose a “hybrid omnibus test” for high‐dicmensional data testing purpose with much weaker requirements. Our hybrid omnibus test is developed under a semiparametric framework where a likelihood function is no longer necessary. Our test is a version of a frequentist‐Bayesian hybrid score‐type test for a generalized partially linear single‐index model, which has a link function being a function of a set of variables through a generalized partially linear single index. We propose an efficient score based on estimating equations, define local tests, and then construct our hybrid omnibus test using local tests. We compare our approach with an empirical‐likelihood ratio test and Bayesian inference based on Bayes factors, using simulation studies. Our simulation results suggest that our approach outperforms the others, in terms of type I error, power, and computational cost in both the low‐ and high‐dimensional cases. The advantage of our approach is demonstrated by applying it to genetic pathway data for type II diabetes mellitus.  相似文献   

13.
Lin S 《Human heredity》2002,53(2):103-112
We have previously proposed a confidence set approach for finding tightly linked genomic regions under the setting of parametric linkage analysis. In this article, we extend the confidence set approach to nonparametric linkage analysis of affected sib pair (ASP) data based on their identity-by-descent (IBD) information. Two well-known statistics in nonparametric linkage analysis, the Two-IBD test (proportion of ASPs sharing two alleles IBD), and the Mean test (average number of alleles shared IBD in the ASPs), are used for constructing confidence sets. Some numerical analyses as well as a simulation study were carried out to demonstrate the utility of the methods. Our results show that the fundamental advantages of the confidence set approach in parametric linkage analysis are retained when the method is generalized to nonparametric analysis. Our study on the accuracy of confidence sets, in terms of choice of tests, underlying disease incidence data, and amount of data available, leads us to conclude, among other things, that the Mean test outperforms the Two-IBD test in most situations, with the reverse being true only for traits with small additive variance. Although we describe how to construct confidence sets based on only two familiar tests, one can construct confidence sets similarly using other allele sharing statistics.  相似文献   

14.
Paired data arises in a wide variety of applications where often the underlying distribution of the paired differences is unknown. When the differences are normally distributed, the t‐test is optimum. On the other hand, if the differences are not normal, the t‐test can have substantially less power than the appropriate optimum test, which depends on the unknown distribution. In textbooks, when the normality of the differences is questionable, typically the non‐parametric Wilcoxon signed rank test is suggested. An adaptive procedure that uses the Shapiro‐Wilk test of normality to decide whether to use the t‐test or the Wilcoxon signed rank test has been employed in several studies. Faced with data from heavy tails, the U.S. Environmental Protection Agency (EPA) introduced another approach: it applies both the sign and t‐tests to the paired differences, the alternative hypothesis is accepted if either test is significant. This paper investigates the statistical properties of a currently used adaptive test, the EPA's method and suggests an alternative technique. The new procedure is easy to use and generally has higher empirical power, especially when the differences are heavy‐tailed, than currently used methods.  相似文献   

15.
Behavioural studies are commonly plagued with data that violate the assumptions of parametric statistics. Consequently, classic nonparametric methods (e.g. rank tests) and novel distribution-free methods (e.g. randomization tests) have been used to a great extent by behaviourists. However, the robustness of such methods in terms of statistical power and type I error have seldom been evaluated. This probably reflects the fact that empirical methods, such as Monte Carlo approaches, are required to assess these concerns. In this study we show that analytical methods cannot always be used to evaluate the robustness of statistical tests, but rather Monte Carlo approaches must be employed. We detail empirical protocols for estimating power and type I error rates for parametric, nonparametric and randomization methods, and demonstrate their application for an analysis of variance and a regression/correlation analysis design. Together, this study provides a framework from which behaviourists can compare the reliability of different methods for data analysis, serving as a basis for selecting the most appropriate statistical test given the characteristics of data at hand. Copyright 2001 The Association for the Study of Animal Behaviour.  相似文献   

16.
BRIDGET PRATT  BEBE LOFF 《Bioethics》2013,27(4):208-214
Health research has been identified as a vehicle for advancing global justice in health. However, in bioethics, issues of global justice are mainly discussed within an ongoing debate on the conditions under which international clinical research is permissible. As a result, current ethical guidance predominantly links one type of international research (biomedical) to advancing one aspect of health equity (access to new treatments). International guidelines largely fail to connect international research to promoting broader aspects of health equity – namely, healthier social environments and stronger health systems. Bioethical frameworks such as the human development approach do consider how international clinical research is connected to the social determinants of health but, again, do so to address the question of when international clinical research is permissible. It is suggested that the narrow focus of this debate is shaped by high‐income countries' economic strategies. The article further argues that the debate's focus obscures a stronger imperative to consider how other types of international research might advance justice in global health. Bioethics should consider the need for non‐clinical health research and its contribution to advancing global justice.  相似文献   

17.
Indirect gradient analysis, or ordination, is primarily a method of exploratory data analysis. However, to support biological interpretations of resulting axes as vegetation gradients, or later confirmatory analyses and statistical tests, these axes need to be stable or at least robust into minor sampling effects. We develop a computer-intensive bootstrap (resampling) approach to estimate sampling effects on solutions from nonlinear ordination.We apply this approach to simulated data and to three forest data sets from North Carolina, USA and examine the resulting patterns of local and global instability in detrended correspondence analysis (DCA) solutions. We propose a bootstrap coefficient, scaled rank variance (SRV), to estimate remaining instability in species ranks after rotating axes to a common global orientation. In analysis of simulated data, bootstrap SRV was generally consistent with an equivalent estimate from repeated sampling. In an example using field data SRV, bootstrapped DCA showed good recovery of the order of common species along the first two axes, but poor recovery of later axes. We also suggest some criteria to use with the SRV to decide how many axes to retain and attempt to interpret.Abbreviations DCA= detrended correspondence analysis - SRV= scaled rank variance  相似文献   

18.
Lei Xu  Jun Shao 《Biometrics》2009,65(4):1175-1183
Summary In studies with longitudinal or panel data, missing responses often depend on values of responses through a subject‐level unobserved random effect. Besides the likelihood approach based on parametric models, there exists a semiparametric method, the approximate conditional model (ACM) approach, which relies on the availability of a summary statistic and a linear or polynomial approximation to some random effects. However, two important issues must be addressed in applying ACM. The first is how to find a summary statistic and the second is how to estimate the parameters in the original model using estimates of parameters in ACM. Our study is to address these two issues. For the first issue, we derive summary statistics under various situations. For the second issue, we propose to use a grouping method, instead of linear or polynomial approximation to random effects. Because the grouping method is a moment‐based approach, the conditions we assumed in deriving summary statistics are weaker than the existing ones in the literature. When the derived summary statistic is continuous, we propose to use a classification tree method to obtain an approximate summary statistic for grouping. Some simulation results are presented to study the finite sample performance of the proposed method. An application is illustrated using data from the study of Modification of Diet in Renal Disease.  相似文献   

19.
We explore the estimation of uncertainty in evolutionary parameters using a recently devised approach for resampling entire additive genetic variance–covariance matrices ( G ). Large‐sample theory shows that maximum‐likelihood estimates (including restricted maximum likelihood, REML) asymptotically have a multivariate normal distribution, with covariance matrix derived from the inverse of the information matrix, and mean equal to the estimated G . This suggests that sampling estimates of G from this distribution can be used to assess the variability of estimates of G , and of functions of G . We refer to this as the REML‐MVN method. This has been implemented in the mixed‐model program WOMBAT. Estimates of sampling variances from REML‐MVN were compared to those from the parametric bootstrap and from a Bayesian Markov chain Monte Carlo (MCMC) approach (implemented in the R package MCMCglmm). We apply each approach to evolvability statistics previously estimated for a large, 20‐dimensional data set for Drosophila wings. REML‐MVN and MCMC sampling variances are close to those estimated with the parametric bootstrap. Both slightly underestimate the error in the best‐estimated aspects of the G matrix. REML analysis supports the previous conclusion that the G matrix for this population is full rank. REML‐MVN is computationally very efficient, making it an attractive alternative to both data resampling and MCMC approaches to assessing confidence in parameters of evolutionary interest.  相似文献   

20.
Till now, multivariate reference regions have played only a marginal role in the practice of clinical chemistry and laboratory medicine. The major reason for this fact is that such regions are traditionally determined by means of concentration ellipsoids of multidimensional Gaussian distributions yielding reference limits which do not allow statements about possible outlyingness of measurements taken in specific diagnostic tests from a given patient or subject. As a promising way around this difficulty we propose to construct multivariate reference regions as p-dimensional rectangles or (in the one-sided case) rectangular half-spaces whose edges determine univariate percentile ranges of the same probability content in each marginal distribution. In a first step, the corresponding notion of a quantile of a p-dimensional probability distribution of any type and shape is made mathematically precise. Subsequently, both parametric and nonparametric procedures of estimating such a quantile are described. Furthermore, results on sample-size calculation for reference-centile studies based on the proposed definition of multivariate quantiles are presented generalizing the approach of Jennen-Steinmetz and Wellek.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号