首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
Perlman MD  Wu L 《Biometrics》2004,60(1):276-280
Testing problems with multivariate one-sided alternative hypotheses are common in clinical trials with multiple endpoints. In the case of comparing two treatments, treatment 1 is often preferred if it is superior for at least one of the endpoints and not biologically inferior for the remaining endpoints. Bloch et al. (2001, Biometrics57, 1039-1047) propose an intersection-union test (IUT) for this testing problem, but their test does not utilize the appropriate multivariate one-sided test. In this note we modify their test by an alternative IUT that does utilize the appropriate one-sided test. Empirical and graphical evidence show that the proposed test is more appropriate for this testing problem.  相似文献   

2.
Till now, multivariate reference regions have played only a marginal role in the practice of clinical chemistry and laboratory medicine. The major reason for this fact is that such regions are traditionally determined by means of concentration ellipsoids of multidimensional Gaussian distributions yielding reference limits which do not allow statements about possible outlyingness of measurements taken in specific diagnostic tests from a given patient or subject. As a promising way around this difficulty we propose to construct multivariate reference regions as p-dimensional rectangles or (in the one-sided case) rectangular half-spaces whose edges determine univariate percentile ranges of the same probability content in each marginal distribution. In a first step, the corresponding notion of a quantile of a p-dimensional probability distribution of any type and shape is made mathematically precise. Subsequently, both parametric and nonparametric procedures of estimating such a quantile are described. Furthermore, results on sample-size calculation for reference-centile studies based on the proposed definition of multivariate quantiles are presented generalizing the approach of Jennen-Steinmetz and Wellek.  相似文献   

3.
Functional data are smooth, often continuous, random curves, which can be seen as an extreme case of multivariate data with infinite dimensionality. Just as componentwise inference for multivariate data naturally performs feature selection, subsetwise inference for functional data performs domain selection. In this paper, we present a unified testing framework for domain selection on populations of functional data. In detail, p-values of hypothesis tests performed on pointwise evaluations of functional data are suitably adjusted for providing control of the familywise error rate (FWER) over a family of subsets of the domain. We show that several state-of-the-art domain selection methods fit within this framework and differ from each other by the choice of the family over which the control of the FWER is provided. In the existing literature, these families are always defined a priori. In this work, we also propose a novel approach, coined thresholdwise testing, in which the family of subsets is instead built in a data-driven fashion. The method seamlessly generalizes to multidimensional domains in contrast to methods based on a priori defined families. We provide theoretical results with respect to consistency and control of the FWER for the methods within the unified framework. We illustrate the performance of the methods within the unified framework on simulated and real data examples and compare their performance with other existing methods.  相似文献   

4.
The novel two-step serologic sensitive/less sensitive testing algorithm for detecting recent HIV seroconversion (STARHS) provides a simple and practical method to estimate HIV-1 incidence using cross-sectional HIV seroprevalence data. STARHS has been used increasingly in epidemiologic studies. However, the uncertainty of incidence estimates using this algorithm has not been well described, especially for high risk groups or when missing data is present because a fraction of sensitive enzyme immunoassay (EIA) positive specimens are not tested by the less sensitive EIA. Ad hoc methods used in practice provide incorrect confidence limits and thus may jeopardize statistical inference. In this report, we propose maximum likelihood and Bayesian methods for correctly estimating the uncertainty in incidence estimates obtained using prevalence data with a fraction missing, and extend the methods to regression settings. Using a study of injection drug users participating in a drug detoxification program in New York city as an example, we demonstrated the impact of underestimating the uncertainty in incidence estimates using ad hoc methods. Our methods can be applied to estimate the incidence of other diseases from prevalence data using similar testing algorithms when missing data is present.  相似文献   

5.
Exposure to air pollution is associated with increased morbidity and mortality. Recent technological advancements permit the collection of time-resolved personal exposure data. Such data are often incomplete with missing observations and exposures below the limit of detection, which limit their use in health effects studies. In this paper, we develop an infinite hidden Markov model for multiple asynchronous multivariate time series with missing data. Our model is designed to include covariates that can inform transitions among hidden states. We implement beam sampling, a combination of slice sampling and dynamic programming, to sample the hidden states, and a Bayesian multiple imputation algorithm to impute missing data. In simulation studies, our model excels in estimating hidden states and state-specific means and imputing observations that are missing at random or below the limit of detection. We validate our imputation approach on data from the Fort Collins Commuter Study. We show that the estimated hidden states improve imputations for data that are missing at random compared to existing approaches. In a case study of the Fort Collins Commuter Study, we describe the inferential gains obtained from our model including improved imputation of missing data and the ability to identify shared patterns in activity and exposure among repeated sampling days for individuals and among distinct individuals.  相似文献   

6.
Becker T  Knapp M 《Human heredity》2005,59(4):185-189
In the context of haplotype association analysis of unphased genotype data, methods based on Monte-Carlo simulations are often used to compensate for missing or inappropriate asymptotic theory. Moreover, such methods are an indispensable means to deal with multiple testing problems. We want to call attention to a potential trap in this usually useful approach: The simulation approach may lead to strongly inflated type I errors in the presence of different missing rates between cases and controls, depending on the chosen test statistic. Here, we consider four different testing strategies for haplotype analysis of case-control data. We recommend to interpret results for data sets with non-comparable distributions of missing genotypes with special caution, in case the test statistic is based on inferred haplotypes per individual. Moreover, our results are important for the conduction and interpretation of genome-wide association studies.  相似文献   

7.
Resampling-based multiple testing methods that control the Familywise Error Rate in the strong sense are presented. It is shown that no assumptions whatsoever on the data-generating process are required to obtain a reasonably powerful and flexible class of multiple testing procedures. Improvements are obtained with mild assumptions. The methods are applicable to gene expression data in particular, but more generally to any multivariate, multiple group data that may be character or numeric. The role of the disputed "subset pivotality" condition is clarified.  相似文献   

8.
基于多元统计分析中对样本完整性的要求,为了在分析中不抛弃大量不完整的化石标本或者不大大减少变量,创建了一种恢复标本残缺数据的方法。本方法基于线性回归理论,假设同类标本个体之间的区别仅仅是大小的区别,形状的区别可以忽略不计,因此,在同类标本中,可以用一件标本的已知测量数据预测另一件标本的残缺测量数据。在多件标本的情况下,对某件标本的某个残缺数据的预测结果是用其他标本分别进行预测所得值的加权平均,加权系数的选取与每件标本的保存完好程度相关。用现生马属头骨及肢骨标本做的数据试验证明,该方法具有良好的稳定性,对标本的种类、数量及残缺值的多少均不敏感,对于尺寸较大的标本或数值较大的数据的预测效果要比对尺寸较小的标本或数值较小的数据的预测效果要好。与传统的线性回归方法的不同之处在于,本方法利用的是样本(即标本)间的线性相关性,传统方法利用的是变量(即测量项)间的线性相关性。在通常情况下,样本间的线性相关程度要优于变量间的线性相关程度。本方法简单实用,在对化石标本进行统计分析,特别是多元统计分析中具有良好的应用前景。  相似文献   

9.
The presence of missing values in gel-based proteomics data represents a real challenge if an objective statistical analysis is pursued. Different methods to handle missing values were evaluated and their influence is discussed on the selection of important proteins through multivariate techniques. The evaluated methods consisted of directly dealing with them during the multivariate analysis with the nonlinear estimation by iterative partial least squares (NIPALS) algorithm or imputing them by using either k-nearest neighbor or Bayesian principal component analysis (BPCA) before carrying out the multivariate analysis. These techniques were applied to data obtained from gels stained with classical postrunning dyes and from DIGE gels. Before applying the multivariate techniques, the normality and homoscedasticity assumptions on which parametric tests are based on were tested in order to perform a sound statistical analysis. From the three tested methods to handle missing values in our datasets, BPCA imputation of missing values showed to be the most consistent method.  相似文献   

10.
In this paper we propose a multivariate extension of family-based association tests based on generalized estimating equations. The test can be applied to multiple phenotypes and to phenotypic data obtained in longitudinal studies without making any distributional assumptions for the phenotypic observations. Methods for handling missing phenotypic information are discussed. Further, we compare the power of the multivariate test with permutation tests and with using separate tests for each outcome which are adjusted for multiple testing. Application of the proposed test to an asthma study illustrates the power of the approach.  相似文献   

11.
Summary Permutation tests based on distances among multivariate observations have found many applications in the biological sciences. Two major testing frameworks of this kind are multiresponse permutation procedures and pseudo‐F tests arising from a distance‐based extension of multivariate analysis of variance. In this article, we derive conditions under which these two frameworks are equivalent. The methods and equivalence results are illustrated by reanalyzing an ecological data set and by a novel application to functional magnetic resonance imaging data.  相似文献   

12.
The Apolipoprotein-E (Apo-E) gene, a gene that produces proteins which help to regulate lipid levels in the bloodstream, is of interest in the study of cardiovascular diseases. An approach to making inferences about the genetic effects of the Apo-E gene has been developed by Glickman and Gagnon (2002). The framework describes the role of genetic and risk factors on the onset ages of multiple diseases, and accounts for the possibility that an individual was censored for reasons related to the diseases of interest. The framework also allows for missing genetic information, so that subjects censored prior to genetic sampling, and therefore missing such information, may still be included in the analysis. We apply an extension to this framework to the original cohort of the Framingham Heart Study for measuring the effects of different Apo-E genotypes on the onset age of various cardiovascular disease events. In particular, we compare the fit of univariate versus multivariate onset age components to the model, whether to incorporate health covariates measured at baseline or at a point later in the study, and whether to assume a heritability model for Apo-E genotype frequencies. The results of the best fitting model are presented.  相似文献   

13.
C S Davis  L J Wei 《Biometrics》1988,44(4):1005-1018
In comparing the effectiveness of two treatments, suppose that nondecreasing repeated measurements of the same characteristic are scheduled to be taken over a common set of time points for each study subject. A class of univariate one-sided global asymptotically distribution-free tests is proposed to test the equality of the two treatments. The test procedures allow different patterns of missing observations in the two groups to be compared, although the missing data mechanism is required to be independent of the observations in each treatment group. Test-based point and interval estimators of the global treatment difference are given. Multiple inference procedures are also provided to examine the time trend of treatment differences over the entire study. The proposed methods are illustrated by an example from a bladder cancer study.  相似文献   

14.
W W Piegorsch 《Biometrics》1990,46(2):309-316
Dichotomous response models are common in many experimental settings. Often, concomitant explanatory variables are recorded, and a generalized linear model, such as a logit model, is fit. In some cases, interest in specific model parameters is directed only at one-sided departures from some null effect. In these cases, procedures can be developed for testing the null effect against a one-sided alternative. These include Bonferroni-type adjustments of univariate Wald tests, and likelihood ratio tests that employ inequality-constrained multivariate theory. This paper examines such tests of significance. Monte Carlo evaluations are undertaken to examine the small-sample properties of the various procedures. The procedures are seen to perform fairly well, generally achieving their nominal sizes at total sample sizes near 100 experimental units. Extensions to the problem of one-sided tests against a control or standard are also considered.  相似文献   

15.
Tang NS  Tang ML 《Biometrics》2002,58(4):972-980
In this article, we consider small-sample statistical inference for rate ratio (RR) in a correlated 2 x 2 table with a structural zero in one of the off-diagonal cells. Existing Wald's test statistic and logarithmic transformation test statistic will be adopted for this purpose. Hypothesis testing and confidence interval construction based on large-sample theory will be reviewed first. We then propose reliable small-sample exact unconditional procedures for hypothesis testing and confidence interval construction. We present empirical results to evince the better confidence interval performance of our proposed exact unconditional procedures over the traditional large-sample procedures in small-sample designs. Unlike the findings given in Lui (1998, Biometrics 54, 706-711), our empirical studies show that the existing asymptotic procedures may not attain a prespecified confidence level even in moderate sample-size designs (e.g., n = 50). Our exact unconditional procedures on the other hand do not suffer from this problem. Hence, the asymptotic procedures should be applied with caution. We propose two approximate unconditional confidence interval construction methods that outperform the existing asymptotic ones in terms of coverage probability and expected interval width. Also, we empirically demonstrate that the approximate unconditional tests are more powerful than their associated exact unconditional tests. A real data set from a two-step tuberculosis testing study is used to illustrate the methodologies.  相似文献   

16.
Metric data are usually assessed on a continuous scale with good precision, but sometimes agricultural researchers cannot obtain precise measurements of a variable. Values of such a variable cannot then be expressed as real numbers (e.g., 1.51 or 2.56), but often can be represented by intervals into which the values fall (e.g., from 1 to 2 or from 2 to 3). In this situation, statisticians talk about censoring and censored data, as opposed to missing data, where no information is available at all. Traditionally, in agriculture and biology, three methods have been used to analyse such data: (a) when intervals are narrow, some form of imputation (e.g., mid‐point imputation) is used to replace the interval and traditional methods for continuous data are employed (such as analyses of variance [ANOVA] and regression); (b) for time‐to‐event data, the cumulative proportions of individuals that experienced the event of interest are analysed, instead of the individual observed times‐to‐event; (c) when intervals are wide and many individuals are collected, non‐parametric methods of data analysis are favoured, where counts are considered instead of the individual observed value for each sample element. In this paper, we show that these methods may be suboptimal: The first one does not respect the process of data collection, the second leads to unreliable standard errors (SEs), while the third does not make full use of all the available information. As an alternative, methods of survival analysis for censored data can be useful, leading to reliable inferences and sound hypotheses testing. These methods are illustrated using three examples from plant and crop sciences.  相似文献   

17.
Geometric morphometric methods constitute a powerful and precise tool for the quantification of morphological differences. The use of geometric morphometrics in palaeontology is very often limited by missing data. Shape analysis methods based on landmarks are very sensible but until now have not been adapted to this kind of dataset. To analyze the prospective utility of this method for fossil taxa, we propose a model based on prosimian cranial morphology in which we test two methods of missing data reconstruction. These consist of generating missing-data in a dataset (by increments of five percent) and estimating missing data using two multivariate methods. Estimates were found to constitute a useful tool for the analysis of partial datasets (to a certain extent). These results are promising for future studies of morphological variation in fossil taxa.  相似文献   

18.
IntroductionKinetic compartmental analysis is frequently used to compute physiologically relevant quantitative values from time series of images. In this paper, a new approach based on Bayesian analysis to obtain information about these parameters is presented and validated.Materials and methodsThe closed-form of the posterior distribution of kinetic parameters is derived with a hierarchical prior to model the standard deviation of normally distributed noise. Markov chain Monte Carlo methods are used for numerical estimation of the posterior distribution. Computer simulations of the kinetics of F18-fluorodeoxyglucose (FDG) are used to demonstrate drawing statistical inferences about kinetic parameters and to validate the theory and implementation. Additionally, point estimates of kinetic parameters and covariance of those estimates are determined using the classical non-linear least squares approach.Results and discussionPosteriors obtained using methods proposed in this work are accurate as no significant deviation from the expected shape of the posterior was found (one-sided P > 0.08). It is demonstrated that the results obtained by the standard non-linear least-square methods fail to provide accurate estimation of uncertainty for the same data set (P < 0.0001).ConclusionsThe results of this work validate new methods for a computer simulations of FDG kinetics. Results show that in situations where the classical approach fails in accurate estimation of uncertainty, Bayesian estimation provides an accurate information about the uncertainties in the parameters. Although a particular example of FDG kinetics was used in the paper, the methods can be extended for different pharmaceuticals and imaging modalities.  相似文献   

19.
Trend test based on cross-classified data in dose-response has been a central problem in medicine. Most of existing test methods are known to only fit to binary response variables. However, the approaches for binary response tables may suffer from the lack of a clear choice for dichotomization. For multivariate response with ordered categories, some studies have been done for simple stochastic order, likelihood ratio order and so on. However, methods of statistical inference on increasing convex order for more than two multinomial populations have not been fully developed. For testing the increasing convex order alternative, this article provides a model-free test method which can be used in the case of two-way tables and stratified data. Two real examples will be used to illustrate how to apply our test method.  相似文献   

20.
In studies of morphology, methods for comparing amounts of variability are often important. Three different ways of utilizing determinants of covariance matrices for testing for surplus variability in a hypothesis sample compared to a reference sample are presented: an F-test based on standardized generalized variances, a parametric bootstrap based on draws on Wishart matrices, and a nonparametric bootstrap. The F-test based on standardized generalized variances and the Wishart-based bootstrap are applicable when multivariate normality can be assumed. These methods can be applied with only summary data available. However, the nonparametric bootstrap can be applied with multivariate nonnormally distributed data as well as multivariate normally distributed data, and small sample sizes. Therefore, this method is preferable when raw data are available. Three craniometric samples are used to present the methods. A Hungarian Zalavár sample and an Austrian Berg sample are compared to a Norwegian Oslo sample, the latter employed as reference sample. In agreement with a previous study, it is shown that the Zalavár sample does not represent surplus variability, whereas the Berg sample does represent such a surplus variability.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号