首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 78 毫秒
1.
Problem statement: Breast cancer screening in women of younger age has been controversial. The screening sensitivities, transition probabilities and sojourn time distributions are estimated for females aged 40–49 years and 50–59 years separately, using the Canadian National Breast Screening Study (CNBSS) data. The purpose is to estimate the lead time distribution and the probability of not detecting the cancer early. Approach: Within the 40–49-year-old and 50–59-year-old cohorts separately, the age-independent statistical model was applied. Bayesian estimators along with 95% highest probability density (HPD) credible intervals (CI) were calculated. Bayesian hypothesis testing was used to compare the parameter estimates of the two cohorts. The lead time density was also estimated for both the 40–49 and 50–59-year-old cohorts. Results: The screening sensitivity, transition probability of the disease, and mean sojourn time were all found to increase with age. For the 40–49-year-old and 50–59-year-old cohorts, the posterior mean sensitivities were 0.70 (95% HPD-CI: 0.46, 0.93) and 0.77 (0.61, 0.93), respectively. The posterior mean transition probabilities were 0.0023 (0.0018, 0.0027) and 0.0031 (0.0024, 0.0038), while the posterior mean sojourn times were 2.55 (1.56, 4.26) years and 3.15 (2.12, 4.96) years. Bayes factors for the ratio of posterior probabilities that the respective parameter was larger vs. smaller in the 50–59-year-old cohort were estimated to be 2.09, 40.8 and 3.0 for the sensitivity, transition probability, and mean sojourn time, respectively. All three Bayes factors were larger than two, indicating greater than 2:1 odds in favor of the hypothesis that each of these parameters was greater in the 50–59-year-old cohort. The estimated mean lead times were 0.83 years and 0.96 years if the two cohorts were screened annually. Conclusions: The increase in sensitivity corresponds to an increase in the mean sojourn time. Breast cancer in younger women is more difficult to detect by screening tests and is more aggressive than breast cancer in older women. Women aged 50–59 tend to benefit more from screening compared with women aged 40–49.  相似文献   

2.
This paper uses the analysis of a data set to examine a number of issues in Bayesian statistics and the application of MCMC methods. The data concern the selectivity of fishing nets and logistic regression is used to relate the size of a fish to the probability it will be retained or escape from a trawl net. Hierarchical models relate information from different trawls and posterior distributions are determined using MCMC. Centring data is shown to radically reduce autocorrelation in chains and Rao‐Blackwellisation and chain‐thinning are found to have little effect on parameter estimates. The results of four convergence diagnostics are compared and the sensitivity of the posterior distribution to the prior distribution is examined using a novel method. Nested models are fitted to the data and compared using intrinsic Bayes factors, pseudo‐Bayes factors and credible intervals.  相似文献   

3.
Hu B  Ji Y  Tsui KW 《Biometrics》2008,64(4):1223-1230
SUMMARY: Inverse dose-response estimation refers to the inference of an effective dose of some agent that gives a desired probability of response, say 0.5. We consider inverse dose response for two agents, an application that has not received much attention in the literature. Through the posterior profiling technique (Hsu, 1995, The Canadian Journal of Statistics 23, 399-410), we propose a Bayesian method in which we approximate the marginal posterior distribution of an effective dose using a profile posterior distribution, and obtain the maximum a posteriori (MAP) estimate for the effective dose. We then employ an adaptive direction sampling algorithm to obtain the highest posterior density (HPD) credible region for the effective dose. Using the MAP and HPD estimates, investigators will be able to simultaneously calibrate the levels of two agents in dose-response studies. We illustrate our proposed Bayesian method through a simulation study and two practical examples.  相似文献   

4.
Kolassa JE  Tanner MA 《Biometrics》1999,55(4):1291-1294
This article presents an algorithm for small-sample conditional confidence regions for two or more parameters for any discrete regression model in the generalized linear interactive model family. Regions are constructed by careful inversion of conditional hypothesis tests. This method presupposes the use of approximate or exact techniques for enumerating the sample space for some components of the vector of sufficient statistics conditional on other components. Such enumeration may be performed exactly or by exact or approximate Monte Carlo, including the algorithms of Kolassa and Tanner (1994, Journal of the American Statistical Association 89, 697-702; 1999, Biometrics 55, 246-251). This method also assumes that one can compute certain conditional probabilities for a fixed value of the parameter vector. Because of a property of exponential families, one can use this set of conditional probabilities to directly compute the conditional probabilities associated with any other value of the vector of the parameters of interest. This observation dramatically reduces the computational effort required to invert the hypothesis test to obtain the confidence region. To construct a region with confidence level 1 - alpha, the algorithm begins with a grid of values for the parameters of interest. For each parameter vector on the grid (corresponding to the current null hypothesis), one transforms the initial set of conditional probabilities using exponential tilting and then calculates the p value for this current null hypothesis. The confidence region is the set of parameter values for which the p value is at least alpha.  相似文献   

5.
The objective of this study was to develop methods to estimate the optimal threshold of a longitudinal biomarker and its credible interval when the diagnostic test is based on a criterion that reflects a dynamic progression of that biomarker. Two methods are proposed: one parametric and one non‐parametric. In both the cases, the Bayesian inference was used to derive the posterior distribution of the optimal threshold from which an estimate and a credible interval could be obtained. A numerical study shows that the bias of the parametric method is low and the coverage probability of the credible interval close to the nominal value, with a small coverage asymmetry in some cases. This is also true for the non‐parametric method in case of large sample sizes. Both the methods were applied to estimate the optimal prostate‐specific antigen nadir value to diagnose prostate cancer recurrence after a high‐intensity focused ultrasound treatment. The parametric method can also be applied to non‐longitudinal biomarkers.  相似文献   

6.
Problems involving thousands of null hypotheses have been addressed by estimating the local false discovery rate (LFDR). A previous LFDR approach to reporting point and interval estimates of an effect-size parameter uses an estimate of the prior distribution of the parameter conditional on the alternative hypothesis. That estimated prior is often unreliable, and yet strongly influences the posterior intervals and point estimates, causing the posterior intervals to differ from fixed-parameter confidence intervals, even for arbitrarily small estimates of the LFDR. That influence of the estimated prior manifests the failure of the conditional posterior intervals, given the truth of the alternative hypothesis, to match the confidence intervals. Those problems are overcome by changing the posterior distribution conditional on the alternative hypothesis from a Bayesian posterior to a confidence posterior. Unlike the Bayesian posterior, the confidence posterior equates the posterior probability that the parameter lies in a fixed interval with the coverage rate of the coinciding confidence interval. The resulting confidence-Bayes hybrid posterior supplies interval and point estimates that shrink toward the null hypothesis value. The confidence intervals tend to be much shorter than their fixed-parameter counterparts, as illustrated with gene expression data. Simulations nonetheless confirm that the shrunken confidence intervals cover the parameter more frequently than stated. Generally applicable sufficient conditions for correct coverage are given. In addition to having those frequentist properties, the hybrid posterior can also be motivated from an objective Bayesian perspective by requiring coherence with some default prior conditional on the alternative hypothesis. That requirement generates a new class of approximate posteriors that supplement Bayes factors modified for improper priors and that dampen the influence of proper priors on the credibility intervals. While that class of posteriors intersects the class of confidence-Bayes posteriors, neither class is a subset of the other. In short, two first principles generate both classes of posteriors: a coherence principle and a relevance principle. The coherence principle requires that all effect size estimates comply with the same probability distribution. The relevance principle means effect size estimates given the truth of an alternative hypothesis cannot depend on whether that truth was known prior to observing the data or whether it was learned from the data.  相似文献   

7.
Ploidy patterns can be summarized in the form of a vector of proportions representing the frequency of occurrence of DNA contents in specified intervals. Data represented in this way can be analyzed statistically using the multinomial distribution. Properties of the multinomial distribution and computational difficulties that arise in its application are considered. Special problems involved in formulating hypothesis tests and confidence regions for multivariate discrete distributions are discussed in the context of evaluating vectors of proportions representing DNA ploidy patterns. Construction of tailor-made critical regions for detecting specific types of deviations from normal ploidy patterns is proposed, and a detailed example is given. In this example, two critical regions are compared: a standard critical region, consisting of cases whose probability of occurrence is small, and a diagnostic critical region, consisting of cases considered to be clinically indicative of abnormality. Advantages of the diagnostic critical region are noted.  相似文献   

8.
Statistical analysis on landmark-based shape spaces has diverse applications in morphometrics, medical diagnostics, machine vision and other areas. These shape spaces are non-Euclidean quotient manifolds. To conduct nonparametric inferences, one may define notions of centre and spread on this manifold and work with their estimates. However, it is useful to consider full likelihood-based methods, which allow nonparametric estimation of the probability density. This article proposes a broad class of mixture models constructed using suitable kernels on a general compact metric space and then on the planar shape space in particular. Following a Bayesian approach with a nonparametric prior on the mixing distribution, conditions are obtained under which the Kullback-Leibler property holds, implying large support and weak posterior consistency. Gibbs sampling methods are developed for posterior computation, and the methods are applied to problems in density estimation and classification with shape-based predictors. Simulation studies show improved estimation performance relative to existing approaches.  相似文献   

9.
Assessment of the reliability of a given phylogenetic hypothesis is an important step in phylogenetic analysis. Historically, the nonparametric bootstrap procedure has been the most frequently used method for assessing the support for specific phylogenetic relationships. The recent employment of Bayesian methods for phylogenetic inference problems has resulted in clade support being expressed in terms of posterior probabilities. We used simulated data and the four-taxon case to explore the relationship between nonparametric bootstrap values (as inferred by maximum likelihood) and posterior probabilities (as inferred by Bayesian analysis). The results suggest a complex association between the two measures. Three general regions of tree space can be identified: (1) the neutral zone, where differences between mean bootstrap and mean posterior probability values are not significant, (2) near the two-branch corner, and (3) deep in the two-branch corner. In the last two regions, significant differences occur between mean bootstrap and mean posterior probability values. Whether bootstrap or posterior probability values are higher depends on the data in support of alternative topologies. Examination of star topologies revealed that both bootstrap and posterior probability values differ significantly from theoretical expectations; in particular, there are more posterior probability values in the range 0.85-1 than expected by theory. Therefore, our results corroborate the findings of others that posterior probability values are excessively high. Our results also suggest that extrapolations from single topology branch-length studies are unlikely to provide any general conclusions regarding the relationship between bootstrap and posterior probability values.  相似文献   

10.
MOTIVATION: Gene genealogies offer a powerful context for inferences about the evolutionary process based on presently segregating DNA variation. In many cases, it is the distribution of population parameters, marginalized over the effectively infinite-dimensional tree space, that is of interest. Our evolutionary forest (EF) algorithm uses Monte Carlo methods to generate posterior distributions of population parameters. A novel feature is the updating of parameter values based on a probability measure defined on an ensemble of histories (a forest of genealogies), rather than a single tree. RESULTS: The EF algorithm generates samples from the correct marginal distribution of population parameters. Applied to actual data from closely related fruit fly species, it rapidly converged to posterior distributions that closely approximated the exact posteriors generated through massive computational effort. Applied to simulated data, it generated credible intervals that covered the actual parameter values in accordance with the nominal probabilities. AVAILABILITY: A C++ implementation of this method is freely accessible at http://www.isds.duke.edu/~scl13  相似文献   

11.
Summary For a general multiple loop feedback inhibition system in which the end product can inhibit any or all of the intermediate reactions it is shown that biologically significant behaviour is always confined to a bounded region of reaction space containing a unique equilibrium. By explicit construction of a Liapunov function for the general n dimensional differential equation it is shown that some values of reaction parameters cause the concentration vector to approach the equilibrium asymptotically for all physically realizable initial conditions. As the parameter values change, periodic solutions can appear within the bounded region. Some information about these periodic solutions can be obtained from the Hopf bifurcation theorem. Alternatively, if specific parameter values are known a numerical method can be used to find periodic solutions and determine their stability by locating a zero of the displacement map. The single loop Goodwin oscillator is analysed in detail. The methods are then used to treat an oscillator with two feedback loops and it is found that oscillations are possible even if both Hill coefficients are equal to one.  相似文献   

12.
Inferential structure determination uses Bayesian theory to combine experimental data with prior structural knowledge into a posterior probability distribution over protein conformational space. The posterior distribution encodes everything one can say objectively about the native structure in the light of the available data and additional prior assumptions and can be searched for structural representatives. Here an analogy is drawn between the posterior distribution and the canonical ensemble of statistical physics. A statistical mechanics analysis assesses the complexity of a structure calculation globally in terms of ensemble properties. Analogs of the free energy and density of states are introduced; partition functions evaluate the consistency of prior assumptions with data. Critical behavior is observed with dwindling restraint density, which impairs structure determination with too sparse data. However, prior distributions with improved realism ameliorate the situation by lowering the critical number of observations. An in-depth analysis of various experimentally accessible structural parameters and force field terms will facilitate a statistical approach to protein structure determination with sparse data that avoids bias as much as possible.  相似文献   

13.
Hans C  Dunson DB 《Biometrics》2005,61(4):1018-1026
In regression applications with categorical predictors, interest often focuses on comparing the null hypothesis of homogeneity to an ordered alternative. This article proposes a Bayesian approach for addressing this problem in the setting of normal linear and probit regression models. The regression coefficients are assigned a conditionally conjugate prior density consisting of mixtures of point masses at 0 and truncated normal densities, with a (possibly unknown) changepoint parameter included to accommodate umbrella ordering. Two strategies of prior elicitation are considered: (1) a Bayesian Bonferroni approach in which the probability of the global null hypothesis is specified and local hypotheses are considered independent; and (2) an approach which treats these probabilities as random. A single Gibbs sampling chain can be used to obtain posterior probabilities for the different hypotheses and to estimate regression coefficients and predictive quantities either by model averaging or under the preferred hypothesis. The methods are applied to data from a carcinogenesis study.  相似文献   

14.
Bochkina N  Richardson S 《Biometrics》2007,63(4):1117-1125
We consider the problem of identifying differentially expressed genes in microarray data in a Bayesian framework with a noninformative prior distribution on the parameter quantifying differential expression. We introduce a new rule, tail posterior probability, based on the posterior distribution of the standardized difference, to identify genes differentially expressed between two conditions, and we derive a frequentist estimator of the false discovery rate associated with this rule. We compare it to other Bayesian rules in the considered settings. We show how the tail posterior probability can be extended to testing a compound null hypothesis against a class of specific alternatives in multiclass data.  相似文献   

15.
Single species difference population models can show complex dynamics such as periodicity and chaos under certain circumstances, but usually only when rates of intrinsic population growth or other life history parameter are unrealistically high. Single species models with Allee effects (positive density dependence at low density) have also been shown to exhibit complex dynamics when combined with over-compensatory density dependence or a narrow fertility window. Here we present a simple two-stage model with Allee effects which shows large amplitude periodic fluctuations for some initial conditions, without these requirements. Periodicity arises out of a tension between the critical equilibrium of each stage, i.e. when the initial population vector is such that the adult stage is above the critical value, while the juvenile stage is below the critical value. Within this area of parameter space, the range of initial conditions giving rise to periodic dynamics is driven mainly by adult mortality rates. Periodic dynamics become more important as adult mortality increases up to a certain point, after which periodic dynamics are replaced by extinction. This model has more realistic life history parameter values than most 'chaotic' models. Conditions for periodic dynamics might arise in some marine species which are exploited (high adult mortality) leading to recruitment limitation (low juvenile density) and might be an additional source of extinction risk.  相似文献   

16.
Cho H  Ibrahim JG  Sinha D  Zhu H 《Biometrics》2009,65(1):116-124
We propose Bayesian case influence diagnostics for complex survival models. We develop case deletion influence diagnostics for both the joint and marginal posterior distributions based on the Kullback-Leibler divergence (K-L divergence). We present a simplified expression for computing the K-L divergence between the posterior with the full data and the posterior based on single case deletion, as well as investigate its relationships to the conditional predictive ordinate. All the computations for the proposed diagnostic measures can be easily done using Markov chain Monte Carlo samples from the full data posterior distribution. We consider the Cox model with a gamma process prior on the cumulative baseline hazard. We also present a theoretical relationship between our case-deletion diagnostics and diagnostics based on Cox's partial likelihood. A simulated data example and two real data examples are given to demonstrate the methodology.  相似文献   

17.
Reversible-jump Markov chain Monte Carlo (RJ-MCMC) is a technique for simultaneously evaluating multiple related (but not necessarily nested) statistical models that has recently been applied to the problem of phylogenetic model selection. Here we use a simulation approach to assess the performance of this method and compare it to Akaike weights, a measure of model uncertainty that is based on the Akaike information criterion. Under conditions where the assumptions of the candidate models matched the generating conditions, both Bayesian and AIC-based methods perform well. The 95% credible interval contained the generating model close to 95% of the time. However, the size of the credible interval differed with the Bayesian credible set containing approximately 25% to 50% fewer models than an AIC-based credible interval. The posterior probability was a better indicator of the correct model than the Akaike weight when all assumptions were met but both measures performed similarly when some model assumptions were violated. Models in the Bayesian posterior distribution were also more similar to the generating model in their number of parameters and were less biased in their complexity. In contrast, Akaike-weighted models were more distant from the generating model and biased towards slightly greater complexity. The AIC-based credible interval appeared to be more robust to the violation of the rate homogeneity assumption. Both AIC and Bayesian approaches suggest that substantial uncertainty can accompany the choice of model for phylogenetic analyses, suggesting that alternative candidate models should be examined in analysis of phylogenetic data. [AIC; Akaike weights; Bayesian phylogenetics; model averaging; model selection; model uncertainty; posterior probability; reversible jump.].  相似文献   

18.
The jaw lever system in ungulates: a new model   总被引:6,自引:0,他引:6  
In ungulates the distance from the jaw joint to the last molar is approximately the same as that of the grinding tooth row length. An hypothesis is presented that attempts to explain why ungulate grinding teeth are positioned where they are along the jaw. If the jaw joint on the balancing side serves as the fulcrum in anisognathus animals, a region along the jaw is defined, corresponding to where teeth are actually found, where rather simple muscle action can apply equal forces at any point. The ability to produce the same force at any point in this region, as such, may not be the critical feature since anterior or posterior to this region, tooth placement produces either a relatively inefficient or an unstable condition.  相似文献   

19.
Capture-recapture studies are frequently used to monitor the status and trends of wildlife populations. Detection histories from individual animals are used to estimate probability of detection and abundance or density. The accuracy of abundance and density estimates depends on the ability to model factors affecting detection probability. Non-spatial capture-recapture models have recently evolved into spatial capture-recapture models that directly include the effect of distances between an animal’s home range centre and trap locations on detection probability. Most studies comparing non-spatial and spatial capture-recapture biases focussed on single year models and no studies have compared the accuracy of demographic parameter estimates from open population models. We applied open population non-spatial and spatial capture-recapture models to three years of grizzly bear DNA-based data from Banff National Park and simulated data sets. The two models produced similar estimates of grizzly bear apparent survival, per capita recruitment, and population growth rates but the spatial capture-recapture models had better fit. Simulations showed that spatial capture-recapture models produced more accurate parameter estimates with better credible interval coverage than non-spatial capture-recapture models. Non-spatial capture-recapture models produced negatively biased estimates of apparent survival and positively biased estimates of per capita recruitment. The spatial capture-recapture grizzly bear population growth rates and 95% highest posterior density averaged across the three years were 0.925 (0.786–1.071) for females, 0.844 (0.703–0.975) for males, and 0.882 (0.779–0.981) for females and males combined. The non-spatial capture-recapture population growth rates were 0.894 (0.758–1.024) for females, 0.825 (0.700–0.948) for males, and 0.863 (0.771–0.957) for both sexes. The combination of low densities, low reproductive rates, and predominantly negative population growth rates suggest that Banff National Park’s population of grizzly bears requires continued conservation-oriented management actions.  相似文献   

20.
Ayres KL  Balding DJ 《Genetics》2001,157(1):413-423
We describe a Bayesian approach to analyzing multilocus genotype or haplotype data to assess departures from gametic (linkage) equilibrium. Our approach employs a Markov chain Monte Carlo (MCMC) algorithm to approximate the posterior probability distributions of disequilibrium parameters. The distributions are computed exactly in some simple settings. Among other advantages, posterior distributions can be presented visually, which allows the uncertainties in parameter estimates to be readily assessed. In addition, background knowledge can be incorporated, where available, to improve the precision of inferences. The method is illustrated by application to previously published datasets; implications for multilocus forensic match probabilities and for simple association-based gene mapping are also discussed.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号