首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 109 毫秒
1.
Introduction

The Monte Carlo technique is widely used and recommended for including uncertainties LCA. Typically, 1000 or 10,000 runs are done, but a clear argument for that number is not available, and with the growing size of LCA databases, an excessively high number of runs may be a time-consuming thing. We therefore investigate if a large number of runs are useful, or if it might be unnecessary or even harmful.

Probability theory

We review the standard theory or probability distributions for describing stochastic variables, including the combination of different stochastic variables into a calculation. We also review the standard theory of inferential statistics for estimating a probability distribution, given a sample of values. For estimating the distribution of a function of probability distributions, two major techniques are available, analytical, applying probability theory and numerical, using Monte Carlo simulation. Because the analytical technique is often unavailable, the obvious way-out is Monte Carlo. However, we demonstrate and illustrate that it leads to overly precise conclusions on the values of estimated parameters, and to incorrect hypothesis tests.

Numerical illustration

We demonstrate the effect for two simple cases: one system in a stand-alone analysis and a comparative analysis of two alternative systems. Both cases illustrate that statistical hypotheses that should not be rejected in fact are rejected in a highly convincing way, thus pointing out a fundamental flaw.

Discussion and conclusions

Apart form the obvious recommendation to use larger samples for estimating input distributions, we suggest to restrict the number of Monte Carlo runs to a number not greater than the sample sizes used for the input parameters. As a final note, when the input parameters are not estimated using samples, but through a procedure, such as the popular pedigree approach, the Monte Carlo approach should not be used at all.

  相似文献   

2.
This paper develops a deterministic model of frequency distributions for energy imparted (total energy deposition) in small volumes similar to DNA molecules from high-energy ions of interest for space radiation protection and cancer therapy. Frequency distributions for energy imparted are useful for considering radiation quality and for modeling biological damage produced by ionizing radiation. For high-energy ions, secondary electron (delta-ray) tracks originating from a primary ion track make dominant contributions to energy deposition events in small volumes. Our method uses the distribution of electrons produced about an ion's path and incorporates results from Monte Carlo simulation of electron tracks to predict frequency distributions for ions, including their dependence on radial distance. The contribution from primary ion events is treated using an impact parameter formalism of spatially restricted linear energy transfer (LET) and energy-transfer straggling. We validate our model by comparing it directly to results from Monte Carlo simulations for proton and alpha-particle tracks. We show for the first time frequency distributions of energy imparted in DNA structures by several high-energy ions such as cosmic-ray iron ions. Our comparison with results from Monte Carlo simulations at low energies indicates the accuracy of the method.  相似文献   

3.
Accurate estimates of population parameters are vital for estimating extinction risk. Such parameters, however, are typically not available for threatened populations. We used a recently developed software tool based on Markov Chain Monte Carlo methods for carrying out Bayesian inference (the BUGS package) to estimate four demographic parameters; the intrinsic growth rate, the strength of density dependence, and the demographic and environmental variance, in three species of small temperate passerines from two sets of time series data taken from a dipper and a song sparrow population, and from previously obtained frequentist estimates of the same parameters in the great tit. By simultaneously modeling variation in these demographic parameters across species and using the resulting distributions as priors in the estimation for individual species, we improve the estimates for each individual species. This framework also allows us to make probabilistic statements about plausible parameter values for small passerines temperate birds in general which is often critically needed in management of species for which little or no data are available. We also discuss how our work relates to recently developed theory on dynamic stochastic population models, and finally note some important differences between frequentist and Bayesian methods.  相似文献   

4.
We describe a Monte Carlo simulation of the within-host dynamics of human immunodeficiency virus 1 (HIV-1). The simulation proceeds at the level of individual T-cells and virions in a small volume of plasma, thus capturing the inherent stochasticity in viral replication, mutation and T-cell infection. When cell lifetimes are distributed exponentially in the Monte Carlo approach, our simulation results are in perfect agreement with the predictions of the corresponding systems of differential equations from the literature. The Monte Carlo model, however, uniquely allows us to estimate the natural variability in important parameters such as the T-cell count, viral load, and the basic reproductive ratio, in both the presence and absence of drug therapy. The simulation also yields the probability that an infection will not become established after exposure to a viral inoculum of a given size. Finally, we extend the Monte Carlo approach to include distributions of cell lifetimes that are less-dispersed than exponential.  相似文献   

5.
Monte Carlo simulation has commonly been used in phylogenetic studies to test different tree-reconstruction methods, and consequently, its application for testing evolutionary models can be considered as a natural extension of this usage. Repetitive simulation of a given evolutionary process, under the restrictions imposed by the model to be tested, along a determinate tree topology allow the estimate of probability distributions for the desired parameters. Next, the phylogenetic tree can be reconstructed again without the constraints of the model, and the parameter of interest, derived from this tree, can be compared to the corresponding probability distribution derived from the restricted, simulated trees. As an example we have used Monte Carlo simulation to test the constancy of evolutionary rates in a set of cytochrome-c protein sequences. Correspondence to: J. Dopazo  相似文献   

6.

Background, aim, and scope

Uncertainty information is essential for the proper use of life cycle assessment (LCA) and environmental assessments in decision making. So far, parameter uncertainty propagation has mainly been studied using Monte Carlo techniques that are relatively computationally heavy to conduct, especially for the comparison of multiple scenarios, often limiting its use to research or to inventory only. Furthermore, Monte Carlo simulations do not automatically assess the sensitivity and contribution to overall uncertainty of individual parameters. The present paper aims to develop and apply to both inventory and impact assessment an explicit and transparent analytical approach to uncertainty. This approach applies Taylor series expansions to the uncertainty propagation of lognormally distributed parameters.

Materials and methods

We first apply the Taylor series expansion method to analyze the uncertainty propagation of a single scenario, in which case the squared geometric standard deviation of the final output is determined as a function of the model sensitivity to each input parameter and the squared geometric standard deviation of each parameter. We then extend this approach to the comparison of two or more LCA scenarios. Since in LCA it is crucial to account for both common inventory processes and common impact assessment characterization factors among the different scenarios, we further develop the approach to address this dependency. We provide a method to easily determine a range and a best estimate of (a) the squared geometric standard deviation on the ratio of the two scenario scores, “A/B”, and (b) the degree of confidence in the prediction that the impact of scenario A is lower than B (i.e., the probability that A/B<1). The approach is tested on an automobile case study and resulting probability distributions of climate change impacts are compared to classical Monte Carlo distributions.

Results

The probability distributions obtained with the Taylor series expansion lead to results similar to the classical Monte Carlo distributions, while being substantially simpler; the Taylor series method tends to underestimate the 2.5% confidence limit by 1-11% and the 97.5% limit by less than 5%. The analytical Taylor series expansion easily provides the explicit contributions of each parameter to the overall uncertainty. For the steel front end panel, the factor contributing most to the climate change score uncertainty is the gasoline consumption (>75%). For the aluminum panel, the electricity and aluminum primary production, as well as the light oil consumption, are the dominant contributors to the uncertainty. The developed approach for scenario comparisons, differentiating between common and independent parameters, leads to results similar to those of a Monte Carlo analysis; for all tested cases, we obtained a good concordance between the Monte Carlo and the Taylor series expansion methods regarding the probability that one scenario is better than the other.

Discussion

The Taylor series expansion method addresses the crucial need of accounting for dependencies in LCA, both for common LCI processes and common LCIA characterization factors. The developed approach in Eq. 8, which differentiates between common and independent parameters, estimates the degree of confidence in the prediction that scenario A is better than B, yielding results similar to those found with Monte Carlo simulations.

Conclusions

The probability distributions obtained with the Taylor series expansion are virtually equivalent to those from a classical Monte Carlo simulation, while being significantly easier to obtain. An automobile case study on an aluminum front end panel demonstrated the feasibility of this method and illustrated its simultaneous and consistent application to both inventory and impact assessment. The explicit and innovative analytical approach, based on Taylor series expansions of lognormal distributions, provides the contribution to the uncertainty from each parameter and strongly reduces calculation time.  相似文献   

7.
Parameter inference and model selection are very important for mathematical modeling in systems biology. Bayesian statistics can be used to conduct both parameter inference and model selection. Especially, the framework named approximate Bayesian computation is often used for parameter inference and model selection in systems biology. However, Monte Carlo methods needs to be used to compute Bayesian posterior distributions. In addition, the posterior distributions of parameters are sometimes almost uniform or very similar to their prior distributions. In such cases, it is difficult to choose one specific value of parameter with high credibility as the representative value of the distribution. To overcome the problems, we introduced one of the population Monte Carlo algorithms, population annealing. Although population annealing is usually used in statistical mechanics, we showed that population annealing can be used to compute Bayesian posterior distributions in the approximate Bayesian computation framework. To deal with un-identifiability of the representative values of parameters, we proposed to run the simulations with the parameter ensemble sampled from the posterior distribution, named “posterior parameter ensemble”. We showed that population annealing is an efficient and convenient algorithm to generate posterior parameter ensemble. We also showed that the simulations with the posterior parameter ensemble can, not only reproduce the data used for parameter inference, but also capture and predict the data which was not used for parameter inference. Lastly, we introduced the marginal likelihood in the approximate Bayesian computation framework for Bayesian model selection. We showed that population annealing enables us to compute the marginal likelihood in the approximate Bayesian computation framework and conduct model selection depending on the Bayes factor.  相似文献   

8.
Data-driven methods for personalizing treatment assignment have garnered much attention from clinicians and researchers. Dynamic treatment regimes formalize this through a sequence of decision rules that map individual patient characteristics to a recommended treatment. Observational studies are commonly used for estimating dynamic treatment regimes due to the potentially prohibitive costs of conducting sequential multiple assignment randomized trials. However, estimating a dynamic treatment regime from observational data can lead to bias in the estimated regime due to unmeasured confounding. Sensitivity analyses are useful for assessing how robust the conclusions of the study are to a potential unmeasured confounder. A Monte Carlo sensitivity analysis is a probabilistic approach that involves positing and sampling from distributions for the parameters governing the bias. We propose a method for performing a Monte Carlo sensitivity analysis of the bias due to unmeasured confounding in the estimation of dynamic treatment regimes. We demonstrate the performance of the proposed procedure with a simulation study and apply it to an observational study examining tailoring the use of antidepressant medication for reducing symptoms of depression using data from Kaiser Permanente Washington.  相似文献   

9.
Schoen DJ  Clegg MT 《Genetics》1986,112(4):927-945
Estimation of mating system parameters in plant populations typically employs family-structured samples of progeny genotypes. These estimation models postulate a mixture of self-fertilization and random outcrossing. One assumption of such models concerns the distribution of pollen genotypes among eggs within single maternal families. Previous applications of the mixed mating model to mating system estimation have assumed that pollen genotypes are sampled randomly from the total population in forming outcrossed progeny within families. In contrast, the one-pollen parent model assumes that outcrossed progeny within a family share a single-pollen parent genotype. Monte Carlo simulations of family-structured sampling were carried out to examine the consequences of violations of the different assumptions of the two models regarding the distribution of pollen genotypes among eggs. When these assumptions are violated, estimates of mating system parameters may be significantly different from their true values and may exhibit distributions which depart from normality. Monte Carlo methods were also used to examine the utility of the bootstrap resampling algorithm for estimating the variances of mating system parameters. The bootstrap method gives variance estimates that approximate empirically determined values. When applied to data from two plant populations which differ in pollen genotype distributions within families, the two estimation procedures exhibit the same behavior as that seen with the simulated data.  相似文献   

10.
We introduce a Bayesian method for estimating hidden population substructure using multilocus molecular markers and geographical information provided by the sampling design. The joint posterior distribution of the substructure and allele frequencies of the respective populations is available in an analytical form when the number of populations is small, whereas an approximation based on a Markov chain Monte Carlo simulation approach can be obtained for a moderate or large number of populations. Using the joint posterior distribution, posteriors can also be derived for any evolutionary population parameters, such as the traditional fixation indices. A major advantage compared to most earlier methods is that the number of populations is treated here as an unknown parameter. What is traditionally considered as two genetically distinct populations, either recently founded or connected by considerable gene flow, is here considered as one panmictic population with a certain probability based on marker data and prior information. Analyses of previously published data on the Moroccan argan tree (Argania spinosa) and of simulated data sets suggest that our method is capable of estimating a population substructure, while not artificially enforcing a substructure when it does not exist. The software (BAPS) used for the computations is freely available from http://www.rni.helsinki.fi/~mjs.  相似文献   

11.
Four computational methods for estimating mean fecundity are compared by Monte Carlo simulation. One of the four methods is the simple expedient of estimating fecundity at sample mean length, a method known to be downwardly biassed. The Monte Carlo study shows that the other three methods reduce bias and provide worthwhile efficiency gains. For small samples, the most efficient of the four methods is a 'bias adjustment', proposed here, that uses easily calculated sample statistics. For large samples, a numerical integration method has the highest efficiency. The fourth method, a 'direct summation' procedure which can be done easily in many statistical or spreadsheet programs, performs well for all sample sizes.  相似文献   

12.
A new design for estimating the distribution of time to pregnancy is proposed and investigated. The design is based on recording current durations in a cross-sectional sample of women, leading to statistical problems similar to estimating renewal time distributions from backward recurrence times. Non-parametric estimation is studied in some detail and a parametric approach is indicated. The results are illustrated on Monte Carlo simulations and on data from a recent European collaborative study. The role and applicability of this approach is discussed.  相似文献   

13.
We investigate the distribution of sizes of fragments obtained from the amplified fragment length polymorphism (AFLP) marker technique. We find that empirical distributions obtained in two plant species, Phaseolus lunatus and Lolium perenne, are consistent with the expected distributions obtained from analytical theory and from numerical simulations. Our results indicate that the size distribution is strongly asymmetrical, with a much higher proportion of small than large fragments, that it is not influenced by the number of selective nucleotides nor by genome size but that it may vary with genome-wide GC-content, with a higher proportion of small fragments in cases of lower GC-content when considering the standard AFLP protocol with the enzyme MseI. Results from population samples of the two plant species show that there is a negative relationship between AFLP fragment size and fragment population frequency. Monte Carlo simulations reveal that size homoplasy, arising from pulling together nonhomologous fragments of the same size, generates patterns similar to those observed in P. lunatus and L. perenne because of the asymmetry of the size distribution. We discuss the implications of these results in the context of estimating genetic diversity with AFLP markers.  相似文献   

14.
We present a method for estimating the parameters in random effects models for survival data when covariates are subject to missingness. Our method is more general than the usual frailty model as it accommodates a wide range of distributions for the random effects, which are included as an offset in the linear predictor in a manner analogous to that used in generalized linear mixed models. We propose using a Monte Carlo EM algorithm along with the Gibbs sampler to obtain parameter estimates. This method is useful in reducing the bias that may be incurred using complete-case methods in this setting. The methodology is applied to data from Eastern Cooperative Oncology Group melanoma clinical trials in which observations were believed to be clustered and several tumor characteristics were not always observed.  相似文献   

15.
Recent developments in modeling stream of variation in multistage manufacturing system along with the urgent need for yield enhancement in the semiconductor industry has led to complex large scale simulation problems in design and performance prediction, thus challenging current Monte Carlo (MC) based simulation techniques. MC method prevails in statistical simulation approaches for multi-dimensional cases with general (i.e., non-Gaussian) distributions and/or complex response functions. A method is proposed based on number theory (NT-net) to reduce computing effort and the variability of MC's results in tolerance design and circuit performance simulation. The sampling strategy is improved by introducing NT-net that can provide better convergent rate over MC. The new method is presented and verified using several case studies, including analytical and industrial cases of a filter design and analyses of a four-bar mechanism. Results indicate a 90–95% reduction of computation effort with significant improvement in accuracy that can be achieved by the proposed technique.  相似文献   

16.
Construction of confidence intervals or regions is an important part of statistical inference. The usual approach to constructing a confidence interval for a single parameter or confidence region for two or more parameters requires that the distribution of estimated parameters is known or can be assumed. In reality, the sampling distributions of parameters of biological importance are often unknown or difficult to be characterized. Distribution-free nonparametric resampling methods such as bootstrapping and permutation have been widely used to construct the confidence interval for a single parameter. There are also several parametric (ellipse) and nonparametric (convex hull peeling, bagplot and HPDregionplot) methods available for constructing confidence regions for two or more parameters. However, these methods have some key deficiencies including biased estimation of the true coverage rate, failure to account for the shape of the distribution inherent in the data and difficulty to implement. The purpose of this paper is to develop a new distribution-free method for constructing the confidence region that is based only on a few basic geometrical principles and accounts for the actual shape of the distribution inherent in the real data. The new method is implemented in an R package, distfree.cr/R. The statistical properties of the new method are evaluated and compared with those of the other methods through Monte Carlo simulation. Our new method outperforms the other methods regardless of whether the samples are taken from normal or non-normal bivariate distributions. In addition, the superiority of our method is consistent across different sample sizes and different levels of correlation between the two variables. We also analyze three biological data sets to illustrate the use of our new method for genomics and other biological researches.  相似文献   

17.
A new version of the test particle method for determining the chemical potential by Monte Carlo simulations is proposed. The method, applicable to any fluid at any density, combines the Widom's test particle insertion method with the ideas of the scaled particle theory, gradual insertion method and multistage sampling. Its applicability is exemplified by evaluating the chemical potential of the hard sphere fluid at a very high density in semi-grand-canonical and grand-canonical ensembles. A theory estimating the efficiency (i.e. statistical errors) of the method is proposed and the results are compared with the Widom's and gradual insertion methods, and the analytic results.  相似文献   

18.
The present paper reports the results of a Monte Carlo simulation study to examine the performance of several approximate confidence intervals for the Relative Risk Ratio (RRR) parameter in an epidemiologic study, involving two groups of individuals. The first group consists of n1 individuals, called the experimental group, who are exposed to some carcinogen, say radiation, whose effect on the incidence of some form of cancer, say skin cancer, is being investigated. The second group consists of n2 individuals (called the control group) who are not exposed to the carcinogen. Two cases are considered in which the life times (or time to cancer) in the two groups follow (i) the exponential and (ii) the Weibull distributions. The case when the life times follow a Rayleigh distribution follows as a particular case. A general random censorship model is considered in which the life times of the individuals are censored on the right by random censoring times following (i) the exponential and (ii) the Weibull distributions. The Relative Risk Ratio parameter in the study is defined as the ratio of the hazard rates in the two distributions of the times to cancer. Approximate confidence intervals are constructed for the RRR parameter using its maximum likelihood estimator (m.l.e) and several other methods, including a method due to FIELLER. SPROTT'S (1973) and Cox's (1953) suggestions, as well as the Box-Cox (1964) transformation, are also utilized to construct approximate confidence intervals. The performance of these confidence intervals in small samples is investigated by means of some Monte Carlo simulations based on 500 random samples. Our simulation study indicates that many of these confidence intervals perform quite well in samples of size 10 and 15, in terms of the coverage probability and expected length of the interval.  相似文献   

19.
Yuan Y  Johnson VE 《Biometrics》2012,68(1):156-164
This article proposes methodology for assessing goodness of fit in Bayesian hierarchical models. The methodology is based on comparing values of pivotal discrepancy measures (PDMs), computed using parameter values drawn from the posterior distribution, to known reference distributions. Because the resulting diagnostics can be calculated from standard output of Markov chain Monte Carlo algorithms, their computational costs are minimal. Several simulation studies are provided, each of which suggests that diagnostics based on PDMs have higher statistical power than comparable posterior-predictive diagnostic checks in detecting model departures. The proposed methodology is illustrated in a clinical application; an application to discrete data is described in supplementary material.  相似文献   

20.
Stubbendick AL  Ibrahim JG 《Biometrics》2003,59(4):1140-1150
This article analyzes quality of life (QOL) data from an Eastern Cooperative Oncology Group (ECOG) melanoma trial that compared treatment with ganglioside vaccination to treatment with high-dose interferon. The analysis of this data set is challenging due to several difficulties, namely, nonignorable missing longitudinal responses and baseline covariates. Hence, we propose a selection model for estimating parameters in the normal random effects model with nonignorable missing responses and covariates. Parameters are estimated via maximum likelihood using the Gibbs sampler and a Monte Carlo expectation maximization (EM) algorithm. Standard errors are calculated using the bootstrap. The method allows for nonmonotone patterns of missing data in both the response variable and the covariates. We model the missing data mechanism and the missing covariate distribution via a sequence of one-dimensional conditional distributions, allowing the missing covariates to be either categorical or continuous, as well as time-varying. We apply the proposed approach to the ECOG quality-of-life data and conduct a small simulation study evaluating the performance of the maximum likelihood estimates. Our results indicate that a patient treated with the vaccine has a higher QOL score on average at a given time point than a patient treated with high-dose interferon.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号