首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 46 毫秒
1.
Summary The validity of limiting dilution assays can be compromised or negated by the use of statistical methodology which does not consider all issues surrounding the biological process. This study critically evaluates statistical methods for estimating the mean frequency of responding cells in multiple sample limiting dilution assays. We show that methods that pool limiting dilution assay data, or samples, are unable to estimate the variance appropriately. In addition, we use Monte Carlo simulations to evaluate an unweighted mean of the maximum likelihood estimator, an unweighted mean based on the jackknife estimator, and a log transform of the maximum likelihood estimator. For small culture replicate size, the log transform outperforms both unweighted mean procedures. For moderate culture replicate size, the unweighted mean based on the jackknife produces the most acceptable results. This study also addresses the important issue of experimental design in multiple sample limiting dilution assays. In particular, we demonstrate that optimization of multiple sample limiting dilution assays is achieved by increasing the number of biological samples at the expense of repeat cultures.  相似文献   

2.
Beaumont MA 《Genetics》2003,164(3):1139-1160
This article introduces a new general method for genealogical inference that samples independent genealogical histories using importance sampling (IS) and then samples other parameters with Markov chain Monte Carlo (MCMC). It is then possible to more easily utilize the advantages of importance sampling in a fully Bayesian framework. The method is applied to the problem of estimating recent changes in effective population size from temporally spaced gene frequency data. The method gives the posterior distribution of effective population size at the time of the oldest sample and at the time of the most recent sample, assuming a model of exponential growth or decline during the interval. The effect of changes in number of alleles, number of loci, and sample size on the accuracy of the method is described using test simulations, and it is concluded that these have an approximately equivalent effect. The method is used on three example data sets and problems in interpreting the posterior densities are highlighted and discussed.  相似文献   

3.
Introduction

The Monte Carlo technique is widely used and recommended for including uncertainties LCA. Typically, 1000 or 10,000 runs are done, but a clear argument for that number is not available, and with the growing size of LCA databases, an excessively high number of runs may be a time-consuming thing. We therefore investigate if a large number of runs are useful, or if it might be unnecessary or even harmful.

Probability theory

We review the standard theory or probability distributions for describing stochastic variables, including the combination of different stochastic variables into a calculation. We also review the standard theory of inferential statistics for estimating a probability distribution, given a sample of values. For estimating the distribution of a function of probability distributions, two major techniques are available, analytical, applying probability theory and numerical, using Monte Carlo simulation. Because the analytical technique is often unavailable, the obvious way-out is Monte Carlo. However, we demonstrate and illustrate that it leads to overly precise conclusions on the values of estimated parameters, and to incorrect hypothesis tests.

Numerical illustration

We demonstrate the effect for two simple cases: one system in a stand-alone analysis and a comparative analysis of two alternative systems. Both cases illustrate that statistical hypotheses that should not be rejected in fact are rejected in a highly convincing way, thus pointing out a fundamental flaw.

Discussion and conclusions

Apart form the obvious recommendation to use larger samples for estimating input distributions, we suggest to restrict the number of Monte Carlo runs to a number not greater than the sample sizes used for the input parameters. As a final note, when the input parameters are not estimated using samples, but through a procedure, such as the popular pedigree approach, the Monte Carlo approach should not be used at all.

  相似文献   

4.
Wu XW  Sung SS 《Proteins》1999,34(3):295-302
A new approach to efficiently calculate solvent effect in computer simulation of macromolecular systems has been developed. Explicit solvent molecules are included in the simulation to provide a mean solvation force for the solute conformational search. Simulations of an alanine dipeptide in aqueous solution showed that the new approach is significantly more efficient than conventional molecular dynamics method in conformational search, mainly because the mean solvation force reduced the solvent damping effect. This approach allows the solute and solvent to be simulated separately with different methods. For the macromolecule, the rigid fragment constraint dynamics method we developed previously allows large time-steps. For the solvent, a combination of a modified force-bias Monte Carlo method and a preferential sampling can efficiently sample the conformational space. A folding simulation of a 16-residue peptide in water showed high efficiency of the new approach.  相似文献   

5.
The paper deals with the statistical estimation of mean triangles of landmark data. For the model introduced by Bookstein (1986) three methods of estimating the “ideal” triangle are compared: the maximum likelihood method based on the exact distribution given in Mardia and Dryden (1989a), a moment method and an iterative algorithm yielding a mean triangle in the sense of Fréchet. These methods are compared by Monte Carlo simulation applied also to models with variances greater than those required for Bookstein's normal approximation.  相似文献   

6.

Background

The group testing method has been proposed for the detection and estimation of genetically modified plants (adventitious presence of unwanted transgenic plants, AP). For binary response variables (presence or absence), group testing is efficient when the prevalence is low, so that estimation, detection, and sample size methods have been developed under the binomial model. However, when the event is rare (low prevalence <0.1), and testing occurs sequentially, inverse (negative) binomial pooled sampling may be preferred.

Methodology/Principal Findings

This research proposes three sample size procedures (two computational and one analytic) for estimating prevalence using group testing under inverse (negative) binomial sampling. These methods provide the required number of positive pools (), given a pool size (k), for estimating the proportion of AP plants using the Dorfman model and inverse (negative) binomial sampling. We give real and simulated examples to show how to apply these methods and the proposed sample-size formula. The Monte Carlo method was used to study the coverage and level of assurance achieved by the proposed sample sizes. An R program to create other scenarios is given in Appendix S2.

Conclusions

The three methods ensure precision in the estimated proportion of AP because they guarantee that the width (W) of the confidence interval (CI) will be equal to, or narrower than, the desired width (), with a probability of . With the Monte Carlo study we found that the computational Wald procedure (method 2) produces the more precise sample size (with coverage and assurance levels very close to nominal values) and that the samples size based on the Clopper-Pearson CI (method 1) is conservative (overestimates the sample size); the analytic Wald sample size method we developed (method 3) sometimes underestimated the optimum number of pools.  相似文献   

7.
Three numerical techniques for generating thermally accessible configurations of globular proteins are considered; these techniques are the molecular dynamics method, the Metropolis Monte Carlo method, and a modified Monte Carlo method which takes account of the forces acting on the protein atoms. The molecular dynamics method is shown to be more efficient than either of the Monte Carlo methods. Because it may be necessary to use Monte Carlo methods in certain important types of sampling problems, the behavior of these methods is examined in some detail. It is found that an acceptance ratio close to 1/6 yields optimum efficiency for the Metropolis method, in contrast to what is often assumed. This result, together with the overall inefficiency of the Monte Carlo methods, appears to arise from the anisotropic forces acting on the protein atoms due to their covalent bonding. Possible ways of improving the Monte Carlo methods are suggested.  相似文献   

8.
敖长林  周领  焦扬  王世雪 《生态学报》2016,36(3):854-862
双边界二分式条件价值评估法是当前广泛应用于评估生态环境和自然资源价值的陈述偏好法。为了研究双边界二分式CVM中,不同分布中初始投标值数量及样本容量对WTP的影响,应用蒙特卡洛模拟和支付意愿函数模型相结合的方法,用概率分位数设计投标值,分别选择数据来源不同的Weibull分布和对数Logistic分布,动态模拟不同初始投标值数量和样本容量对WTP期望值产生的影响。模拟结果表明,对不同的WTP概率分布,初始投标值数量和样本容量对WTP的影响趋势是一致的,当初始投标值数量和样本容量分别大于5和500时,初始投标值数量和样本容量的增加,对WTP估计值的影响程度越来越小。研究结果表明,在二分式CVM研究中,当以WTP期望为代表值时,其初始投标值数量和样本容量至少为5和500。研究结论为二分式CVM问卷设计中投标值数量及样本容量的设定提供参考依据。  相似文献   

9.
Xie W  Lewis PO  Fan Y  Kuo L  Chen MH 《Systematic biology》2011,60(2):150-160
The marginal likelihood is commonly used for comparing different evolutionary models in Bayesian phylogenetics and is the central quantity used in computing Bayes Factors for comparing model fit. A popular method for estimating marginal likelihoods, the harmonic mean (HM) method, can be easily computed from the output of a Markov chain Monte Carlo analysis but often greatly overestimates the marginal likelihood. The thermodynamic integration (TI) method is much more accurate than the HM method but requires more computation. In this paper, we introduce a new method, steppingstone sampling (SS), which uses importance sampling to estimate each ratio in a series (the "stepping stones") bridging the posterior and prior distributions. We compare the performance of the SS approach to the TI and HM methods in simulation and using real data. We conclude that the greatly increased accuracy of the SS and TI methods argues for their use instead of the HM method, despite the extra computation needed.  相似文献   

10.
Recently, several statistical methods for estimating fine-scale recombination rates using population samples have been developed. However, currently available methods that can be applied to large-scale data are limited to approximated likelihoods. Here, we developed a full-likelihood Markov chain Monte Carlo method for estimating recombination rate under a Bayesian framework. Genealogies underlying a sampling of chromosomes are effectively modelled by using marginal individual single nucleotide polymorphism genealogies related through an ancestral recombination graph. The method is compared with two existing composite-likelihood methods using simulated data.Simulation studies show that our method performs well for different simulation scenarios. The method is applied to two human population genetic variation datasets that have been studied by sperm typing. Our results are consistent with the estimates from sperm crossover analysis.  相似文献   

11.
This study considers the effects of sample size on estimates of three parasitological indices (prevalence, mean abundance and mean intensity) in four different host–parasite systems, each showing a different pattern of infection. Monte Carlo simulation procedures were used in order to obtain an estimation of the parasitological indices, as well as their variance and bias, based on samples of different size. Although results showed that mean values of all indices were similar irrespective of sample size, estimates of prevalence were not significantly affected by sample size whereas mean abundance and mean intensity were affected in at least one sample. Underestimation of values was more perceptible in small (<40) sample sizes. Distribution of the estimated values revealed a different arrangement according to the host–parasite system and to the parasitological parameter. Monte Carlo simulation procedures are, therefore, suggested to be included in studies concerning estimation of parasitological parameters.  相似文献   

12.
Pooling of DNA samples can significantly reduce the effort of population studies with DNA markers. I present a statistical model and numerical method for estimating gene frequency when pooled DNA is assayed for the presence/absence of alleles. Analytical and Monte‐Carlo methods examined estimation variance and bias, and hence optimal pool size, under a triangular allele frequency distribution. For gene frequency of rarer alleles, the optimal number of pooled individuals is approximately the inverse of the gene frequency. For heterozygosity, the optimal pool is approximately half the allele number; this results in pools containing, on average, 60% of possible alleles.  相似文献   

13.
The MC simulation program MOCCA and the combined methods of Distance Geometry and Molecular Dynamics are utilised for structural studies of four isomers of the bee venom toxin apamin. For the MC strategy the conformational space is reduced to torsional degrees of freedom. The study compares the efficiency of both simulation strategies for structure determination of bicyclic peptides and examines the limits of the Monte Carlo method. MOCCA shows a lower efficiency as compared to the combined methods of Distance Geometry and Molecular Dynamics for the structure determination of the bicyclic isomers of apamin.Electronic Supplementary Material available.  相似文献   

14.
Summary In this article, we propose a family of semiparametric transformation models with time‐varying coefficients for recurrent event data in the presence of a terminal event such as death. The new model offers great flexibility in formulating the effects of covariates on the mean functions of the recurrent events among survivors at a given time. For the inference on the proposed models, a class of estimating equations is developed and asymptotic properties of the resulting estimators are established. In addition, a lack‐of‐fit test is provided for assessing the adequacy of the model, and some tests are presented for investigating whether or not covariate effects vary with time. The finite‐sample behavior of the proposed methods is examined through Monte Carlo simulation studies, and an application to a bladder cancer study is also illustrated.  相似文献   

15.
Genetic assignment methods use genotype likelihoods to draw inference about where individuals were or were not born, potentially allowing direct, real-time estimates of dispersal. We used simulated data sets to test the power and accuracy of Monte Carlo resampling methods in generating statistical thresholds for identifying F0 immigrants in populations with ongoing gene flow, and hence for providing direct, real-time estimates of migration rates. The identification of accurate critical values required that resampling methods preserved the linkage disequilibrium deriving from recent generations of immigrants and reflected the sampling variance present in the data set being analysed. A novel Monte Carlo resampling method taking into account these aspects was proposed and its efficiency was evaluated. Power and error were relatively insensitive to the frequency assumed for missing alleles. Power to identify F0 immigrants was improved by using large sample size (up to about 50 individuals) and by sampling all populations from which migrants may have originated. A combination of plotting genotype likelihoods and calculating mean genotype likelihood ratios (DLR) appeared to be an effective way to predict whether F0 immigrants could be identified for a particular pair of populations using a given set of markers.  相似文献   

16.
Estimating the age of the common ancestor of a sample of DNA sequences   总被引:10,自引:3,他引:7  
We present a simple Monte Carlo method for estimating the age of the most recent common ancestor (MRCA) of a sample of DNA sequences. We show that Templeton's (1993) estimator of the age of the MRCA based on the maximum number of nucleotide differences between two sequences in a sample is inaccurate, and we demonstrate the new method by reanalyzing a sample of DNA sequences from human Y chromosomes and a sample of human Alu sequences.   相似文献   

17.
Abstract

Taboo-based Monte Carlo search which restricts the sampling of the region near an old configuration, is developed. In this procedure, Monte Carlo simulation and random search method are combined to improve the sampling efficiency. The feasibility of this method is tested on global optimization of a continuous model function, melting of the 256 Lennard-Jones particles at T? = 0.680 and ρ? = 0.850 and polypeptides (alanine dipeptide and Metenkephalin). From the comparison of results for the model function between our method and other methods, we find the increase of convergence rate and the high possibility of escaping from the local energy minima. The results of the Lennard-Jones solids and polypeptides show that the convergence property to reach the equilibrium state is better than that of others. It is also found that no significant bias in ensemble distribution is detected, though taboo-based Monte Carlo search does not sample the correct ensemble distribution owing to the restriction of the sampling of the region near an old configuration.  相似文献   

18.
Multiple lower limits of quantification (MLOQs) result if various laboratories are involved in the analysis of concentration data and some observations are too low to be quantified. For normally distributed data under MLOQs there exists only the multiple regression method of Helsel to estimate the mean and variance. We propose a simple imputation method and two new maximum likelihood estimation methods: the multiple truncated sample method and the multiple censored sample method. A simulation study is conducted to compare the performances of the newly introduced methods to Helsel's via the criteria root mean squared error (RMSE) and bias of the parameter estimates. Two and four lower limits of quantification (LLOQs), various amounts of unquantifiable observations and two sample sizes are studied. Furthermore, the robustness is investigated under model misspecification. The methods perform with decreasing accuracy for increasing rates of unquantified observations. Increasing sample sizes lead to smaller bias. There is almost no change in the performance between two and four LLOQs. The magnitude of the variance impairs the performance of all methods. For a smaller variance, the multiple censored sample method leads to superior estimates regarding the RMSE and bias, whereas Helsel's method is superior regarding the bias for a larger variance. Under model misspecification, Helsel's method was inferior to the other methods. Estimating the mean, the multiple censored sample method performed better, whereas the multiple truncated sample method performs best in estimating the variance. Summarizing, for a large sample size and normally distributed data we recommend to use Helsel's method. Otherwise, the multiple censored sample method should be used to obtain estimates of the mean and variance of data including MLOQs.  相似文献   

19.
DNA extracts from enumerated target organism cells (calibrator samples) have been used for estimating Enterococcus cell equivalent densities in surface waters by a comparative cycle threshold (Ct) qPCR analysis method. To compare surface water Enterococcus density estimates from different studies by this approach, either a consistent source of calibrator cells must be used or the estimates must account for any differences in target sequence recoveries from different sources of calibrator cells. In this report we describe two methods for estimating target sequence recoveries from whole cell calibrator samples based on qPCR analyses of their serially diluted DNA extracts and most probable number (MPN) calculation. The first method employed a traditional MPN calculation approach. The second method employed a Bayesian hierarchical statistical modeling approach and a Monte Carlo Markov Chain (MCMC) simulation method to account for the uncertainty in these estimates associated with different individual samples of the cell preparations, different dilutions of the DNA extracts and different qPCR analytical runs. The two methods were applied to estimate mean target sequence recoveries per cell from two different lots of a commercially available source of enumerated Enterococcus cell preparations. The mean target sequence recovery estimates (and standard errors) per cell from Lot A and B cell preparations by the Bayesian method were 22.73 (3.4) and 11.76 (2.4), respectively, when the data were adjusted for potential false positive results. Means were similar for the traditional MPN approach which cannot comparably assess uncertainty in the estimates. Cell numbers and estimates of recoverable target sequences in calibrator samples prepared from the two cell sources were also used to estimate cell equivalent and target sequence quantities recovered from surface water samples in a comparative Ct method. Our results illustrate the utility of the Bayesian method in accounting for uncertainty, the high degree of precision attainable by the MPN approach and the need to account for the differences in target sequence recoveries from different calibrator sample cell sources when they are used in the comparative Ct method.  相似文献   

20.
Rarefaction methods have been introduced into population genetics (from ecology) for predicting and comparing the allelic richness of future samples (or sometimes populations) on the basis of currently available samples, possibly of different sizes. Here, we focus our attention on one such problem: Predicting which population is most likely to yield the future sample having the highest allelic richness. (This problem can arise when we want to construct a core collection from a larger germplasm collection.) We use extensive simulations to compare the performance of the Monte Carlo rarefaction (repeated random subsampling) method with a simple Bayesian approach we have developed-which is based on the Ewens sampling distribution. We found that neither this Bayesian method nor the (Monte Carlo) rarefaction method performed uniformly better than the other. We also examine briefly some of the other motivations offered for these methods and try to make sense of them from a Bayesian point of view.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号