首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.

Background  

Inference of population stratification and individual admixture from genetic markers is an integrative part of a study in diverse situations, such as association mapping and evolutionary studies. Bayesian methods have been proposed for population stratification and admixture inference using multilocus genotypes and widely used in practice. However, these Bayesian methods demand intensive computation resources and may run into convergence problem in Markov Chain Monte Carlo based posterior samplings.  相似文献   

2.

Background  

Statistical approaches for protein design are relevant in the field of molecular evolutionary studies. In recent years, new, so-called structurally constrained (SC) models of protein-coding sequence evolution have been proposed, which use statistical potentials to assess sequence-structure compatibility. In a previous work, we defined a statistical framework for optimizing knowledge-based potentials especially suited to SC models. Our method used the maximum likelihood principle and provided what we call the joint potentials. However, the method required numerical estimations by the use of computationally heavy Markov Chain Monte Carlo sampling algorithms.  相似文献   

3.

Background  

For gene expression data obtained from a time-course microarray experiment, Liu et al. [1] developed a new algorithm for clustering genes with similar expression profiles over time. Performance of their proposal was compared with three other methods including the order-restricted inference based methodology of Peddada et al. [2, 3]. In this note we point out several inaccuracies in Liu et al. [1] and conclude that the order-restricted inference based methodology of Peddada et al. (programmed in the software ORIOGEN) indeed operates at the desired nominal Type 1 error level, an important feature of a statistical decision rule, while being computationally substantially faster than indicated by Liu et al. [1].  相似文献   

4.

Background  

Wise et al. introduced a rank-based statistical technique for meta-analysis of genome scans, the Genome Scan Meta-Analysis (GSMA) method. Levinson et al. recently described two generalizations of the GSMA statistic: (i) a weighted version of the GSMA statistic, so that different studies could be ascribed different weights for analysis; and (ii) an order statistic approach, reflecting the fact that a GSMA statistic can be computed for each chromosomal region or bin width across the various genome scan studies.  相似文献   

5.
6.
We propose a reduced ODE model for the mechanical activation of cardiac myofilaments, which is based on explicit spatial representation of nearest-neighbour interactions. Our model is derived from the cooperative Markov Chain model of Washio et al. (Cell Mol Bioeng 5(1):113–126, 2012), under the assumption of conditional independence of specific sets of events. This physically motivated assumption allows to drastically reduce the number of degrees of freedom, thus resulting in a significantly large computational saving. Indeed, the original Markov Chain model involves a huge number of degrees of freedom (order of \(10^{21}\)) and is solved by means of the Monte Carlo method, which notoriously reaches statistical convergence in a slow fashion. With our reduced model, instead, numerical simulations can be carried out by solving a system of ODEs, reducing the computational time by more than 10, 000 times. Moreover, the reduced model is accurate with respect to the original Markov Chain model. We show that the reduced model is capable of reproducing physiological steady-state force–calcium and force–length relationships with the observed asymmetry in apparent cooperativity near the calcium level producing half activation. Finally, we also report good qualitative and quantitative agreement with experimental measurements under dynamic conditions.  相似文献   

7.

Background and aims

Continuous vegetable cultivation in greenhouses can easily induce soil degradation, which considerably affects the development of sustainable vegetable production. Recently, the reductive soil disinfestation (RSD) is widely used as an alternative to chemical soil disinfestations to improve degraded greenhouse vegetable soils. Considering the importance of nitrogen (N) for plant growth and environment effect, the internal N transformation processes and rates should be well investigated in degraded vegetable soils treated by RSD, but few works have been undertaken.

Methods

Three RSD-treated and three untreated degraded vegetable soils were chosen and a 15?N tracing incubation experiment differentially labeled with 15NH4NO3 or NH4 15NO3 was conducted at 25 °C under 50 % water holding capacity (WHC) for 96 h. Soil gross N transformation rates were calculated using a 15?N tracing model combined with Markov Chain Monte Carlo Metropolis algorithm (Müller et al. 2007), while the emissions of N2O and NO were also measured.

Results

RSD could significantly enhance the soil microbial NH4 + immobilization rate, the heterotrophic and autotrophic nitrification rates, and the NO3 ? turnover time. The ratio of heterotrophic nitrification to total inorganic N supply rate (mineralization + heterotrophic nitrification) increased greatly from 5.4 % in untreated vegetable soil to 56.1 % in treated vegetable soil. In addition, low release potential of NO and N2O was observed in RSD-treated vegetable soil, due to the decrease in the NO and N2O product ratios from heterotrophic and autotrophic nitrifications. These significant differences in gross N transformation rates, the supply processes and capacity of inorganic N, and the NO and N2O emissions between untreated and treated vegetable soils could be explained by the elimination of accumulated NO3 ?, increased pH, and decreased electrical conductivity (EC) caused by RSD. Noticeably, the NO3 ? consumption rates were still significantly lower than the NO3 ? production rates in RSD-treated vegetable soil.

Conclusions

Except for improving soil chemical properties, RSD could significantly alter the supply processes of inorganic N and reduce the release potential of N2O and NO in RSD-treated degraded vegetable soil. In order to retard the re-occurrence of NO3 ? accumulation, acidification and salinization and to promote the long-term productivity of greenhouse vegetable fields, the rational use of N fertilizer should be paid great attention to farmers in vegetable cultivation.  相似文献   

8.

Background

In general, the individual patient-level data (IPD) collected in clinical trials are not available to independent researchers to conduct economic evaluations; researchers only have access to published survival curves and summary statistics. Thus, methods that use published survival curves and summary statistics to reproduce statistics for economic evaluations are essential. Four methods have been identified: two traditional methods 1) least squares method, 2) graphical method; and two recently proposed methods by 3) Hoyle and Henley, 4) Guyot et al. The four methods were first individually reviewed and subsequently assessed regarding their abilities to estimate mean survival through a simulation study.

Methods

A number of different scenarios were developed that comprised combinations of various sample sizes, censoring rates and parametric survival distributions. One thousand simulated survival datasets were generated for each scenario, and all methods were applied to actual IPD. The uncertainty in the estimate of mean survival time was also captured.

Results

All methods provided accurate estimates of the mean survival time when the sample size was 500 and a Weibull distribution was used. When the sample size was 100 and the Weibull distribution was used, the Guyot et al. method was almost as accurate as the Hoyle and Henley method; however, more biases were identified in the traditional methods. When a lognormal distribution was used, the Guyot et al. method generated noticeably less bias and a more accurate uncertainty compared with the Hoyle and Henley method.

Conclusions

The traditional methods should not be preferred because of their remarkable overestimation. When the Weibull distribution was used for a fitted model, the Guyot et al. method was almost as accurate as the Hoyle and Henley method. However, if the lognormal distribution was used, the Guyot et al. method was less biased compared with the Hoyle and Henley method.  相似文献   

9.
A Bayesian model-based clustering approach is proposed for identifying differentially expressed genes in meta-analysis. A Bayesian hierarchical model is used as a scientific tool for combining information from different studies, and a mixture prior is used to separate differentially expressed genes from non-differentially expressed genes. Posterior estimation of the parameters and missing observations are done by using a simple Markov chain Monte Carlo method. From the estimated mixture model, useful measure of significance of a test such as the Bayesian false discovery rate (FDR), the local FDR (Efron et al., 2001), and the integration-driven discovery rate (IDR; Choi et al., 2003) can be easily computed. The model-based approach is also compared with commonly used permutation methods, and it is shown that the model-based approach is superior to the permutation methods when there are excessive under-expressed genes compared to over-expressed genes or vice versa. The proposed method is applied to four publicly available prostate cancer gene expression data sets and simulated data sets.  相似文献   

10.
The software tool PBEAM provides a parallel implementation of the BEAM, which is the first algorithm for large scale epistatic interaction mapping, including genome-wide studies with hundreds of thousands of markers. BEAM describes markers and their interactions with a Bayesian partitioning model and computes the posterior probability of each marker sets via Markov Chain Monte Carlo (MCMC). PBEAM takes the advantage of simulating multiple Markov chains simultaneously. This design can efficiently reduce ~n-fold execution time in the circumstance of n CPUs. The implementation of PBEAM is based on MPI libraries.

Availability

PBEAM is available for download at http://bioinfo.au.tsinghua.edu.cn/pbeam/  相似文献   

11.
Martins H  Villesen P 《PloS one》2011,6(3):e14745

Background

Endogenous retroviruses (ERVs) are genetic fossils of ancient retroviral integrations that remain in the genome of many organisms. Most loci are rendered non-functional by mutations, but several intact retroviral genes are known in mammalian genomes. Some have been adopted by the host species, while the beneficial roles of others remain unclear. Besides the obvious possible immunogenic impact from transcribing intact viral genes, endogenous retroviruses have also become an interesting and useful tool to study phylogenetic relationships. The determination of the integration time of these viruses has been based upon the assumption that both 5′ and 3′ Long Terminal Repeats (LTRs) sequences are identical at the time of integration, but evolve separately afterwards. Similar approaches have been using either a constant evolutionary rate or a range of rates for these viral loci, and only single species data. Here we show the advantages of using different approaches.

Results

We show that there are strong advantages in using multiple species data and state-of-the-art phylogenetic analysis. We incorporate both simple phylogenetic information and Monte Carlo Markov Chain (MCMC) methods to date the integrations of these viruses based on a relaxed molecular clock approach over a Bayesian phylogeny model and applied them to several selected ERV sequences in primates. These methods treat each ERV locus as having a distinct evolutionary rate for each LTR, and make use of consensual speciation time intervals between primates to calibrate the relaxed molecular clocks.

Conclusions

The use of a fixed rate produces results that vary considerably with ERV family and the actual evolutionary rate of the sequence, and should be avoided whenever multi-species phylogenetic data are available. For genome-wide studies, the simple phylogenetic approach constitutes a better alternative, while still being computationally feasible.  相似文献   

12.

Background  

In recent years there has been a trend of leaving the strict molecular clock in order to infer dating of speciations and other evolutionary events. Explicit modeling of substitution rates and divergence times makes formulation of informative prior distributions for branch lengths possible. Models with birth-death priors on tree branching and auto-correlated or iid substitution rates among lineages have been proposed, enabling simultaneous inference of substitution rates and divergence times. This problem has, however, mainly been analysed in the Markov chain Monte Carlo (MCMC) framework, an approach requiring computation times of hours or days when applied to large phylogenies.  相似文献   

13.
14.

Background  

Interaction of a drug or chemical with a biological system can result in a gene-expression profile or signature characteristic of the event. Using a suitably robust algorithm these signatures can potentially be used to connect molecules with similar pharmacological or toxicological properties by gene expression profile. Lamb et al first proposed the Connectivity Map [Lamb et al (2006), Science 313, 1929–1935] to make successful connections among small molecules, genes, and diseases using genomic signatures.  相似文献   

15.
Inference of protein functions is one of the most important aims of modern biology. To fully exploit the large volumes of genomic data typically produced in modern-day genomic experiments, automated computational methods for protein function prediction are urgently needed. Established methods use sequence or structure similarity to infer functions but those types of data do not suffice to determine the biological context in which proteins act. Current high-throughput biological experiments produce large amounts of data on the interactions between proteins. Such data can be used to infer interaction networks and to predict the biological process that the protein is involved in. Here, we develop a probabilistic approach for protein function prediction using network data, such as protein-protein interaction measurements. We take a Bayesian approach to an existing Markov Random Field method by performing simultaneous estimation of the model parameters and prediction of protein functions. We use an adaptive Markov Chain Monte Carlo algorithm that leads to more accurate parameter estimates and consequently to improved prediction performance compared to the standard Markov Random Fields method. We tested our method using a high quality S.cereviciae validation network with 1622 proteins against 90 Gene Ontology terms of different levels of abstraction. Compared to three other protein function prediction methods, our approach shows very good prediction performance. Our method can be directly applied to protein-protein interaction or coexpression networks, but also can be extended to use multiple data sources. We apply our method to physical protein interaction data from S. cerevisiae and provide novel predictions, using 340 Gene Ontology terms, for 1170 unannotated proteins and we evaluate the predictions using the available literature.  相似文献   

16.

Objective

Set up a framework for evaluating automatic segmentation methods of tumour volumes on PET images.

Patient and methods

This study was performed with PET images of 18 patients with non-Hodgkin's lymphoma. One target lesion per patient was pointed out. Each lesion was then three times manually delineated by five experts. Four automatic methods (the application of a threshold of 42% of the maximum SUV, the MIP-based method, the Daisne et al. method and the Nestle et al. method) were evaluated by comparison with the set of manual delineations.

Results

From the manual delineations, we have concluded to a moderate intra-operator variability and to a reduced interoperator reproducibility. From statistical tests performed on various quantitative criteria, there was no significant difference between the MIP-based method, the Daisne et al. method and the Nestle et al. one. The application of a threshold of 42% of the maximum SUV appears to be less efficient.

Conclusion

This work proposes a comparison and an evaluation protocol for segmentation methods. The generated data set will be distributed online for the community to simplify the evaluation of any new method of segmentation.  相似文献   

17.

Purpose  

Previous methods of estimating characterization factors (CFs) of metals in life cycle impact assessment (LCIA) models were based on multimedia fate, exposure, and effect models originally developed to address the potential impacts of organic chemicals. When applied to metals, the models neglect the influence of ambient chemistry on metal speciation, bioavailability and toxicity. Gandhi et al. (2010) presented a new method of calculating CFs for freshwater ecotoxicity that addresses these metal-specific issues. In this paper, we compared and assessed the consequences of using the new method versus currently available LCIA models for calculating freshwater ecotoxicity, as applied to two case studies previously examined by Gloria et al. (2006): (1) the production of copper (Cu) pipe and (2) a zinc (Zn) gutter system.  相似文献   

18.

Background  

Recently, Hill et al. [1] implemented a new software package--called SPRIT--which aims at calculating the minimum number of horizontal gene transfer events that is needed to simultaneously explain the evolution of two rooted binary phylogenetic trees on the same set of taxa. To this end, SPRIT computes the closely related so-called rooted subtree prune and regraft distance between two phylogenies. However, calculating this distance is an NP-hard problem and exact algorithms are often only applicable to small- or medium-sized problem instances. Trying to overcome this problem, Hill et al. propose a divide-and-conquer approach to speed up their algorithm and conjecture that this approach can be used to compute the rooted subtree prune and regraft distance exactly.  相似文献   

19.

Background  

Recent years have seen the emergence of genome annotation methods based on the phylo-grammar, a probabilistic model combining continuous-time Markov chains and stochastic grammars. Previously, phylo-grammars have required considerable effort to implement, limiting their adoption by computational biologists.  相似文献   

20.

Background  

The use of chromatography coupled with mass spectrometry (MS) analysis is a powerful approach to identify proteins, owing to its capacity to fractionate molecules according to different chemical features. The first protein expression map of vascular smooth muscle cells (VSMC) was published in 2001 and since then other papers have been produced. The most detailed two-dimensional polyacrylamide gel electrophoresis (2D-PAGE) map was presented by Mayr et al who identified 235 proteins, corresponding to the 154 most abundant unique proteins in mouse aortic VSMC. A chromatographic approach aimed at fractionating the VSMC proteome has never been used before.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号