首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
《Ecological Complexity》2007,4(1-2):1-12
Scientists investigate the dynamics of complex systems with quantitative models, employing them to synthesize knowledge, to explain observations, and to forecast future system behavior. Complete specification of systems is impossible, so models must be simplified abstractions. Thus, the art of modeling involves deciding which system elements to include and determining how they should be represented. We view modeling as search through a space of candidate models that is guided by model objectives, theoretical knowledge, and empirical data. In this contribution, we introduce a method for representing process-based models that facilitates the discovery of structures that explain observed behavior. This representation casts dynamic systems as interacting sets of processes that act on entities. Using this approach, a modeler first encodes relevant ecological knowledge into a library of generic entities and processes, then instantiates these theoretical components, and finally assembles candidate models from these elements. We illustrate this methodology with a model of the Ross Sea ecosystem.  相似文献   

2.
MOTIVATION: We consider models useful for learning an evolutionary or phylogenetic tree from data consisting of DNA sequences corresponding to the leaves of the tree. In particular, we consider a general probabilistic model described in Siepel and Haussler that we call the phylogenetic-HMM model which generalizes the classical probabilistic models of Neyman and Felsenstein. Unfortunately, computing the likelihood of phylogenetic-HMM models is intractable. We consider several approximations for computing the likelihood of such models including an approximation introduced in Siepel and Haussler, loopy belief propagation and several variational methods. RESULTS: We demonstrate that, unlike the other approximations, variational methods are accurate and are guaranteed to lower bound the likelihood. In addition, we identify a particular variational approximation to be best-one in which the posterior distribution is variationally approximated using the classic Neyman-Felsenstein model. The application of our best approximation to data from the cystic fibrosis transmembrane conductance regulator gene region across nine eutherian mammals reveals a CpG effect.  相似文献   

3.
In this paper we consider the recent advances in methodology for modeling of prokaryotic communities evolution and new features of the software package "Haploid evolutionary constructor" (http://evol-constructor.bionet.nsc.ru). We show the principles of building complex computer models in our software tool. These models describe several levels of biological organization: genetic, metabolic, population, ecological. New features of the haploid evolutionary constructor include the modeling of gene networks and phage infections.  相似文献   

4.
The ABC model postulates that expression combinations of three classes of genes (A, B and C) specify the four floral organs at early stages of flower development. This classic model provides a solid framework to study flower development and has been the foundation for multiple studies in different plant species, as well as for new evolutionary hypotheses. Nevertheless, it has been shown that in spite of being necessary, these three gene classes are not sufficient for flower organ specification. Rather, flower organ specification depends on complex interactions of several genes, and probably other non-genetic factors. Being useful to study systems of complex interactions, mathematical and computational models have enlightened the origin of the A, B and C stereotyped and robust expression patterns and the process of early flower morphogenesis. Here, we present a brief introduction to basic modeling concepts and techniques and review the results that these models have rendered for the particular case of the Arabidopsis thaliana flower organ specification. One of the main results is the uncovering of a robust functional module that is sufficient to recover the gene configurations characterizing flower organ primordia. Another key result is that the temporal sequence with which such gene configurations are attained may be recovered only by modeling the aforementioned functional module as a noisy or stochastic system. Finally, modeling approaches enable testable predictions regarding the role of non-genetic factors (noise, mechano-elastic forces, etc.) in development. These predictions, along with some perspectives for future work, are also reviewed and discussed.  相似文献   

5.
BIOCHAM (the BIOCHemical Abstract Machine) is a software environment for modeling biochemical systems. It is based on two aspects: (1) the analysis and simulation of boolean, kinetic and stochastic models and (2) the formalization of biological properties in temporal logic. BIOCHAM provides tools and languages for describing protein networks with a simple and straightforward syntax, and for integrating biological properties into the model. It then becomes possible to analyze, query, verify and maintain the model with respect to those properties. For kinetic models, BIOCHAM can search for appropriate parameter values in order to reproduce a specific behavior observed in experiments and formalized in temporal logic. Coupled with other methods such as bifurcation diagrams, this search assists the modeler/biologist in the modeling process. AVAILABILITY: BIOCHAM (v. 2.5) is a free software available for download, with example models, at http://contraintes.inria.fr/BIOCHAM/.  相似文献   

6.
7.
The identification of metabolic regulation is a major concern in metabolic engineering. Metabolic regulation phenomena depend on intracellular compounds such as enzymes, metabolites and cofactors. A complete understanding of metabolic regulation requires quantitative information about these compounds under in vivo conditions. This quantitative knowledge in combination with the known network of metabolic pathways allows the construction of mathematical models that describe the dynamic changes in metabolite concentrations over time. Rapid sampling combined with pulse experiments is a useful tool for the identification of metabolic regulation owing to the transient data they provide. Enzymatic tests in combination with ESI-LC-MS (Electrospray Ionization Liquid Chromatographic Tandem Mass Spectrometry) and HPLC measurements have been used to identify up to 30 metabolites and nucleotides from rapid sampling experiments. A metabolic modeling tool (MMT) that is built on a relational database was developed specifically for analysis of rapid sampling experiments. The tool allows to construct complex pathway models with information stored in the relational database. Parameter fitting and simulation algorithms for the resulting system of Ordinary Differential Equations (ODEs) are part of MMT. Additionally explicit sensitivity functions are calculated. The integration of all necessary algorithms in one tool allows fast model analysis and comparison. Complex models have been developed to describe the central metabolic pathways of Escherichia coli during a glucose pulse experiment.  相似文献   

8.
Summary .  Joint modeling of a primary response and a longitudinal process via shared random effects is widely used in many areas of application. Likelihood-based inference on joint models requires model specification of the random effects. Inappropriate model specification of random effects can compromise inference. We present methods to diagnose random effect model misspecification of the type that leads to biased inference on joint models. The methods are illustrated via application to simulated data, and by application to data from a study of bone mineral density in perimenopausal women and data from an HIV clinical trial.  相似文献   

9.
A translation of Systems Biology Markup Language (SBML) into a process algebra is proposed in order to allow the formal specification, the simulation and the formal analysis of biological models. Beta-binders, a language with a quantitative stochastic extension, is chosen for the translation. The proposed translation focuses on the main components of SBML models, as species and reactions. Furthermore, it satisfies the compositional property, i.e. the translation of the whole model is obtained by composing the translation of the subcomponents. An automatic translator tool of SBML models into Beta-binders has been implemented as well. Finally, the translation of a simple model is reported.  相似文献   

10.
The widely used “Maxent” software for modeling species distributions from presence‐only data (Phillips et al., Ecological Modelling, 190, 2006, 231) tends to produce models with high‐predictive performance but low‐ecological interpretability, and implications of Maxent's statistical approach to variable transformation, model fitting, and model selection remain underappreciated. In particular, Maxent's approach to model selection through lasso regularization has been shown to give less parsimonious distribution models—that is, models which are more complex but not necessarily predictively better—than subset selection. In this paper, we introduce the MIAmaxent R package, which provides a statistical approach to modeling species distributions similar to Maxent's, but with subset selection instead of lasso regularization. The simpler models typically produced by subset selection are ecologically more interpretable, and making distribution models more grounded in ecological theory is a fundamental motivation for using MIAmaxent. To that end, the package executes variable transformation based on expected occurrence–environment relationships and contains tools for exploring data and interrogating models in light of knowledge of the modeled system. Additionally, MIAmaxent implements two different kinds of model fitting: maximum entropy fitting for presence‐only data and logistic regression (GLM) for presence–absence data. Unlike Maxent, MIAmaxent decouples variable transformation, model fitting, and model selection, which facilitates methodological comparisons and gives the modeler greater flexibility when choosing a statistical approach to a given distribution modeling problem.  相似文献   

11.
12.
Dust storm has serious disastrous impacts on environment, human health, and assets. The developments and applications of dust storm models have contributed significantly to better understand and predict the distribution, intensity and structure of dust storms. However, dust storm simulation is a data and computing intensive process. To improve the computing performance, high performance computing has been widely adopted by dividing the entire study area into multiple subdomains and allocating each subdomain on different computing nodes in a parallel fashion. Inappropriate allocation may introduce imbalanced task loads and unnecessary communications among computing nodes. Therefore, allocation is a key factor that may impact the efficiency of parallel process. An allocation algorithm is expected to consider the computing cost and communication cost for each computing node to minimize total execution time and reduce overall communication cost for the entire simulation. This research introduces three algorithms to optimize the allocation by considering the spatial and communicational constraints: 1) an Integer Linear Programming (ILP) based algorithm from combinational optimization perspective; 2) a K-Means and Kernighan-Lin combined heuristic algorithm (K&K) integrating geometric and coordinate-free methods by merging local and global partitioning; 3) an automatic seeded region growing based geometric and local partitioning algorithm (ASRG). The performance and effectiveness of the three algorithms are compared based on different factors. Further, we adopt the K&K algorithm as the demonstrated algorithm for the experiment of dust model simulation with the non-hydrostatic mesoscale model (NMM-dust) and compared the performance with the MPI default sequential allocation. The results demonstrate that K&K method significantly improves the simulation performance with better subdomain allocation. This method can also be adopted for other relevant atmospheric and numerical modeling.  相似文献   

13.
The covarion hypothesis of molecular evolution proposes that selective pressures on an amino acid or nucleotide site change through time, thus causing changes of evolutionary rate along the edges of a phylogenetic tree. Several kinds of Markov models for the covarion process have been proposed. One model, proposed by Huelsenbeck (2002), has 2 substitution rate classes: the substitution process at a site can switch between a single variable rate, drawn from a discrete gamma distribution, and a zero invariable rate. A second model, suggested by Galtier (2001), assumes rate switches among an arbitrary number of rate classes but switching to and from the invariable rate class is not allowed. The latter model allows for some sites that do not participate in the rate-switching process. Here we propose a general covarion model that combines features of both models, allowing evolutionary rates not only to switch between variable and invariable classes but also to switch among different rates when they are in a variable state. We have implemented all 3 covarion models in a maximum likelihood framework for amino acid sequences and tested them on 23 protein data sets. We found significant likelihood increases for all data sets for the 3 models, compared with a model that does not allow site-specific rate switches along the tree. Furthermore, we found that the general model fit the data better than the simpler covarion models in the majority of the cases, highlighting the complexity in modeling the covarion process. The general covarion model can be used for comparing tree topologies, molecular dating studies, and the investigation of protein adaptation.  相似文献   

14.

Purpose

Life cycle assessment (LCA) practitioners face many challenges in their efforts to describe, share, review, and revise their product system models, and to reproduce the models and results of others. Current life cycle inventory modeling techniques have weaknesses in the areas of describing model structure, documenting the use of proxy or non-ideal data, specifying allocation, and including modeler’s observations and assumptions—all affecting how the study is interpreted and limiting the reuse of models. Moreover, LCA software systems manage modeling information in different and sometimes non-compatible ways. Practitioners must also deal with licensing, privacy/confidentiality of data, and other issues around data access which impact how a model can be shared.

Methods

This letter was prepared by a working group of the North American Life Cycle Assessment Advisory Group to support the UNEP-SETAC Life Cycle Initiative’s Flagship Activity on Data, Methods, and Product Sustainability Information. The aim of the working group is to define a roadmap of the technical advances needed to achieve easier LCA model sharing and improve replicability of LCA results among different users in a way that is independent of the LCA software used to compute the results and does not infringe on any licensing restrictions or confidentiality requirements. This is intended to be a consensus document providing the state of the art in this area, with milestones for research and implementation needed to resolve current issues.

Results and Conclusions

The roadmap identifies fifteen milestones in three areas: “describing model contents,” “describing model structure,” and “collaborative use of models.” The milestones should support researchers and software developers in advancing practitioners’ abilities to share and review product system models.
  相似文献   

15.
The choice of a probabilistic model to describe sequence evolution can and should be justified. Underfitting the data through the use of overly simplistic models may miss out on interesting phenomena and lead to incorrect inferences. Overfitting the data with models that are too complex may ascribe biological meaning to statistical artifacts and result in falsely significant findings. We describe a likelihood-based approach for evolutionary model selection. The procedure employs a genetic algorithm (GA) to quickly explore a combinatorially large set of all possible time-reversible Markov models with a fixed number of substitution rates. When applied to stem RNA data subject to well-understood evolutionary forces, the models found by the GA 1) capture the expected overall rate patterns a priori; 2) fit the data better than the best available models based on a priori assumptions, suggesting subtle substitution patterns not previously recognized; 3) cannot be rejected in favor of the general reversible model, implying that the evolution of stem RNA sequences can be explained well with only a few substitution rate parameters; and 4) perform well on simulated data, both in terms of goodness of fit and the ability to estimate evolutionary rates. We also investigate the utility of several distance measures for comparing and contrasting inferred evolutionary models. Using widely available small computer clusters, our approach allows, for the first time, to evaluate the performance of existing RNA evolutionary models by comparing them with a large pool of candidate models and to validate common modeling assumptions. In addition, the new method provides the foundation for rigorous selection and comparison of substitution models for other types of sequence data.  相似文献   

16.
Hypothesis generation in observational, biomedical data science often starts with computing an association or identifying the statistical relationship between a dependent and an independent variable. However, the outcome of this process depends fundamentally on modeling strategy, with differing strategies generating what can be called “vibration of effects” (VoE). VoE is defined by variation in associations that often lead to contradictory results. Here, we present a computational tool capable of modeling VoE in biomedical data by fitting millions of different models and comparing their output. We execute a VoE analysis on a series of widely reported associations (e.g., carrot intake associated with eyesight) with an extended additional focus on lifestyle exposures (e.g., physical activity) and components of the Framingham Risk Score for cardiovascular health (e.g., blood pressure). We leveraged our tool for potential confounder identification, investigating what adjusting variables are responsible for conflicting models. We propose modeling VoE as a critical step in navigating discovery in observational data, discerning robust associations, and cataloging adjusting variables that impact model output.

COVID positivity and vitamin D intake, red meat and heart disease; how can we discern when biomedical associations are reliable and when they are susceptible to our own arbitrary choices and assumptions? This study presents “quantvoe,” a software package for exploring the entirety of possible findings due to the multiverse of associations possible.  相似文献   

17.
Insertions and deletions (indels) are common molecular evolutionary events. However, probabilistic models for indel evolution are under-developed due to their computational complexity. Here, we introduce several improvements to indel modeling: 1) While previous models for indel evolution assumed that the rates and length distributions of insertions and deletions are equal, here we propose a richer model that explicitly distinguishes between the two; 2) we introduce numerous summary statistics that allow approximate Bayesian computation-based parameter estimation; 3) we develop a method to correct for biases introduced by alignment programs, when inferring indel parameters from empirical data sets; and 4) using a model-selection scheme, we test whether the richer model better fits biological data compared with the simpler model. Our analyses suggest that both our inference scheme and the model-selection procedure achieve high accuracy on simulated data. We further demonstrate that our proposed richer model better fits a large number of empirical data sets and that, for the majority of these data sets, the deletion rate is higher than the insertion rate.  相似文献   

18.
In recent years, hybrid neural network approaches, which combine mechanistic and neural network models, have received considerable attention. These approaches are potentially very efficient for obtaining more accurate predictions of process dynamics by combining mechanistic and neural network models in such a way that the neural network model properly accounts for unknown and nonlinear parts of the mechanistic model. In this work, a full-scale coke-plant wastewater treatment process was chosen as a model system. Initially, a process data analysis was performed on the actual operational data by using principal component analysis. Next, a simplified mechanistic model and a neural network model were developed based on the specific process knowledge and the operational data of the coke-plant wastewater treatment process, respectively. Finally, the neural network was incorporated into the mechanistic model in both parallel and serial configurations. Simulation results showed that the parallel hybrid modeling approach achieved much more accurate predictions with good extrapolation properties as compared with the other modeling approaches even in the case of process upset caused by, for example, shock loading of toxic compounds. These results indicate that the parallel hybrid neural modeling approach is a useful tool for accurate and cost-effective modeling of biochemical processes, in the absence of other reasonably accurate process models.  相似文献   

19.
The complexity of the global organization and internal structure of motifs in higher eukaryotic organisms raises significant challenges for motif detection techniques. To achieve successful de novo motif detection, it is necessary to model the complex dependencies within and among motifs and to incorporate biological prior knowledge. In this paper, we present LOGOS, an integrated LOcal and GlObal motif Sequence model for biopolymer sequences, which provides a principled framework for developing, modularizing, extending and computing expressive motif models for complex biopolymer sequence analysis. LOGOS consists of two interacting submodels: HMDM, a local alignment model capturing biological prior knowledge and positional dependency within the motif local structure; and HMM, a global motif distribution model modeling frequencies and dependencies of motif occurrences. Model parameters can be fit using training motifs within an empirical Bayesian framework. A variational EM algorithm is developed for de novo motif detection. LOGOS improves over existing models that ignore biological priors and dependencies in motif structures and motif occurrences, and demonstrates superior performance on both semi-realistic test data and cis-regulatory sequences from yeast and Drosophila genomes with regard to sensitivity, specificity, flexibility and extensibility.  相似文献   

20.
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号