首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
Experimental techniques with high temporal and spatial resolution extend our knowledge of how biological macromolecules self-organise and function. Here, we provide an illustration of the convergence between simulation and experiment made possible by techniques such as triplet-triplet energy transfer and fluorescence quenching with long-lifetime and fast-quenching fluorescent probes. These techniques have recently been used to determine the average time needed for two residues in a peptide or protein segment to form a contact. The timescale of this process is accessible to computer simulation, providing a microscopic interpretation of the data and yielding new insight into the disordered state of proteins. Conversely, such experimental data also provide a test of the validity of alternative choices for the molecular models used in simulations, indicating their possible deficiencies. We carried out simulations of peptides of various composition and length using several models. End-to-end contact formation rates and their dependence on peptide length agree with experimental estimates for some sequences and some force fields but not for others. The deviations are due to artefactual structuring of some peptides, which is not observed when an atomistic model for the solvation water is used. Simulations show that the observed experimental rates are compatible with considerably different distributions of the end-to-end distance; for realistic models, these are never Gaussian but indicative of a rugged energy landscape.  相似文献   

2.
SUMMARY: VISDA (Visual Statistical Data Analyzer) is a caBIG analytical tool for cluster modeling, visualization and discovery that has met silver-level compatibility under the caBIG initiative. Being statistically principled and visually interfaced, VISDA exploits both hierarchical statistics modeling and human gift for pattern recognition to allow a progressive yet interactive discovery of hidden clusters within high dimensional and complex biomedical datasets. The distinctive features of VISDA are particularly useful for users across the cancer research and broader research communities to analyze complex biological data. AVAILABILITY: http://gforge.nci.nih.gov/projects/visda/  相似文献   

3.
A huge amount of important biomedical information is hidden in the bulk of research articles in biomedical fields. At the same time, the publication of databases of biological information and of experimental datasets generated by high-throughput methods is in great expansion, and a wealth of annotated gene databases, chemical, genomic (including microarray datasets), clinical and other types of data repositories are now available on the Web. Thus a current challenge of bioinformatics is to develop targeted methods and tools that integrate scientific literature, biological databases and experimental data for reducing the time of database curation and for accessing evidence, either in the literature or in the datasets, useful for the analysis at hand. Under this scenario, this article reviews the knowledge discovery systems that fuse information from the literature, gathered by text mining, with microarray data for enriching the lists of down and upregulated genes with elements for biological understanding and for generating and validating new biological hypothesis. Finally, an easy to use and freely accessible tool, GeneWizard, that exploits text mining and microarray data fusion for supporting researchers in discovering gene-disease relationships is described.  相似文献   

4.
Systems biology aims to provide a holistic and in many cases dynamic picture of biological function and malfunction, in case of disease. Technology developments in the generation of genome-wide datasets and massive improvements in computer power now allow to obtain new insights into complex biological networks and to copy nature by computing these interactions and their kinetics and by generating in silico models of cells, tissues and organs. The expectations are high that systems biology will pave the way to the identification of novel disease genes, to the selection of successful drug candidates—that do not fail in clinical studies due to toxicity or lack of human efficacy—and finally to a more successful discovery of novel therapeutics. However, further research is necessary to fully unleash the potential of systems biology. Within this review we aim to highlight the most important and promising top-down and bottom-up systems biology applications in drug discovery.  相似文献   

5.
Computational techniques have been adopted in medi-cal and biological systems for a long time. There is no doubt that the development and application of computational methods will render great help in better understanding biomedical and biological functions. Large amounts of datasets have been produced by biomedical and biological experiments and simulations. In order for researchers to gain knowledge from origi- nal data, nontrivial transformation is necessary, which is regarded as a critical link in the chain of knowledge acquisition, sharing, and reuse. Challenges that have been encountered include: how to efficiently and effectively represent human knowledge in formal computing models, how to take advantage of semantic text mining techniques rather than traditional syntactic text mining, and how to handle security issues during the knowledge sharing and reuse. This paper summarizes the state-of-the-art in these research directions. We aim to provide readers with an introduction of major computing themes to be applied to the medical and biological research.  相似文献   

6.
7.
Protein-protein interactions (PPIs) trigger a wide range of biological signaling pathways that are crucial for biomedical research and drug discovery. Various techniques have been used to study specific proteins, including affinity chromatography, activity-based probes, affinity-based probes and photo-affinity labeling (PAL). PAL has become one of the most powerful strategies to study PPIs. Traditional photocrosslinkers are used in PAL, including benzophenone, aryl azide, and diazirine. Upon photoirradiation, these photocrosslinkers (Pls) generate highly reactive species that react with adjacent molecules, resulting in a direct covalent modification. This review introduces recent examples of chemical proteomics study using PAL for PPIs.  相似文献   

8.
Accuracy of results from mathematical and computer models of biological systems is often complicated by the presence of uncertainties in experimental data that are used to estimate parameter values. Current mathematical modeling approaches typically use either single-parameter or local sensitivity analyses. However, these methods do not accurately assess uncertainty and sensitivity in the system as, by default, they hold all other parameters fixed at baseline values. Using techniques described within we demonstrate how a multi-dimensional parameter space can be studied globally so all uncertainties can be identified. Further, uncertainty and sensitivity analysis techniques can help to identify and ultimately control uncertainties. In this work we develop methods for applying existing analytical tools to perform analyses on a variety of mathematical and computer models. We compare two specific types of global sensitivity analysis indexes that have proven to be among the most robust and efficient. Through familiar and new examples of mathematical and computer models, we provide a complete methodology for performing these analyses, in both deterministic and stochastic settings, and propose novel techniques to handle problems encountered during these types of analyses.  相似文献   

9.
Hypothesis generation in observational, biomedical data science often starts with computing an association or identifying the statistical relationship between a dependent and an independent variable. However, the outcome of this process depends fundamentally on modeling strategy, with differing strategies generating what can be called “vibration of effects” (VoE). VoE is defined by variation in associations that often lead to contradictory results. Here, we present a computational tool capable of modeling VoE in biomedical data by fitting millions of different models and comparing their output. We execute a VoE analysis on a series of widely reported associations (e.g., carrot intake associated with eyesight) with an extended additional focus on lifestyle exposures (e.g., physical activity) and components of the Framingham Risk Score for cardiovascular health (e.g., blood pressure). We leveraged our tool for potential confounder identification, investigating what adjusting variables are responsible for conflicting models. We propose modeling VoE as a critical step in navigating discovery in observational data, discerning robust associations, and cataloging adjusting variables that impact model output.

COVID positivity and vitamin D intake, red meat and heart disease; how can we discern when biomedical associations are reliable and when they are susceptible to our own arbitrary choices and assumptions? This study presents “quantvoe,” a software package for exploring the entirety of possible findings due to the multiverse of associations possible.  相似文献   

10.
We present our experience of building biological databases. Such databases have most aspects in common with other complex databases in other fields. We do not believe that biological data are that different from complex data in other fields. Our experience has led us to emphasise simplicity and conservative technology choices when building these databases. This is a short paper of advice that we hope is useful to people designing their own biological database.  相似文献   

11.
For the average biologist, hands-on literature mining currently means a keyword search in PubMed. However, methods for extracting biomedical facts from the scientific literature have improved considerably, and the associated tools will probably soon be used in many laboratories to automatically annotate and analyse the growing number of system-wide experimental data sets. Owing to the increasing body of text and the open-access policies of many journals, literature mining is also becoming useful for both hypothesis generation and biological discovery. However, the latter will require the integration of literature and high-throughput data, which should encourage close collaborations between biologists and computational linguists.  相似文献   

12.
Federated Learning enables machine learning across multiple sources of data and alleviates the risk of leaking private information between partners thereby encouraging knowledge sharing and collaborative modelling. Hence, Federated Learning opens the ways to a new generation of improved models. Domains involving molecular informatics, like Drug Discovery, are progressively adopting Federated Learning; this review describes the main projects and applications of Federated Learning for molecular discovery with a special focus on their benefits and the remaining challenges. All the studies demonstrate a real benefit of Federated Learning, namely the improvement of the performance of models as well as their applicability domain thanks to knowledge aggregation. The selected publications also reveal several remaining challenges to be addressed to fully exploit Federated Learning.  相似文献   

13.
Each diploid organism has two alleles at every gene locus. In sexual organisms such as most plants, animals and fungi, the two alleles in an individual may be genetically very different from each other. DNA sequence data from individual alleles (called a haplotype) can provide powerful information to address a variety of biological questions and guide many practical applications. The advancement in molecular technology and computational tools in the last decade has made obtaining large-scale haplotypes feasible. This review summarizes the two basic approaches for obtaining haplotypes and discusses the associated techniques and methods. The first approach is to experimentally obtain diploid sequence information and then use computer algorithms to infer haplotypes. The second approach is to obtain haplotype sequences directly through experimentation. The advantages and disadvantages of each approach are discussed. I then discussed a specific example on how the direct approach was used to obtain haplotype information to address several fundamental biological questions of a pathogenic yeast. With increasing sophistication in both bioinformatics tools and high-throughput molecular techniques, haplotype analysis is becoming an integrated component in biomedical research.  相似文献   

14.
Microarray technologies, which can measure tens of thousands of gene expression values simultaneously in a single experiment, have become a common research method for biomedical researchers. Computational tools to analyze microarray data for biological discovery are needed. In this paper, we investigate the feasibility of using formal concept analysis (FCA) as a tool for microarray data analysis. The method of FCA builds a (concept) lattice from the experimental data together with additional biological information. For microarray data, each vertex of the lattice corresponds to a subset of genes that are grouped together according to their expression values and some biological information related to gene function. The lattice structure of these gene sets might reflect biological relationships in the dataset. Similarities and differences between experiments can then be investigated by comparing their corresponding lattices according to various graph measures. We apply our method to microarray data derived from influenza-infected mouse lung tissue and healthy controls. Our preliminary results show the promise of our method as a tool for microarray data analysis.  相似文献   

15.
MOTIVATION: A very promising approach in drug discovery involves the integration of available biomedical data through mathematical modelling and data mining. We have developed a method called optimization program for drug discovery (OPDD) that allows new enzyme targets to be identified in enzymopathies through the integration of metabolic models and biomedical data in a mathematical optimization program. The method involves four steps: (i) collection of the necessary information about the metabolic system and disease; (ii) translation of the information into mathematical terms; (iii) computation of the optimization programs prioritizing the solutions that propose the inhibition of a reduced number of enzymes and (iv) application of additional biomedical criteria to select and classify the solutions. Each solution consists of a set of predicted values for metabolites, initial substrates and enzyme activities, which describe a biologically acceptable steady state of the system that shifts the pathologic state towards a healthy state. RESULTS: The OPDD was used to detect target enzymes in an enzymopathy, the human hyperuricemia. An existing S-system model and bibliographic information about the disease were used. The method detected six single-target enzyme solutions involving dietary modification, one of them coinciding with the conventional clinical treatment using allopurinol. The OPDD detected a large number of possible solutions involving two enzyme targets. All except one contained one of the previously detected six enzyme targets. The purpose of this work was not to obtain solutions for direct clinical implementation but to illustrate how increasing levels of biomedical information can be integrated together with mathematical models in drug discovery. SUPPLEMENTARY INFORMATION: Supplementary data are available at Bioinformatics online.  相似文献   

16.
As the amount of biological data in the public domain grows, so does the range of modeling and analysis techniques employed in systems biology. In recent years, a number of theoretical computer science developments have enabled modeling methodology to keep pace. The growing interest in systems biology in executable models and their analysis has necessitated the borrowing of terms and methods from computer science, such as formal analysis, model checking, static analysis, and runtime verification. Here, we discuss the most important and exciting computational methods and tools currently available to systems biologists. We believe that a deeper understanding of the concepts and theory highlighted in this review will produce better software practice, improved investigation of complex biological processes, and even new ideas and better feedback into computer science.  相似文献   

17.
Miao H  Dykes C  Demeter LM  Wu H 《Biometrics》2009,65(1):292-300
Summary .  Many biological processes and systems can be described by a set of differential equation (DE) models. However, literature in statistical inference for DE models is very sparse. We propose statistical estimation, model selection, and multimodel averaging methods for HIV viral fitness experiments in vitro that can be described by a set of nonlinear ordinary differential equations (ODE). The parameter identifiability of the ODE models is also addressed. We apply the proposed methods and techniques to experimental data of viral fitness for HIV-1 mutant 103N. We expect that the proposed modeling and inference approaches for the DE models can be widely used for a variety of biomedical studies.  相似文献   

18.
A knowledge model for analysis and simulation of regulatory networks   总被引:5,自引:0,他引:5  
MOTIVATION: In order to aid in hypothesis-driven experimental gene discovery, we are designing a computer application for the automatic retrieval of signal transduction data from electronic versions of scientific publications using natural language processing (NLP) techniques, as well as for visualizing and editing representations of regulatory systems. These systems describe both signal transduction and biochemical pathways within complex multicellular organisms, yeast, and bacteria. This computer application in turn requires the development of a domain-specific ontology, or knowledge model. RESULTS: We introduce an ontological model for the representation of biological knowledge related to regulatory networks in vertebrates. We outline a taxonomy of the concepts, define their 'whole-to-part' relationships, describe the properties of major concepts, and outline a set of the most important axioms. The ontology is partially realized in a computer system designed to aid researchers in biology and medicine in visualizing and editing a representation of a signal transduction system.  相似文献   

19.
Understanding epigenetic processes holds immense promise for medical applications. Advances in Machine Learning (ML) are critical to realize this promise. Previous studies used epigenetic data sets associated with the germline transmission of epigenetic transgenerational inheritance of disease and novel ML approaches to predict genome-wide locations of critical epimutations. A combination of Active Learning (ACL) and Imbalanced Class Learning (ICL) was used to address past problems with ML to develop a more efficient feature selection process and address the imbalance problem in all genomic data sets. The power of this novel ML approach and our ability to predict epigenetic phenomena and associated disease is suggested. The current approach requires extensive computation of features over the genome. A promising new approach is to introduce Deep Learning (DL) for the generation and simultaneous computation of novel genomic features tuned to the classification task. This approach can be used with any genomic or biological data set applied to medicine. The application of molecular epigenetic data in advanced machine learning analysis to medicine is the focus of this review.  相似文献   

20.
Natural products have immense therapeutic potential not only due to their structural variation and complexity but also due to their range of biological activities. Research based on natural products has led to the discovery of molecules with biomedical and pharmaceutical applications in different therapeutic areas like cancer, inflammation responses, diabetes, and infectious diseases. There are still several challenges to be overcome in natural product drug discovery research programs and the challenge of high throughput screening of natural substances is one of them. Bioactivity screening is an integral part of the drug discovery process and several in vitro and in vivo biological models are now available for this purpose. Among other well-reported biological models, the zebrafish (Danio rerio) is emerging as an important in vivo model for preclinical studies of synthetic molecules in different therapeutic areas. Zebrafish embryos have a short reproductive cycle, show ease of maintenance at high densities in the laboratory and administration of drugs is a straightforward procedure. The embryos are optically transparent, allowing for the visualization of drug effects on internal organs during the embryogenesis process. In this review, we illustrate the importance of using zebrafish as an important biological model in the discovery of bioactive drugs from natural sources.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号