首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 62 毫秒
1.
Post ‘omic’ era has resulted in the development of many primary, secondary and derived databases. Many analytical and visualization bioinformatics tools have been developed to manage and analyze the data available through large sequencing projects. Availability of heterogeneous databases and tools make it difficult for researchers to access information from varied sources and run different bioinformatics tools to get desired analysis done. Building integrated bioinformatics platforms is one of the most challenging tasks that bioinformatics community is facing. Integration of various databases, tools and algorithm is a challenging problem to deal with. This article describes the bioinformatics analysis workflow management systems that are developed in the area of gene sequence analysis and phylogeny. This article will be useful for biotechnologists, molecular biologists, computer scientists and statisticians engaged in computational biology and bioinformatics research.  相似文献   

2.
With an exponential growth in applications identifying protein post‐translational modifications via mass spectrometry, discovery and presentation of motifs surrounding those modification sites have become increasingly desirable. Despite a few tools being designed, there is still a scarcity of effective and polyfunctional software for such analysis and illustrations. In this study, a versatile and user‐friendly web tool is developed, motifeR, for extracting and visualizing statistically significant motifs from large datasets. Particularly, several functions are also integrated for processing multi‐modification sites enrichment. Public datasets are applied to test their usability, indicating that some concurrent modification sites may form motifs and that peptides with low location probability may be not identified randomly and can be included to support motif discovery. In addition, for human phosphoproteomics datasets, the characterization of differential kinase signaling networks can be estimated and modeled by combining kinase‐substrate relations based on the NetworKIN database as an optional feature for users. The motifeR toolkit can be conveniently operated by any scientific community or individuals, even those without any bioinformatics background and is freely available at https://www.omicsolution.org/wukong/motifeR .  相似文献   

3.
With the proliferation of high-throughput technologies, genome-level data analysis has become common in molecular biology. Bioinformaticians are developing extensive resources to annotate and mine biological features from high-throughput data. The underlying database management systems for most bioinformatics software are based on a relational model. Modern non-relational databases offer an alternative that has flexibility, scalability, and a non-rigid design schema. Moreover, with an accelerated development pace, non-relational databases like CouchDB can be ideal tools to construct bioinformatics utilities. We describe CouchDB by presenting three new bioinformatics resources: (a) geneSmash, which collates data from bioinformatics resources and provides automated gene-centric annotations, (b) drugBase, a database of drug-target interactions with a web interface powered by geneSmash, and (c) HapMap-CN, which provides a web interface to query copy number variations from three SNP-chip HapMap datasets. In addition to the web sites, all three systems can be accessed programmatically via web services.  相似文献   

4.
ABSTRACT

Introduction: Discovery proteomics for cancer research generates complex datasets of diagnostic, prognostic, and therapeutic significance in human cancer. With the advent of high-resolution mass spectrometers, able to identify thousands of proteins in complex biological samples, only the application of bioinformatics can lead to the interpretation of data which can be relevant for cancer research.

Areas covered: Here, we give an overview of the current bioinformatic tools used in cancer proteomics. Moreover, we describe their applications in cancer proteomics studies of cell lines, serum, and tissues, highlighting recent results and critically evaluating their outcomes.

Expert opinion: The use of bioinformatic tools is a fundamental step in order to manage the large amount of proteins (from hundreds to thousands) that can be identified and quantified in a cancer biological samples by proteomics. To handle this challenge and obtain useful data for translational medicine, it is important the combined use of different bioinformatic tools. Moreover, a particular attention to the global experimental design, and the integration of multidisciplinary skills are essential for best setting of tool parameters and best interpretation of bioinformatics output.  相似文献   

5.
Evolution of web services in bioinformatics   总被引:4,自引:0,他引:4  
Bioinformaticians have developed large collections of tools to make sense of the rapidly growing pool of molecular biological data. Biological systems tend to be complex and in order to understand them, it is often necessary to link many data sets and use more than one tool. Therefore, bioinformaticians have experimented with several strategies to try to integrate data sets and tools. Owing to the lack of standards for data sets and the interfaces of the tools this is not a trivial task. Over the past few years building services with web-based interfaces has become a popular way of sharing the data and tools that have resulted from many bioinformatics projects. This paper discusses the interoperability problem and how web services are being used to try to solve it, resulting in the evolution of tools with web interfaces from HTML/web form-based tools not suited for automatic workflow generation to a dynamic network of XML-based web services that can easily be used to create pipelines.  相似文献   

6.
7.
Bioinformatics is a central discipline in modern life sciences aimed at describing the complex properties of living organisms starting from large-scale data sets of cellular constituents such as genes and proteins. In order for this wealth of information to provide useful biological knowledge, databases and software tools for data collection, analysis and interpretation need to be developed. In this paper, we review recent advances in the design and implementation of bioinformatics resources devoted to the study of metals in biological systems, a research field traditionally at the heart of bioinorganic chemistry. We show how metalloproteomes can be extracted from genome sequences, how structural properties can be related to function, how databases can be implemented, and how hints on interactions can be obtained from bioinformatics.  相似文献   

8.
9.
Flow cytometry (FCM) is an analytical tool widely used for cancer and HIV/AIDS research, and treatment, stem cell manipulation and detecting microorganisms in environmental samples. Current data standards do not capture the full scope of FCM experiments and there is a demand for software tools that can assist in the exploration and analysis of large FCM datasets. We are implementing a standardized approach to capturing, analyzing, and disseminating FCM data that will facilitate both more complex analyses and analysis of datasets that could not previously be efficiently studied. Initial work has focused on developing a community-based guideline for recording and reporting the details of FCM experiments. Open source software tools that implement this standard are being created, with an emphasis on facilitating reproducible and extensible data analyses. As well, tools for electronic collaboration will assist the integrated access and comprehension of experiments to empower users to collaborate on FCM analyses. This coordinated, joint development of bioinformatics standards and software tools for FCM data analysis has the potential to greatly facilitate both basic and clinical research--impacting a notably diverse range of medical and environmental research areas.  相似文献   

10.
Quantitative trait locus (QTL) analysis is a powerful method for localizing disease genes, but identifying the causal gene remains difficult. Rodent models of disease facilitate QTL gene identification, and causal genes underlying rodent QTL are often associated with the corresponding human diseases. Recently developed bioinformatics methods, including comparative genomics, combined cross analysis, interval-specific and genome-wide haplotype analysis, followed by sequence and expression analysis, each facilitated by public databases, provide new tools for narrowing rodent QTLs. Here we discuss each tool, illustrate its application and generate a bioinformatics strategy for narrowing QTLs. Combining these bioinformatics tools with classical experimental methods should accelerate QTL gene identification.  相似文献   

11.
Modern biological and chemical studies rely on life science databases as well as sophisticated software tools (e.g., homology search tools, modeling and visualization tools). These tools often have to be combined and integrated in order to support a given study. SIBIOS (System for the Integration of Bioinformatics Services) serves this purpose. The services are both life science database search services and software tools. The task engine is the core component of SIBIOS. It supports the execution of dynamic workflows that incorporate multiple bioinformatics services. The architecture of SIBIOS, the approaches to addressing the heterogeneity as well as interoperability of bioinformatics services, including data integration are presented in this paper.  相似文献   

12.
Toward understanding the origin and evolution of cellular organisms   总被引:1,自引:0,他引:1  
In this era of high‐throughput biology, bioinformatics has become a major discipline for making sense out of large‐scale datasets. Bioinformatics is usually considered as a practical field developing databases and software tools for supporting other fields, rather than a fundamental scientific discipline for uncovering principles of biology. The KEGG resource that we have been developing is a reference knowledge base for biological interpretation of genome sequences and other high‐throughput data. It is now one of the most utilized biological databases because of its practical values. For me personally, KEGG is a step toward understanding the origin and evolution of cellular organisms.  相似文献   

13.
园艺植物分子育种相关生物信息资源及其应用   总被引:5,自引:0,他引:5  
园艺植物分子育种中,生物信息技术是一项新技术.GenBank、EMBL、DDBJ、Swiss-Prot等数据库及其序列查询系统、序列比对软件和序列提交软件是园艺植物分子育种中的重要生物信息资源.本文综述了这些生物信息资源,以及它们在克隆新基因、预测新序列功能、鉴定种质资源和进行系谱分析等方面的应用.  相似文献   

14.
园艺植物分子育种中, 生物信息技术是一项新技术。GenBank、EMBL、DDBJ、Swiss-Prot等数据库及其序列查询系统、序列比对软件和序列提交软件是园艺植物分子育种中的重要生物信息资源。本文综述了这些生物信息资源, 以及它们在克隆新基因、预测新序列功能、鉴定种质资源和进行系谱分析等方面的应用。  相似文献   

15.
During the last years gene interaction networks are increasingly being used for the assessment and interpretation of biological measurements. Knowledge of the interaction partners of an unknown protein allows scientists to understand the complex relationships between genetic products, helps to reveal unknown biological functions and pathways, and get a more detailed picture of an organism''s complexity. Being able to measure all protein interactions under all relevant conditions is virtually impossible. Hence, computational methods integrating different datasets for predicting gene interactions are needed. However, when integrating different sources one has to account for the fact that some parts of the information may be redundant, which may lead to an overestimation of the true likelihood of an interaction. Our method integrates information derived from three different databases (Bioverse, HiMAP and STRING) for predicting human gene interactions. A Bayesian approach was implemented in order to integrate the different data sources on a common quantitative scale. An important assumption of the Bayesian integration is independence of the input data (features). Our study shows that the conditional dependency cannot be ignored when combining gene interaction databases that rely on partially overlapping input data. In addition, we show how the correlation structure between the databases can be detected and we propose a linear model to correct for this bias. Benchmarking the results against two independent reference data sets shows that the integrated model outperforms the individual datasets. Our method provides an intuitive strategy for weighting the different features while accounting for their conditional dependencies.  相似文献   

16.
Patel RK  Jain M 《PloS one》2012,7(2):e30619
Next generation sequencing (NGS) technologies provide a high-throughput means to generate large amount of sequence data. However, quality control (QC) of sequence data generated from these technologies is extremely important for meaningful downstream analysis. Further, highly efficient and fast processing tools are required to handle the large volume of datasets. Here, we have developed an application, NGS QC Toolkit, for quality check and filtering of high-quality data. This toolkit is a standalone and open source application freely available at http://www.nipgr.res.in/ngsqctoolkit.html. All the tools in the application have been implemented in Perl programming language. The toolkit is comprised of user-friendly tools for QC of sequencing data generated using Roche 454 and Illumina platforms, and additional tools to aid QC (sequence format converter and trimming tools) and analysis (statistics tools). A variety of options have been provided to facilitate the QC at user-defined parameters. The toolkit is expected to be very useful for the QC of NGS data to facilitate better downstream analysis.  相似文献   

17.
A huge amount of important biomedical information is hidden in the bulk of research articles in biomedical fields. At the same time, the publication of databases of biological information and of experimental datasets generated by high-throughput methods is in great expansion, and a wealth of annotated gene databases, chemical, genomic (including microarray datasets), clinical and other types of data repositories are now available on the Web. Thus a current challenge of bioinformatics is to develop targeted methods and tools that integrate scientific literature, biological databases and experimental data for reducing the time of database curation and for accessing evidence, either in the literature or in the datasets, useful for the analysis at hand. Under this scenario, this article reviews the knowledge discovery systems that fuse information from the literature, gathered by text mining, with microarray data for enriching the lists of down and upregulated genes with elements for biological understanding and for generating and validating new biological hypothesis. Finally, an easy to use and freely accessible tool, GeneWizard, that exploits text mining and microarray data fusion for supporting researchers in discovering gene-disease relationships is described.  相似文献   

18.
19.
The requirements for bioinformatics resources to support genome research in farm animals is reviewed.The resources developed to meet these needs are described. Resource databases and associated tools have been developed to handle experimental data. Several of these systems serve the needs of multinational collaborations. Genome databases have been established to provide contemporary summaries of the status of genome maps in a range of farm and domestic animals along with experimental details and citations. New resources and tools will be required to address the informatics needs of emerging technologies such as microarrays. However, continued investment is also required to maintain the currency and utility of the current systems, especially the genome databases.  相似文献   

20.
Current trends in the development of methods for monitoring, modeling and controlling biological production systems are reviewed from a bioengineering perspective. The ability to measure intracellular conditions in bioprocesses using genomics and other bioinformatics tools is addressed. Devices provided via micromachining techniques and new real-time optical technology are other novel methods that may facilitate biosystem engineering. Mathematical modeling of data obtained from bioinformatics or real-time monitoring methods are necessary in order to handle the dense flows of data that are generated. Furthermore, control methods must be able to cope with these data flows in efficient ways that can be implemented in plant-wide computer communication systems.Mini-review for the proceedings of the M3C conference  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号