首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
Biological imaging software tools   总被引:1,自引:0,他引:1  
Few technologies are more widespread in modern biological laboratories than imaging. Recent advances in optical technologies and instrumentation are providing hitherto unimagined capabilities. Almost all these advances have required the development of software to enable the acquisition, management, analysis and visualization of the imaging data. We review each computational step that biologists encounter when dealing with digital images, the inherent challenges and the overall status of available software for bioimage informatics, focusing on open-source options.  相似文献   

2.
Sarkar IN  Trizna M 《PloS one》2011,6(7):e14689
With the volume of molecular sequence data that is systematically being generated globally, there is a need for centralized resources for data exploration and analytics. DNA Barcode initiatives are on track to generate a compendium of molecular sequence-based signatures for identifying animals and plants. To date, the range of available data exploration and analytic tools to explore these data have only been available in a boutique form--often representing a frustrating hurdle for many researchers that may not necessarily have resources to install or implement algorithms described by the analytic community. The Barcode of Life Data Portal (BDP) is a first step towards integrating the latest biodiversity informatics innovations with molecular sequence data from DNA barcoding. Through establishment of community driven standards, based on discussion with the Data Analysis Working Group (DAWG) of the Consortium for the Barcode of Life (CBOL), the BDP provides an infrastructure for incorporation of existing and next-generation DNA barcode analytic applications in an open forum.  相似文献   

3.
The need for open, reproducible science is of growing concern in the twenty-first century, with multiple initiatives like the widely supported FAIR principles advocating for data to be Findable, Accessible, Interoperable and Reusable. Plant ecological and evolutionary studies are not exempt from the need to ensure that the data upon which their findings are based are accessible and allow for replication in accordance with the FAIR principles. However, it is common that the collection and curation of herbarium specimens, a foundational aspect of studies involving plants, is neglected by authors. Without publicly available specimens, huge numbers of studies that rely on the field identification of plants are fundamentally not reproducible. We argue that the collection and public availability of herbarium specimens is not only good botanical practice but is also fundamental in ensuring that plant ecological and evolutionary studies are replicable, and thus scientifically sound. Data repositories that adhere to the FAIR principles must make sure that the original data are traceable to and re-examinable at their empirical source. In order to secure replicability, and adherence to the FAIR principles, substantial changes need to be brought about to restore the practice of collecting and curating specimens, to educate students of their importance, and to properly fund the herbaria which house them.  相似文献   

4.
ContinuousFlex is a user-friendly open-source software package for analyzing continuous conformational variability of macromolecules in cryo electron microscopy (cryo-EM) and cryo electron tomography (cryo-ET) data. In 2019, ContinuousFlex became available as a plugin for Scipion, an image processing software package extensively used in the cryo-EM field. Currently, ContinuousFlex contains software for running (1) recently published methods HEMNMA-3D, TomoFlow, and NMMD; (2) earlier published methods HEMNMA and StructMap; and (3) methods for simulating cryo-EM and cryo-ET data with conformational variability and methods for data preprocessing. It also includes external software for molecular dynamics simulation (GENESIS) and normal mode analysis (ElNemo), used in some of the mentioned methods. The HEMNMA software has been presented in the past, but not the software of other methods. Besides, ContinuousFlex currently also offers a deep learning extension of HEMNMA, named DeepHEMNMA. In this article, we review these methods in the context of the ContinuousFlex package, developed to facilitate their use by the community.  相似文献   

5.
Abstract

The use of modern data science has recently emerged as a promising new path to tackling the complex challenges involved in the creation of next-generation chemistry and materials. However, despite the appeal of this potentially transformative development, the chemistry community has yet to incorporate it as a central tool in every-day work. Our research program is designed to enable and advance this emerging research approach. It is centred around the creation of a software ecosystem that brings together physics-based modelling, high-throughput in silico screening and data analytics (i.e. the use of machine learning and informatics for the validation, mining and modelling of chemical data). This cyberinfrastructure is devised to offer a comprehensive set of data science techniques and tools as well as a general-purpose scope to make it as versatile and widely applicable as possible. It also emphasises user-friendliness to make it accessible to the community at large. It thus provides the means for the large-scale exploration of chemical space and for a better understanding of the hidden mechanisms that determine the properties of complex chemical systems. Such insights can dramatically accelerate, streamline and ultimately transform the way chemical research is conducted. Aside from serving as a production-level tool, our cyberinfrastructure is also designed to facilitate and assess methodological innovation. Both the software and method development work are driven by concrete molecular design problems, which also allow us to assess the efficacy of the overall cyberinfrastructure.  相似文献   

6.
With the continuous development of medical image informatics technology, more and more high-throughput quantitative data could be extracted from digital medical images, which has resulted in a new kind of omics-Radiomics. In recent years, in addition to genomics, proteomics and metabolomics, radiomic has attracted the interest of more and more researchers. Compared to other omics, radiomics can be perfectly integrated with clinical data, even with the pathology and molecular biomarker, so that the study can be closer to the clinical reality and more revealing of the tumor development. Mass data will also be generated in this process. Machine learning, due to its own characteristics, has a unique advantage in processing massive radiomic data. By analyzing mass amounts of data with strong clinical relevance, people can construct models that more accurately reflect tumor development and progression, thereby providing the possibility of personalized and sequential treatment of patients. As one of the cancer types whose treatment and diagnosis rely on imaging examination, radiomics has a very broad application prospect in head and neck cancers (HNC). Until now, there have been some notable results in HNC. In this review, we will introduce the concepts and workflow of radiomics and machine learning and their current applications in head and neck cancers, as well as the directions and applications of artificial intelligence in the treatment and diagnosis of HNC.  相似文献   

7.
The age of big data is poised to revolutionize vegetation science. As online resources continue to grow, vegetation ecologists will need a growing set of computational skills to advance vegetation science in the digital age. Two papers in this issue of the Journal of Vegetation Science (Wiser 2016, Sandel et al. 2016) illustrate the resources available and use of big data to explore challenging ecological questions.  相似文献   

8.
《Biophysical journal》2020,118(9):2077-2085
Genomics is a sequence-based informatics science and a three-dimensional-structure-based material science. However, in practice, most genomics researchers utilize sequence-based informatics approaches or three-dimensional-structure-based material science techniques, not both. This division is, at least in part, the result of historical developments rather than a fundamental necessity. The underlying computational tools, experimental techniques, and theoretical models were developed independently. The primary result presented here is a framework for the unification of informatics- and physics-based data associated with DNA, nucleosomes, and chromatin. The framework is based on the mathematical representation of geometrically exact rods and the generalization of DNA basepair step parameters. Data unification enables researchers to integrate computational, experimental, and theoretical approaches for the study of chromatin biology. The framework can be implemented using model-view-controller design principles, existing genome browsers, and existing molecular visualization tools. We developed a minimal, web-based genome dashboard, G-Dash-min, and applied it to two simple examples to demonstrate the usefulness of data unification and proof of concept. Genome dashboards developed using the framework and design principles presented here are extensible and customizable and are therefore more broadly applicable than the examples presented. We expect a number of purpose-specific genome dashboards to emerge as a novel means of investigating structure-function relationships for genomes that range from basepairs to entire chromosomes and for generating, validating, and testing mechanistic hypotheses.  相似文献   

9.
Given the rapid growth of artificial intelligence (AI) applications in radiotherapy and the related transformations toward the data-driven healthcare domain, this article summarizes the need and usage of the FAIR (Findable, Accessible, Interoperable, Reusable) data principles in radiotherapy. This work introduces the FAIR data concept, presents practical and relevant use cases and the future role of the different parties involved. The goal of this article is to provide guidance and potential applications of FAIR to various radiotherapy stakeholders, focusing on the central role of medical physicists.  相似文献   

10.
Spectral searching has drawn increasing interest as an alternative to sequence-database searching in proteomics. We developed and validated an open-source software toolkit, SpectraST, to enable proteomics researchers to build spectral libraries and to integrate this promising approach in their data-analysis pipeline. It allows individual researchers to condense raw data into spectral libraries, summarizing information about observed proteomes into a concise and retrievable format for future data analyses.  相似文献   

11.
PDBx/mmCIF, Protein Data Bank Exchange (PDBx) macromolecular Crystallographic Information Framework (mmCIF), has become the data standard for structural biology. With its early roots in the domain of small-molecule crystallography, PDBx/mmCIF provides an extensible data representation that is used for deposition, archiving, remediation, and public dissemination of experimentally determined three-dimensional (3D) structures of biological macromolecules by the Worldwide Protein Data Bank (wwPDB, wwpdb.org). Extensions of PDBx/mmCIF are similarly used for computed structure models by ModelArchive (modelarchive.org), integrative/hybrid structures by PDB-Dev (pdb-dev.wwpdb.org), small angle scattering data by Small Angle Scattering Biological Data Bank SASBDB (sasbdb.org), and for models computed generated with the AlphaFold 2.0 deep learning software suite (alphafold.ebi.ac.uk). Community-driven development of PDBx/mmCIF spans three decades, involving contributions from researchers, software and methods developers in structural sciences, data repository providers, scientific publishers, and professional societies. Having a semantically rich and extensible data framework for representing a wide range of structural biology experimental and computational results, combined with expertly curated 3D biostructure data sets in public repositories, accelerates the pace of scientific discovery. Herein, we describe the architecture of the PDBx/mmCIF data standard, tools used to maintain representations of the data standard, governance, and processes by which data content standards are extended, plus community tools/software libraries available for processing and checking the integrity of PDBx/mmCIF data. Use cases exemplify how the members of the Worldwide Protein Data Bank have used PDBx/mmCIF as the foundation for its pipeline for delivering Findable, Accessible, Interoperable, and Reusable (FAIR) data to many millions of users worldwide.  相似文献   

12.
Understanding patterns of human evolution across space and time requires synthesizing data collected by independent research teams, and this effort is part of a larger trend to develop cyber infrastructure and e‐science initiatives. 1 At present, paleoanthropology cannot easily answer basic questions about the total number of fossils and artifacts that have been discovered, or exactly how those items were collected. In this paper, we examine the methodological challenges to data integration, with the hope that mitigating the technical obstacles will further promote data sharing. At a minimum, data integration efforts must document what data exist and how the data were collected (discovery), after which we can begin standardizing data collection practices with the aim of achieving combined analyses (synthesis). This paper outlines a digital data collection system for paleoanthropology. We review the relevant data management principles for a general audience and supplement this with technical details drawn from over 15 years of paleontological and archeological field experience in Africa and Europe. The system outlined here emphasizes free open‐source software (FOSS) solutions that work on multiple computer platforms; it builds on recent advances in open‐source geospatial software and mobile computing.  相似文献   

13.
生物多样性信息学:一个正在兴起的新方向及其关键技术   总被引:5,自引:0,他引:5  
生物多样性科学和生物信息学是生命科学中两个极为重要也是十分活跃的交叉学科,生物多样性信息学则是目前正在兴起的一个新方向,基发展必将进一步深化信息技术在生物多样性研究中的应用。本文简要介绍了国内外该领域的主要目标与进展,讨论了有关的关键技术(如数据库间的互操作与数字图书馆),并列出了两个原型系统(Species2000和GBIF)和其他相关系统的网址。  相似文献   

14.
Dental Informatics (DI) is the application of computer and information science to improve dental practice, research, education, and program administration. To support the growth of this emerging discipline, we created the Dental Informatics Online Community (DIOC). The DIOC provides a dedicated professional home for DI researchers and serves as an open, common, and worldwide forum for all individuals interested in the field. It was created and is maintained by the Center for Dental Informatics at the University of Pittsburgh School of Dental Medicine, independent from any professional association, corporate interest or funding source. The DIOC's Website provides many useful features, such as a learning center, publication archive, member and project directories, and the Current Awareness Service (CAS). The CAS automatically notifies members about new information added to the Community. Notifications are individualized according to a member's profile and activities on the site. The DIOC is a research-oriented online community which provides resources in the dental informatics and dental technology field, as well as a way to establish social connections to share ideas, problems and research opportunities. Member and activity growth since the community's inception in 2005 have been steady, but future sustainability of the community depends on many factors.  相似文献   

15.
陈建平  许哲平 《广西植物》2022,42(Z1):52-61
标本数字化建设是生物多样性保护和利用的重要工作基础,通过标本数据的整合分析,在生物分类学、生态学、生物工程、生物保护、粮食安全、生物多样性评估、教学教育和人类社会活动等方面提供数据支撑。为了了解全球标本数字化建设工作的现状以及数据共享的策略与技术发展趋势,该文分别调查梳理了北美洲、南美洲、欧洲、非洲、亚洲和大洋洲地区的标本数字化和平台建设情况,对标本数据共享现状和趋势从数据使用协议、新技术新方法和公众科学等方面进行了对比和分析,并为中国国内的标本数字化工作提出了工作建议,包括:(1)加强标本数字化建设、管理和动态更新方面的协同机制建设,确保实物资源和数字化资源信息同步;(2)加强数据整理和发布,促进数据质量的提升,充分开放数据使用协议,减少数据使用的阻碍;(3)加强对新技术的学习和引入,特别是开源软件、机器学习和人工智能技术的应用,能够在标签快速识别、自动鉴定和属性数据提取等方面发挥作用;(4)加强区域和国际合作,推动数据的整合应用;(5)推动公众科学项目发展,促进野外采集、室内整理、在线纠错、数据产品研发等工作的开展。  相似文献   

16.
Biological measurements frequently involve measuring parameters as a function of time, space, or frequency. Later, during the analysis phase of the study, the researcher splits the recorded data trace into smaller sections, analyzes each section separately by finding a mean or fitting against a specified function, and uses the analysis results in the study. Here, we present the software that allows to analyze these data traces in a manner that ensures repeatability of the analysis and simplifies the application of FAIR (findability, accessibility, interoperability, and reusability) principles in such studies. At the same time, it simplifies the routine data analysis pipeline and gives access to a fast overview of the analysis results. For that, the software supports reading the raw data, processing the data as specified in the protocol, and storing all intermediate results in the laboratory database. The software can be extended by study- or hardware-specific modules to provide the required data import and analysis facilities. To simplify the development of the data entry web interfaces, that can be used to enter data describing the experiments, we released a web framework with an example implementation of such a site. The software is covered by open-source license and is available through several online channels.  相似文献   

17.
18.
《Genomics》2023,115(2):110584
Cardiovascular disease (CVD) is the leading cause of mortality and loss of disability adjusted life years (DALYs) globally. CVDs like Heart Failure (HF) and Atrial Fibrillation (AF) are associated with physical effects on the heart muscles. As a result of the complex nature, progression, inherent genetic makeup, and heterogeneity of CVDs, personalized treatments are believed to be critical. Rightful application of artificial intelligence (AI) and machine learning (ML) approaches can lead to new insights into CVDs for providing better personalized treatments with predictive analysis and deep phenotyping. In this study we focused on implementing AI/ML techniques on RNA-seq driven gene-expression data to investigate genes associated with HF, AF, and other CVDs, and predict disease with high accuracy. The study involved generating RNA-seq data derived from the serum of consented CVD patients. Next, we processed the sequenced data using our RNA-seq pipeline and applied GVViZ for gene-disease data annotation and expression analysis. To achieve our research objectives, we developed a new Findable, Accessible, Intelligent, and Reproducible (FAIR) approach that includes a five-level biostatistical evaluation, primarily based on the Random Forest (RF) algorithm. During our AI/ML analysis, we have fitted, trained, and implemented our model to classify and distinguish high-risk CVD patients based on their age, gender, and race. With the successful execution of our model, we predicted the association of highly significant HF, AF, and other CVDs genes with demographic variables.  相似文献   

19.

Background

Systems Biology research tools, such as Cytoscape, have greatly extended the reach of genomic research. By providing platforms to integrate data with molecular interaction networks, researchers can more rapidly begin interpretation of large data sets collected for a system of interest. BioNetBuilder is an open-source client-server Cytoscape plugin that automatically integrates molecular interactions from all major public interaction databases and serves them directly to the user's Cytoscape environment. Until recently however, chicken and other eukaryotic model systems had little interaction data available.

Results

Version 2.0 of BioNetBuilder includes a redesigned synonyms resolution engine that enables transfer and integration of interactions across species; this engine translates between alternate gene names as well as between orthologs in multiple species. Additionally, BioNetBuilder is now implemented to be part of the Gaggle, thereby allowing seamless communication of interaction data to any software implementing the widely used Gaggle software. Using BioNetBuilder, we constructed a chicken interactome possessing 72,000 interactions among 8,140 genes directly in the Cytoscape environment. In this paper, we present a tutorial on how to do so and analysis of a specific use case.

Conclusion

BioNetBuilder 2.0 provides numerous user-friendly systems biology tools that were otherwise inaccessible to researchers in chicken genomics, as well as other model systems. We provide a detailed tutorial spanning all required steps in the analysis. BioNetBuilder 2.0, the tools for maintaining its data bases, standard operating procedures for creating local copies of its back-end data bases, as well as all of the Gaggle and Cytoscape codes required, are open-source and freely available at http://err.bio.nyu.edu/cytoscape/bionetbuilder/.
  相似文献   

20.
This review summarizes important work in open-source bioinformatics software that has occurred over the past couple of years. The survey is intended to illustrate how programs and toolkits whose source code has been developed or released under an Open Source license have changed informatics-heavy areas of life science research. Rather than creating a comprehensive list of all tools developed over the last 2-3 years, we use a few selected projects encompassing toolkit libraries, analysis tools, data analysis environments and interoperability standards to show how freely available and modifiable open-source software can serve as the foundation for building important applications, analysis workflows and resources.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号