共查询到20条相似文献,搜索用时 15 毫秒
1.
Kümmel A Selzer P Beibel M Gubler H Parker CN Gabriel D 《Journal of biomolecular screening》2011,16(3):338-347
High-content screening (HCS) is increasingly used in biomedical research generating multivariate, single-cell data sets. Before scoring a treatment, the complex data sets are processed (e.g., normalized, reduced to a lower dimensionality) to help extract valuable information. However, there has been no published comparison of the performance of these methods. This study comparatively evaluates unbiased approaches to reduce dimensionality as well as to summarize cell populations. To evaluate these different data-processing strategies, the prediction accuracies and the Z' factors of control compounds of a HCS cell cycle data set were monitored. As expected, dimension reduction led to a lower degree of discrimination between control samples. A high degree of classification accuracy was achieved when the cell population was summarized on well level using percentile values. As a conclusion, the generic data analysis pipeline described here enables a systematic review of alternative strategies to analyze multiparametric results from biological systems. 相似文献
2.
Andrea C Pfeifer Daniel Kaschek Julie Bachmann Ursula Klingmüller Jens Timmer 《BMC systems biology》2010,4(1):106
Background
High-quality quantitative data is a major limitation in systems biology. The experimental data used in systems biology can be assigned to one of the following categories: assays yielding average data of a cell population, high-content single cell measurements and high-throughput techniques generating single cell data for large cell populations. For modeling purposes, a combination of data from different categories is highly desirable in order to increase the number of observable species and processes and thereby maximize the identifiability of parameters. 相似文献3.
Jansen RC 《Nature reviews. Genetics》2003,4(2):145-151
4.
5.
Recent years have seen a sharp increase in the development of deep learning and artificial intelligence-based molecular informatics. There has been a growing interest in applying deep learning to several subfields, including the digital transformation of synthetic chemistry, extraction of chemical information from the scientific literature, and AI in natural product-based drug discovery. The application of AI to molecular informatics is still constrained by the fact that most of the data used for training and testing deep learning models are not available as FAIR and open data. As open science practices continue to grow in popularity, initiatives which support FAIR and open data as well as open-source software have emerged. It is becoming increasingly important for researchers in the field of molecular informatics to embrace open science and to submit data and software in open repositories. With the advent of open-source deep learning frameworks and cloud computing platforms, academic researchers are now able to deploy and test their own deep learning models with ease. With the development of new and faster hardware for deep learning and the increasing number of initiatives towards digital research data management infrastructures, as well as a culture promoting open data, open source, and open science, AI-driven molecular informatics will continue to grow. This review examines the current state of open data and open algorithms in molecular informatics, as well as ways in which they could be improved in future. 相似文献
6.
Bioinformatics tools for proteomics, also called proteome informatics tools, span today a large panel of very diverse applications ranging from simple tools to compare protein amino acid compositions to sophisticated software for large-scale protein structure determination. This review considers the available and ready to use tools that can help end-users to interpret, validate and generate biological information from their experimental data. It concentrates on bioinformatics tools for 2-DE analysis, for LC followed by MS analysis, for protein identification by PMF, by peptide fragment fingerprinting and by de novo sequencing and for data quantitation with MS data. It also discloses initiatives that propose to automate the processes of MS analysis and enhance the quality of the obtained results. 相似文献
7.
Background
High content screening (HCS) is a powerful method for the exploration of cellular signalling and morphology that is rapidly being adopted in cancer research. HCS uses automated microscopy to collect images of cultured cells. The images are subjected to segmentation algorithms to identify cellular structures and quantitate their morphology, for hundreds to millions of individual cells. However, image analysis may be imperfect, especially for "HCS-unfriendly" cell lines whose morphology is not well handled by current image segmentation algorithms. We asked if segmentation errors were common for a clinically relevant cell line, if such errors had measurable effects on the data, and if HCS data could be improved by automated identification of well-segmented cells. 相似文献8.
High-content screening studies of mitotic checkpoints are important for identifying cancer targets and developing novel cancer-specific therapies. A crucial step in such a study is to determine the stage of cell cycle. Due to the overwhelming number of cells assayed in a high-content screening experiment and the multiple factors that need to be taken into consideration for accurate determination of mitotic subphases, an automated classifier is necessary. In this article, the authors describe in detail a support vector machine (SVM) classifier that they have implemented to recognize various mitotic subphases. In contrast to previous studies to recognize subcellular patterns, they used only low-resolution cell images and a few parameters that can be calculated inexpensively with off-the-shelf image-processing software. The performance of the SVM was evaluated with a cross-validation method and was shown to be comparable to that of a human expert. 相似文献
9.
10.
Nielsen DA Leidner M Haynes C Krauthammer M Kreek MJ 《Molecular diagnosis & therapy》2007,11(1):15-19
The Biology of Addictive Diseases-Database (BiolAD-DB) system is a research bioinformatics system for archiving, analyzing, and processing of complex clinical and genetic data. The database schema employs design principles for handling complex clinical information, such as response items in genetic questionnaires. Data access and validation is provided by the BiolAD-DB client application, which features a data validation engine tightly coupled to a graphical user interface. Data integrity is provided by the password-protected BiolAD-DB SQL compliant server and database. BiolAD-DB tools further provide functionalities for generating customized reports and views. The BiolAD-DB system schema, client, and installation instructions are freely available at http://www.rockefeller.edu/biolad-db/. 相似文献
11.
Soberón J Peterson AT 《Philosophical transactions of the Royal Society of London. Series B, Biological sciences》2004,359(1444):689-698
Recently, advances in information technology and an increased willingness to share primary biodiversity data are enabling unprecedented access to it. By combining presences of species data with electronic cartography via a number of algorithms, estimating niches of species and their areas of distribution becomes feasible at resolutions one to three orders of magnitude higher than it was possible a few years ago. Some examples of the power of that technique are presented. For the method to work, limitations such as lack of high-quality taxonomic determination, precise georeferencing of the data and availability of high-quality and updated taxonomic treatments of the groups must be overcome. These are discussed, together with comments on the potential of these biodiversity informatics techniques not only for fundamental studies but also as a way for developing countries to apply state of the art bioinformatic methods and large quantities of data, in practical ways, to tackle issues of biodiversity management. 相似文献
12.
Saez-Rodriguez J Goldsipe A Muhlich J Alexopoulos LG Millard B Lauffenburger DA Sorger PK 《Bioinformatics (Oxford, England)》2008,24(6):840-847
MOTIVATION: Linking experimental data to mathematical models in biology is impeded by the lack of suitable software to manage and transform data. Model calibration would be facilitated and models would increase in value were it possible to preserve links to training data along with a record of all normalization, scaling, and fusion routines used to assemble the training data from primary results. RESULTS: We describe the implementation of DataRail, an open source MATLAB-based toolbox that stores experimental data in flexible multi-dimensional arrays, transforms arrays so as to maximize information content, and then constructs models using internal or external tools. Data integrity is maintained via a containment hierarchy for arrays, imposition of a metadata standard based on a newly proposed MIDAS format, assignment of semantically typed universal identifiers, and implementation of a procedure for storing the history of all transformations with the array. We illustrate the utility of DataRail by processing a newly collected set of approximately 22 000 measurements of protein activities obtained from cytokine-stimulated primary and transformed human liver cells. AVAILABILITY: DataRail is distributed under the GNU General Public License and available at http://code.google.com/p/sbpipeline/ 相似文献
13.
Wong L 《Briefings in bioinformatics》2002,3(4):389-404
The process of building a new database relevant to some field of study in biomedicine involves transforming, integrating and cleansing multiple data sources, as well as adding new material and annotations. This paper reviews some of the requirements of a general solution to this data integration problem. Several representative technologies and approaches to data integration in biomedicine are surveyed. Then some interesting features that separate the more general data integration technologies from the more specialised ones are highlighted. 相似文献
14.
15.
16.
RNA interference (RNAi) is a recent advance that provides the possibility to reduce the expression of specific target genes in cultured mammalian cells with potential applications on a genome-wide scale. However, to achieve this, robust methodologies that allow automated and efficient delivery of small interfering RNAs (siRNAs) into living cultured cells and reliable quality control of siRNA function must be in place. Here we describe the production of cell arrays for reverse transfection of tissue culture cells with siRNA and plasmid DNA suitable for subsequent high-content screening microscopy applications. All the necessary transfection components are mixed prior to the robotic spotting on noncoated chambered coverglass tissue culture dishes, which are ideally suited for time-lapse microscopy applications in living cells. The addition of fibronectin to the spotting solution improves cell adherence. After cell seeding, no further cell culture manipulations, such as medium changes or the addition of 7 serum, are needed. Adaptation of the cell density improves autofocus performance for high-quality data acquisition and cell recognition. The co-transfection of a nonspecific fluorescently labeled DNA oligomer with the specific siRNA helps to mark each successfully transfected cell and cell cluster. We demonstrate such an siRNA cell array in a microscope-based functional assay in living cells to determine the effect of various siRNA oligonucleotides against endogenous targets on cellular secretion. 相似文献
17.
Sonnhammer EL 《Genome biology》2005,6(1):301
A report on the fourth Cold Spring Harbor Laboratory/Wellcome Trust Conference on Genome Informatics, Hinxton, UK, 22-26 September 2004. 相似文献
18.
19.
《International journal of bio-medical computing》1986,18(1):25-28
A simple microcomputer program written in Microsoft Basic estimates pharmacokinetic parameters using the coordinate search technique to minimize the sum of squared errors. The program developed for portable computers combines a plot of data and curve fitting so as to find rapidly the initial parameters with the subsequent optimization of the parameter estimate. 相似文献
20.
Allan R Brasier 《BioTechniques》2002,32(1):100-2, 104, 106, 108-9
High-density oligonucleotide arrays are widely employed for detecting global changes in gene expression profiles of cells or tissues exposed to specific stimuli. Presented with large amounts of data, investigators can spend significant amounts of time analyzing and interpreting this array data. In our application of GeneChip arrays to analyze changes in gene expression in viral-infected epithelium, we have needed to develop additional computational tools that may be of utility to other investigators using this methodology. Here, I describe two executable programs to facilitate data extraction and multiple data point analysis. These programs run in a virtual DOS environment on Microsoft Windows 95/98/2K operating systems on a desktop PC. Both programs can be freely downloaded from the BioTechniques Software Library (www.BioTechniques.com). The first program, Retriever, extracts primary data from an array experiment contained in an Affymetrix textfile using user-inputted individual identification strings (e.g., the probe set identification numbers). With specific data retrieved for individual genes, hybridization profiles can be examined and data normalized. The second program, CompareTable, is used to facilitate comparison analysis of two experimental replicates. CompareTable compares two lists of genes, identifies common entries, extracts their data, and writes an output text file containing only those genes present in both of the experiments. The output files generated by these two programs can be opened and manipulated by any software application recognizing tab-delimited text files (e.g., Microsoft NotePad or Excel). 相似文献