共查询到20条相似文献,搜索用时 0 毫秒
1.
Casper J Albers Ritsert C Jansen Jan Kok Oscar P Kuipers Sacha AFT van Hijum 《BMC bioinformatics》2006,7(1):205-14
Background
Simulation of DNA-microarray data serves at least three purposes: (i) optimizing the design of an intended DNA microarray experiment, (ii) comparing existing pre-processing and processing methods for best analysis of a given DNA microarray experiment, (iii) educating students, lab-workers and other researchers by making them aware of the many factors influencing DNA microarray experiments. 相似文献2.
SUMMARY: QuickLIMS is a Microsoft Access-based laboratory information and management system, capable of processing all information for microarray production. The program's operational flow is protocol-based, dynamically adapting to changes of the process. It interacts with the laboratory robot and with other database systems over the network, and it represents a complete solution for the management of the entire manufacturing process. AVAILABILITY AND SUPPLEMENTARY INFORMATION: http://www.dkfz-heidelberg.de/kompl_genome/Other/QuickLims/index.html 相似文献
3.
Normalization of cDNA microarray data 总被引:43,自引:0,他引:43
Normalization means to adjust microarray data for effects which arise from variation in the technology rather than from biological differences between the RNA samples or between the printed probes. This paper describes normalization methods based on the fact that dye balance typically varies with spot intensity and with spatial position on the array. Print-tip loess normalization provides a well-tested general purpose normalization method which has given good results on a wide range of arrays. The method may be refined by using quality weights for individual spots. The method is best combined with diagnostic plots of the data which display the spatial and intensity trends. When diagnostic plots show that biases still remain in the data after normalization, further normalization steps such as plate-order normalization or scale-normalization between the arrays may be undertaken. Composite normalization may be used when control spots are available which are known to be not differentially expressed. Variations on loess normalization include global loess normalization and two-dimensional normalization. Detailed commands are given to implement the normalization techniques using freely available software. 相似文献
4.
Sacha?AFT?van Hijum Anne?de Jong Richard?JS?Baerends Harma?A?Karsens Naomi?E?Kramer Rasmus?Larsen Chris?D?den Hengst Casper?J?Albers Jan?Kok Oscar?P?Kuipers
Background
In research laboratories using DNA-microarrays, usually a number of researchers perform experiments, each generating possible sources of error. There is a need for a quick and robust method to assess data quality and sources of errors in DNA-microarray experiments. To this end, a novel and cost-effective validation scheme was devised, implemented, and employed. 相似文献5.
MOTIVATION: Detailed comparison and analysis of the output of DNA gene expression arrays from multiple samples require global normalization of the measured individual gene intensities from the different hybridizations. This is needed for accounting for variations in array preparation and sample hybridization conditions. RESULTS: Here, we present a simple, robust and accurate procedure for the global normalization of datasets generated with single-channel DNA arrays based on principal component analysis. The procedure makes minimal assumptions about the data and performs well in cases where other standard procedures produced biased estimates. It is also insensitive to data transformation, filtering (thresholding) and pre-screening. 相似文献
6.
7.
Normalization and analysis of DNA microarray data by self-consistency and local regression
下载免费PDF全文

Background
With the advent of DNA hybridization microarrays comes the remarkable ability, in principle, to simultaneously monitor the expression levels of thousands of genes. The quantiative comparison of two or more microarrays can reveal, for example, the distinct patterns of gene expression that define different cellular phenotypes or the genes induced in the cellular response to insult or changing environmental conditions. Normalization of the measured intensities is a prerequisite of such comparisons, and indeed, of any statistical analysis, yet insufficient attention has been paid to its systematic study. The most straightforward normalization techniques in use rest on the implicit assumption of linear response between true expression level and output intensity. We find that these assumptions are not generally met, and that these simple methods can be improved. 相似文献8.
Shouyong Peng Artyom A Alekseyenko Erica Larschan Mitzi I Kuroda Peter J Park 《BMC bioinformatics》2007,8(1):219
Background
Chromatin immunoprecipitation on tiling arrays (ChIP-chip) has been widely used to investigate the DNA binding sites for a variety of proteins on a genome-wide scale. However, several issues in the processing and analysis of ChIP-chip data have not been resolved fully, including the effect of background (mock control) subtraction and normalization within and across arrays. 相似文献9.
10.
Analysis of data from the analytical ultracentrifuge by nonlinear least-squares techniques. 总被引:37,自引:1,他引:37
下载免费PDF全文

Least-squares analysis of experimental data from the analytical ultracentrifuge is discussed in detail, with particular attention to the use of interference optics in studying nonideal self-associating macromolecular systems. Several samples are given that describe the application of the technique, the expected precision of the results, and some of its limitations. A FORTRAN IV computer program is available from the authors. 相似文献
11.
Normalization in the fitting of data by iterative methods. Application to tracer kinetics and enzyme kinetics 总被引:1,自引:4,他引:1
J. H. Ottaway 《The Biochemical journal》1973,134(3):729-736
1. The normalization of biochemical data to weight them appropriately for parameter estimation is considered, with reference particularly to data from tracer kinetics and enzyme kinetics. If the data are in replicate, it is recommended that the sum of squared deviations for each experimental variable at each time or concentration point is divided by the local variance at that point. 2. If there is only one observation for each variable at each sampling point, normalization may still be required if the observations cover more than one order of magnitude, but there is no absolute criterion for judging the effect of the weighting that is produced. The goodness of fit that is produced by minimizing the weighted sum of squares of deviations must be judged subjectively. It is suggested that the goodness of fit may be regarded as satisfactory if the data points are distributed uniformly on either side of the fitted curve. A chi-square test may be used to decide whether the distribution is abnormal. The proportion of the residual variance associated with points on one or other side of the fitted curve may also be taken into account, because this gives an indication of the sensitivity of the residual variance to movement of the curve away from particular data points. These criteria for judging the effect of weighting are only valid if the model equation may reasonably be expected to apply to all the data points. 3. On this basis, normalizing by dividing the deviation for each data point by the experimental observation or by the equivalent value calculated by the model equation may both be shown to produce a consistent bias for numerically small observations, the former biasing the curve towards the smallest observations, the latter tending to produce a curve that is above the numerically smaller data points. It was found that dividing each deviation by the mean of observed and calculated variable appropriate to it produces a weighting that is fairly free from bias as judged by the criteria mentioned above. This normalization factor was tested on published data from both tracer kinetics and enzyme kinetics. 相似文献
12.
13.
Alžběta Gardlo Age K. Smilde Karel Hron Marcela Hrdá Radana Karlíková David Friedecký Tomáš Adam 《Metabolomics : Official journal of the Metabolomic Society》2016,12(7):117
Introduction
One of the body fluids often used in metabolomics studies is urine. The concentrations of metabolites in urine are affected by hydration status of an individual, resulting in dilution differences. This requires therefore normalization of the data to correct for such differences. Two normalization techniques are commonly applied to urine samples prior to their further statistical analysis. First, AUC normalization aims to normalize a group of signals with peaks by standardizing the area under the curve (AUC) within a sample to the median, mean or any other proper representation of the amount of dilution. The second approach uses specific end-product metabolites such as creatinine and all intensities within a sample are expressed relative to the creatinine intensity.Objectives
Another way of looking at urine metabolomics data is by realizing that the ratios between peak intensities are the information-carrying features. This opens up possibilities to use another class of data analysis techniques designed to deal with such ratios: compositional data analysis. The aim of this paper is to develop PARAFAC modeling of three-way urine metabolomics data in the context of compositional data analysis and compare this with standard normalization techniques.Methods
In the compositional data analysis approach, special coordinate systems are defined to deal with the ratio problem. In essence, it comes down to using other distance measures than the Euclidian Distance that is used in the conventional analysis of metabolomic data.Results
We illustrate using this type of approach in combination with three-way methods (i.e. PARAFAC) of a longitudinal urine metabolomics study and two simulations. In both cases, the advantage of the compositional approach is established in terms of improved interpretability of the scores and loadings of the PARAFAC model.Conclusion
For urine metabolomics studies, we advocate the use of compositional data analysis approaches. They are easy to use, well established and proof to give reliable results.14.
DNA microarray is a powerful tool for high-throughput analysis of biological systems. Various computational tools have been created to facilitate the analysis of the large volume of data produced in DNA microarray experiments. Normalization is a critical step for obtaining data that are reliable and usable for subsequent analysis such as identification of differentially expressed genes and clustering. A variety of normalization methods have been proposed over the past few years, but no methods are still perfect. Various assumptions are often taken in the process of normalization. Therefore, the knowledge of underlying assumption and principle of normalization would be helpful for the correct analysis of microarray data. We present a review of normalization techniques from single-labeled platforms such as the Affymetrix GeneChip array to dual-labeled platforms like spotted array focusing on their principles and assumptions. 相似文献
15.
Kernel and nonlinear canonical correlation analysis 总被引:8,自引:0,他引:8
We review a neural implementation of the statistical technique of Canonical Correlation Analysis (CCA) and extend it to nonlinear CCA. We then derive the method of kernel-based CCA and compare these two methods on real and artificial data sets before using both on the Blind Separation of Sources. 相似文献
16.
In the preceding companion article in this issue, an optical dye and a nitroxide radical were combined in a new dual function probe, 5-SLE. In this report, it is demonstrated that time-resolved optical anisotropy and electron paramagnetic resonance (EPR) data can be combined in a single analysis to measure rotational dynamics. Rigid-limit and rotational diffusion models for simulating nitroxide EPR data have been incorporated into a general non-linear least-squares procedure based on the Marquardt-Levenberg algorithm. Simultaneous fits to simulated time-resolved fluorescence anisotropy and linear EPR data, together with simultaneous fits to experimental time-resolved phosphorescence anisotropy decays and saturation transfer EPR (ST-EPR) spectra of 5-SLE noncovalently bound to bovine serum albumin (BSA) have been performed. These results demonstrate that data from optical and EPR experiments can be combined and globally fit to a single dynamic model. 相似文献
17.
With the increasing amount of DNA sequence data available from natural populations, new computational methods are needed to efficiently process raw sequences into formats that are applicable to a variety of analytical methods. One highly successful approach to inferring aspects of demographic history is grounded in coalescent theory. Many of these methods restrict themselves to perfectly tree-like genealogies (i.e. regions with no observed recombination), because theoretical difficulties prevent ready statistical evaluation of recombining regions. However, determining which recombination-filtered dataset to analyze from a larger recombination-rich genomic region is a non-trivial problem. Current applications primarily aim to quantify recombination rates (rather than produce optimal recombination-filtered blocks), require significant manual intervention, and are impractical for multiple genomic datasets in high-throughput, automated research environments. Here, we present a fast, simple and automatable command-line program that extracts optimal recombination-filtered blocks (no four-gamete violations) from recombination-rich genomic re-sequence data. Availability: http://hammerlab.biosci.arizona.edu/software.html. 相似文献
18.
Ali Mohammadian Seyed Javad Mowla Elahe Elahi Mahmood Tavallaei Mohammad Reza Nourani Yu Liang 《Biotechnology letters》2013,35(6):843-851
Low-density quantitative real-time PCR (qPCR) arrays are often used to profile expression patterns of microRNAs in various biological milieus. To achieve accurate analysis of expression of miRNAs, non-biological sources of variation in data should be removed through precise normalization of data. We have systematically compared the performance of 19 normalization methods on different subsets of a real miRNA qPCR dataset that covers 40 human tissues. After robustly modeling the mean squared error (MSE) in normalized data, we demonstrate lower variability between replicates is achieved using various methods not applied to high-throughput miRNA qPCR data yet. Normalization methods that use splines or wavelets smoothing to estimate and remove Cq dependent non-linearity between pairs of samples best reduced the MSE of differences in Cq values of replicate samples. These methods also retained between-group variability in different subsets of the dataset. 相似文献
19.
Normalization and integration of large-scale metabolomics data using support vector regression 总被引:1,自引:0,他引:1
Xiaotao Shen Xiaoyun Gong Yuping Cai Yuan Guo Jia Tu Hao Li Tao Zhang Jialin Wang Fuzhong Xue Zheng-Jiang Zhu 《Metabolomics : Official journal of the Metabolomic Society》2016,12(5):89
Introduction
Untargeted metabolomics studies for biomarker discovery often have hundreds to thousands of human samples. Data acquisition of large-scale samples has to be divided into several batches and may span from months to as long as several years. The signal drift of metabolites during data acquisition (intra- and inter-batch) is unavoidable and is a major confounding factor for large-scale metabolomics studies.Objectives
We aim to develop a data normalization method to reduce unwanted variations and integrate multiple batches in large-scale metabolomics studies prior to statistical analyses.Methods
We developed a machine learning algorithm-based method, support vector regression (SVR), for large-scale metabolomics data normalization and integration. An R package named MetNormalizer was developed and provided for data processing using SVR normalization.Results
After SVR normalization, the portion of metabolite ion peaks with relative standard deviations (RSDs) less than 30 % increased to more than 90 % of the total peaks, which is much better than other common normalization methods. The reduction of unwanted analytical variations helps to improve the performance of multivariate statistical analyses, both unsupervised and supervised, in terms of classification and prediction accuracy so that subtle metabolic changes in epidemiological studies can be detected.Conclusion
SVR normalization can effectively remove the unwanted intra- and inter-batch variations, and is much better than other common normalization methods.20.
《Cell cycle (Georgetown, Tex.)》2013,12(23):4285-4286