首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
SNOMAD is a collection of algorithms for the normalization and standardization of gene expression datasets derived from diverse biological and technological sources. In addition to conventional transformations and visualization tools, SNOMAD includes two non-linear transformations which correct for bias and variance which are non-uniformly distributed across the range of microarray element signal intensities: (1). Local mean normalization; and (2). Local variance correction (Z-score generation using a locally calculated standard deviation).  相似文献   

2.
Normalization of expression levels applied to microarray data can help in reducing measurement error. Different methods, including cyclic loess, quantile normalization and median or mean normalization, have been utilized to normalize microarray data. Although there is considerable literature regarding normalization techniques for mRNA microarray data, there are no publications comparing normalization techniques for microRNA (miRNA) microarray data, which are subject to similar sources of measurement error. In this paper, we compare the performance of cyclic loess, quantile normalization, median normalization and no normalization for a single-color microRNA microarray dataset. We show that the quantile normalization method works best in reducing differences in miRNA expression values for replicate tissue samples. By showing that the total mean squared error are lowest across almost all 36 investigated tissue samples, we are assured that the bias correction provided by quantile normalization is not outweighed by additional error variance that can arise from a more complex normalization method. Furthermore, we show that quantile normalization does not achieve these results by compression of scale.  相似文献   

3.
In this paper, fluorescent microarray images and various analysis techniques are described to improve the microarray data acquisition processes. Signal intensities produced by rarely expressed genes are initially correctly detected, but they are often lost in corrections for background, log or ratio. Our analyses indicate that a simple correlation between the mean and median signal intensities may be the best way to eliminate inaccurate microarray signals. Unlike traditional quality control methods, the low intensity signals are retained and inaccurate signals are eliminated in this mean and median correlation. With larger amounts of microarray data being generated, it becomes increasingly more difficult to analyze data on a visual basis. Our method allows for the automatic quantitative determination of accurate and reliable signals, which can then be used for normalization. We found that a mean to median correlation of 85% or higher not only retains more data than current methods, but the retained data is more accurate than traditional thresholds or common spot flagging algorithms. We have also found that by using pin microtapping and microvibrations, we can control spot quality independent from initial PCR volume.  相似文献   

4.

Background  

Analysis of DNA microarray data usually begins with a normalization step where intensities of different arrays are adjusted to the same scale so that the intensity levels from different arrays can be compared with one other. Both simple total array intensity-based as well as more complex "local intensity level" dependent normalization methods have been developed, some of which are widely used. Much less developed methods for microarray data analysis include those that bypass the normalization step and therefore yield results that are not confounded by potential normalization errors.  相似文献   

5.
6.
Affymetrix high-density oligonucleotide array is a tool that has the capacity to simultaneously measure the abundance of thousands of mRNA sequences in biological samples. In order to allow direct array-to-array comparisons, normalization is a necessity. When deciding on an appropriate normalization procedure there are a couple questions that need to be addressed, e.g., on which level should the normalization be performed: On the level of feature intensities or on the level of expression indexes? Should all features/expression indexes be used or can we choose a subset of features likely to be unregulated? Another question is how to actually perform the normalization: normalize using the overall mean intensity or use a smooth normalization curve? Most of the currently used normalization methods are linear; e.g., the normalization method implemented in the Affymetrix software GeneChip is based on the overall mean intensity. However, along with alternative methods of summarizing feature intensities into an expression index, nonlinear methods have recently started to appear. For many of these alternative methods, the natural choice is to normalize on the level of feature intensities, either using all feature intensities or only perfect match intensities. In this report, a nonlinear normalization procedure aimed for normalizing feature intensities is proposed.  相似文献   

7.
Systematic variations can occur at various steps of a cDNA microarray experiment and affect the measurement of gene expression levels. Accepted standards integrated into every cDNA microarray analysis can assess these variabilities and aid the interpretation of cDNA microarray experiments from different sources. A universally applicable approach to evaluate parameters such as input and output ratios, signal linearity, hybridization specificity and consistency across an array, as well as normalization strategies, is the utilization of exogenous control genes as spike-in and negative controls. We suggest that the use of such control sets, together with a sufficient number of experimental repeats, in-depth statistical analysis and thorough data validation should be made mandatory for the publication of cDNA microarray data.  相似文献   

8.
MOTIVATION: Microarray data are susceptible to a wide-range of artifacts, many of which occur on physical scales comparable to the spatial dimensions of the array. These artifacts introduce biases that are spatially correlated. The ability of current methodologies to detect and correct such biases is limited. RESULTS: We introduce a new approach for analyzing spatial artifacts, termed 'conditional residual analysis for microarrays' (CRAM). CRAM requires a microarray design that contains technical replicates of representative features and a limited number of negative controls, but is free of the assumptions that constrain existing analytical procedures. The key idea is to extract residuals from sets of matched replicates to generate residual images. The residual images reveal spatial artifacts with single-feature resolution. Surprisingly, spatial artifacts were found to coexist independently as additive and multiplicative errors. Efficient procedures for bias estimation were devised to correct the spatial artifacts on both intensity scales. In a survey of 484 published single-channel datasets, variance fell 4- to 12-fold in 5% of the datasets after bias correction. Thus, inclusion of technical replicates in a microarray design affords benefits far beyond what one might expect with a conventional 'n = 5' averaging, and should be considered when designing any microarray for which randomization is feasible. AVAILABILITY: CRAM is implemented as version 2 of the hoptag software package for R, which is included in the Supplementary information.  相似文献   

9.
10.
The standard (STD) 5 × 5 hybrid median filter (HMF) was previously described as a nonparametric local backestimator of spatially arrayed microtiter plate (MTP) data. As such, the HMF is a useful tool for mitigating global and sporadic systematic error in MTP data arrays. Presented here is the first known HMF correction of a primary screen suffering from systematic error best described as gradient vectors. Application of the STD 5 × 5 HMF to the primary screen raw data reduced background signal deviation, thereby improving the assay dynamic range and hit confirmation rate. While this HMF can correct gradient vectors, it does not properly correct periodic patterns that may present in other screening campaigns. To address this issue, 1 × 7 median and a row/column 5 × 5 hybrid median filter kernels (1 × 7 MF and RC 5 × 5 HMF) were designed ad hoc, to better fit periodic error patterns. The correction data show periodic error in simulated MTP data arrays is reduced by these alternative filter designs and that multiple corrective filters can be combined in serial operations for progressive reduction of complex error patterns in a MTP data array.  相似文献   

11.
12.
Hu J  He X 《Biometrics》2007,63(1):50-59
In microarray experiments, removal of systematic variations resulting from array preparation or sample hybridization conditions is crucial to ensure sensible results from the ensuing data analysis. For example, quantile normalization is routinely used in the treatment of both oligonucleotide and cDNA microarray data, even though there might be some loss of information in the normalization process. We recognize that the ideal normalization, if it ever exists, would aim to keep the maximal amount of gene profile information with the lowest possible noise. With this objective in mind, we propose a valuable enhancement to quantile normalization, and demonstrate through three Affymetrix experiments that the enhanced normalization can result in better performance in detecting and ranking differentially expressed genes across experimental conditions.  相似文献   

13.
New normalization methods for cDNA microarray data   总被引:7,自引:0,他引:7  
MOTIVATION: The focus of this paper is on two new normalization methods for cDNA microarrays. After the image analysis has been performed on a microarray and before differentially expressed genes can be detected, some form of normalization must be applied to the microarrays. Normalization removes biases towards one or other of the fluorescent dyes used to label each mRNA sample allowing for proper evaluation of differential gene expression. RESULTS: The two normalization methods that we present here build on previously described non-linear normalization techniques. We extend these techniques by firstly introducing a normalization method that deals with smooth spatial trends in intensity across microarrays, an important issue that must be dealt with. Secondly we deal with normalization of a new type of cDNA microarray experiment that is coming into prevalence, the small scale specialty or 'boutique' array, where large proportions of the genes on the microarrays are expected to be highly differentially expressed. AVAILABILITY: The normalization methods described in this paper are available via http://www.pi.csiro.au/gena/ in a software suite called tRMA: tools for R Microarray Analysis upon request of the authors. Images and data used in this paper are also available via the same link.  相似文献   

14.

Background  

In two-channel competitive genomic hybridization microarray experiments, the ratio of the two fluorescent signal intensities at each spot on the microarray is commonly used to infer the relative amounts of the test and reference sample DNA levels. This ratio may be influenced by systematic measurement effects from non-biological sources that can introduce biases in the estimated ratios. These biases should be removed before drawing conclusions about the relative levels of DNA. The performance of existing gene expression microarray normalization strategies has not been evaluated for removing systematic biases encountered in array-based comparative genomic hybridization (CGH), which aims to detect single copy gains and losses typically in samples with heterogeneous cell populations resulting in only slight shifts in signal ratios. The purpose of this work is to establish a framework for correcting the systematic sources of variation in high density CGH array images, while maintaining the true biological variations.  相似文献   

15.
T Conway  B Kraus  D L Tucker  D J Smalley  A F Dorman  L McKibben 《BioTechniques》2002,32(1):110, 112-4, 116, 118-9
Microsoft Windows-based computers have evolved to the point that they provide sufficient computational and visualization power for robust analysis of DNA array data. In fact, smaller laboratories might prefer to carry out some or all of their analyses and visualization in a Windows environment, rather than alternative platforms such as UNIX. We have developed a series of manually executed macros written in Visual Basic for Microsoft Excel spreadsheets, that allows for rapid and comprehensive gene expression data analysis. The first macro assigns gene names to spots on the DNA array and normalizes individual hybridizations by expressing the signal intensity for each gene as a percentage of the sum of all gene intensities. The second macro streamlines statistical consideration of the confidence in individual gene measurements for sets of experimental replicates by calculating probability values with the Student's t test. The third macro introduces a threshold value, calculates expression ratios between experimental conditions, and calculates the standard deviation of the mean of the log ratio values. Selected columns of data are copied by a fourth macro to create a processed data set suitable for entry into a Microsoft Access database. An Access database structure is described that allows simple queries across multiple experiments and export of data into third-party data visualization software packages. These analysis tools can be used in their present form by others working with commercial E. coli membrane arrays, or they may be adapted for use with other systems. The Excel spreadsheets with embedded Visual Basic macros and detailed instructions for their use are available at http://www.ou.edu/microarray.  相似文献   

16.
High-density functional gene arrays have become a powerful tool for environmental microbial detection and characterization. However, microarray data normalization and comparison for this type of microarray remain a challenge in environmental microbiology studies because some commonly used normalization methods (e.g., genomic DNA) for the study of pure cultures are not applicable. In this study, we developed a common oligonucleotide reference standard (CORS) method to address this problem. A unique 50-mer reference oligonucleotide probe was selected to co-spot with gene probes for each array feature. The complementary sequence was synthesized and labeled for use as the reference target, which was then spiked and cohybridized with each sample. The signal intensity of this reference target was used for microarray data normalization and comparison. The optimal amount or concentration were determined to be ca. 0.5 to 2.5% of a gene probe for the reference probe and ca. 0.25 to 1.25 fmol/μl for the reference target based on our evaluation with a pilot array. The CORS method was then compared to dye swap and genomic DNA normalization methods using the Desulfovibrio vulgaris whole-genome microarray, and significant linear correlations were observed. This method was then applied to a functional gene array to analyze soil microbial communities, and the results demonstrated that the variation of signal intensities among replicates based on the CORS method was significantly lower than the total intensity normalization method. The developed CORS provides a useful approach for microarray data normalization and comparison for studies of complex microbial communities.Microarray-based technology has become a robust genomic tool to detect, track, and profile hundreds to thousands of different microbial populations simultaneously in complex environments such as soils and sediments. For example, GeoChip, a comprehensive functional gene array, has been developed for investigating biogeochemical, ecological, and environmental processes (12, 18, 23, 27, 29, 32). Although a massive amount of microarray data can be generated rapidly, one of the bottlenecks in using microarrays for environmental microbial community studies is the lack of an appropriate standard for data comparison and normalization (6). Currently, it is difficult to compare microarray data across different sites, experiments, laboratories, and/or time periods (10). This limits the power of the technology to address ecological and environmental questions.In pure culture-based functional genomics studies, genomic DNAs (gDNAs) have been used as a common reference for hybridizations in which the same amount of gDNAs are used to cohybridize with each target cDNA sample and then to normalize different target cDNAs based on the gDNA standard (4, 5, 8, 9, 19, 21, 23). Several normalization methods such as scale normalization, quantile normalization, and Lowess normalization have been used for gene expression studies (2). Using the gDNA standard method can minimize or eliminate differences in target cDNA quantity, spot morphology, uneven hybridization, labeling, and sequence-specific hybridization behaviors (5), and this allows the comparison of microarray data across different sites, laboratories, experiments, and/or times. The main rationale for gDNA as a common reference is that it provides complete coverage for all genes represented on the array because the DNA composition from a particular organism should be identical across different treatment samples even though RNA expression is different (8). However, this approach is not applicable to microbial community studies because not all communities have identical DNA compositions. Pooling of equal amounts of gDNA or RNA from every target sample to make a common sample could be used as an alternative reference for cohybridization (1, 22). However, the disadvantage of the sample pooling approach is that samples do not provide large amounts of DNA or RNA in a reliable and reproducible way. For example, groundwater samples usually have a very low biomass and thus would not provide enough DNA for pooling. In addition, the sample pool itself is uncharacterized, and gene abundance may be diluted out so that insufficient DNA is present to result in a positive signal some array features, especially for those genes in low abundance. Moreover, a new sample pool would be required for every new experiment, making comparison across experiments difficult. Thus, other approaches need to be developed for microbial community studies.Dudley et al. (7) used a 25-mer oligonucleotide that matched a small portion of the parental EST clone vector contained in every PCR product printed on the array for normalization of pure culture RNA expression. Although the oligonucleotide generated a stable hybridization signal on every array feature, this method requires a universal sequence tag as a “capture” sequence, limiting its general use in microbial community studies. Thus, in the present study, we developed a common oligonucleotide reference standard (CORS) approach by co-spotting a common oligonucleotide with each array feature to improve the accuracy and comparability of microarray data for microbial community studies. This method was evaluated by using a pilot array, a whole-genome array, and a functional gene array, and all results demonstrate that the developed CORS is a reliable and reproducible method for microarray data normalization and comparison for microbial community studies.  相似文献   

17.
Do JH  Choi DK 《Molecules and cells》2006,22(3):254-261
DNA microarray is a powerful tool for high-throughput analysis of biological systems. Various computational tools have been created to facilitate the analysis of the large volume of data produced in DNA microarray experiments. Normalization is a critical step for obtaining data that are reliable and usable for subsequent analysis such as identification of differentially expressed genes and clustering. A variety of normalization methods have been proposed over the past few years, but no methods are still perfect. Various assumptions are often taken in the process of normalization. Therefore, the knowledge of underlying assumption and principle of normalization would be helpful for the correct analysis of microarray data. We present a review of normalization techniques from single-labeled platforms such as the Affymetrix GeneChip array to dual-labeled platforms like spotted array focusing on their principles and assumptions.  相似文献   

18.

Background  

Array-based comparative genomic hybridization (array-CGH) is a recently developed technique for analyzing changes in DNA copy number. As in all microarray analyses, normalization is required to correct for experimental artifacts while preserving the true biological signal. We investigated various sources of systematic variation in array-CGH data and identified two distinct types of spatial effect of no biological relevance as the predominant experimental artifacts: continuous spatial gradients and local spatial bias. Local spatial bias affects a large proportion of arrays, and has not previously been considered in array-CGH experiments.  相似文献   

19.
Chaotic mixer improves microarray hybridization   总被引:3,自引:0,他引:3  
Hybridization is an important aspect of microarray experimental design which influences array signal levels and the repeatability of data within an array and across different arrays. Current methods typically require 24h and use target inefficiently. In these studies, we compare hybridization signals obtained in conventional static hybridization, which depends on diffusional target delivery, with signals obtained in a dynamic hybridization chamber, which employs a fluid mixer based on chaotic advection theory to deliver targets across a conventional glass slide array. Microarrays were printed with a pattern of 102 identical probe spots containing a 65-mer oligonucleotide capture probe. Hybridization of a 725-bp fluorescently labeled target was used to measure average target hybridization levels, local signal-to-noise ratios, and array hybridization uniformity. Dynamic hybridization for 1h with 1 or 10ng of target DNA increased hybridization signal intensities approximately threefold over a 24-h static hybridization. Similarly, a 10- or 60-min dynamic hybridization of 10ng of target DNA increased hybridization signal intensities fourfold over a 24h static hybridization. In time course studies, static hybridization reached a maximum within 8 to 12h using either 1 or 10ng of target. In time course studies using the dynamic hybridization chamber, hybridization using 1ng of target increased to a maximum at 4h and that using 10ng of target did not vary over the time points tested. In comparison to static hybridization, dynamic hybridization reduced the signal-to-noise ratios threefold and reduced spot-to-spot variation twofold. Therefore, we conclude that dynamic hybridization based on a chaotic mixer design improves both the speed of hybridization and the maximum level of hybridization while increasing signal-to-noise ratios and reducing spot-to-spot variation.  相似文献   

20.
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号