首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 46 毫秒
1.

Background  

We propose a statistically principled baseline correction method, derived from a parametric smoothing model. It uses a score function to describe the key features of baseline distortion and constructs an optimal baseline curve to maximize it. The parameters are determined automatically by using LOWESS (locally weighted scatterplot smoothing) regression to estimate the noise variance.  相似文献   

2.

Background  

Analysis of DNA microarray data usually begins with a normalization step where intensities of different arrays are adjusted to the same scale so that the intensity levels from different arrays can be compared with one other. Both simple total array intensity-based as well as more complex "local intensity level" dependent normalization methods have been developed, some of which are widely used. Much less developed methods for microarray data analysis include those that bypass the normalization step and therefore yield results that are not confounded by potential normalization errors.  相似文献   

3.
4.

Background  

Low-level processing and normalization of microarray data are most important steps in microarray analysis, which have profound impact on downstream analysis. Multiple methods have been suggested to date, but it is not clear which is the best. It is therefore important to further study the different normalization methods in detail and the nature of microarray data in general.  相似文献   

5.

Background  

The quality of microarray data can seriously affect the accuracy of downstream analyses. In order to reduce variability and enhance signal reproducibility in these data, many normalization methods have been proposed and evaluated, most of which are for data obtained from cDNA microarrays and Affymetrix GeneChips. CodeLink Bioarrays are a newly emerged, single-color oligonucleotide microarray platform. To date, there are no reported studies that evaluate normalization methods for CodeLink Bioarrays.  相似文献   

6.

Background  

RT-qPCR analysis is a widely used method for the analysis of mRNA expression throughout the field of mesenchymal stromal cell (MSC) research. Comparison between MSC studies, both in vitro and in vivo, are challenging due to the varied methods of RT-qPCR data normalization and analysis. Therefore, this study focuses on putative housekeeping genes for the normalization of RT-qPCR data between heterogeneous commercially available human MSC, compared with more homogeneous populations of MSC such as MIAMI and RS-1 cells.  相似文献   

7.

Background  

Illumina Infinium whole genome genotyping (WGG) arrays are increasingly being applied in cancer genomics to study gene copy number alterations and allele-specific aberrations such as loss-of-heterozygosity (LOH). Methods developed for normalization of WGG arrays have mostly focused on diploid, normal samples. However, for cancer samples genomic aberrations may confound normalization and data interpretation. Therefore, we examined the effects of the conventionally used normalization method for Illumina Infinium arrays when applied to cancer samples.  相似文献   

8.

Background  

It is well known that the normalization step of microarray data makes a difference in the downstream analysis. All normalization methods rely on certain assumptions, so differences in results can be traced to different sensitivities to violation of the assumptions. Illustrating the lack of robustness, in a striking spike-in experiment all existing normalization methods fail because of an imbalance between up- and down-regulated genes. This means it is still important to develop a normalization method that is robust against violation of the standard assumptions  相似文献   

9.

Background  

To cancel experimental variations, microarray data must be normalized prior to analysis. Where an appropriate model for statistical data distribution is available, a parametric method can normalize a group of data sets that have common distributions. Although such models have been proposed for microarray data, they have not always fit the distribution of real data and thus have been inappropriate for normalization. Consequently, microarray data in most cases have been normalized with non-parametric methods that adjust data in a pair-wise manner. However, data analysis and the integration of resultant knowledge among experiments have been difficult, since such normalization concepts lack a universal standard.  相似文献   

10.

Background  

Normalization is a basic step in microarray data analysis. A proper normalization procedure ensures that the intensity ratios provide meaningful measures of relative expression values.  相似文献   

11.

Background  

In the microarray experiment, many undesirable systematic variations are commonly observed. Normalization is the process of removing such variation that affects the measured gene expression levels. Normalization plays an important role in the earlier stage of microarray data analysis. The subsequent analysis results are highly dependent on normalization. One major source of variation is the background intensities. Recently, some methods have been employed for correcting the background intensities. However, all these methods focus on defining signal intensities appropriately from foreground and background intensities in the image analysis. Although a number of normalization methods have been proposed, no systematic methods have been proposed using the background intensities in the normalization process.  相似文献   

12.
13.

Background  

Gene expression analysis based on comparison of electrophoretic patterns is strongly dependent on the accuracy of DNA fragment sizing. The current normalization strategy based on molecular weight markers has limited accuracy because marker peaks are often masked by intense peaks nearby. Cumulative errors in fragment lengths cause problems in the alignment of same-length fragments across different electropherograms, especially for small fragments (< 100 bp). For accurate comparison of electrophoretic patterns, further inspection and normalization of electrophoretic data after fragment sizing by conventional strategies is needed.  相似文献   

14.
15.

Background

High-throughput sequencing, such as ribonucleic acid sequencing (RNA-seq) and chromatin immunoprecipitation sequencing (ChIP-seq) analyses, enables various features of organisms to be compared through tag counts. Recent studies have demonstrated that the normalization step for RNA-seq data is critical for a more accurate subsequent analysis of differential gene expression. Development of a more robust normalization method is desirable for identifying the true difference in tag count data.

Results

We describe a strategy for normalizing tag count data, focusing on RNA-seq. The key concept is to remove data assigned as potential differentially expressed genes (DEGs) before calculating the normalization factor. Several R packages for identifying DEGs are currently available, and each package uses its own normalization method and gene ranking algorithm. We compared a total of eight package combinations: four R packages (edgeR, DESeq, baySeq, and NBPSeq) with their default normalization settings and with our normalization strategy. Many synthetic datasets under various scenarios were evaluated on the basis of the area under the curve (AUC) as a measure for both sensitivity and specificity. We found that packages using our strategy in the data normalization step overall performed well. This result was also observed for a real experimental dataset.

Conclusion

Our results showed that the elimination of potential DEGs is essential for more accurate normalization of RNA-seq data. The concept of this normalization strategy can widely be applied to other types of tag count data and to microarray data.  相似文献   

16.

Background  

Accuracy in quantitative real-time RT-PCR is dependent on high quality RNA, consistent cDNA synthesis, and validated stable reference genes for data normalization. Reference genes used for normalization impact the results generated from expression studies and, hence, should be evaluated prior to use across samples and treatments. Few statistically validated reference genes have been reported in grapevine. Moreover, success in isolating high quality RNA from grapevine tissues is typically limiting due to low pH, and high polyphenolic and polysaccharide contents.  相似文献   

17.
18.

Background  

Non-biological factors give rise to unwanted variations in cDNA microarray data. There are many normalization methods designed to remove such variations. However, to date there have been few published systematic evaluations of these techniques for removing variations arising from dye biases in the context of downstream, higher-order analytical tasks such as classification.  相似文献   

19.
20.

Introduction

Different normalization methods are available for urinary data. However, it is unclear which method performs best in minimizing error variance on a certain data-set as no generally applicable empirical criteria have been established so far.

Objectives

The main aim of this study was to develop an applicable and formally correct algorithm to decide on the normalization method without using phenotypic information.

Methods

We proved mathematically for two classical measurement error models that the optimal normalization method generates the highest correlation between the normalized urinary metabolite concentrations and its blood concentrations or, respectively, its raw urinary concentrations. We then applied the two criteria to the urinary 1H-NMR measured metabolomic data from the Study of Health in Pomerania (SHIP-0; n?=?4068) under different normalization approaches and compared the results with in silico experiments to explore the effects of inflated error variance in the dilution estimation.

Results

In SHIP-0, we demonstrated consistently that probabilistic quotient normalization based on aligned spectra outperforms all other tested normalization methods. Creatinine normalization performed worst, while for unaligned data integral normalization seemed to most reasonable. The simulated and the actual data were in line with the theoretical modeling, underlining the general validity of the proposed criteria.

Conclusions

The problem of choosing the best normalization procedure for a certain data-set can be solved empirically. Thus, we recommend applying different normalization procedures to the data and comparing their performances via the statistical methodology explicated in this work. On the basis of classical measurement error models, the proposed algorithm will find the optimal normalization method.
  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号