首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.

Background

Gene expression microarray has been the primary biomarker platform ubiquitously applied in biomedical research, resulting in enormous data, predictive models, and biomarkers accrued. Recently, RNA-seq has looked likely to replace microarrays, but there will be a period where both technologies co-exist. This raises two important questions: Can microarray-based models and biomarkers be directly applied to RNA-seq data? Can future RNA-seq-based predictive models and biomarkers be applied to microarray data to leverage past investment?

Results

We systematically evaluated the transferability of predictive models and signature genes between microarray and RNA-seq using two large clinical data sets. The complexity of cross-platform sequence correspondence was considered in the analysis and examined using three human and two rat data sets, and three levels of mapping complexity were revealed. Three algorithms representing different modeling complexity were applied to the three levels of mappings for each of the eight binary endpoints and Cox regression was used to model survival times with expression data. In total, 240,096 predictive models were examined.

Conclusions

Signature genes of predictive models are reciprocally transferable between microarray and RNA-seq data for model development, and microarray-based models can accurately predict RNA-seq-profiled samples; while RNA-seq-based models are less accurate in predicting microarray-profiled samples and are affected both by the choice of modeling algorithm and the gene mapping complexity. The results suggest continued usefulness of legacy microarray data and established microarray biomarkers and predictive models in the forthcoming RNA-seq era.

Electronic supplementary material

The online version of this article (doi:10.1186/s13059-014-0523-y) contains supplementary material, which is available to authorized users.  相似文献   

2.

Background

Improved maps of species distributions are important for effective management of wildlife under increasing anthropogenic pressures. Recent advances in lidar and radar remote sensing have shown considerable potential for mapping forest structure and habitat characteristics across landscapes. However, their relative efficacies and integrated use in habitat mapping remain largely unexplored. We evaluated the use of lidar, radar and multispectral remote sensing data in predicting multi-year bird detections or prevalence for 8 migratory songbird species in the unfragmented temperate deciduous forests of New Hampshire, USA.

Methodology and Principal Findings

A set of 104 predictor variables describing vegetation vertical structure and variability from lidar, phenology from multispectral data and backscatter properties from radar data were derived. We tested the accuracies of these variables in predicting prevalence using Random Forests regression models. All data sets showed more than 30% predictive power with radar models having the lowest and multi-sensor synergy (“fusion”) models having highest accuracies. Fusion explained between 54% and 75% variance in prevalence for all the birds considered. Stem density from discrete return lidar and phenology from multispectral data were among the best predictors. Further analysis revealed different relationships between the remote sensing metrics and bird prevalence. Spatial maps of prevalence were consistent with known habitat preferences for the bird species.

Conclusion and Significance

Our results highlight the potential of integrating multiple remote sensing data sets using machine-learning methods to improve habitat mapping. Multi-dimensional habitat structure maps such as those generated from this study can significantly advance forest management and ecological research by facilitating fine-scale studies at both stand and landscape level.  相似文献   

3.
P Gao  X Zhou  ZN Wang  YX Song  LL Tong  YY Xu  ZY Yue  HM Xu 《PloS one》2012,7(7):e42015

Objective

Over the past decades, many studies have used data mining technology to predict the 5-year survival rate of colorectal cancer, but there have been few reports that compared multiple data mining algorithms to the TNM classification of malignant tumors (TNM) staging system using a dataset in which the training and testing data were from different sources. Here we compared nine data mining algorithms to the TNM staging system for colorectal survival analysis.

Methods

Two different datasets were used: 1) the National Cancer Institute''s Surveillance, Epidemiology, and End Results dataset; and 2) the dataset from a single Chinese institution. An optimization and prediction system based on nine data mining algorithms as well as two variable selection methods was implemented. The TNM staging system was based on the 7th edition of the American Joint Committee on Cancer TNM staging system.

Results

When the training and testing data were from the same sources, all algorithms had slight advantages over the TNM staging system in predictive accuracy. When the data were from different sources, only four algorithms (logistic regression, general regression neural network, Bayesian networks, and Naïve Bayes) had slight advantages over the TNM staging system. Also, there was no significant differences among all the algorithms (p>0.05).

Conclusions

The TNM staging system is simple and practical at present, and data mining methods are not accurate enough to replace the TNM staging system for colorectal cancer survival prediction. Furthermore, there were no significant differences in the predictive accuracy of all the algorithms when the data were from different sources. Building a larger dataset that includes more variables may be important for furthering predictive accuracy.  相似文献   

4.

Background

Genetic selection has been successful in achieving increased production in dairy cattle; however, corresponding declines in fitness traits have been documented. Selection for fitness traits is more difficult, since they have low heritabilities and are influenced by various non-genetic factors. The objective of this paper was to investigate the predictive ability of two-stage and single-step genomic selection methods applied to health data collected from on-farm computer systems in the U.S.

Methods

Implementation of single-trait and two-trait sire models was investigated using BayesA and single-step methods for mastitis and somatic cell score. Variance components were estimated. The complete dataset was divided into training and validation sets to perform model comparison. Estimated sire breeding values were used to estimate the number of daughters expected to develop mastitis. Predictive ability of each model was assessed by the sum of χ2 values that compared predicted and observed numbers of daughters with mastitis and the proportion of wrong predictions.

Results

According to the model applied, estimated heritabilities of liability to mastitis ranged from 0.05 (SD=0.02) to 0.11 (SD=0.03) and estimated heritabilities of somatic cell score ranged from 0.08 (SD=0.01) to 0.18 (SD=0.03). Posterior mean of genetic correlation between mastitis and somatic cell score was equal to 0.63 (SD=0.17). The single-step method had the best predictive ability. Conversely, the smallest number of wrong predictions was obtained with the univariate BayesA model. The best model fit was found for single-step and pedigree-based models. Bivariate single-step analysis had a better predictive ability than bivariate BayesA; however, the latter led to the smallest number of wrong predictions.

Conclusions

Genomic data improved our ability to predict animal breeding values. Performance of genomic selection methods depends on a multitude of factors. Heritability of traits and reliability of genotyped individuals has a large impact on the performance of genomic evaluation methods. Given the current characteristics of producer-recorded health data, single-step methods have several advantages compared to two-step methods.  相似文献   

5.

Aim

Sustained virologic response (SVR) can be attained with boceprevir plus peginterferon alfa and ribavirin (PR) in up to 68% of patients, and short duration therapy is possible if plasma HCV RNA levels are undetectable at treatment week 8 (TW8 response). We have developed predictive models for SVR, and TW8 response using data from boceprevir clinical trials.

Methods

Regression models were built to predict TW8 response and SVR. Separate models were built for TW8 and SVR using baseline variables only, and compared to models with baseline variables plus HCV RNA change after 4 weeks of PR (TW4 delta). Predictive accuracy was assessed by c-statistics, calibration curves, and decision curve analyses. Nomograms were developed to create clinical decision support tools. Models were externally validated using independent data.

Results

The models that included TW4 delta produced the best discrimination ability. The predictive factors for TW8 response (n = 856) were TW4 delta, race, platelet count and ALT. The predictive factors for SVR (n = 522) were TW4 delta, HCV-subtype, gender, BMI, RBV dose and platelet count. The discrimination abilities of these models were excellent (C-statistics = 0.88, 0.80 respectively). Baseline models for TW8 response (n = 444) and SVR (n = 197) had weaker discrimination ability (C-statistic = 0.76, 0.69). External validation confirmed the predictive accuracy of the week 4 models.

Conclusions

Models incorporating baseline and treatment week 4 data provide excellent prediction of TW8 response and SVR, and support the clinical utility of the lead-in phase of PR. The nomograms are suitable for point-of-care use to inform individual patient and physician decision-making.  相似文献   

6.

Background

The inhalation of allergens by allergic asthmatics results in the early asthmatic response (EAR), which is characterized by acute airway obstruction beginning within a few minutes. The EAR is the earliest indicator of the pathological progression of allergic asthma. Because the molecular mechanism underlying the EAR is not fully defined, this study will contribute to a better understanding of asthma.

Methods

In order to gain insight into the molecular basis of the EAR, we examined changes in protein expression patterns in the lung tissue of asthmatic rats during the EAR using 2-DE/MS-based proteomic techniques. Bioinformatic analysis of the proteomic data was then performed using PPI Spider and KEGG Spider to investigate the underlying molecular mechanism.

Results

In total, 44 differentially expressed protein spots were detected in the 2-DE gels. Of these 44 protein spots, 42 corresponded to 36 unique proteins successfully identified using mass spectrometry. During subsequent bioinformatic analysis, the gene ontology classification, the protein-protein interaction networking and the biological pathway exploration demonstrated that the identified proteins were mainly involved in glycolysis, calcium binding and mitochondrial activity. Using western blot and semi-quantitative RT-PCR, we confirmed the changes in expression of five selected proteins, which further supports our proteomic and bioinformatic analyses.

Conclusions

Our results reveal that the allergen-induced EAR in asthmatic rats is associated with glycolysis, calcium binding and mitochondrial activity, which could establish a functional network in which calcium binding may play a central role in promoting the progression of asthma.  相似文献   

7.
8.
Shi HY  Lee KT  Lee HH  Ho WH  Sun DP  Wang JJ  Chiu CC 《PloS one》2012,7(4):e35781

Background

Since most published articles comparing the performance of artificial neural network (ANN) models and logistic regression (LR) models for predicting hepatocellular carcinoma (HCC) outcomes used only a single dataset, the essential issue of internal validity (reproducibility) of the models has not been addressed. The study purposes to validate the use of ANN model for predicting in-hospital mortality in HCC surgery patients in Taiwan and to compare the predictive accuracy of ANN with that of LR model.

Methodology/Principal Findings

Patients who underwent a HCC surgery during the period from 1998 to 2009 were included in the study. This study retrospectively compared 1,000 pairs of LR and ANN models based on initial clinical data for 22,926 HCC surgery patients. For each pair of ANN and LR models, the area under the receiver operating characteristic (AUROC) curves, Hosmer-Lemeshow (H-L) statistics and accuracy rate were calculated and compared using paired T-tests. A global sensitivity analysis was also performed to assess the relative significance of input parameters in the system model and the relative importance of variables. Compared to the LR models, the ANN models had a better accuracy rate in 97.28% of cases, a better H-L statistic in 41.18% of cases, and a better AUROC curve in 84.67% of cases. Surgeon volume was the most influential (sensitive) parameter affecting in-hospital mortality followed by age and lengths of stay.

Conclusions/Significance

In comparison with the conventional LR model, the ANN model in the study was more accurate in predicting in-hospital mortality and had higher overall performance indices. Further studies of this model may consider the effect of a more detailed database that includes complications and clinical examination findings as well as more detailed outcome data.  相似文献   

9.

Background

In silico models have recently been created in order to predict which genetic variants are more likely to contribute to the risk of a complex trait given their functional characteristics. However, there has been no comprehensive review as to which type of predictive accuracy measures and data visualization techniques are most useful for assessing these models.

Methods

We assessed the performance of the models for predicting risk using various methodologies, some of which include: receiver operating characteristic (ROC) curves, histograms of classification probability, and the novel use of the quantile-quantile plot. These measures have variable interpretability depending on factors such as whether the dataset is balanced in terms of numbers of genetic variants classified as risk variants versus those that are not.

Results

We conclude that the area under the curve (AUC) is a suitable starting place, and for models with similar AUCs, violin plots are particularly useful for examining the distribution of the risk scores.

Electronic supplementary material

The online version of this article (doi:10.1186/s12864-015-1616-z) contains supplementary material, which is available to authorized users.  相似文献   

10.

Background

Genomic prediction is becoming a daily tool for plant breeders. It makes use of genotypic information to make predictions used for selection decisions. The accuracy of the predictions depends on the number of genotypes used in the calibration; hence, there is a need of combining data across years. A proper phenotypic analysis is a crucial prerequisite for accurate calibration of genomic prediction procedures. We compared stage-wise approaches to analyse a real dataset of a multi-environment trial (MET) in rye, which was connected between years only through one check, and used different spatial models to obtain better estimates, and thus, improved predictive abilities for genomic prediction. The aims of this study were to assess the advantage of using spatial models for the predictive abilities of genomic prediction, to identify suitable procedures to analyse a MET weakly connected across years using different stage-wise approaches, and to explore genomic prediction as a tool for selection of models for phenotypic data analysis.

Results

Using complex spatial models did not significantly improve the predictive ability of genomic prediction, but using row and column effects yielded the highest predictive abilities of all models. In the case of MET poorly connected between years, analysing each year separately and fitting year as a fixed effect in the genomic prediction stage yielded the most realistic predictive abilities. Predictive abilities can also be used to select models for phenotypic data analysis. The trend of the predictive abilities was not the same as the traditionally used Akaike information criterion, but favoured in the end the same models.

Conclusions

Making predictions using weakly linked datasets is of utmost interest for plant breeders. We provide an example with suggestions on how to handle such cases. Rather than relying on checks we show how to use year means across all entries for integrating data across years. It is further shown that fitting of row and column effects captures most of the heterogeneity in the field trials analysed.

Electronic supplementary material

The online version of this article (doi:10.1186/1471-2164-15-646) contains supplementary material, which is available to authorized users.  相似文献   

11.

Background and Objectives

In Brazil, as in many other affected countries, a large proportion of visceral leishmaniasis (VL) occurs in remote locations and treatment is often performed on basis of clinical suspicion. This study aimed at developing predictive models to help with the clinical management of VL in patients with suggestive clinical of disease.

Methods

Cases of VL (n = 213) had the diagnosis confirmed by parasitological method, non-cases (n = 119) presented suggestive clinical presentation of VL but a negative parasitological diagnosis and a firm diagnosis of another disease. The original data set was divided into two samples for generation and validation of the prediction models. Prediction models based on clinical signs and symptoms, results of laboratory exams and results of five different serological tests, were developed by means of logistic regression and classification and regression trees (CART). From these models, clinical-laboratory and diagnostic prediction scores were generated. The area under the receiver operator characteristic curve, sensitivity, specificity, and positive predictive value were used to evaluate the models'' performance.

Results

Based on the variables splenomegaly, presence of cough and leukopenia and on the results of five serological tests it was possible to generate six predictive models using logistic regression, showing sensitivity ranging from 90.1 to 99.0% and specificity ranging from 53.0 to 97.2%. Based on the variables splenomegaly, leukopenia, cough, age and weight loss and on the results of five serological tests six predictive models were generated using CART with sensitivity ranging from 90.1 to 97.2% and specificity ranging from 68.4 to 97.4%. The models composed of clinical-laboratory variables and the rk39 rapid test showed the best performance.

Conclusion

The predictive models showed to be a potential useful tool to assist healthcare systems and control programs in their strategical choices, contributing to more efficient and more rational allocation of healthcare resources.  相似文献   

12.

Background

Most studies on genomic prediction with reference populations that include multiple lines or breeds have used linear models. Data heterogeneity due to using multiple populations may conflict with model assumptions used in linear regression methods.

Methods

In an attempt to alleviate potential discrepancies between assumptions of linear models and multi-population data, two types of alternative models were used: (1) a multi-trait genomic best linear unbiased prediction (GBLUP) model that modelled trait by line combinations as separate but correlated traits and (2) non-linear models based on kernel learning. These models were compared to conventional linear models for genomic prediction for two lines of brown layer hens (B1 and B2) and one line of white hens (W1). The three lines each had 1004 to 1023 training and 238 to 240 validation animals. Prediction accuracy was evaluated by estimating the correlation between observed phenotypes and predicted breeding values.

Results

When the training dataset included only data from the evaluated line, non-linear models yielded at best a similar accuracy as linear models. In some cases, when adding a distantly related line, the linear models showed a slight decrease in performance, while non-linear models generally showed no change in accuracy. When only information from a closely related line was used for training, linear models and non-linear radial basis function (RBF) kernel models performed similarly. The multi-trait GBLUP model took advantage of the estimated genetic correlations between the lines. Combining linear and non-linear models improved the accuracy of multi-line genomic prediction.

Conclusions

Linear models and non-linear RBF models performed very similarly for genomic prediction, despite the expectation that non-linear models could deal better with the heterogeneous multi-population data. This heterogeneity of the data can be overcome by modelling trait by line combinations as separate but correlated traits, which avoids the occasional occurrence of large negative accuracies when the evaluated line was not included in the training dataset. Furthermore, when using a multi-line training dataset, non-linear models provided information on the genotype data that was complementary to the linear models, which indicates that the underlying data distributions of the three studied lines were indeed heterogeneous.

Electronic supplementary material

The online version of this article (doi:10.1186/s12711-014-0075-3) contains supplementary material, which is available to authorized users.  相似文献   

13.

Background

There are two ways that statistical methods can learn from biomedical data. One way is to learn classifiers to identify diseases and to predict outcomes using the training dataset with established diagnosis for each sample. When the training dataset is not available the task can be to mine for presence of meaningful groups (clusters) of samples and to explore underlying data structure (unsupervised learning).

Results

We investigated the proteomic profiles of the cytosolic fraction of human liver samples using two-dimensional electrophoresis (2DE). Samples were resected upon surgical treatment of hepatic metastases in colorectal cancer. Unsupervised hierarchical clustering of 2DE gel images (n = 18) revealed a pair of clusters, containing 11 and 7 samples. Previously we used the same specimens to measure biochemical profiles based on cytochrome P450-dependent enzymatic activities and also found that samples were clearly divided into two well-separated groups by cluster analysis. It turned out that groups by enzyme activity almost perfectly match to the groups identified from proteomic data. Of the 271 reproducible spots on our 2DE gels, we selected 15 to distinguish the human liver cytosolic clusters. Using MALDI-TOF peptide mass fingerprinting, we identified 12 proteins for the selected spots, including known cancer-associated species.

Conclusions/Significance

Our results highlight the importance of hierarchical cluster analysis of proteomic data, and showed concordance between results of biochemical and proteomic approaches. Grouping of the human liver samples and/or patients into differing clusters may provide insights into possible molecular mechanism of drug metabolism and creates a rationale for personalized treatment.  相似文献   

14.

Background

Non-alcoholic fatty liver disease (NAFLD) is a prevalent and rapidly increasing disease worldwide; however, no widely accepted screening models to assess the risk of NAFLD are available. Therefore, we aimed to develop and validate a self-assessment score for NAFLD in the general population using two independent cohorts.

Methods

The development cohort comprised 15676 subjects (8313 males and 7363 females) who visited the National Health Insurance Service Ilsan Hospital in Korea in 2008–2010. Anthropometric, clinical, and laboratory data were examined during regular health check-ups and fatty liver diagnosed by abdominal ultrasound. Logistic regression analysis was conducted to determine predictors of prevalent NAFLD and to derive risk scores/models. We validated our models and compared them with other existing methods using an external cohort (N = 66868).

Results

The simple self-assessment score consists of age, sex, waist circumference, body mass index, history of diabetes and dyslipidemia, alcohol intake, physical activity and menopause status, which are independently associated with NAFLD, and has a value of 0–15. A cut-off point of ≥8 defined 58% of males and 36% of females as being at high-risk of NAFLD, and yielded a sensitivity of 80% in men (77% in women), a specificity of 67% (81%), a positive predictive value of 72% (63%), a negative predictive value of 76% (89%) and an AUC of 0.82 (0.88). Comparable results were obtained using the validation dataset. The comprehensive NAFLD score, which includes additional laboratory parameters, has enhanced discrimination ability, with an AUC of 0.86 for males and 0.91 for females. Both simple and comprehensive NAFLD scores were significantly increased in subjects with higher fatty liver grades or severity of liver conditions (e.g., simple steatosis, steatohepatitis).

Conclusions

The new non–laboratory-based self-assessment score may be useful for identifying individuals at high-risk of NAFLD. Further studies are warranted to evaluate the utility and feasibility of the scores in various settings.  相似文献   

15.

Objective

The objective of this discovery-level investigation was to use mass spectrometry to identify low mass compounds in bronchoalveolar lavage fluid from lung transplant recipients that associate with bronchiolitis obliterans syndrome.

Experimental Design

Bronchoalveolar lavage fluid samples from lung transplant recipients were evaluated for small molecules using ESI-TOF mass spectrometry and correlated to the development of bronchiolitis obliterans syndrome. Peptides associated with samples from persons with bronchiolitis obliterans syndrome and controls were identified separately by MS/MS analysis.

Results

The average bronchoalveolar lavage fluid MS spectrum profile of individuals that developed bronchiolitis obliterans syndrome differed greatly compared to controls. Controls demonstrated close inter-sample correlation (R = 0.97+/−0.02, average+/−SD) while bronchiolitis obliterans syndrome showed greater heterogeneity (R = 0.86+/−0.09, average+/−SD). We identified 89 features that were predictive of developing BOS grade 1 and 66 features predictive of developing BOS grade 2 or higher. Fractions from MS analysis were pooled and evaluated for peptide content. Nearly 10-fold more peptides were found in bronchiolitis obliterans syndrome relative to controls. C-terminal residues suggested trypsin-like specificity among controls compared to elastase-type enzymes among those with bronchiolitis obliterans syndrome.

Conclusions

Bronchoalveolar lavage fluid from individuals with bronchiolitis obliterans syndrome has an increase in low mass components detected by mass spectrometry. Many of these features were peptides that likely result from elevated neutrophil elastase activity.  相似文献   

16.

Objectives

Rotator cuff tear is a common cause of shoulder diseases. Correct diagnosis of rotator cuff tears can save patients from further invasive, costly and painful tests. This study used predictive data mining and Bayesian theory to improve the accuracy of diagnosing rotator cuff tears by clinical examination alone.

Methods

In this retrospective study, 169 patients who had a preliminary diagnosis of rotator cuff tear on the basis of clinical evaluation followed by confirmatory MRI between 2007 and 2011 were identified. MRI was used as a reference standard to classify rotator cuff tears. The predictor variable was the clinical assessment results, which consisted of 16 attributes. This study employed 2 data mining methods (ANN and the decision tree) and a statistical method (logistic regression) to classify the rotator cuff diagnosis into “tear” and “no tear” groups. Likelihood ratio and Bayesian theory were applied to estimate the probability of rotator cuff tears based on the results of the prediction models.

Results

Our proposed data mining procedures outperformed the classic statistical method. The correction rate, sensitivity, specificity and area under the ROC curve of predicting a rotator cuff tear were statistical better in the ANN and decision tree models compared to logistic regression. Based on likelihood ratios derived from our prediction models, Fagan''s nomogram could be constructed to assess the probability of a patient who has a rotator cuff tear using a pretest probability and a prediction result (tear or no tear).

Conclusions

Our predictive data mining models, combined with likelihood ratios and Bayesian theory, appear to be good tools to classify rotator cuff tears as well as determine the probability of the presence of the disease to enhance diagnostic decision making for rotator cuff tears.  相似文献   

17.

Purpose

To identify non-invasive clinical parameters to predict urodynamic bladder outlet obstruction (BOO) in patients with benign prostatic hyperplasia (BPH) using causal Bayesian networks (CBN).

Subjects and Methods

From October 2004 to August 2013, 1,381 eligible BPH patients with complete data were selected for analysis. The following clinical variables were considered: age, total prostate volume (TPV), transition zone volume (TZV), prostate specific antigen (PSA), maximum flow rate (Qmax), and post-void residual volume (PVR) on uroflowmetry, and International Prostate Symptom Score (IPSS). Among these variables, the independent predictors of BOO were selected using the CBN model. The predictive performance of the CBN model using the selected variables was verified through a logistic regression (LR) model with the same dataset.

Results

Mean age, TPV, and IPSS were 6.2 (±7.3, SD) years, 48.5 (±25.9) ml, and 17.9 (±7.9), respectively. The mean BOO index was 35.1 (±25.2) and 477 patients (34.5%) had urodynamic BOO (BOO index ≥40). By using the CBN model, we identified TPV, Qmax, and PVR as independent predictors of BOO. With these three variables, the BOO prediction accuracy was 73.5%. The LR model showed a similar accuracy (77.0%). However, the area under the receiver operating characteristic curve of the CBN model was statistically smaller than that of the LR model (0.772 vs. 0.798, p = 0.020).

Conclusions

Our study demonstrated that TPV, Qmax, and PVR are independent predictors of urodynamic BOO.  相似文献   

18.
19.

Background

Bypass of foregut secreted factors promoting insulin resistance is hypothesized to be one of the mechanisms by which resolution of type 2 diabetes (T2D) follows roux-en-y gastric bypass (GBP) surgery.

Aim

To identify insulin resistance-associated proteins and metabolites which decrease more after GBP than after sleeve gastrectomy (SG) prior to diabetes remission.

Methods

Fasting plasma from 15 subjects with T2D undergoing GBP or SG was analyzed by proteomic and metabolomic methods 3 days before and 3 days after surgery. Subjects were matched for age, BMI, metformin therapy and glycemic control. Insulin resistance was calculated using homeostasis model assessment (HOMA-IR). For proteomics, samples were depleted of abundant plasma proteins, digested with trypsin and labeled with iTRAQ isobaric tags prior to liquid chromatography-tandem mass spectrometry analysis. Metabolomic analysis was performed using gas chromatography-mass spectrometry. The effect of the respective bariatric surgery on identified proteins and metabolites was evaluated using two-way analysis of variance and appropriate post-hoc tests.

Results

HOMA-IR improved, albeit not significantly, in both groups after surgery. Proteomic analysis yielded seven proteins which decreased significantly after GBP only, including Fetuin-A and Retinol binding protein 4, both previously linked to insulin resistance. Significant decrease in Fetuin-A and Retinol binding protein 4 after GBP was confirmed using ELISA and immunoassay. Metabolomic analysis identified significant decrease of citrate, proline, histidine and decanoic acid specifically after GBP.

Conclusion

Greater early decrease was seen for Fetuin-A, Retinol binding protein 4, and several metabolites after GBP compared to SG, preceding significant weight loss. This may contribute to enhanced T2D remission observed following foregut bypass procedures.  相似文献   

20.

Background

The theory has been put forward that if a null hypothesis is true, P-values should follow a Uniform distribution. This can be used to check the validity of randomisation.

Method

The theory was tested by simulation for two sample t tests for data from a Normal distribution and a Lognormal distribution, for two sample t tests which are not independent, and for chi-squared and Fisher’s exact test using small and using large samples.

Results

For the two sample t test with Normal data the distribution of P-values was very close to the Uniform. When using Lognormal data this was no longer true, and the distribution had a pronounced mode. For correlated tests, even using data from a Normal distribution, the distribution of P-values varied from simulation run to simulation run, but did not look close to Uniform in any realisation. For binary data in a small sample, only a few probabilities were possible and distribution was very uneven. With a sample of two groups of 1,000 observations, there was great unevenness in the histogram and a poor fit to the Uniform.

Conclusions

The notion that P-values for comparisons of groups using baseline data in randomised clinical trials should follow a Uniform distribution if the randomisation is valid has been found to be true only in the context of independent variables which follow a Normal distribution, not for Lognormal data, correlated variables, or binary data using either chi-squared or Fisher’s exact tests. This should not be used as a check for valid randomisation.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号