首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
To fulfill existing guidelines, applicants that aim to place their genetically modified (GM) insect‐resistant crop plants on the market are required to provide data from field experiments that address the potential impacts of the GM plants on nontarget organisms (NTO's). Such data may be based on varied experimental designs. The recent EFSA guidance document for environmental risk assessment (2010) does not provide clear and structured suggestions that address the statistics of field trials on effects on NTO's. This review examines existing practices in GM plant field testing such as the way of randomization, replication, and pseudoreplication. Emphasis is placed on the importance of design features used for the field trials in which effects on NTO's are assessed. The importance of statistical power and the positive and negative aspects of various statistical models are discussed. Equivalence and difference testing are compared, and the importance of checking the distribution of experimental data is stressed to decide on the selection of the proper statistical model. While for continuous data (e.g., pH and temperature) classical statistical approaches – for example, analysis of variance (ANOVA) – are appropriate, for discontinuous data (counts) only generalized linear models (GLM) are shown to be efficient. There is no golden rule as to which statistical test is the most appropriate for any experimental situation. In particular, in experiments in which block designs are used and covariates play a role GLMs should be used. Generic advice is offered that will help in both the setting up of field testing and the interpretation and data analysis of the data obtained in this testing. The combination of decision trees and a checklist for field trials, which are provided, will help in the interpretation of the statistical analyses of field trials and to assess whether such analyses were correctly applied.  相似文献   

2.
为了评估转基因抗草甘膦除草剂大豆的食用安全性,以20%的比例将转基因抗草甘膦除草剂大豆GTS40-3-2和其亲本非转基因大豆A5403豆粕分别添加到基础饲料中喂养两代Sprague-Dawley(SD)大鼠,采用定性、定量PCR和ELISA方法检测转基因大豆成分相关基因和蛋白在长期饲喂的大鼠体内代谢残留状况。结果表明,大鼠喂养转基因大豆豆粕后,除了大鼠肠粪和盲肠内容物检测到有转基因成分的残留,肠道菌群和实质脏器均未发现相关基因和蛋白。结果提示,长期饲喂转基因抗草甘膦除草剂大豆GTS40-3-2与亲本A5403大豆豆粕对SD大鼠具有同样的食用安全性。  相似文献   

3.
4.
The central challenge from the Precautionary Principle to statistical methodology is to help delineate (preferably quantitatively) the possibility that some exposure is hazardous, even in cases where this is not established beyond reasonable doubt. The classical approach to hypothesis testing is unhelpful, because lack of significance can be due either to uninformative data or to genuine lack of effect (the Type II error problem). Its inversion, bioequivalence testing, might sometimes be a model for the Precautionary Principle in its ability to ‘prove the null hypothesis.’ Current procedures for setting safe exposure levels are essentially derived from these classical statistical ideas, and we outline how uncertainties in the exposure and response measurements affect the No Observed Adverse Effect Level (NOAEL), the Benchmark approach and the “Hockey Stick” model. A particular problem concerns model uncertainty: usually these procedures assume that the class of models describing dose/response is known with certainty; this assumption is however often violated, perhaps particularly often when epidemiological data form the source of the risk assessment, and regulatory authorities have occasionally resorted to some average based on competing models. The recent methodology of Bayesian model averaging might be a systematic version of this, but is this an arena for the Precautionary Principle to come into play?  相似文献   

5.
Most regulations worldwide stipulate that a new genetically modified (GM) crop event has to be compared to its closest non-GM counterpart as a corner stone of the pre-market risk assessment. To this end the GM crop and its comparator should be grown in field trials for a phenotypic comparison as well as for subsequent detailed analysis of the composition of the two crop varieties. A more in-depth globally harmonised approach for the conduct of these field trials is lacking. Only a few countries have formulated detailed protocols for the set-up of GM field trials. In some countries, commercial non-GM reference varieties need to be included in a field study to compile reliable data that indicate the range of natural variation for the compounds tested at the specific location. Detailed analysis of pre-market assessment reports have so far not shown the added value of including these reference varieties in the field trials. In all cases where specific values were found to be outside of the range of the reference varieties, it proved possible to draw conclusions on the part of the pre-market risk assessment that relates to the compositional analysis, on the basis of already available compositional data. With the increasing quality of several databases on compositional data of a growing number of crop species, it seems unlikely that reference varieties will become more important on future occasions. It was furthermore investigated whether this part of the risk assessment can be related to field trial requirements for variety registration with the explicit intention of reducing the data burden on producers of new GM plant varieties. Field trials for variety registration so far include an assessment of phenotypic characteristics that do not cover safety aspects, with the exception of establishment of the glycoalkaloid content in potatoes in the Netherlands and Sweden. It may, however, under certain conditions be relatively easy to exchange data from compositional measurements between variety registration and GM testing procedures, thus laying a foundation for testing the feasibility of combining both pre-market assessment procedures in a single pre-market evaluation path.  相似文献   

6.
In this paper, we introduce a Bayesian statistical model for the analysis of functional data observed at several time points. Examples of such data include the Michigan growth study where we wish to characterize the shape changes of human mandible profiles. The form of the mandible is often used by clinicians as an aid in predicting the mandibular growth. However, whereas many studies have demonstrated the changes in size that may occur during the period of pubertal growth spurt, shape changes have been less well investigated. Considering a group of subjects presenting normal occlusion, in this paper we thus describe a Bayesian functional ANOVA model that provides information about where and when the shape changes of the mandible occur during different stages of development. The model is developed by defining the notion of predictive process models for Gaussian process (GP) distributions used as priors over the random functional effects. We show that the predictive approach is computationally appealing and that it is useful to analyze multivariate functional data with unequally spaced observations that differ among subjects and times. Graphical posterior summaries show that our model is able to provide a biological interpretation of the morphometric findings and that they comprehensively describe the shape changes of the human mandible profiles. Compared with classical cephalometric analysis, this paper represents a significant methodological advance for the study of mandibular shape changes in two dimensions.  相似文献   

7.
针对生物信息学中序列模体的显著性检验问题,提出了一种基于极大似然准则的贝叶斯假设检验方法.将模体的显著性检验转化为多项分布的拟合优度检验问题,选取Dirichlet分布作为多项分布的先验分布并采用Newton-Raphson算法估计Dirichlet分布的超参数,使得数据的预测分布达到最大.应用贝叶斯定理得到贝叶斯因子进行模型选择,用于评价模体检验的统计显著性,这种方法克服了传统多项分布检验中构造检验统计量并计算其在零假设下确切分布的困难.选择JASPAR数据库中107个转录因子结合位点和100组随机模拟数据进行实验,采用皮尔逊积矩相关系数作为评价检验质量的一个标准,发现实验结果好于传统的模体检验的一些方法.  相似文献   

8.

Background  

Modern biology has shifted from "one gene" approaches to methods for genomic-scale analysis like microarray technology, which allow simultaneous measurement of thousands of genes. This has created a need for tools facilitating interpretation of biological data in "batch" mode. However, such tools often leave the investigator with large volumes of apparently unorganized information. To meet this interpretation challenge, gene-set, or cluster testing has become a popular analytical tool. Many gene-set testing methods and software packages are now available, most of which use a variety of statistical tests to assess the genes in a set for biological information. However, the field is still evolving, and there is a great need for "integrated" solutions.  相似文献   

9.
This paper considers the statistical analysis of entomological count data from field experiments with genetically modified (GM) plants. Such trials are carried out to assess environmental safety. Potential effects on nontarget organisms (NTOs), as indicators of biodiversity, are investigated. The European Food Safety Authority (EFSA) gives broad guidance on the environmental risk assessment (ERA) of GM plants. Field experiments must contain suitable comparator crops as a benchmark for the assessment of designated endpoints. In this paper, a detailed protocol is proposed to perform data analysis for the purpose of assessing environmental safety. The protocol includes the specification of a list of endpoints and their hierarchical relations, the specification of intended levels of data analysis, and the specification of provisional limits of concern to decide on the need for further investigation. The protocol emphasizes a graphical representation of estimates and confidence intervals for the ratio of mean abundances for the GM plant and its comparator crop. Interpretation relies mainly on equivalence testing in which confidence intervals are compared with the limits of concern. The proposed methodology is illustrated with entomological count data resulting from multiyear, multilocation field trials. A cisgenically modified potato line (with enhanced resistance to late blight disease) was compared to the original conventional potato variety in the Netherlands and Ireland in two successive years (2013, 2014). It is shown that the protocol encompasses alternative schemes for safety assessment resulting from different research questions and/or expert choices. Graphical displays of equivalence testing at several hierarchical levels and their interpretation are presented for one of these schemes. The proposed approaches should be of help in the ERA of GM or other novel plants.  相似文献   

10.
The interpretation of data-driven experiments in genomics often involves a search for biological categories that are enriched for the responder genes identified by the experiments. However, knowledge bases such as the Gene Ontology (GO) contain hundreds or thousands of categories with very high overlap between categories. Thus, enrichment analysis performed on one category at a time frequently returns large numbers of correlated categories, leaving the choice of the most relevant ones to the user''s; interpretation.Here we present model-based gene set analysis (MGSA) that analyzes all categories at once by embedding them in a Bayesian network, in which gene response is modeled as a function of the activation of biological categories. Probabilistic inference is used to identify the active categories. The Bayesian modeling approach naturally takes category overlap into account and avoids the need for multiple testing corrections met in single-category enrichment analysis. On simulated data, MGSA identifies active categories with up to 95% precision at a recall of 20% for moderate settings of noise, leading to a 10-fold precision improvement over single-category statistical enrichment analysis. Application to a gene expression data set in yeast demonstrates that the method provides high-level, summarized views of core biological processes and correctly eliminates confounding associations.  相似文献   

11.
Numerous statistical methods have been developed for analyzing high‐dimensional data. These methods often focus on variable selection approaches but are limited for the purpose of testing with high‐dimensional data. They are often required to have explicit‐likelihood functions. In this article, we propose a “hybrid omnibus test” for high‐dicmensional data testing purpose with much weaker requirements. Our hybrid omnibus test is developed under a semiparametric framework where a likelihood function is no longer necessary. Our test is a version of a frequentist‐Bayesian hybrid score‐type test for a generalized partially linear single‐index model, which has a link function being a function of a set of variables through a generalized partially linear single index. We propose an efficient score based on estimating equations, define local tests, and then construct our hybrid omnibus test using local tests. We compare our approach with an empirical‐likelihood ratio test and Bayesian inference based on Bayes factors, using simulation studies. Our simulation results suggest that our approach outperforms the others, in terms of type I error, power, and computational cost in both the low‐ and high‐dimensional cases. The advantage of our approach is demonstrated by applying it to genetic pathway data for type II diabetes mellitus.  相似文献   

12.
The classical multiple testing model remains an important practical area of statistics with new approaches still being developed. In this paper we develop a new multiple testing procedure inspired by a method sometimes used in a problem with a different focus. Namely, the inference after model selection problem. We note that solutions to that problem are often accomplished by making use of a penalized likelihood function. A classic example is the Bayesian information criterion (BIC) method. In this paper we construct a generalized BIC method and evaluate its properties as a multiple testing procedure. The procedure is applicable to a wide variety of statistical models including regression, contrasts, treatment versus control, change point, and others. Numerical work indicates that, in particular, for sparse models the new generalized BIC would be preferred over existing multiple testing procedures.  相似文献   

13.
自1996年第1例转基因作物在美国商业化种植, 其在全球的种植面积一直处于持续、快速增长的趋势。2010年, 全球转基因作物种植总面积达1.48×108 hm2, 所种植的转基因作物主要是耐除草剂和抗虫作物, 其中耐除草剂作物占种植总面积的81%。耐除草剂作物的种植为杂草的高效控制提供了新的手段, 但其可能带来的生态环境风险也引起了全世界各国的广泛关注和争议。该文在总结归纳前人研究的基础上, 针对耐除草剂作物的基因漂移、杂草化及对生物多样性的影响等当前人们普遍关注的环境风险问题, 系统讨论了相关的风险评价程序和方法, 概括和分析了当前耐除草剂作物的环境风险研究进展和管理措施, 以期为我国转基因耐除草剂作物的开发、风险评价及管理提供依据。  相似文献   

14.

Background  

Gene Ontology (GO) terms are often used to assess the results of microarray experiments. The most common way to do this is to perform Fisher's exact tests to find GO terms which are over-represented amongst the genes declared to be differentially expressed in the analysis of the microarray experiment. However, due to the high degree of dependence between GO terms, statistical testing is conservative, and interpretation is difficult.  相似文献   

15.

Background

This article describes classical and Bayesian interval estimation of genetic susceptibility based on random samples with pre-specified numbers of unrelated cases and controls.

Results

Frequencies of genotypes in cases and controls can be estimated directly from retrospective case-control data. On the other hand, genetic susceptibility defined as the expected proportion of cases among individuals with a particular genotype depends on the population proportion of cases (prevalence). Given this design, prevalence is an external parameter and hence the susceptibility cannot be estimated based on only the observed data. Interval estimation of susceptibility that can incorporate uncertainty in prevalence values is explored from both classical and Bayesian perspective. Similarity between classical and Bayesian interval estimates in terms of frequentist coverage probabilities for this problem allows an appealing interpretation of classical intervals as bounds for genetic susceptibility. In addition, it is observed that both the asymptotic classical and Bayesian interval estimates have comparable average length. These interval estimates serve as a very good approximation to the "exact" (finite sample) Bayesian interval estimates. Extension from genotypic to allelic susceptibility intervals shows dependency on phenotype-induced deviations from Hardy-Weinberg equilibrium.

Conclusions

The suggested classical and Bayesian interval estimates appear to perform reasonably well. Generally, the use of exact Bayesian interval estimation method is recommended for genetic susceptibility, however the asymptotic classical and approximate Bayesian methods are adequate for sample sizes of at least 50 cases and controls.  相似文献   

16.
A screening method aimed at identifying potential human carcinogens using either animal cancer bioassays or short-term genotoxic assays has 4 possible results: true positive, true negative, false positive and false negative. Such a categorisation is superficially similar to the results of hypothesis testing in a statistical analysis. In this latter case the false positive rate is determined by the significance level of the test and the false negative rate by the statistical power of the test. Although the two types of categorisation appear somewhat similar, different statistical issues are involved in their interpretation. Statistical methods appropriate for the analysis of the results of a series of assays include the use of Bayes' theorem and multivariate methods such as clustering techniques for the selection of batteries of short-term test capable of a better prediction of potential carcinogens. The conclusions drawn from such studies are dependent upon the estimates of values of sensitivity and specificity used, the choice of statistical method and the nature of the data set. The statistical issues resulting from the analysis of specific genotoxicity experiments involve the choice of suitable experimental designs and appropriate analyses together with the relationship of statistical significance to biological importance. The purpose of statistical analysis should increasingly be to estimate and explore effects rather than for formal hypothesis testing.  相似文献   

17.
Split-test Bonferroni correction for QEEG statistical maps   总被引:2,自引:0,他引:2  
With statistical testing, corrections for multiple comparisons, such as Bonferroni adjustments, have given rise to controversies in the scientific community, because of their negative impact on statistical power. This impact is especially problematic for high-multidimensional data, such as multi-electrode brain recordings. With brain imaging data, a reliable method is needed to assess statistical significance of the data without losing statistical power. Conjunction analysis allows the combination of significance and consistency of an effect. Through a balanced combination of information from retest experiments (multiple trials split testing), we present an intuitively appealing, novel approach for brain imaging conjunction. The method is then tested and validated on synthetic data followed by a real-world test on QEEG data from patients with Alzheimer’s disease. This latter application requires both reliable type-I error and type-II error rates, because of the poor signal-to-noise ratio inherent in EEG signals.  相似文献   

18.
Several penalization approaches have been developed to identify homogeneous subgroups based on a regression model with subject-specific intercepts in subgroup analysis. These methods often apply concave penalty functions to pairwise comparisons of the intercepts, such that the subjects with similar intercept values are assigned to the same group, which is very similar to the procedure of the penalization approaches for variable selection. Since the Bayesian methods are commonly used in variable selection, it is worth considering the corresponding approaches to subgroup analysis in the Bayesian framework. In this paper, a Bayesian hierarchical model with appropriate prior structures is developed for the pairwise differences of intercepts based on a regression model with subject-specific intercepts, which can automatically detect and identify homogeneous subgroups. A Gibbs sampling algorithm is also provided to select the hyperparameter and estimate the intercepts and coefficients of the covariates simultaneously, which is computationally efficient for pairwise comparisons compared to the time-consuming procedures for parameter estimation of the penalization methods (e.g., alternating direction method of multiplier) in the case of large sample sizes. The effectiveness and usefulness of the proposed Bayesian method are evaluated through simulation studies and analysis of a Cleveland Heart Disease Dataset.  相似文献   

19.
Use of nuclear magnetic resonance (NMR)-based metabonomics to search for human disease biomarkers is becoming increasingly common. For many researchers, the ultimate goal is translation from biomarker discovery to clinical application. Studies typically involve investigators from diverse educational and training backgrounds, including physicians, academic researchers, and clinical staff. In evaluating potential biomarkers, clinicians routinely use statistical significance testing language, whereas academicians typically use multivariate statistical analysis techniques that do not perform statistical significance evaluation. In this article, we outline an approach to integrate statistical significance testing with conventional principal components analysis data representation. A decision tree algorithm is introduced to select and apply appropriate statistical tests to loadings plot data, which are then heat map color-coded according to P score, enabling direct visual assessment of statistical significance. A multiple comparisons correction must be applied to determine P scores from which reliable inferences can be made. Knowledge of means and standard deviations of statistically significant buckets enabled computation of effect sizes and study sizes for a given statistical power. Methods were demonstrated using data from a previous study. Integrated metabonomics data assessment methodology should facilitate translation of NMR-based metabonomics discovery of human disease biomarkers to clinical use.  相似文献   

20.

Background  

With the increasing amount of data generated in molecular genetics laboratories, it is often difficult to make sense of results because of the vast number of different outcomes or variables studied. Examples include expression levels for large numbers of genes and haplotypes at large numbers of loci. It is then natural to group observations into smaller numbers of classes that allow for an easier overview and interpretation of the data. This grouping is often carried out in multiple steps with the aid of hierarchical cluster analysis, each step leading to a smaller number of classes by combining similar observations or classes. At each step, either implicitly or explicitly, researchers tend to interpret results and eventually focus on that set of classes providing the "best" (most significant) result. While this approach makes sense, the overall statistical significance of the experiment must include the clustering process, which modifies the grouping structure of the data and often removes variation.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号