首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
Xu S  Yonash N  Vallejo RL  Cheng HH 《Genetica》1998,104(2):171-178
A typical problem in mapping quantitative trait loci (QTLs) comes from missing QTL genotype. A routine method for parameter estimation involving missing data is the mixture model maximum likelihood method. We developed an alternative QTL mapping method that describes a mixture of several distributions by a single model with a heterogeneous residual variance. The two methods produce similar results, but the heterogeneous residual variance method is computationally much faster than the mixture model approach. In addition, the new method can automatically generate sampling variances of the estimated parameters. We derive the new method in the context of QTL mapping for binary traits in a F2 population. Using the heterogeneous residual variance model, we identified a QTL on chromosome IV that controls Marek's disease susceptibility in chickens. The QTL alone explains 7.2% of the total disease variation. This revised version was published online in July 2006 with corrections to the Cover Date.  相似文献   

2.
Discussing the organization of developmental switches, West-Eberhard (2003) proposed the use of regression residual plots to locate the neutral point of the switch, which is characterized by maximum genotypic influence on the resulting phenotype. However, statistical artifacts due to measurement error in nonlinear models might account for a substantial proportion of the increase of residuals variance at the switch point. Simulations based on field data show that increases in residual variance occur as artifacts when normal amounts of measurement error are present, even in absence of any genotypic variance. The results suggest that interpretation of nonlinear variation in thresold traits is problematic and requires considering this statistical effect. A method to estimate the weight of genotypic contribution to residual variance is proposed, and its assumptions and limitations are discussed.  相似文献   

3.
The aim of this study was to compare the variance component approach for QTL linkage mapping in half-sib designs to the simple regression method. Empirical power was determined by Monte Carlo simulation in granddaughter designs. The factors studied (base values in parentheses) included the number of sires (5) and sons per sire (80), ratio of QTL variance to total genetic variance (λ = 0.1), marker spacing (10 cM), and QTL allele frequency (0.5). A single bi-allelic QTL and six equally spaced markers with six alleles each were simulated. Empirical power using the regression method was 0.80, 0.92 and 0.98 for 5, 10, and 20 sires, respectively, versus 0.88, 0.98 and 0.99 using the variance component method. Power was 0.74, 0.80, 0.93, and 0.95 using regression versus 0.77, 0.88, 0.94, and 0.97 using the variance component method for QTL variance ratios (λ) of 0.05, 0.1, 0.2, and 0.3, respectively. Power was 0.79, 0.85, 0.80 and 0.87 using regression versus 0.80, 0.86, 0.88, and 0.85 using the variance component method for QTL allele frequencies of 0.1, 0.3, 0.5, and 0.8, respectively. The log10 of type I error profiles were quite flat at close marker spacing (1 cM), confirming the inability to fine-map QTL by linkage analysis in half-sib designs. The variance component method showed slightly more potential than the regression method in QTL mapping.  相似文献   

4.
Mao Y  Xu S 《Genetical research》2004,83(3):159-168
Many quantitative traits are measured as percentages. As a result, the assumption of a normal distribution for the residual errors of such percentage data is often violated. However, most quantitative trait locus (QTL) mapping procedures assume normality of the residuals. Therefore, proper data transformation is often recommended before statistical analysis is conducted. We propose the probit transformation to convert percentage data into variables with a normal distribution. The advantage of the probit transformation is that it can handle measurement errors with heterogeneous variance and correlation structure in a statistically sound manner. We compared the results of this data transformation with other transformations and found that this method can substantially increase the statistical power of QTL detection. We develop the QTL mapping procedure based on the maximum likelihood methodology implemented via the expectation-maximization algorithm. The efficacy of the new method is demonstrated using Monte Carlo simulation.  相似文献   

5.
Interval mapping by simple regression is a powerful method for the detection of quantitative trait loci (QTLs) in line crosses such as F2 populations. Due to the ease of computation of the regression approach, relatively complex models with multiple fixed effects, interactions between QTLs or between QTLs and fixed effects can easily be accommodated. However, polygenic effects, which are not targeted in QTL analysis, cannot be treated as random effects in a least squares analysis. In a cross between true inbred lines this is of no consequence, as the polygenic effect contributes just to the residual variance. In a cross between outbred lines, however, if a trait has high polygenic heritability, the additive polygenic effect has a large influence on variation in the population. Here we extend the fixed model for the regression interval mapping method to a mixed model using an animal model. This makes it possible to use not only the observations from progeny (e.g. F2), but also those from the parents (F1) to evaluate QTLs and polygenic effects. We show how the animal model using parental observations can be applied to an outbred cross and so increase the power and accuracy of QTL analysis. Three estimation methods, i.e. regression and an animal model either with or without parental observations, are applied to simulated data. The animal model using parental observations is shown to have advantages in estimating QTL position and additive genotypic value, especially when the polygenic heritability is large and the number of progeny per parent is small.  相似文献   

6.
Zimmer D  Mayer M  Reinsch N 《Genetics》2011,187(1):261-270
Methodology for mapping quantitative trait loci (QTL) has focused primarily on treating the QTL as a fixed effect. These methods differ from the usual models of genetic variation that treat genetic effects as random. Computationally expensive methods that allow QTL to be treated as random have been explicitly developed for additive genetic and dominance effects. By extending these methods with a variance component method (VCM), multiple QTL can be mapped. We focused on an F(2) crossbred population derived from inbred lines and estimated effects for each individual and their corresponding marker-derived genetic covariances. We present extensions to pairwise epistatic effects, which are computationally intensive because a great many individual effects must be estimated. But by replacing individual genetic effects with average genetic effects for each marker class, genetic covariances are approximated. This substantially reduces the computational burden by reducing the dimensions of covariance matrices of genetic effects, resulting in a remarkable gain in the speed of estimating the variance components and evaluating the residual log-likelihood. Preliminary results from simulations indicate competitiveness of the reduced model with multiple-interval mapping, regression interval mapping, and VCM with individual genetic effects in its estimated QTL positions and experimental power.  相似文献   

7.
The classical F‐test in the one‐way random effects ANOVA model is extended to solve the long outstanding problem of testing the between‐group variance on values also different from zero. This is done first for homoscedastic and heteroscedastic cases in not necessarily balanced models and secondly for balanced homoscedastic models. By simulation, the tests are shown to attain acceptable significance levels and high power even in data that do not follow the usual ANOVA model. An important application of the tests is given by the heterogeneity questions concerning the treatment effects across studies in meta‐analysis.  相似文献   

8.
S Eguchi  M Matsuura 《Biometrics》1990,46(2):415-426
A new method of testing the Hardy-Weinberg equilibrium in the human leukocyte antigen (HLA) system is proposed and applied to real data. The derivation is based on the maximum likelihood method and closely related to standard regression theory. The test statistic has a closed representation of residual sum of squares by a projection mapping of data onto the estimated regression plane. Under the Hardy-Weinberg law the noniterative estimates for the gene frequencies are suggested by the use of the projection mapping. The test statistic and gene frequency estimates are shown to be asymptotically equivalent to the maximum likelihood method and to be more efficient than the other suggested test statistic when there are more than two identified alleles.  相似文献   

9.
Luo ZW  Tao SH  Zeng ZB 《Genetics》2000,156(1):457-467
Three approaches are proposed in this study for detecting or estimating linkage disequilibrium between a polymorphic marker locus and a locus affecting quantitative genetic variation using the sample from random mating populations. It is shown that the disequilibrium over a wide range of circumstances may be detected with a power of 80% by using phenotypic records and marker genotypes of a few hundred individuals. Comparison of ANOVA and regression methods in this article to the transmission disequilibrium test (TDT) shows that, given the genetic variance explained by the trait locus, the power of TDT depends on the trait allele frequency, whereas the power of ANOVA and regression analyses is relatively independent from the allelic frequency. The TDT method is more powerful when the trait allele frequency is low, but much less powerful when it is high. The likelihood analysis provides reliable estimation of the model parameters when the QTL variance is at least 10% of the phenotypic variance and the sample size of a few hundred is used. Potential use of these estimates in mapping the trait locus is also discussed.  相似文献   

10.
The ability to properly assess and accurately phenotype true differences in feed efficiency among dairy cows is key to the development of breeding programs for improving feed efficiency. The variability among individuals in feed efficiency is commonly characterised by the residual intake approach. Residual feed intake is represented by the residuals of a linear regression of intake on the corresponding quantities of the biological functions that consume (or release) energy. However, the residuals include both, model fitting and measurement errors as well as any variability in cow efficiency. The objective of this study was to isolate the individual animal variability in feed efficiency from the residual component. Two separate models were fitted, in one the standard residual energy intake (REI) was calculated as the residual of a multiple linear regression of lactation average net energy intake (NEI) on lactation average milk energy output, average metabolic BW, as well as lactation loss and gain of body condition score. In the other, a linear mixed model was used to simultaneously fit fixed linear regressions and random cow levels on the biological traits and intercept using fortnight repeated measures for the variables. This method split the predicted NEI in two parts: one quantifying the population mean intercept and coefficients, and one quantifying cow-specific deviations in the intercept and coefficients. The cow-specific part of predicted NEI was assumed to isolate true differences in feed efficiency among cows. NEI and associated energy expenditure phenotypes were available for the first 17 fortnights of lactation from 119 Holstein cows; all fed a constant energy-rich diet. Mixed models fitting cow-specific intercept and coefficients to different combinations of the aforementioned energy expenditure traits, calculated on a fortnightly basis, were compared. The variance of REI estimated with the lactation average model represented only 8% of the variance of measured NEI. Among all compared mixed models, the variance of the cow-specific part of predicted NEI represented between 53% and 59% of the variance of REI estimated from the lactation average model or between 4% and 5% of the variance of measured NEI. The remaining 41% to 47% of the variance of REI estimated with the lactation average model may therefore reflect model fitting errors or measurement errors. In conclusion, the use of a mixed model framework with cow-specific random regressions seems to be a promising method to isolate the cow-specific component of REI in dairy cows.  相似文献   

11.
Kao CH 《Genetics》2000,156(2):855-865
The differences between maximum-likelihood (ML) and regression (REG) interval mapping in the analysis of quantitative trait loci (QTL) are investigated analytically and numerically by simulation. The analytical investigation is based on the comparison of the solution sets of the ML and REG methods in the estimation of QTL parameters. Their differences are found to relate to the similarity between the conditional posterior and conditional probabilities of QTL genotypes and depend on several factors, such as the proportion of variance explained by QTL, relative QTL position in an interval, interval size, difference between the sizes of QTL, epistasis, and linkage between QTL. The differences in mean squared error (MSE) of the estimates, likelihood-ratio test (LRT) statistics in testing parameters, and power of QTL detection between the two methods become larger as (1) the proportion of variance explained by QTL becomes higher, (2) the QTL locations are positioned toward the middle of intervals, (3) the QTL are located in wider marker intervals, (4) epistasis between QTL is stronger, (5) the difference between QTL effects becomes larger, and (6) the positions of QTL get closer in QTL mapping. The REG method is biased in the estimation of the proportion of variance explained by QTL, and it may have a serious problem in detecting closely linked QTL when compared to the ML method. In general, the differences between the two methods may be minor, but can be significant when QTL interact or are closely linked. The ML method tends to be more powerful and to give estimates with smaller MSEs and larger LRT statistics. This implies that ML interval mapping can be more accurate, precise, and powerful than REG interval mapping. The REG method is faster in computation, especially when the number of QTL considered in the model is large. Recognizing the factors affecting the differences between REG and ML interval mapping can help an efficient strategy, using both methods in QTL mapping to be outlined.  相似文献   

12.
In most QTL mapping studies, phenotypes are assumed to follow normal distributions. Deviations from this assumption may lead to detection of false positive QTL. To improve the robustness of Bayesian QTL mapping methods, the normal distribution for residuals is replaced with a skewed Student-t distribution. The latter distribution is able to account for both heavy tails and skewness, and both components are each controlled by a single parameter. The Bayesian QTL mapping method using a skewed Student-t distribution is evaluated with simulated data sets under five different scenarios of residual error distributions and QTL effects.  相似文献   

13.
Haley CS  Knott SA 《Heredity》1992,69(4):315-324
The use of flanking marker methods has proved to be a powerful tool for the mapping of quantitative trait loci (QTL) in the segregating generations derived from crosses between inbred lines. Methods to analyse these data, based on maximum-likelihood, have been developed and provide good estimates of QTL effects in some situations. Maximum-likelihood methods are, however, relatively complex and can be computationally slow. In this paper we develop methods for mapping QTL based on multiple regression which can be applied using any general statistical package. We use the example of mapping in an F(2) population and show that these regression methods produce very similar results to those obtained using maximum likelihood. The relative simplicity of the regression methods means that models with more than a single QTL can be explored and we give examples of two lined loci and of two interacting loci. Other models, for example with more than two QTL, with environmental fixed effects, with between family variance or for threshold traits, could be fitted in a similar way. The ease, speed of application and generality of regression methods for flanking marker analysis, and the good estimates they obtain, suggest that they should provide the method of choice for the analysis of QTL mapping data from inbred line crosses.  相似文献   

14.
Methodologies for segregation analysis and QTL mapping in plants   总被引:1,自引:0,他引:1  
Zhang YM  Gai J 《Genetica》2009,136(2):311-318
Most characters of biological interest and economic importance are quantitative traits. To uncover the genetic architecture of quantitative traits, two approaches have become popular in China. One is the establishment of an analytical model for mixed major-gene plus polygenes inheritance and the other the discovery of quantitative trait locus (QTL). Here we review our progress employing these two approaches. First, we proposed joint segregation analysis of multiple generations for mixed major-gene plus polygenes inheritance. Second, we extended the multilocus method of Lander and Green (1987), Jiang and Zeng (1997) to a more generalized approach. Our methodology handles distorted, dominant and missing markers, including the effect of linked segregation distortion loci on the estimation of map distance. Finally, we developed several QTL mapping methods. In the Bayesian shrinkage estimation (BSE) method, we suggested a method to test the significance of QTL effects and studied the effect of the prior distribution of the variance of QTL effect on QTL mapping. To reduce running time, a penalized maximum likelihood method was adopted. To mine novel genes in crop inbred lines generated in the course of normal crop breeding work, three methods were introduced. If a well-documented genealogical history of the lines is available, two-stage variance component analysis and multi-QTL Haseman-Elston regression were suggested; if unavailable, multiple loci in silico mapping was proposed.  相似文献   

15.

Background

Information for mapping of quantitative trait loci (QTL) comes from two sources: linkage disequilibrium (non-random association of allele states) and cosegregation (non-random association of allele origin). Information from LD can be captured by modeling conditional means and variances at the QTL given marker information. Similarly, information from cosegregation can be captured by modeling conditional covariances. Here, we consider a Bayesian model based on gene frequency (BGF) where both conditional means and variances are modeled as a function of the conditional gene frequencies at the QTL. The parameters in this model include these gene frequencies, additive effect of the QTL, its location, and the residual variance. Bayesian methodology was used to estimate these parameters. The priors used were: logit-normal for gene frequencies, normal for the additive effect, uniform for location, and inverse chi-square for the residual variance. Computer simulation was used to compare the power to detect and accuracy to map QTL by this method with those from least squares analysis using a regression model (LSR).

Results

To simplify the analysis, data from unrelated individuals in a purebred population were simulated, where only LD information contributes to map the QTL. LD was simulated in a chromosomal segment of 1 cM with one QTL by random mating in a population of size 500 for 1000 generations and in a population of size 100 for 50 generations. The comparison was studied under a range of conditions, which included SNP density of 0.1, 0.05 or 0.02 cM, sample size of 500 or 1000, and phenotypic variance explained by QTL of 2 or 5%. Both 1 and 2-SNP models were considered. Power to detect the QTL for the BGF, ranged from 0.4 to 0.99, and close or equal to the power of the regression using least squares (LSR). Precision to map QTL position of BGF, quantified by the mean absolute error, ranged from 0.11 to 0.21 cM for BGF, and was better than the precision of LSR, which ranged from 0.12 to 0.25 cM.

Conclusions

In conclusion given a high SNP density, the gene frequency model can be used to map QTL with considerable accuracy even within a 1 cM region.  相似文献   

16.
Summary A time‐specific log‐linear regression method on quantile residual lifetime is proposed. Under the proposed regression model, any quantile of a time‐to‐event distribution among survivors beyond a certain time point is associated with selected covariates under right censoring. Consistency and asymptotic normality of the regression estimator are established. An asymptotic test statistic is proposed to evaluate the covariate effects on the quantile residual lifetimes at a specific time point. Evaluation of the test statistic does not require estimation of the variance–covariance matrix of the regression estimators, which involves the probability density function of the survival distribution with censoring. Simulation studies are performed to assess finite sample properties of the regression parameter estimator and test statistic. The new regression method is applied to a breast cancer data set with long‐term follow‐up to estimate the patients' median residual lifetimes, adjusting for important prognostic factors.  相似文献   

17.
A modified algorithm for the improvement of composite interval mapping   总被引:27,自引:0,他引:27       下载免费PDF全文
Li H  Ye G  Wang J 《Genetics》2007,175(1):361-374
Composite interval mapping (CIM) is the most commonly used method for mapping quantitative trait loci (QTL) with populations derived from biparental crosses. However, the algorithm implemented in the popular QTL Cartographer software may not completely ensure all its advantageous properties. In addition, different background marker selection methods may give very different mapping results, and the nature of the preferred method is not clear. A modified algorithm called inclusive composite interval mapping (ICIM) is proposed in this article. In ICIM, marker selection is conducted only once through stepwise regression by considering all marker information simultaneously, and the phenotypic values are then adjusted by all markers retained in the regression equation except the two markers flanking the current mapping interval. The adjusted phenotypic values are finally used in interval mapping (IM). The modified algorithm has a simpler form than that used in CIM, but a faster convergence speed. ICIM retains all advantages of CIM over IM and avoids the possible increase of sampling variance and the complicated background marker selection process in CIM. Extensive simulations using two genomes and various genetic models indicated that ICIM has increased detection power, a reduced false detection rate, and less biased estimates of QTL effects.  相似文献   

18.
Xu C  Li Z  Xu S 《Genetics》2005,169(2):1045-1059
Joint mapping for multiple quantitative traits has shed new light on genetic mapping by pinpointing pleiotropic effects and close linkage. Joint mapping also can improve statistical power of QTL detection. However, such a joint mapping procedure has not been available for discrete traits. Most disease resistance traits are measured as one or more discrete characters. These discrete characters are often correlated. Joint mapping for multiple binary disease traits may provide an opportunity to explore pleiotropic effects and increase the statistical power of detecting disease loci. We develop a maximum-likelihood method for mapping multiple binary traits. We postulate a set of multivariate normal disease liabilities, each contributing to the phenotypic variance of one disease trait. The underlying liabilities are linked to the binary phenotypes through some underlying thresholds. The new method actually maps loci for the variation of multivariate normal liabilities. As a result, we are able to take advantage of existing methods of joint mapping for quantitative traits. We treat the multivariate liabilities as missing values so that an expectation-maximization (EM) algorithm can be applied here. We also extend the method to joint mapping for both discrete and continuous traits. Efficiency of the method is demonstrated using simulated data. We also apply the new method to a set of real data and detect several loci responsible for blast resistance in rice.  相似文献   

19.
Martinez VA  Hill WG  Knott SA 《Heredity》2002,88(6):423-431
The power to detect quantitative trait loci (QTL) using the double haploid (DH), full-sib (FS) and hierarchical (HI) designs implemented in outbred fish populations was assessed for interval mapping using deterministic methods. The predictions were tested using simulation. The DH design was most efficient for the range of designs and parameters considered and was most beneficial when the FS design was not very powerful. The difference between the design was largest for a low amount of residual genetic variation. Accounting for an increase of the environmental variance due to the genetic constitution of the double haploid progeny changed the magnitude of the power, but the ranking of the designs remained the same. As large full sib family sizes can be obtained in fish, the practical value of HI designs as a strategy for increasing the power of QTL mapping experiments is limited when compared with the FS design. Overall, the results suggested that the DH design could be a very useful tool for QTL mapping in fish, and of particular importance when the effect of the QTL is low and the residual genetic variation from other chromosomes can be controlled by using multiple markers.  相似文献   

20.
Constraints arise naturally in many scientific experiments/studies such as in, epidemiology, biology, toxicology, etc. and often researchers ignore such information when analyzing their data and use standard methods such as the analysis of variance (ANOVA). Such methods may not only result in a loss of power and efficiency in costs of experimentation but also may result poor interpretation of the data. In this paper we discuss constrained statistical inference in the context of linear mixed effects models that arise naturally in many applications, such as in repeated measurements designs, familial studies and others. We introduce a novel methodology that is broadly applicable for a variety of constraints on the parameters. Since in many applications sample sizes are small and/or the data are not necessarily normally distributed and furthermore error variances need not be homoscedastic (i.e. heterogeneity in the data) we use an empirical best linear unbiased predictor (EBLUP) type residual based bootstrap methodology for deriving critical values of the proposed test. Our simulation studies suggest that the proposed procedure maintains the desired nominal Type I error while competing well with other tests in terms of power. We illustrate the proposed methodology by re-analyzing a clinical trial data on blood mercury level. The methodology introduced in this paper can be easily extended to other settings such as nonlinear and generalized regression models.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号