首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
Zhao and Tsiatis (1997) consider the problem of estimation of the distribution of the quality-adjusted lifetime when the chronological survival time is subject to right censoring. The quality-adjusted lifetime is typically defined as a weighted sum of the times spent in certain states up until death or some other failure time. They propose an estimator and establish the relevant asymptotics under the assumption of independent censoring. In this paper we extend the data structure with a covariate process observed until the end of follow-up and identify the optimal estimation problem. Because of the curse of dimensionality, no globally efficient nonparametric estimators, which have a good practical performance at moderate sample sizes, exist. Given a correctly specified model for the hazard of censoring conditional on the observed quality-of-life and covariate processes, we propose a closed-form one-step estimator of the distribution of the quality-adjusted lifetime whose asymptotic variance attains the efficiency bound if we can correctly specify a lower-dimensional working model for the conditional distribution of quality-adjusted lifetime given the observed quality-of-life and covariate processes. The estimator remains consistent and asymptotically normal even if this latter submodel is misspecified. The practical performance of the estimators is illustrated with a simulation study. We also extend our proposed one-step estimator to the case where treatment assignment is confounded by observed risk factors so that this estimator can be used to test a treatment effect in an observational study.  相似文献   

2.
We consider models of nucleotidic substitution processes where the rate of substitution at a given site depends on the state of the neighbours of the site. We first estimate the time elapsed between an ancestral sequence at stationarity and a present sequence. Second, assuming that two sequences are issued from a common ancestral sequence at stationarity, we estimate the time since divergence. In the simplest non-trivial case of a Jukes-Cantor model with CpG influence, we provide and justify mathematically consistent estimators in these two settings. We also provide asymptotic confidence intervals, valid for nucleotidic sequences of finite length, and we compute explicit formulas for the estimators and for their confidence intervals. In the general case of an RN model with YpR influence, we extend these results under a proviso, namely that the equation defining the estimator has a unique solution.  相似文献   

3.
In capture-recapture and mark-resight surveys, movements of individuals both within and between sampling periods can alter the susceptibility of individuals to detection over the region of sampling. In these circumstances spatially explicit capture-recapture (SECR) models, which incorporate the observed locations of individuals, allow population density and abundance to be estimated while accounting for differences in detectability of individuals. In this paper I propose two Bayesian SECR models, one for the analysis of recaptures observed in trapping arrays and another for the analysis of recaptures observed in area searches. In formulating these models I used distinct submodels to specify the distribution of individual home-range centers and the observable recaptures associated with these individuals. This separation of ecological and observational processes allowed me to derive a formal connection between Bayes and empirical Bayes estimators of population abundance that has not been established previously. I showed that this connection applies to every Poisson point-process model of SECR data and provides theoretical support for a previously proposed estimator of abundance based on recaptures in trapping arrays. To illustrate results of both classical and Bayesian methods of analysis, I compared Bayes and empirical Bayes esimates of abundance and density using recaptures from simulated and real populations of animals. Real populations included two iconic datasets: recaptures of tigers detected in camera-trap surveys and recaptures of lizards detected in area-search surveys. In the datasets I analyzed, classical and Bayesian methods provided similar – and often identical – inferences, which is not surprising given the sample sizes and the noninformative priors used in the analyses.  相似文献   

4.
This paper presents the Bayes estimators of the Poisson distribution function based on complete and truncated data under a natural conjugate prior. Laplace transform of the incomplete gamma function and the Gauss hypergeometric function have been employed in order to overcome the intractability of the integrals. Numerical examples from biosciences are given to illustrate the results. A Monte Carlo study has been carried out to compare Bayes estimators under complete data with the corresponding maximum liklihood estimators.  相似文献   

5.
Most current models of sequence evolution assume that all sites of a protein evolve under the same substitution process, characterized by a 20 x 20 substitution matrix. Here, we propose to relax this assumption by developing a Bayesian mixture model that allows the amino-acid replacement pattern at different sites of a protein alignment to be described by distinct substitution processes. Our model, named CAT, assumes the existence of distinct processes (or classes) differing by their equilibrium frequencies over the 20 residues. Through the use of a Dirichlet process prior, the total number of classes and their respective amino-acid profiles, as well as the affiliations of each site to a given class, are all free variables of the model. In this way, the CAT model is able to adapt to the complexity actually present in the data, and it yields an estimate of the substitutional heterogeneity through the posterior mean number of classes. We show that a significant level of heterogeneity is present in the substitution patterns of proteins, and that the standard one-matrix model fails to account for this heterogeneity. By evaluating the Bayes factor, we demonstrate that the standard model is outperformed by CAT on all of the data sets which we analyzed. Altogether, these results suggest that the complexity of the pattern of substitution of real sequences is better captured by the CAT model, offering the possibility of studying its impact on phylogenetic reconstruction and its connections with structure-function determinants.  相似文献   

6.
We introduce new robust small area estimation procedures basedon area-level models. We first find influence functions correspondingto each individual area-level observation by measuring the divergencebetween the posterior density functions of regression coefficientswith and without that observation. Next, based on these influencefunctions, properly standardized, we propose some new robustBayes and empirical Bayes small area estimators. The mean squarederrors and estimated mean squared errors of these estimatorsare also found. A small simulation study compares the performanceof the robust and the regular empirical Bayes estimators. Whenthe model variance is larger than the sample variance, the proposedrobust empirical Bayes estimators are superior.  相似文献   

7.
C R Rao 《Biometrics》1975,31(2):545-554
Empirical Bayes procedure is employed in simultaneous estimation of vector parameters from a number of Gauss-Markoff linear models. It is shown that with respect to quadratic loss function, empirical Bayes estimators are better than least squares estimators. While estimating the parameter for a particular linear model, a suggestion has been made for distinguishing between the loss due to decision maker and the loss due to individual. A method has been proposed but not fully studied to achieve balance between the two losses. Finally the problem of predicting future observations in a linear model has been considered.  相似文献   

8.
Böhning D  Kuhnert R 《Biometrics》2006,62(4):1207-1215
This article is about modeling count data with zero truncation. A parametric count density family is considered. The truncated mixture of densities from this family is different from the mixture of truncated densities from the same family. Whereas the former model is more natural to formulate and to interpret, the latter model is theoretically easier to treat. It is shown that for any mixing distribution leading to a truncated mixture, a (usually different) mixing distribution can be found so that the associated mixture of truncated densities equals the truncated mixture, and vice versa. This implies that the likelihood surfaces for both situations agree, and in this sense both models are equivalent. Zero-truncated count data models are used frequently in the capture-recapture setting to estimate population size, and it can be shown that the two Horvitz-Thompson estimators, associated with the two models, agree. In particular, it is possible to achieve strong results for mixtures of truncated Poisson densities, including reliable, global construction of the unique NPMLE (nonparametric maximum likelihood estimator) of the mixing distribution, implying a unique estimator for the population size. The benefit of these results lies in the fact that it is valid to work with the mixture of truncated count densities, which is less appealing for the practitioner but theoretically easier. Mixtures of truncated count densities form a convex linear model, for which a developed theory exists, including global maximum likelihood theory as well as algorithmic approaches. Once the problem has been solved in this class, it might readily be transformed back to the original problem by means of an explicitly given mapping. Applications of these ideas are given, particularly in the case of the truncated Poisson family.  相似文献   

9.
The implementation of genetic groups in BLUP evaluations accounts for different expectations of breeding values in base animals. Notwithstanding, many feasible structures of genetic groups exist and there are no analytical tools described to compare them easily. In this sense, the recent development of a simple and stable procedure to calculate the Bayes factor between nested competing models allowed us to develop a new approach of that method focused on compared models with different structures of random genetic groups. The procedure is based on a reparameterization of the model in terms of intraclass correlation of genetic groups. The Bayes factor can be easily calculated from the output of a Markov chain Monte Carlo sampling by averaging conditional densities at the null intraclass correlation. It compares two nested models, a model with a given structure of genetic groups against a model without genetic groups. The calculation of the Bayes factor between different structures of genetic groups can be quickly and easily obtained from the Bayes factor between the nested models. We applied this approach to a weaning weight data set of the Bruna dels Pirineus beef cattle, comparing several structures of genetic groups, and the final results showed that the preferable structure was an only group for unknown dams and different groups for unknown sires for each year of calving.  相似文献   

10.
The increasing diversity and heterogeneity of molecular data for phylogeny estimation has led to development of complex models and model-based estimators. Here, we propose a penalized likelihood (PL) framework in which the levels of complexity in the underlying model can be smoothly controlled. We demonstrate the PL framework for a four-taxon tree case and investigate its properties. The PL framework yields an estimator in which the majority of currently employed estimators such as the maximum-parsimony estimator, homogeneous likelihood estimator, gamma mixture likelihood estimator, etc., become special cases of a single family of PL estimators. Furthermore, using the appropriate penalty function, the complexity of the underlying models can be partitioned into separately controlled classes allowing flexible control of model complexity.  相似文献   

11.
We consider lifetime data involving pairs of study individuals with more than one possible cause of failure for each individual. Non-parametric estimation of cause-specific distribution functions is considered under independent censoring. Properties of the estimators are discussed and an illustration of their application is given.  相似文献   

12.
Simple discrete-time estimators which allow the on-line estimation of the kinetic rates from the measurements of components' concentrations inside a bioreactor are proposed. In fact, the proposed estimators are obtained by a direct forward Euler discretization of continuous-time estimators. The design of the estimators in the continuous as well as in the discrete-time does not require or assume any model for the kinetic rates. One of the main characteristics of these estimators lies in the easiness of their calibration. We here emphasize on the performances of the discrete version of these estimators, whose stability and convergence are proved under the same conditions as in the continuous case with an additional mild assumption on the sampling time. Simulation and real-life experiments results corresponding to the discrete estimation are given. The accuracy of the obtained estimates as well as the easiness of the estimators' implementation do constitute reliable and powerful arguments for their use, in particular in adaptive control schemes.  相似文献   

13.
Plant disease is responsible for major losses in agriculture throughout the world. Diseases are often spread by insect organisms that transmit a bacterium, virus, or other pathogen. To assess disease epidemics, plant pathologists often use multiple-vector-transfers. In such contexts, groups of insect vectors are moved from an infected source to each of n test plants that will then be observed for developing symptoms of infection. The purpose of this paper is to present new estimators for p, the probability of pathogen transmission for an individual vector, motivated from an empirical Bayesian approach. We specifically investigate four such estimators, characterize their small-sample properties, and propose new credible intervals for p. These estimators remove the need to specify hyperparameters a priori and are shown to be easier to compute than the classical Bayes estimators proposed by Chaubey and Li (1995, Journal of Official Statistics 11, 1035-1046) and Chick (1996, Biometrics 52, 1055-1062). Furthermore, some of these estimators are shown to have better frequentist properties than the commonly used maximum likelihood estimator and to provide a smaller Bayes risk than the estimator proposed by Burrows (1987, Phytopathology 77, 363-365).  相似文献   

14.
We discuss the importance of non-reversible evolutionary models when analyzing context-dependence. Given the inherent non-reversible nature of the well-known CpG-methylation-deamination process in mammalian evolution, non-reversible context-dependent evolutionary models may be well able to accurately model such a process. In particular, the lack of constraints on non-reversible substitution models might allow for more accurate estimation of context-dependent substitution parameters. To demonstrate this, we have developed different time-homogeneous context-dependent evolutionary models to analyze a large genomic dataset of primate ancestral repeats based on existing independent evolutionary models. We have calculated the difference in model fit for each of these models using Bayes Factors obtained via thermodynamic integration. We find that non-reversible context-dependent models can drastically increase model fit when compared to independent models and this on two primate non-coding datasets. Further, we show that further improvements are possible by clustering similar parameters across contexts.  相似文献   

15.
Codon-based substitution models are routinely used to measure selective pressures acting on protein-coding genes. To this effect, the nonsynonymous to synonymous rate ratio (dN/dS = omega) is estimated. The proportion of amino-acid sites potentially under positive selection, as indicated by omega > 1, is inferred by fitting a probability distribution where some sites are permitted to have omega > 1. These sites are then inferred by means of an empirical Bayes or by a Bayes empirical Bayes approach that, respectively, ignores or accounts for sampling errors in maximum-likelihood estimates of the distribution used to infer the proportion of sites with omega > 1. Here, we extend a previous full-Bayes approach to include models with high power and low false-positive rates when inferring sites under positive selection. We propose some heuristics to alleviate the computational burden, and show that (i) full Bayes can be superior to empirical Bayes when analyzing a small data set or small simulated data, (ii) full Bayes has only a small advantage over Bayes empirical Bayes with our small test data, and (iii) Bayesian methods appear relatively insensitive to mild misspecifications of the random process generating adaptive evolution in our simulations, but in practice can prove extremely sensitive to model specification. We suggest that the codon model used to detect amino acids under selection should be carefully selected, for instance using Akaike information criterion (AIC).  相似文献   

16.
The discovery that the lipids constituting the plasma membrane are not randomly distributed, but instead are able to form laterally segregated lipid domains with different properties has given hints how the formation of such lipid domains influences and regulates many processes occurring at the plasma membrane. While in model systems these lipid domains can be easily accessed and their properties studied, it is still challenging to determine the properties of cholesterol rich lipid domains, the so called “Rafts”, in the plasma membrane of living cells due to their small size and transient nature. One promising technique to address such issues is fluorescence lifetime imaging (FLIM) microscopy, as spatially resolved images make the visualization of the lateral lipid distribution possible, while at the same time the fluorescence lifetime of a membrane probe yields information about the bilayer structure and organization of the lipids in lipid domains and various properties like preferential protein-protein interactions or the enrichment of membrane probes. This review aims to give an overview of the techniques underlying FLIM probes which can be applied to investigate the formation of lipid domains and their respective properties in model membrane and biological systems. Also a short technical introduction into the techniques of a FLIM microscope is given.  相似文献   

17.
There is a great deal of recent interests in modeling right‐censored clustered survival time data with a possible fraction of cured subjects who are nonsusceptible to the event of interest using marginal mixture cure models. In this paper, we consider a semiparametric marginal mixture cure model for such data and propose to extend an existing generalized estimating equation approach by a new unbiased estimating equation for the regression parameters in the latency part of the model. The large sample properties of the regression effect estimators in both incidence and the latency parts are established. The finite sample properties of the estimators are studied in simulation studies. The proposed method is illustrated with a bone marrow transplantation data and a tonsil cancer data.  相似文献   

18.
Likelihood, parsimony, and heterogeneous evolution   总被引:5,自引:0,他引:5  
Evolutionary rates vary among sites and across the phylogenetic tree (heterotachy). A recent analysis suggested that parsimony can be better than standard likelihood at recovering the true tree given heterotachy. The authors recommended that results from parsimony, which they consider to be nonparametric, be reported alongside likelihood results. They also proposed a mixture model, which was inconsistent but better than either parsimony or standard likelihood under heterotachy. We show that their main conclusion is limited to a special case for the type of model they study. Their mixture model was inconsistent because it was incorrectly implemented. A useful nonparametric model should perform well over a wide range of possible evolutionary models, but parsimony does not have this property. Likelihood-based methods are therefore the best way to deal with heterotachy.  相似文献   

19.
A flexible statistical framework is developed for the analysis of read counts from RNA-Seq gene expression studies. It provides the ability to analyse complex experiments involving multiple treatment conditions and blocking variables while still taking full account of biological variation. Biological variation between RNA samples is estimated separately from the technical variation associated with sequencing technologies. Novel empirical Bayes methods allow each gene to have its own specific variability, even when there are relatively few biological replicates from which to estimate such variability. The pipeline is implemented in the edgeR package of the Bioconductor project. A case study analysis of carcinoma data demonstrates the ability of generalized linear model methods (GLMs) to detect differential expression in a paired design, and even to detect tumour-specific expression changes. The case study demonstrates the need to allow for gene-specific variability, rather than assuming a common dispersion across genes or a fixed relationship between abundance and variability. Genewise dispersions de-prioritize genes with inconsistent results and allow the main analysis to focus on changes that are consistent between biological replicates. Parallel computational approaches are developed to make non-linear model fitting faster and more reliable, making the application of GLMs to genomic data more convenient and practical. Simulations demonstrate the ability of adjusted profile likelihood estimators to return accurate estimators of biological variability in complex situations. When variation is gene-specific, empirical Bayes estimators provide an advantageous compromise between the extremes of assuming common dispersion or separate genewise dispersion. The methods developed here can also be applied to count data arising from DNA-Seq applications, including ChIP-Seq for epigenetic marks and DNA methylation analyses.  相似文献   

20.
Uncovering pathways of tumor progression is an important topic in cancer research that has led to numerous studies, including several pathway models proposed and investigated by Sontag and Axelrod [Progression of heterogeneous breast tumors. J. Theoret. Biol. 210, 107-119, 2005]. In their comparative studies, the authors focused on relative goodness of fits of the various pathways, but a simple test revealed that even the "best" model did not provide an adequate explanation for the observed breast tumor data. The heterogeneous nature of breast tumors leads to the question of whether more than one (i.e., a combination of) pathway models are needed in order to explain the observed data. In the current paper, we address this question based on the finite mixture modeling framework and utilizing the four pathways proposed in Sontag and Axelrod as our individual pathway models. The expectation-maximization algorithm was used to derive estimates for the mixing proportions of the mixture models. Indeed, a two-pathway mixture provides a dramatic improvement over any of the single pathway models for explaining the data derived under either the Van Nuys or the Holland system. In particular, for data graded under the Van Nuys system, the mixture model was shown to be consistent with the observed data at the 1% significant level.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号