首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
    
We propose a likelihood-based model for correlated count data that display under- or overdispersion within units (e.g. subjects). The model is capable of handling correlation due to clustering and/or serial correlation, in the presence of unbalanced, missing or unequally spaced data. A family of distributions based on birth-event processes is used to model within-subject underdispersion. A computational approach is given to overcome a parameterization difficulty with this family, and this allows use of common Markov Chain Monte Carlo software (e.g. WinBUGS) for estimation. Application of the model to daily counts of asthma inhaler use by children shows substantial within-subject underdispersion, between-subject heterogeneity and correlation due to both clustering of measurements within subjects and serial correlation of longitudinal measurements. The model provides a major improvement over Poisson longitudinal models, and diagnostics show that the model fits well.  相似文献   

2.
    
  相似文献   

3.
    
Little is known about the human intra‐individual metabolic profile changes over an extended period of time. Here, we introduce a novel concept suggesting that children even at a very young age can be categorized in terms of metabolic state as they advance in development. The hidden Markov models were used as a method for discovering the underlying progression in the metabolic state. We applied the methodology to study metabolic trajectories in children between birth and 4 years of age, based on a series of samples selected from a large birth cohort study. We found multiple previously unknown age‐ and gender‐related metabolome changes of potential medical significance. Specifically, we found that the major developmental state differences between girls and boys are attributed to sphingolipids. In addition, we demonstrated the feasibility of state‐based alignment of personal metabolic trajectories. We show that children have different development rates at the level of metabolome and thus the state‐based approach may be advantageous when applying metabolome profiling in search of markers for subtle (patho)physiological changes.  相似文献   

4.
    
Variability in ecological community composition is often analyzed by recording the presence or abundance of taxa in sample units, calculating a symmetric matrix of pairwise distances or dissimilarities among sample units and then mapping the resulting matrix to a low‐dimensional representation through methods collectively called ordination. Unconstrained ordination only uses taxon composition data, without any environmental or experimental covariates, to infer latent compositional gradients associated with the sampling units. Commonly, such distance‐based methods have been used for ordination, but recently there has been a shift toward model‐based approaches. Model‐based unconstrained ordinations are commonly formulated using a Bayesian latent factor model that permits uncertainty assessment for parameters, including the latent factors that correspond to gradients in community composition. While model‐based methods have the additional benefit of addressing uncertainty in the estimated gradients, typically the current practice is to report point estimates without summarizing uncertainty. To demonstrate the uncertainty present in model‐based unconstrained ordination, the well‐known spider and dune data sets were analyzed and shown to have large uncertainty in the ordination projections. Hence to understand the factors that contribute to the uncertainty, simulation studies were conducted to assess the impact of additional sampling units or species to help inform future ordination studies that seek to minimize variability in the latent factors. Accurate reporting of uncertainty is an important part of transparency in the scientific process; thus, a model‐based approach that accounts for uncertainty is valuable. An R package, UncertainOrd , contains visualization tools that accurately represent estimates of the gradients in community composition in the presence of uncertainty.  相似文献   

5.
6.
This paper presents procedures for implementing the EM algorithm to compute REML estimates of variance covariance components in Gaussian mixed models for longitudinal data analysis. The class of models considered includes random coefficient factors, stationary time processes and measurement errors. The EM algorithm allows separation of the computations pertaining to parameters involved in the random coefficient factors from those pertaining to the time processes and errors. The procedures are illustrated with Pothoff and Roy''s data example on growth measurements taken on 11 girls and 16 boys at four ages. Several variants and extensions are discussed.  相似文献   

7.
    
  相似文献   

8.
    
Patrick LeBlanc  Li Ma 《Biometrics》2023,79(3):2321-2332
Mixed-membership (MM) models such as latent Dirichlet allocation (LDA) have been applied to microbiome compositional data to identify latent subcommunities of microbial species. These subcommunities are informative for understanding the biological interplay of microbes and for predicting health outcomes. However, microbiome compositions typically display substantial cross-sample heterogeneities in subcommunity compositions—that is, the variability in the proportions of microbes in shared subcommunities across samples—which is not accounted for in prior analyses. As a result, LDA can produce inference, which is highly sensitive to the specification of the number of subcommunities and often divides a single subcommunity into multiple artificial ones. To address this limitation, we incorporate the logistic-tree normal (LTN) model into LDA to form a new MM model. This model allows cross-sample variation in the composition of each subcommunity around some “centroid” composition that defines the subcommunity. Incorporation of auxiliary Pólya-Gamma variables enables a computationally efficient collapsed blocked Gibbs sampler to carry out Bayesian inference under this model. By accounting for such heterogeneity, our new model restores the robustness of the inference in the specification of the number of subcommunities and allows meaningful subcommunities to be identified.  相似文献   

9.
This paper relates the stability properties of a class of delay-difference equations to those of an associated scalar difference equation. Simple but powerful conditions for testing global stability are presented which are independent of the length of the time delay involved. For models which do not have globally stable equilibria, estimates of stability regions are obtained. Some well known baleen whale models are used to illustrate the results.  相似文献   

10.
A longitudinal approach is proposed to map QTL affecting function-valued traits and to estimate their effect over time. The method is based on fitting mixed random regression models. The QTL allelic effects are modelled with random coefficient parametric curves and using a gametic relationship matrix. A simulation study was conducted in order to assess the ability of the approach to fit different patterns of QTL over time. It was found that this longitudinal approach was able to adequately fit the simulated variance functions and considerably improved the power of detection of time-varying QTL effects compared to the traditional univariate model. This was confirmed by an analysis of protein yield data in dairy cattle, where the model was able to detect QTL with high effect either at the beginning or the end of the lactation, that were not detected with a simple 305 day model.  相似文献   

11.
Identifiability of statistical models is a fundamental regularity condition that is required for valid statistical inference. Investigation of model identifiability is mathematically challenging for complex models such as latent class models. Jones et al. used Goodman's technique to investigate the identifiability of latent class models with applications to diagnostic tests in the absence of a gold standard test. The tool they used was based on examining the singularity of the Jacobian or the Fisher information matrix, in order to obtain insights into local identifiability (ie, there exists a neighborhood of a parameter such that no other parameter in the neighborhood leads to the same probability distribution as the parameter). In this paper, we investigate a stronger condition: global identifiability (ie, no two parameters in the parameter space give rise to the same probability distribution), by introducing a powerful mathematical tool from computational algebra: the Gröbner basis. With several existing well-known examples, we argue that the Gröbner basis method is easy to implement and powerful to study global identifiability of latent class models, and is an attractive alternative to the information matrix analysis by Rothenberg and the Jacobian analysis by Goodman and Jones et al.  相似文献   

12.
    
This article demonstrates the use of mixed effects models for characterizing individual and sample average growth curves based on serial anthropometric data. These models are advancement over conventional general linear regression because they effectively handle the hierarchical nature of serial growth data. Using body weight data on 70 infants in the Born in Bradford study, we demonstrate how a mixed effects model provides a better fit than a conventional regression model. Further, we demonstrate how mixed effects models can be used to explore the influence of environmental factors on the sample average growth curve. Analyzing data from 183 infant boys (aged 3–15 months) from rural South India, we show how maternal education shapes infant growth patterns as early as within the first 6 months of life. The presented analyses highlight the utility of mixed effects models for analyzing serial growth data because they allow researchers to simultaneously predict individual curves, estimate sample average curves, and investigate the effects of environmental exposure variables. Am J Phys Anthropol, 2013. © 2012 Wiley Periodicals, Inc.  相似文献   

13.
    
In this paper, we propose a functional partially linear regression model with latent group structures to accommodate the heterogeneous relationship between a scalar response and functional covariates. The proposed model is motivated by a salinity tolerance study of barley families, whose main objective is to detect salinity tolerant barley plants. Our model is flexible, allowing for heterogeneous functional coefficients while being efficient by pooling information within a group for estimation. We develop an algorithm in the spirit of the K-means clustering to identify latent groups of the subjects under study. We establish the consistency of the proposed estimator, derive the convergence rate and the asymptotic distribution, and develop inference procedures. We show by simulation studies that the proposed method has higher accuracy for recovering latent groups and for estimating the functional coefficients than existing methods. The analysis of the barley data shows that the proposed method can help identify groups of barley families with different salinity tolerant abilities.  相似文献   

14.
This work investigates the insights and understanding which can be deduced from predictive process models for the product quality of a monoclonal antibody based on designed high‐throughput cell culture experiments performed at milliliter (ambr‐15®) scale. The investigated process conditions include various media supplements as well as pH and temperature shifts applied during the process. First, principal component analysis (PCA) is used to show the strong correlation characteristics among the product quality attributes including aggregates, fragments, charge variants, and glycans. Then, partial least square regression (PLS1 and PLS2) is applied to predict the product quality variables based on process information (one by one or simultaneously). The comparison of those two modeling techniques shows that a single (PLS2) model is capable of revealing the interrelationship of the process characteristics to the large set product quality variables. In order to show the dynamic evolution of the process predictability separate models are defined at different time points showing that several product quality attributes are mainly driven by the media composition and, hence, can be decently predicted from early on in the process, while others are strongly affected by process parameter changes during the process. Finally, by coupling the PLS2 models with a genetic algorithm first the model performance can be further improved and, most importantly, the interpretation of the large‐dimensioned process–product‐interrelationship can be significantly simplified. The generally applicable toolset presented in this case study provides a solid basis for decision making and process optimization throughout process development. © 2017 American Institute of Chemical Engineers Biotechnol. Prog., 33:1368–1380, 2017  相似文献   

15.
    
Summary Joint models are used to rigorously explore the relationship between the dynamics of biomarkers and clinical events. In the context of HIV infection, where the multivariate dynamics of HIV‐RNA and CD4 are complex, a mechanistic approach based on a system of nonlinear differential equations naturally takes into account the correlation between the biomarkers. Using data from a randomized clinical trial comparing dual antiretroviral therapy to a single drug regimen, a full maximum likelihood approach is proposed to explore the relationship between the evolution of the biomarkers and the time to a clinical event. The role of each marker as an independent predictor of disease progression is assessed. We show that the joint dynamics of HIV‐RNA and CD4 captures the effect of antiretroviral treatment; the CD4 dynamics alone is found to capture most but not all of the treatment effect.  相似文献   

16.
    
  相似文献   

17.
    
The ultrafine particle measurements in the Augsburger Umweltstudie, a panel study conducted in Augsburg, Germany, exhibit measurement error from various sources. Measurements of mobile devices show classical possibly individual–specific measurement error; Berkson–type error, which may also vary individually, occurs, if measurements of fixed monitoring stations are used. The combination of fixed site and individual exposure measurements results in a mixture of the two error types. We extended existing bias analysis approaches to linear mixed models with a complex error structure including individual–specific error components, autocorrelated errors, and a mixture of classical and Berkson error. Theoretical considerations and simulation results show, that autocorrelation may severely change the attenuation of the effect estimations. Furthermore, unbalanced designs and the inclusion of confounding variables influence the degree of attenuation. Bias correction with the method of moments using data with mixture measurement error partially yielded better results compared to the usage of incomplete data with classical error. Confidence intervals (CIs) based on the delta method achieved better coverage probabilities than those based on Bootstrap samples. Moreover, we present the application of these new methods to heart rate measurements within the Augsburger Umweltstudie: the corrected effect estimates were slightly higher than their naive equivalents. The substantial measurement error of ultrafine particle measurements has little impact on the results. The developed methodology is generally applicable to longitudinal data with measurement error.  相似文献   

18.
    
Instrumental variable methods, which handle unmeasured confounding by targeting the part of the exposure explained by an exogenous variable not subject to confounding, have gained much interest in observational studies. We consider the very frequent setting of estimating the unconfounded effect of an exposure measured at baseline on the subsequent trajectory of an outcome repeatedly measured over time. We didactically explain how to apply the instrumental variable method in such setting by adapting the two-stage classical methodology with (1) the prediction of the exposure according to the instrumental variable, (2) its inclusion into a mixed model to quantify the exposure association with the subsequent outcome trajectory, and (3) the computation of the estimated total variance. A simulation study illustrates the consequences of unmeasured confounding in classical analyses and the usefulness of the instrumental variable approach. The methodology is then applied to 6224 participants of the 3C cohort to estimate the association of type-2 diabetes with subsequent cognitive trajectory, using 42 genetic polymorphisms as instrumental variables. This contribution shows how to handle endogeneity when interested in repeated outcomes, along with a R implementation. However, it should still be used with caution as it relies on instrumental variable assumptions hardly testable in practice.  相似文献   

19.
    
Viable cell concentration (VCC) is an essential parameter that is required to support the efficient cultivation of mammalian cells. Although commonly determined using at-line or off-line analytics, in-line capacitance measurements represent a suitable alternative method for the determination of VCC. In addition, these latter efforts are complimentary with the Food and Drug Administration's initiative for process analytical technologies (PATs). However, current applications for online determination of the VCC often rely on single frequency measurements and corresponding linear regression models. It has been reported that this may be insufficient for application at all stages of a mammalian cell culture processes due to changes in multiple cell parameters over time. Alternatively, dielectric spectroscopy, measuring capacitance at multiple frequencies, in combination with multivariate mathematical models, has proven to be more robust. However, this has only been applied for retrospective data analysis. Here, we present the implementation of an O-PLS model for the online processing of multifrequency capacitance signals and the on-the-fly integration of the models’ VCC results into a supervisory control and data acquisition (SCADA) system commonly used for cultivation observation and control. This system was evaluated using a Chinese hamster ovary (CHO) cell perfusion process.  相似文献   

20.
    
Cell-free protein synthesis (CFPS) is an established method for rapid recombinant protein production. Advantages like short synthesis times and an open reaction environment make CFPS a desirable platform for new and difficult-to-express products. Most recently, interest has grown in using the technology to make larger amounts of material. This has been driven through a variety of reasons from making site specific antibody drug conjugates, to emergency response, to the safe manufacture of toxic biological products. We therefore need robust methods to determine the appropriate reaction conditions for product expression in CFPS. Here we propose a process development strategy for Escherichia coli lysate-based CFPS reactions that can be completed in as little as 48 hr. We observed the most dramatic increases in titer were due to the E. coli strain for the cell extract. Therefore, we recommend identifying a high-producing cell extract for the product of interest as a first step. Next, we manipulated the plasmid concentration, amount of extract, temperature, concentrated reaction mix pH levels, and length of reaction. The influence of these process parameters on titer was evaluated through multivariate data analysis. The process parameters with the highest impact on titer were subsequently included in a design of experiments to determine the conditions that increased titer the most in the design space. This proposed process development strategy resulted in superfolder green fluorescent protein titers of 0.686 g/L, a 38% improvement on the standard operating conditions, and hepatitis B core antigen titers of 0.386 g/L, a 190% improvement.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号