首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
Taking households having at least one infective as standard units and considering both a within-household infection rate and a global infection rate, we propose a Bayesian two level mixing S-I-R (susceptible-infective-removed) counting process model in which the transmission parameters may change over time and the parameters of interest are the within-household infection rate and the removal rate. Customized Markov chain Monte Carlo methods are developed for generating samples from the posterior distribution for inference purpose, based only on the removal times. The numerical performance of this method is examined in a simulation study. Applying this method to 2003 Taiwan SARS data, we find that the within-household infection rate decreases, the removal rate increases and their ratio is less than one and decreases significantly during the epidemic. This method allows the estimation of these parameters during the epidemic. For a rapidly transmitted disease, it provides a method to nearly real-time tracking of infection measures.  相似文献   

2.
3.
We consider the estimation of a nonparametric smooth function of some event time in a semiparametric mixed effects model from repeatedly measured data when the event time is subject to right censoring. The within-subject correlation is captured by both cross-sectional and time-dependent random effects, where the latter is modeled by a nonhomogeneous Ornstein–Uhlenbeck stochastic process. When the censoring probability depends on other variables in the model, which often happens in practice, the event time data are not missing completely at random. Hence, the complete case analysis by eliminating all the censored observations may yield biased estimates of the regression parameters including the smooth function of the event time, and is less efficient. To remedy, we derive the likelihood function for the observed data by modeling the event time distribution given other covariates. We propose a two-stage pseudo-likelihood approach for the estimation of model parameters by first plugging an estimator of the conditional event time distribution into the likelihood and then maximizing the resulting pseudo-likelihood function. Empirical evaluation shows that the proposed method yields negligible biases while significantly reduces the estimation variability. This research is motivated by the project of hormone profile estimation around age at the final menstrual period for the cohort of women in the Michigan Bone Health and Metabolism Study.  相似文献   

4.
5.
The modeling of biomolecular complexes by computational docking using the known structures of their constituents is developing rapidly to become a powerful tool in structural biology. It is especially useful in combination with even limited experimental information describing the interface. Here we demonstrate for the first time the use of diffusion anisotropy in combination with chemical shift perturbation data to drive protein–protein docking. For validation purposes we make use of simulated diffusion anisotropy data. Inclusion of this information, which can be derived from NMR relaxation rates and reports on the orientation of the components of a complex with respect to the rotational diffusion tensor, substantially improves the docking results.  相似文献   

6.
One hundred cases of common bile duct explorations were reviewed in an attempt to obtain information that might give insight into the diagnosis and definitive treatment of choledocholithiasis. Fifty of the hundred patients had common duct stones. Correlations were made between the incidence of choledocholithiasis as proved at operation, and the following factors: Kind and number of choledochal exploratory criteria used, the clinical diagnosis of common duct stones, and the pathologic features of gallbladders removed.The incidence of stones was statistically related to aging.The most frequent choledochal exploratory criteria were common duct dilatation or thickening (63 cases) and history of jaundice (50 cases).The most reliable single criterion in “diagnosing” common duct stones was palpable common or hepatic duct stones, the diagnosis having been correct in 15 of 17 such cases.The most reliable combination of criteria was a history of jaundice, plus palpable stones, with correct diagnosis in all such cases.The clinical diagnosis of choledocholithiasis was correct in only 17 per cent of cases.The correlation of the incidence of common duct stones with the degree of gallbladder disease—that is, acute or chronic—did not provide information that might be helpful in diagnosing choledocholithiasis.The incidence of proven retained common duct stones was 3 per cent, the non-fatal postoperative complication rate was 21 per cent and operative mortality was 1 per cent.  相似文献   

7.
The integration of local agricultural knowledge deepens the understanding of complex phenomena such as the association between climate variability, crop yields and undernutrition. Participatory Sensing (PS) is a concept which enables laymen to easily gather geodata with standard low-cost mobile devices, offering new and efficient opportunities for agricultural monitoring. This study presents a methodological approach for crop height assessment based on PS. In-field crop height variations of a maize field in Heidelberg, Germany, are gathered with smartphones and handheld GPS devices by 19 participants. The comparison of crop height values measured by the participants to reference data based on terrestrial laser scanning (TLS) results in R2 = 0.63 for the handheld GPS devices and R2 = 0.24 for the smartphone-based approach. RMSE for the comparison between crop height models (CHM) derived from PS and TLS data is 10.45 cm (GPS devices) and 14.69 cm (smartphones). Furthermore, the results indicate that incorporating participants’ cognitive abilities in the data collection process potentially improves the quality data captured with the PS approach. The proposed PS methods serve as a fundament to collect agricultural parameters on field-level by incorporating local people. Combined with other methods such as remote sensing, PS opens new perspectives to support agricultural development.  相似文献   

8.

Background

The next generation sequencing technologies substantially increased the throughput of microbial genome sequencing. To functionally annotate newly sequenced microbial genomes, a variety of experimental and computational methods are used. Integration of information from different sources is a powerful approach to enhance such annotation. Functional analysis of microbial genomes, necessary for downstream experiments, crucially depends on this annotation but it is hampered by the current lack of suitable information integration and exploration systems for microbial genomes.

Results

We developed a data warehouse system (INDIGO) that enables the integration of annotations for exploration and analysis of newly sequenced microbial genomes. INDIGO offers an opportunity to construct complex queries and combine annotations from multiple sources starting from genomic sequence to protein domain, gene ontology and pathway levels. This data warehouse is aimed at being populated with information from genomes of pure cultures and uncultured single cells of Red Sea bacteria and Archaea. Currently, INDIGO contains information from Salinisphaera shabanensis, Haloplasma contractile, and Halorhabdus tiamatea - extremophiles isolated from deep-sea anoxic brine lakes of the Red Sea. We provide examples of utilizing the system to gain new insights into specific aspects on the unique lifestyle and adaptations of these organisms to extreme environments.

Conclusions

We developed a data warehouse system, INDIGO, which enables comprehensive integration of information from various resources to be used for annotation, exploration and analysis of microbial genomes. It will be regularly updated and extended with new genomes. It is aimed to serve as a resource dedicated to the Red Sea microbes. In addition, through INDIGO, we provide our Automatic Annotation of Microbial Genomes (AAMG) pipeline. The INDIGO web server is freely available at http://www.cbrc.kaust.edu.sa/indigo.  相似文献   

9.
Neutral community models have shown that limited migration can have a pervasive influence on the taxonomic composition of local communities even when all individuals are assumed of equivalent ecological fitness. Notably, the spatially implicit neutral theory yields a single parameter I for the immigration-drift equilibrium in a local community. In the case of plants, seed dispersal is considered as a defining moment of the immigration process and has attracted empirical and theoretical work. In this paper, we consider a version of the immigration parameter I depending on dispersal limitation from the neighbourhood of a community. Seed dispersal distance is alternatively modelled using a distribution that decreases quickly in the tails (thin-tailed Gaussian kernel) and another that enhances the chance of dispersal events over very long distances (heavily fat-tailed Cauchy kernel). Our analysis highlights two contrasting situations, where I is either mainly sensitive to community size (related to ecological drift) under the heavily fat-tailed kernel or mainly sensitive to dispersal distance under the thin-tailed kernel. We review dispersal distances of rainforest trees from field studies and assess the consistency between published estimates of I based on spatially-implicit models and the predictions of the kernel-based model in tropical forest plots. Most estimates of I were derived from large plots (10–50 ha) and were too large to be accounted for by a Cauchy kernel. Conversely, a fraction of the estimates based on multiple smaller plots (1 ha) appeared too small to be consistent with reported ranges of dispersal distances in tropical forests. Very large estimates may reflect within-plot habitat heterogeneity or estimation problems, while the smallest estimates likely imply other factors inhibiting migration beyond dispersal limitation. Our study underscores the need for interpreting I as an integrative index of migration limitation which, besides the limited seed dispersal, possibly includes habitat filtering or fragmentation.  相似文献   

10.
Neurofeedback may be useful for improving sports performance but few studies have examined this potential. Here we present data of five development players from a major league baseball team. The aims were to evaluate the feasibility of conducting sessions within a professional organization, assess changes in quantitative electroencephalograph (QEEG), NeuroPerformance Profile?, and report qualitative self-report data before and after brain training. The EEG was recorded with 19 electrodes for 20 min of baseline conditions and approximately 21 min of a continuous performance test. The fast Fourier transform analysis provided average cross-spectral matrices for bands delta (1–3.5 Hz), theta (4–7.5 Hz), alpha (8–12 Hz), low beta (13–16 Hz), beta 1 (13–21 Hz), beta 2 (22–32 Hz), and gamma (32–45 Hz) from the pre and post intervention evaluations in the baseline condition of eyes open. The continuous performance test metrics included the errors of omission, errors of commission, response time and response time variability. The 9 scales of the NeuroPerformance Profile? were examined. The QEEG data, CPT data and NeuroPerformance Profile? data were all compared between the pre and post 15 sessions of brain training using a within subject paired t test design corrected for multiple comparisons using false discovery rate method. Following brain training, comparative QEEG, CPT and NeuroPerformance Profile? analyses illustrated significant differences. The QEEG findings of all participants illustrated significant changes within the training parameters but also across other frequency bands and electrode sites. Overall, the positive findings in both objective and subjective measures suggest further inquiry into the utility of brain training for performance enhancement with the specific application of sport is warranted. Particularly QEEG and CPT gains were noted in the areas that correspond to client self-report data demonstrating improvement in attention, decreased intrusive thought patterns and improvements in sleep patterns.  相似文献   

11.
A total of 284 clinical isolates of various species of Enterobacteriaceae, Pseudomonas aeruginosa, P. maltophilia, and Acinetobacter anitratum were tested for susceptibility to carbenicillin by the standardized Bauer-Kirby disc diffusion technique and a microtiter broth dilution method. The data obtained led to the following proposed criteria for the interpretation of the results of disc susceptibility tests. Enterobacteriaceae that yield zones of inhibition equal to or greater than 20 mm in diameter around 50-mug discs of carbenicillin are designated as sensitive to the drug; isolates that yield zones measuring from 18 to 19 mm in diameter are reported as of equivocal (intermediate) susceptibility to the drug, whereas those enterobacterial isolates that are characterized by zones of inhibition of 17 mm or less in diameter are interpreted as resistant to carbenicillin. Isolates of P. aeruginosa, P. maltophilia, and A. anitratum yielding zones of 14 mm or more in diameter around 50-mug discs of carbenicillin are reported as sensitive, whereas those isolates that are characterized by zones of 13 mm or less in diameter are reported as resistant to this drug.  相似文献   

12.
Identification of differentially expressed (DE) genes across two conditions is a common task with microarray. Most existing approaches accomplish this goal by examining each gene separately based on a model and then control the false discovery rate over all genes. We took a different approach that employs a uniform platform to simultaneously depict the dynamics of the gene trajectories for all genes and select differently expressed genes. A new Functional Principal Component (FPC) approach is developed for time-course microarray data to borrow strength across genes. The approach is flexible as the temporal trajectory of the gene expressions is modeled nonparametrically through a set of orthogonal basis functions, and often fewer basis functions are needed to capture the shape of the gene expression trajectory than existing nonparametric methods. These basis functions are estimated from the data reflecting major modes of variation in the data. The correlation structure of the gene expressions over time is also incorporated without any parametric assumptions and estimated from all genes such that the information across other genes can be shared to infer one individual gene. Estimation of the parameters is carried out by an efficient hybrid EM algorithm. The performance of the proposed method across different scenarios was compared favorably in simulation to two-way mixed-effects ANOVA and the EDGE method using B-spline basis function. Application to the real data on C. elegans developmental stages also suggested that FPC analysis combined with hybrid EM algorithm provides a computationally fast and efficient method for identifying DE genes based on time-course microarray data.  相似文献   

13.
14.
Recombination is the main cause of genetic diversity. Thus, errors in this process can lead to chromosomal abnormalities. Recombination events are confined to narrow chromosome regions called hotspots in which characteristic DNA motifs are found. Genomic analyses have shown that both recombination hotspots and DNA motifs are distributed unevenly along human chromosomes and are much more frequent in the subtelomeric regions of chromosomes than in their central parts. Clusters of motifs roughly follow the distribution of recombination hotspots whereas single motifs show a negative correlation with the hotspot distribution. To model the phenomena related to recombination, we carried out computer Monte Carlo simulations of genome evolution. Computer simulations generated uneven distribution of hotspots with their domination in the subtelomeric regions of chromosomes. They also revealed that purifying selection eliminating defective alleles is strong enough to cause such hotspot distribution. After sufficiently long time of simulations, the structure of chromosomes reached a dynamic equilibrium, in which number and global distribution of both hotspots and defective alleles remained statistically unchanged, while their precise positions were shifted. This resembles the dynamic structure of human and chimpanzee genomes, where hotspots change their exact locations but the global distributions of recombination events are very similar.  相似文献   

15.
16.
This paper presents the findings of the Belmont Forum’s survey on Open Data which targeted the global environmental research and data infrastructure community. It highlights users’ perceptions of the term “open data”, expectations of infrastructure functionalities, and barriers and enablers for the sharing of data. A wide range of good practice examples was pointed out by the respondents which demonstrates a substantial uptake of data sharing through e-infrastructures and a further need for enhancement and consolidation. Among all policy responses, funder policies seem to be the most important motivator. This supports the conclusion that stronger mandates will strengthen the case for data sharing.  相似文献   

17.
As clinical and cognitive neuroscience mature, the need for sophisticated neuroimaging analysis becomes more apparent. Multivariate analysis techniques have recently received increasing attention as they have many attractive features that cannot be easily realized by the more commonly used univariate, voxel-wise, techniques. Multivariate approaches evaluate correlation/covariance of activation across brain regions, rather than proceeding on a voxel-by-voxel basis. Thus, their results can be more easily interpreted as a signature of neural networks. Univariate approaches, on the other hand, cannot directly address functional connectivity in the brain. The covariance approach can also result in greater statistical power when compared with univariate techniques, which are forced to employ very stringent, and often overly conservative, corrections for voxel-wise multiple comparisons. Multivariate techniques also lend themselves much better to prospective application of results from the analysis of one dataset to entirely new datasets. Multivariate techniques are thus well placed to provide information about mean differences and correlations with behavior, similarly to univariate approaches, with potentially greater statistical power and better reproducibility checks. In contrast to these advantages is the high barrier of entry to the use of multivariate approaches, preventing more widespread application in the community. To the neuroscientist becoming familiar with multivariate analysis techniques, an initial survey of the field might present a bewildering variety of approaches that, although algorithmically similar, are presented with different emphases, typically by people with mathematics backgrounds. We believe that multivariate analysis techniques have sufficient potential to warrant better dissemination. Researchers should be able to employ them in an informed and accessible manner. The following article attempts to provide a basic introduction with sample applications to simulated and real-world data sets.  相似文献   

18.

Background

In genetic studies of rare complex diseases it is common to ascertain familial data from population based registries through all incident cases diagnosed during a pre-defined enrollment period. Such an ascertainment procedure is typically taken into account in the statistical analysis of the familial data by constructing either a retrospective or prospective likelihood expression, which conditions on the ascertainment event. Both of these approaches lead to a substantial loss of valuable data.

Methodology and Findings

Here we consider instead the possibilities provided by a Bayesian approach to risk analysis, which also incorporates the ascertainment procedure and reference information concerning the genetic composition of the target population to the considered statistical model. Furthermore, the proposed Bayesian hierarchical survival model does not require the considered genotype or haplotype effects be expressed as functions of corresponding allelic effects. Our modeling strategy is illustrated by a risk analysis of type 1 diabetes mellitus (T1D) in the Finnish population-based on the HLA-A, HLA-B and DRB1 human leucocyte antigen (HLA) information available for both ascertained sibships and a large number of unrelated individuals from the Finnish bone marrow donor registry. The heterozygous genotype DR3/DR4 at the DRB1 locus was associated with the lowest predictive probability of T1D free survival to the age of 15, the estimate being 0.936 (0.926; 0.945 95% credible interval) compared to the average population T1D free survival probability of 0.995.

Significance

The proposed statistical method can be modified to other population-based family data ascertained from a disease registry provided that the ascertainment process is well documented, and that external information concerning the sizes of birth cohorts and a suitable reference sample are available. We confirm the earlier findings from the same data concerning the HLA-DR3/4 related risks for T1D, and also provide here estimated predictive probabilities of disease free survival as a function of age.  相似文献   

19.
Genomics is a Big Data science and is going to get much bigger, very soon, but it is not known whether the needs of genomics will exceed other Big Data domains. Projecting to the year 2025, we compared genomics with three other major generators of Big Data: astronomy, YouTube, and Twitter. Our estimates show that genomics is a “four-headed beast”—it is either on par with or the most demanding of the domains analyzed here in terms of data acquisition, storage, distribution, and analysis. We discuss aspects of new technologies that will need to be developed to rise up and meet the computational challenges that genomics poses for the near future. Now is the time for concerted, community-wide planning for the “genomical” challenges of the next decade.We compared genomics with three other major generators of Big Data: astronomy, YouTube, and Twitter. Astronomy has faced the challenges of Big Data for over 20 years and continues with ever-more ambitious studies of the universe. YouTube burst on the scene in 2005 and has sparked extraordinary worldwide interest in creating and sharing huge numbers of videos. Twitter, created in 2006, has become the poster child of the burgeoning movement in computational social science [6], with unprecedented opportunities for new insights by mining the enormous and ever-growing amount of textual data [7]. Particle physics also produces massive quantities of raw data, although the footprint is surprisingly limited since the vast majority of data are discarded soon after acquisition using the processing power that is coupled to the sensors [8]. Consequently, we do not include the domain in full detail here, although that model of rapid filtering and analysis will surely play an increasingly important role in genomics as the field matures.To compare these four disparate domains, we considered the four components that comprise the “life cycle” of a dataset: acquisition, storage, distribution, and analysis (
Data Phase Astronomy Twitter YouTube Genomics
Acquisition 25 zetta-bytes/year0.5–15 billion tweets/year500–900 million hours/year1 zetta-bases/year
Storage 1 EB/year1–17 PB/year1–2 EB/year2–40 EB/year
Analysis In situ data reductionTopic and sentiment miningLimited requirementsHeterogeneous data and analysis
Real-time processingMetadata analysisVariant calling, ~2 trillion central processing unit (CPU) hours
Massive volumesAll-pairs genome alignments, ~10,000 trillion CPU hours
Distribution Dedicated lines from antennae to server (600 TB/s)Small units of distributionMajor component of modern user’s bandwidth (10 MB/s)Many small (10 MB/s) and fewer massive (10 TB/s) data movement
Open in a separate window  相似文献   

20.
Can Replication Save Noisy Microarray Data?   总被引:1,自引:0,他引:1  
Wernisch L 《Comparative and Functional Genomics》2002,3(4):372-374
  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号