首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 46 毫秒
1.
刘山林 《生物多样性》2019,27(5):526-367
近年来DNA条形码技术迅速发展, 产生的条形码的数量及其应用范围都呈指数性增长, 现已广泛用于物种鉴定、食性分析、生物多样性评估等方面。本文重点总结并讨论了构建条形码参考数据库和序列聚类相关的信息分析的技术和方法, 包括: 基于高通量测序(high throughput sequencing, HTS)平台以高效并较低的成本获取条形码序列的方法; 同时还介绍了从原始测序序列到分类操作单元(operational taxonomic units, OTUs)过程中的一些计算逻辑以及被广泛采用的软件和技术。这是一个较新并快速发展的领域, 我们希望本文能为读者提供一个梗概, 了解DNA条形码技术在生物多样性研究应用中的方法和手段。  相似文献   

2.
The genetic analysis of faecal material represents a relatively non-invasive way to study animal diet and has been widely adopted in ecological research. Due to the heterogeneous nature of faecal material the primary obstacle, common to all genetic approaches, is a means to dissect the constituent DNA sequences. Traditionally, bacterial cloning of PCR amplified products was employed; less common has been the use of species-specific quantitative PCR (qPCR) assays. Currently, with the advent of High-Throughput Sequencing (HTS) technologies and indexed primers it has become possible to conduct genetic audits of faecal material to a much greater depth than previously possible. To date, no studies have systematically compared the estimates obtained by HTS with that of qPCR. What are the relative strengths and weaknesses of each technique and how quantitative are deep-sequencing approaches that employ universal primers? Using the locally threatened Little Penguin (Eudyptula minor) as a model organism, it is shown here that both qPCR and HTS techniques are highly correlated and produce strikingly similar quantitative estimates of fish DNA in faecal material, with no statistical difference. By designing four species-specific fish qPCR assays and comparing the data to the same four fish in the HTS data it was possible to directly compare the strengths and weaknesses of both techniques. To obtain reproducible quantitative data one of the key, and often overlooked, steps common to both approaches is ensuring that efficient DNA isolation methods are employed and that extracts are free of inhibitors. Taken together, the methodology chosen for long-term faecal monitoring programs is largely dependent on the complexity of the prey species present and the level of accuracy that is desired. Importantly, these methods should not be thought of as mutually exclusive, as the use of both HTS and qPCR in tandem will generate datasets with the highest fidelity.  相似文献   

3.
Large biological datasets are being produced at a rapid pace and create substantial storage challenges, particularly in the domain of high-throughput sequencing (HTS). Most approaches currently used to store HTS data are either unable to quickly adapt to the requirements of new sequencing or analysis methods (because they do not support schema evolution), or fail to provide state of the art compression of the datasets. We have devised new approaches to store HTS data that support seamless data schema evolution and compress datasets substantially better than existing approaches. Building on these new approaches, we discuss and demonstrate how a multi-tier data organization can dramatically reduce the storage, computational and network burden of collecting, analyzing, and archiving large sequencing datasets. For instance, we show that spliced RNA-Seq alignments can be stored in less than 4% the size of a BAM file with perfect data fidelity. Compared to the previous compression state of the art, these methods reduce dataset size more than 40% when storing exome, gene expression or DNA methylation datasets. The approaches have been integrated in a comprehensive suite of software tools (http://goby.campagnelab.org) that support common analyses for a range of high-throughput sequencing assays.  相似文献   

4.
As higher density formats become more and more common in HTS labs, the expectations for maintaining faster, lower cost screens puts great pressure on traditional 96-well screens. In some cases higher density formats are not compatible with the assay. This seems especially true in cell-based assays. In our case, the nature of the cells' response forced us to remain in 96-well plates. In this paper, we describe the development of a luminescence reporter assay and its performance in two detection modes, flash and glow. The advantages in cost and throughput for each technique are explored, along with automation considerations. An additional new technology, the use of pins for low-volume transfers, is also briefly described because of its dramatic effect on our screen's throughput. However, it will be more thoroughly presented in a future publication. Comparing the technologies available for HTS aids in designing automated systems that meet the unique needs of each assay.  相似文献   

5.
With the increasing democratization of high‐throughput sequencing (HTS) technologies, along with the concomitant increase in sequence yield per dollar, many researchers are exploring HTS for microbial community ecology. Many elements of experimental design can drastically affect the final observed community structure, notably the choice of primers for amplification prior to sequencing. Some targeted microbes can fail to amplify due to primer‐targeted sequence divergence and be omitted from obtained sequences, leading to differences among primer pairs in the sequenced organisms even when targeting the same community. This potential source of taxonomic bias in HTS makes it prudent to investigate how primer choice will affect the sequenced community prior to investing in a costly community‐wide sequencing effort. Here, we use Fluidigm's microfluidic Access Arrays (IFC) followed by Illumina® MiSeq Nano sequencing on a culture‐derived local mock community to demonstrate how this approach allows for a low‐cost combinatorial investigation of primer pairs and experimental samples (up to 48 primer pairs and 48 samples) to determine the most effective primers that maximize obtained communities whilst minimizing taxonomic biases.  相似文献   

6.
The genomic revolution has fundamentally changed how we survey biodiversity on earth. High‐throughput sequencing (“HTS”) platforms now enable the rapid sequencing of DNA from diverse kinds of environmental samples (termed “environmental DNA” or “eDNA”). Coupling HTS with our ability to associate sequences from eDNA with a taxonomic name is called “eDNA metabarcoding” and offers a powerful molecular tool capable of noninvasively surveying species richness from many ecosystems. Here, we review the use of eDNA metabarcoding for surveying animal and plant richness, and the challenges in using eDNA approaches to estimate relative abundance. We highlight eDNA applications in freshwater, marine and terrestrial environments, and in this broad context, we distill what is known about the ability of different eDNA sample types to approximate richness in space and across time. We provide guiding questions for study design and discuss the eDNA metabarcoding workflow with a focus on primers and library preparation methods. We additionally discuss important criteria for consideration of bioinformatic filtering of data sets, with recommendations for increasing transparency. Finally, looking to the future, we discuss emerging applications of eDNA metabarcoding in ecology, conservation, invasion biology, biomonitoring, and how eDNA metabarcoding can empower citizen science and biodiversity education.  相似文献   

7.
Determining the genetic factors in a disease is crucial to elucidating its molecular basis. This task is challenging due to a lack of information on gene function. The integration of large-scale functional genomics data has proven to be an effective strategy to prioritize candidate disease genes. Mitochondrial disorders are a prevalent and heterogeneous class of diseases that are particularly amenable to this approach. Here we explain the application of integrative approaches to the identification of mitochondrial disease genes. We first examine various datasets that can be used to evaluate the involvement of each gene in mitochondrial function. The data integration methodology is then described, accompanied by examples of common implementations. Finally, we discuss how gene networks are constructed using integrative techniques and applied to candidate gene prioritization. Relevant public data resources are indicated. This report highlights the success and potential of data integration as well as its applicability to the search for mitochondrial disease genes.  相似文献   

8.
Meta-analysis offers the opportunity to combine evidence from retrospectively accumulated or prospectively generated data. Meta-analyses may provide summary estimates and can help in detecting and addressing potential inconsistency between the combined datasets. Application of meta-analysis in genetic associations presents considerable potential and several pitfalls. In this review, we present basic principles of meta-analytic methods, adapted for human genome epidemiology. We describe issues that arise in the retrospective or the prospective collection of relevant data through various sources, common traps to consider in the appraisal of evidence and potential biases that may interfere. We describe the relative merits and caveats for common methods used to trace inconsistency across studies along with possible reasons for non-replication of proposed associations. Different statistical models may be employed to combine data and some common misconceptions may arise in the process. Several meta-analysis diagnostics are often applied or misapplied in the literature, and we comment on their use and limitations. An alternative to overcome limitations arising from retrospective combination of data from published studies is to create networks of research teams working in the same field and perform collaborative meta-analyses of individual participant data, ideally on a prospective basis. We discuss the advantages and the challenges inherent in such collaborative approaches. Meta-analysis can be a useful tool in dissecting the genetics of complex diseases and traits, provided its methods are properly applied and interpreted.  相似文献   

9.
10.
With the rapid advances in sequencing technology, the cost of sequencing has dramatically dropped and the scale of sequencing projects has increased accordingly. This has provided the opportunity for the routine use of sequencing techniques in the monitoring of environmental microbes. While metagenomic applications have been routinely applied to better understand the ecology and diversity of microbes, their use in environmental monitoring and bioremediation is increasingly common. In this review we seek to provide an overview of some of the metagenomic techniques used in environmental systems biology, addressing their application and limitation. We will also provide several recent examples of the application of metagenomics to bioremediation. We discuss examples where microbial communities have been used to predict the presence and extent of contamination, examples of how metagenomics can be used to characterize the process of natural attenuation by unculturable microbes, as well as examples detailing the use of metagenomics to understand the impact of biostimulation on microbial communities.  相似文献   

11.
High‐throughput sequencing (HTS) technologies generate millions of sequence reads from DNA/RNA molecules rapidly and cost‐effectively, enabling single investigator laboratories to address a variety of ‘omics’ questions in nonmodel organisms, fundamentally changing the way genomic approaches are used to advance biological research. One major challenge posed by HTS is the complexity and difficulty of data quality control (QC). While QC issues associated with sample isolation, library preparation and sequencing are well known and protocols for their handling are widely available, the QC of the actual sequence reads generated by HTS is often overlooked. HTS‐generated sequence reads can contain various errors, biases and artefacts whose identification and amelioration can greatly impact subsequent data analysis. However, a systematic survey on QC procedures for HTS data is still lacking. In this review, we begin by presenting standard ‘health check‐up’ QC procedures recommended for HTS data sets and establishing what ‘healthy’ HTS data look like. We next proceed by classifying errors, biases and artefacts present in HTS data into three major types of ‘pathologies’, discussing their causes and symptoms and illustrating with examples their diagnosis and impact on downstream analyses. We conclude this review by offering examples of successful ‘treatment’ protocols and recommendations on standard practices and treatment options. Notwithstanding the speed with which HTS technologies – and consequently their pathologies – change, we argue that careful QC of HTS data is an important – yet often neglected – aspect of their application in molecular ecology, and lay the groundwork for developing a HTS data QC ‘best practices’ guide.  相似文献   

12.
13.
High-throughput screening (HTS) has become an essential part of the drug discovery process. Due to the rising requirements for both data quality and quantity, along with increased screening cost and the demand to shorten the time for lead identification, increasing throughput and cost-effectiveness has become a necessity in the hit identification process. The authors present a multiplexed HTS for 2 nuclear receptors, the farnesoid X-activated receptor and the peroxisome proliferator-activated receptor delta in a viable cell-based reporter gene assay. The 2 nuclear receptors were individually transfected into human hepatoma cells, and the transient transfected cell lines were pooled for the multiplexed screen. Hits identified by the multiplexed screen are similar to those identified by the individual receptor screens. Furthermore, the multiplexed screen provides selectivity information if ligands selective for one and not the other receptor are one of the hit criteria. The data demonstrate that multiplexing nuclear receptors can be a simple, efficient, cost-effective, and reliable alternative to traditional HTS of individual targets without compromising data quality.  相似文献   

14.
The transition from manual to robotic high throughput screening (HTS) in the last few years has made it feasible to screen hundreds of thousands of chemical entities against a biological target in less than a month. This rate of HTS has increased the visibility of bottlenecks, one of which is assay optimization. In many organizations, experimental methods are generated by therapeutic teams associated with specific targets and passed on to the HTS group. The resulting assays frequently need to be further optimized to withstand the rigors and time frames inherent in robotic handling. Issues such as protein aggregation, ligand instability, and cellular viability are common variables in the optimization process. The availability of robotics capable of performing rapid random access tasks has made it possible to design optimization experiments that would be either very difficult or impossible for a person to carry out. Our approach to reducing the assay optimization bottleneck has been to unify the highly specific fields of statistics, biochemistry, and robotics. The product of these endeavors is a process we have named automated assay optimization (AAO). This has enabled us to determine final optimized assay conditions, which are often a composite of variables that we would not have arrived at by examining each variable independently. We have applied this approach to both radioligand binding and enzymatic assays and have realized benefits in both time and performance that we would not have predicted a priori. The fully developed AAO process encompasses the ability to download information to a robot and have liquid handling methods automatically created. This evolution in smart robotics has proven to be an invaluable tool for maintaining high-quality data in the context of increasing HTS demands.  相似文献   

15.
With the rapid accumulation of biological omics datasets, decoding the underlying relationships of cross-dataset genes becomes an important issue. Previous studies have attempted to identify differentially expressed genes across datasets. However, it is hard for them to detect interrelated ones. Moreover, existing correlation-based algorithms can only measure the relationship between genes within a single dataset or two multi-modal datasets from the same samples. It is still unclear how to quantify the strength of association of the same gene across two biological datasets with different samples. To this end, we propose Approximate Distance Correlation (ADC) to select interrelated genes with statistical significance across two different biological datasets. ADC first obtains the k most correlated genes for each target gene as its approximate observations, and then calculates the distance correlation (DC) for the target gene across two datasets. ADC repeats this process for all genes and then performs the Benjamini-Hochberg adjustment to control the false discovery rate. We demonstrate the effectiveness of ADC with simulation data and four real applications to select highly interrelated genes across two datasets. These four applications including 21 cancer RNA-seq datasets of different tissues; six single-cell RNA-seq (scRNA-seq) datasets of mouse hematopoietic cells across six different cell types along the hematopoietic cell lineage; five scRNA-seq datasets of pancreatic islet cells across five different technologies; coupled single-cell ATAC-seq (scATAC-seq) and scRNA-seq data of peripheral blood mononuclear cells (PBMC). Extensive results demonstrate that ADC is a powerful tool to uncover interrelated genes with strong biological implications and is scalable to large-scale datasets. Moreover, the number of such genes can serve as a metric to measure the similarity between two datasets, which could characterize the relative difference of diverse cell types and technologies.  相似文献   

16.
17.
Sequence capture of ultraconserved elements (UCEs) associated with massively parallel sequencing has become a common source of nuclear data for studies of animal systematics and phylogeography. However, mitochondrial and microsatellite variation are still commonly used in various kinds of molecular studies, and probably will complement genomic data in years to come. Here we show that besides providing abundant genomic data, UCE sequencing is an excellent source of both sequences for microsatellite loci design and complete mitochondrial genomes with high sequencing depth. Identification of dozens of microsatellite loci and assembly of complete mitogenomes is exemplified here using three species of Poospiza warbling finches from southern and southeastern Brazil. This strategy opens exciting opportunities to simultaneously analyze genome-wide nuclear datasets and traditionally used mtDNA and microsatellite markers in non-model amniotes at no additional cost.  相似文献   

18.
In this paper we review recently-developed extension frailty, quadratic hazard, stochastic process, microsimulation, and linear latent structure models, which have the potential to describe the health effects of human populations exposed to ionizing radiation. We discuss the most common situations for which such models are appropriate. We also provide examples of how to estimate the parameters of these models from datasets of various designs. Carcinogenesis models are reviewed in context of application to epidemiologic data of population exposed to ionizing radiation. We also discuss the ways of how to generalize stochastic process and correlated frailty models for longitudinal and family analyses in radiation epidemiology.  相似文献   

19.
Recognition of the importance of intraspecific variation in ecological processes has been growing, but empirical studies and models of global change have only begun to address this issue in detail. This review discusses sources and patterns of intraspecific trait variation and their consequences for understanding how ecological processes and patterns will respond to global change. We examine how current ecological models and theories incorporate intraspecific variation, review existing data sources that could help parameterize models that account for intraspecific variation in global change predictions, and discuss new data that may be needed. We provide guidelines on when it is most important to consider intraspecific variation, such as when trait variation is heritable or when nonlinear relationships are involved. We also highlight benefits and limitations of different model types and argue that many common modeling approaches such as matrix population models or global dynamic vegetation models can allow a stronger consideration of intraspecific trait variation if the necessary data are available. We recommend that existing data need to be made more accessible, though in some cases, new experiments are needed to disentangle causes of variation.  相似文献   

20.
Gene expression analysis applied to toxicology studies, also referred to as toxicogenomics, is rapidly being embraced by the pharmaceutical industry as a useful tool to identify safer drugs in a quicker, more cost-effective manner. Studies have already demonstrated the benefits of applying gene expression profiling towards drug safety evaluation, both for identifying mechanisms underlying toxicity, as well as for providing a means to identify safety liabilities early in the drug discovery process. Furthermore, toxicogenomics has the potential to better identify and assess adverse drug reactions of new drug candidates or marketed products in humans. While much still remains to be learned about the relevance and the application of gene expression changes in human toxicology, the next few years should see gene expression technologies applied to more stages and more programs of the drug discovery and development process. This review will focus on how toxicogenomics can or has been applied in drug discovery and development, and will discuss some of the challenges that still remain.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号