首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 93 毫秒
1.
ABSTRACT

Bioacoustics researchers can use a computer as a powerful tool to measure, classify, compare and synthetise sounds. Vocalisations on tape are commonly converted to a digital format suitable for a computer by using an analogue to digital converter and then a Fourier transformation. Alternatively, sonagrams can be measured, for example by using a digitising pad or an image analysis system. Correlations and indices of similarity have been used to compare sounds, but variations in both the time and frequency dimensions of a noise are a problem. A solution may be the use of pattern recognition methods such as elastic matching and time warping. These methods are briefly described and assessed.  相似文献   

2.
The concept of reaction similarity has been well studied in terms of the overall transformation associated with a reaction, but not in terms of mechanism. We present the first method to give a quantitative measure of the similarity of reactions based upon their explicit mechanisms. Two approaches are presented to measure the similarity between individual steps of mechanisms: a fingerprint-based approach that incorporates relevant information on each mechanistic step; and an approach based only on bond formation, cleavage and changes in order. The overall similarity for two reaction mechanisms is then calculated using the Needleman-Wunsch alignment algorithm. An analysis of MACiE, a database of enzyme mechanisms, using our measure of similarity identifies some examples of convergent evolution of chemical mechanisms. In many cases, mechanism similarity is not reflected by similarity according to the EC system of enzyme classification. In particular, little mechanistic information is conveyed by the class level of the EC system.  相似文献   

3.
The similarity transformation approach is used to analyze the structural identifiability of the parameters of a nonlinear model of microbial growth in a batch reactor in which only the concentration of microorganisms is measured. It is found that some of the model parameters are unidentifiable from this experiment, thus providing the first example of a real-life nonlinear model that turns out not to be globally identifiable. If it is possible to measure the initial concentration of growth-limiting substrate as well, all model parameters are globally identifiable.  相似文献   

4.
Comparing DNA or protein sequences plays an important role in the functional analysis of genomes. Despite many methods available for sequences comparison, few methods retain the information content of sequences. We propose a new approach, the Yau-Hausdorff method, which considers all translations and rotations when seeking the best match of graphical curves of DNA or protein sequences. The complexity of this method is lower than that of any other two dimensional minimum Hausdorff algorithm. The Yau-Hausdorff method can be used for measuring the similarity of DNA sequences based on two important tools: the Yau-Hausdorff distance and graphical representation of DNA sequences. The graphical representations of DNA sequences conserve all sequence information and the Yau-Hausdorff distance is mathematically proved as a true metric. Therefore, the proposed distance can preciously measure the similarity of DNA sequences. The phylogenetic analyses of DNA sequences by the Yau-Hausdorff distance show the accuracy and stability of our approach in similarity comparison of DNA or protein sequences. This study demonstrates that Yau-Hausdorff distance is a natural metric for DNA and protein sequences with high level of stability. The approach can be also applied to similarity analysis of protein sequences by graphic representations, as well as general two dimensional shape matching.  相似文献   

5.
A prospective approach to addressing carcinogen risk assessment is presented. Fuzzy reasoning is used to assess carcinogenic risk, characterize it, and control it. The approach is inspired by fuzzy control inference that deploys linguistic intelligence as input to a system described numerically through membership functions. Fuzzy-based reasoning to estimate carcinogenic risk provides several advantages as discussed here. The fuzzy reasoning approach has more capabilities than traditional models in dealing with risk agents that are probably carcinogens, possibly carcinogens, not classifiable as carcinogens, and probably not carcinogens. Input–output surfaces are presented for each hazard group to enable fast inferencing. Then, a hypothetical example is given to compare the results of traditional methods and the fuzzy-based approach to estimating the risk of a carcinogen to a human population. Results show similarity in risk characterization with less input information to the fuzzy-based approach. Fuzzy reasoning characterizes risk in more explicit and easy to grasp terms. Two outputs of the inferencing system are risk characterization and risk control or remediation.  相似文献   

6.
PARSIMONY, HOMOLOGY AND THE ANALYSIS OF MULTISTATE CHARACTERS   总被引:2,自引:0,他引:2  
Abstract— The order of states in a transformation series describes an internested set of synapomorphies. States adjacent to each other in the transformation series thus share a degree of homology not found in the other states. Whether the level of homology is relatively apomorphic is determined by rooting the order with outgroup comparison. The analysis of state order is a homology problem and is solved with a two-step process using similarity and congruence with other characters as criteria. Other methods that have been proposed (e.g. transformation series analysis, non-additive analysis, morphocline analysis, ontogenetic analysis) fail to apply both similarity and congruence, and thus cannot be used independently for determining character state order.  相似文献   

7.
Digital signal processing (DSP) techniques for biological sequence analysis continue to grow in popularity due to the inherent digital nature of these sequences. DSP methods have demonstrated early success for detection of coding regions in a gene. Recently, these methods are being used to establish DNA gene similarity. We present the inter-coefficient difference (ICD) transformation, a novel extension of the discrete Fourier transformation, which can be applied to any DNA sequence. The ICD method is a mathematical, alignment-free DNA comparison method that generates a genetic signature for any DNA sequence that is used to generate relative measures of similarity among DNA sequences. We demonstrate our method on a set of insulin genes obtained from an evolutionarily wide range of species, and on a set of avian influenza viral sequences, which represents a set of highly similar sequences. We compare phylogenetic trees generated using our technique against trees generated using traditional alignment techniques for similarity and demonstrate that the ICD method produces a highly accurate tree without requiring an alignment prior to establishing sequence similarity.  相似文献   

8.
9.
Body size has a dominant influence on locomotor performance and the morphology of the locomotor apparatus. In locomotion under the influence of gravity, body mass acts as weight force and is a mechanical variable. Accordingly, the application of biomechanical principles and methods allows a functional understanding of scaling effects in locomotion. This is demonstrated here using leaping primates as an example. With increasing body size, the decreasing ratio of muscle force available for acceleration during takeoff to the body mass that has to be accelerated dictates both the movement pattern and the proportions of the hindlimbs. In an arm-swinging movement, the long, heavy arms of the large-bodied leapers are effectively used to gain additional momentum. A new perspective on decreasing size identifies the absolutely small acceleration distance and time available for propulsion as factors limiting leaping distance and extensively determining locomotor behavior and body proportions. As the mechanical constraints differ according to body size for a given mode of locomotion, a typological approach to morphology in relation to locomotor category is ruled out. Across locomotor categories, dynamic similarity (sensu Alexander) can be expected if the propulsive mechanisms as well as the selective pressures acting upon locomotion are the same.  相似文献   

10.
The assignment of orthologous genes between a pair of genomes is a fundamental and challenging problem in comparative genomics, since many computational methods for solving various biological problems critically rely on bona fide orthologs as input. While it is usually done using sequence similarity search, we recently proposed a new combinatorial approach that combines sequence similarity and genome rearrangement. This paper continues the development of the approach and unites genome rearrangement events and (post-speciation) duplication events in a single framework under the parsimony principle. In this framework, orthologous genes are assumed to correspond to each other in the most parsimonious evolutionary scenario involving both genome rearrangement and (post-speciation) gene duplication. Besides several original algorithmic contributions, the enhanced method allows for the detection of inparalogs. Following this approach, we have implemented a high-throughput system for ortholog assignment on a genome scale, called MSOAR, and applied it to human and mouse genomes. As the result will show, MSOAR is able to find 99 more true orthologs than the INPARANOID program did. In comparison to the iterated exemplar algorithm on simulated data, MSOAR performed favorably in terms of assignment accuracy. We also validated our predicted main ortholog pairs between human and mouse using public ortholog assignment datasets, synteny information, and gene function classification. These test results indicate that our approach is very promising for genome-wide ortholog assignment. Supplemental material and MSOAR program are available at http://msoar.cs.ucr.edu.  相似文献   

11.
MOTIVATION: Protein structure classification has been recognized as one of the most important research issues in protein structure analysis. A substantial number of methods for the classification have been proposed, and several databases have been constructed using these methods. Since some proteins with very similar sequences may exhibit structural diversities, we have proposed PDB-REPRDB: a database of representative protein chains from the Protein Data Bank (PDB), which strategy of selection is based not only on sequence similarity but also on structural similarity. Forty-eight representative sets whose similarity criteria were predetermined were made available over the World Wide Web (WWW). However, the sets were insufficient in number to satisfy users researching protein structures by various methods. RESULT: We have improved the system for PDB-REPRDB so that the user may obtain a quick selection of representative chains from PDB. The selection of representative chains can be dynamically configured according to the user's requirement. The WWW interface provides a large degree of freedom in setting parameters, such as cut-off scores of sequence and structural similarity. This paper describes the method we use to classify chains and select the representatives in the system. We also describe the interface used to set the parameters.  相似文献   

12.
Diverse classes of proteins function through large-scale conformational changes and various sophisticated computational algorithms have been proposed to enhance sampling of these macromolecular transition paths. Because such paths are curves in a high-dimensional space, it has been difficult to quantitatively compare multiple paths, a necessary prerequisite to, for instance, assess the quality of different algorithms. We introduce a method named Path Similarity Analysis (PSA) that enables us to quantify the similarity between two arbitrary paths and extract the atomic-scale determinants responsible for their differences. PSA utilizes the full information available in 3N-dimensional configuration space trajectories by employing the Hausdorff or Fréchet metrics (adopted from computational geometry) to quantify the degree of similarity between piecewise-linear curves. It thus completely avoids relying on projections into low dimensional spaces, as used in traditional approaches. To elucidate the principles of PSA, we quantified the effect of path roughness induced by thermal fluctuations using a toy model system. Using, as an example, the closed-to-open transitions of the enzyme adenylate kinase (AdK) in its substrate-free form, we compared a range of protein transition path-generating algorithms. Molecular dynamics-based dynamic importance sampling (DIMS) MD and targeted MD (TMD) and the purely geometric FRODA (Framework Rigidity Optimized Dynamics Algorithm) were tested along with seven other methods publicly available on servers, including several based on the popular elastic network model (ENM). PSA with clustering revealed that paths produced by a given method are more similar to each other than to those from another method and, for instance, that the ENM-based methods produced relatively similar paths. PSA applied to ensembles of DIMS MD and FRODA trajectories of the conformational transition of diphtheria toxin, a particularly challenging example, showed that the geometry-based FRODA occasionally sampled the pathway space of force field-based DIMS MD. For the AdK transition, the new concept of a Hausdorff-pair map enabled us to extract the molecular structural determinants responsible for differences in pathways, namely a set of conserved salt bridges whose charge-charge interactions are fully modelled in DIMS MD but not in FRODA. PSA has the potential to enhance our understanding of transition path sampling methods, validate them, and to provide a new approach to analyzing conformational transitions.  相似文献   

13.
MOTIVATION: We consider the problem of clustering a population of Comparative Genomic Hybridization (CGH) data samples using similarity based clustering methods. A key requirement for clustering is to avoid using the noisy aberrations in the CGH samples. RESULTS: We develop a dynamic programming algorithm to identify a small set of important genomic intervals called markers. The advantage of using these markers is that the potentially noisy genomic intervals are excluded during the clustering process. We also develop two clustering strategies using these markers. The first one, prototype-based approach, maximizes the support for the markers. The second one, similarity-based approach, develops a new similarity measure called RSim and refines clusters with the aim of maximizing the RSim measure between the samples in the same cluster. Our results demonstrate that the markers we found represent the aberration patterns of cancer types well and they improve the quality of clustering significantly. AVAILABILITY: All software developed in this paper and all the datasets used are available from the authors upon request.  相似文献   

14.
Kinetics of facilitated ion transport through planar bilayer membranes are normally analyzed by electrical conductance methods. The additional use of electrical relaxation techniques, such as voltage jump, is necessary to evaluate individual rate constants. Although electrochemical impedance spectroscopy is recognized as the most powerful of the available electric relaxation techniques, it has rarely been used in connection with these kinetic studies. According to the new approach presented in this work, three steps were followed. First, a kinetic model was proposed that has the distinct quality of being general, i.e., it properly describes both carrier and channel mechanisms of ion transport. Second, the state equations for steady-state and for impedance experiments were derived, exhibiting the input–output representation pertaining to the model’s structure. With the application of a method based on the similarity transformation approach, it was possible to check that the proposed mechanism is distinguishable, i.e., no other model with a different structure exhibits the same input–output behavior for any input as the original. Additionally, the method allowed us to check whether the proposed model is globally identifiable (i.e., whether there is a single set of fit parameters for the model) when analyzed in terms of its impedance response. Thus, our model does not represent a theoretical interpretation of the experimental impedance but rather constitutes the prerequisite to select this type of experiment in order to obtain optimal kinetic identification of the system. Finally, impedance measurements were performed and the results were fitted to the proposed theoretical model in order to obtain the kinetic parameters of the system. The successful application of this approach is exemplified with results obtained for valinomycin–K+ in lipid bilayers supported onto gold substrates, i.e., an arrangement capable of emulating biological membranes.  相似文献   

15.
MOTIVATION: Partial order alignment (POA) has been proposed as a new approach to multiple sequence alignment (MSA), which can be combined with existing methods such as progressive alignment. This is important for addressing problems both in the original version of POA (such as order sensitivity) and in standard progressive alignment programs (such as information loss in complex alignments, especially surrounding gap regions). RESULTS: We have developed a new Partial Order-Partial Order alignment algorithm that optimally aligns a pair of MSAs and which therefore can be applied directly to progressive alignment methods such as CLUSTAL. Using this algorithm, we show the combined Progressive POA alignment method yields results comparable with the best available MSA programs (CLUSTALW, DIALIGN2, T-COFFEE) but is far faster. For example, depending on the level of sequence similarity, aligning 1000 sequences, each 500 amino acids long, took 15 min (at 90% average identity) to 44 min (at 30% identity) on a standard PC. For large alignments, Progressive POA was 10-30 times faster than the fastest of the three previous methods (CLUSTALW). These data suggest that POA-based methods can scale to much larger alignment problems than possible for previous methods. AVAILABILITY: The POA source code is available at http://www.bioinformatics.ucla.edu/poa  相似文献   

16.
Extensive data from multilocus electrophoresis are available for many bacterial populations. In some cases, for example Neisseria gonorrhoeae, these data are consistent with the population being in linkage equilibrium. This raises the following question. What frequency of transformation, or other means of genetic recombination, is needed, relative to mutation, to produce apparent panmixis? Simulation of a finite-population model suggests that, if transformation is at least twenty times as frequent as mutation, the population structure will be indistinguishable from a panmictic one, using the best available data sets. That is, relatively infrequent transformation is sufficient to produce approximate linkage equilibrium.  相似文献   

17.
Ecologists often use dispersion metrics and statistical hypothesis testing to infer processes of community formation such as environmental filtering, competitive exclusion, and neutral species assembly. These metrics have limited power in inferring assembly models because they rely on often‐violated assumptions. Here, we adapt a model of phenotypic similarity and repulsion to simulate the process of community assembly via environmental filtering and competitive exclusion, all while parameterizing the strength of the respective ecological processes. We then use random forests and approximate Bayesian computation to distinguish between these models given the simulated data. We find that our approach is more accurate than using dispersion metrics and accounts for uncertainty in model selection. We also demonstrate that the parameter determining the strength of the assembly processes can be accurately estimated. This approach is available in the R package CAMI; Community Assembly Model Inference. We demonstrate the effectiveness of CAMI using an example of plant communities living on lava flow islands.  相似文献   

18.
Transformation and normalization of oligonucleotide microarray data   总被引:3,自引:0,他引:3  
MOTIVATION: Most methods of analyzing microarray data or doing power calculations have an underlying assumption of constant variance across all levels of gene expression. The most common transformation, the logarithm, results in data that have constant variance at high levels but not at low levels. Rocke and Durbin showed that data from spotted arrays fit a two-component model and Durbin, Hardin, Hawkins, and Rocke, Huber et al. and Munson provided a transformation that stabilizes the variance as well as symmetrizes and normalizes the error structure. We wish to evaluate the applicability of this transformation to the error structure of GeneChip microarrays. RESULTS: We demonstrate in an example study a simple way to use the two-component model of Rocke and Durbin and the data transformation of Durbin, Hardin, Hawkins and Rocke, Huber et al. and Munson on Affymetrix GeneChip data. In addition we provide a method for normalization of Affymetrix GeneChips simultaneous with the determination of the transformation, producing a data set without chip or slide effects but with constant variance and with symmetric errors. This transformation/normalization process can be thought of as a machine calibration in that it requires a few biologically constant replicates of one sample to determine the constant needed to specify the transformation and normalize. It is hypothesized that this constant needs to be found only once for a given technology in a lab, perhaps with periodic updates. It does not require extensive replication in each study. Furthermore, the variance of the transformed pilot data can be used to do power calculations using standard power analysis programs. AVAILABILITY: SPLUS code for the transformation/normalization for four replicates is available from the first author upon request. A program written in C is available from the last author.  相似文献   

19.
Reconstructing the evolutionary history of protein sequences will provide a better understanding of divergence mechanisms of protein superfamilies and their functions. Long-term protein evolution often includes dynamic changes such as insertion, deletion, and domain shuffling. Such dynamic changes make reconstructing protein sequence evolution difficult and affect the accuracy of molecular evolutionary methods, such as multiple alignments and phylogenetic methods. Unfortunately, currently available simulation methods are not sufficiently flexible and do not allow biologically realistic dynamic protein sequence evolution. We introduce a new method, indel-Seq-Gen (iSG), that can simulate realistic evolutionary processes of protein sequences with insertions and deletions (indels). Unlike other simulation methods, iSG allows the user to simulate multiple subsequences according to different evolutionary parameters, which is necessary for generating realistic protein families with multiple domains. iSG tracks all evolutionary events including indels and outputs the "true" multiple alignment of the simulated sequences. iSG can also generate a larger sequence space by allowing the use of multiple related root sequences. With all these functions, iSG can be used to test the accuracy of, for example, multiple alignment methods, phylogenetic methods, evolutionary hypotheses, ancestral protein reconstruction methods, and protein family classification methods. We empirically evaluated the performance of iSG against currently available methods by simulating the evolution of the G protein-coupled receptor and lipocalin protein families. We examined their true multiple alignments, reconstruction of the transmembrane regions and beta-strands, and the results of similarity search against a protein database using the simulated sequences. We also presented an example of using iSG for examining how phylogenetic reconstruction is affected by high indel rates.  相似文献   

20.
微束激光转基因技术研究进展   总被引:9,自引:3,他引:6  
本文叙述了激光微速穿刺法导入外源DNA的基本原理及其有关影响因素等。利用该转化方法,现已成功地获得了多种动植物细胞外源基因的转化。实验证明,激光微束穿刺转化技术简便有效、重复性好、靶体选择性强,对靶体无损伤等优点。随着转化技术自动化水平的提高。光镊技术的渗透和结合,激光微束转化技术将会得到更广泛的应用  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号