首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
Roark DE 《Biophysical chemistry》2004,108(1-3):121-126
Biophysical chemistry experiments, such as sedimentation-equilibrium analyses, require computational techniques to reduce the effects of random errors of the measurement process. The existing approaches have primarily relied on assumption of polynomial models and least-squares approximation. Such models by constraining the data to remove random fluctuations may distort the data and cause loss of information. The better the removal of random errors the greater is the likely introduction of systematic errors through the constraining fit itself. An alternative technique, reverse smoothing, is suggested that makes use of a more model-free approach of exponential smoothing of the first derivative. Exponential smoothing approaches have been generally unsatisfactory because they introduce significant data lag. The approaches given here compensates for the lag defect and appears promising for the smoothing of many experimental data sequences, including the macromolecular concentration data generated by sedimentation-equilibria experiments. Test results on simulated sedimentation-equilibrium data indicate that a 4-fold reduction in error may be typical over standard analyses techniques.  相似文献   

2.
Electronic Data Storage Tags (DSTs) have the potential of providing long term, high-resolution observation data of individual fish in the ecosystem. However, traditional time series analysis methods are limited in extracting the required information from DSTs. The continuous wavelet transform (CWT) is a tool for decomposing a time series into instantaneous frequency versus time characteristics. This allows data signals, which are restricted to specific frequencies (localized frequencies), or frequency components, which are restricted to specific time intervals (time localized) to be extracted and displayed in a highly intuitive way. The aim of this short communication is to present the CWT as a tool for extracting information from DST data. In the first case, synthetically generated data is used to demonstrate how the methodology captures different time–frequency dependent patterns in the time series data. The relevance of the methodology to real field data is demonstrated by an analysis of DST depth data from tagged Barents Sea coastal cod. This work has also enjoyed partial funding from the EU-FP5-Q5RS200200813 CODYSSEY project.  相似文献   

3.
‘Endpoint error’ describes the erratic behavior at the beginning and end of the computed acceleration data which is commonly observed after smoothing and differentiating raw displacement data. To evaluate endpoint error produced by four popular smoothing and differentiating techniques, Lanshammar's (1982, J. Biomechanics 15, 99–105) modification of the Pezzack et al. (1977, J. Biomechanics, 10, 377–382) raw angular displacement data set was truncated at three different locations corresponding to the major peaks in the criterion acceleration curve. Also, for each data subset, three padding conditions were applied. Each data subset was smoothed and differentiated using the Butterworth digital filter, cubic spline, quintic spline, and Fourier series to obtain acceleration values. RMS residual errors were calculated between the computed and criterion accelerations in the endpoint regions. Although no method completely eliminated endpoint error, the results demonstrated clear superiority of the quintic spline over the other three methods in producing accurate acceleration values close to the endpoints of the modified Pezzack et al. (1977) data set. In fact, the quintic spline performed best with non-padded data (cumulative error=48.0 rad s−2). Conversely, when applied to non-padded data, the Butterworth digital filter produced wildly deviating values beginning more than the 10 points from the terminal data point (cumulative error=226.6 rad s−2). Each of the four methods performed better when applied to data subsets padded by linear extrapolation (average cumulative error=68.8 rad s−2) than when applied to analogous subsets padded by reflection (average cumulative error=86.1 rad s−2).  相似文献   

4.
A correlation function that compares each base in a DNA sequence to its various neighbours and which is subsequently processed by Fourier and wavelet transforms has been developed. The procedure has been applied to sequences from the human chromosome 22, to nef genes from various HIV clones and to myosin heavy chain DNA. It permits to readily visualize regular features in DNA which are related to the stability of heteroduplexes formed upon strand slippage.  相似文献   

5.
We present a method of data reduction using a wavelet transform in discriminant analysis when the number of variables is much greater than the number of observations. The method is illustrated with a prostate cancer study, where the sample size is 248, and the number of variables is 48,538 (generated using the ProteinChip technology). Using a discrete wavelet transform, the 48,538 data points are represented by 1271 wavelet coefficients. Information criteria identified 11 of the 1271 wavelet coefficients with the highest discriminatory power. The linear classifier with the 11 wavelet coefficients detected prostate cancer in a separate test set with a sensitivity of 97% and specificity of 100%.  相似文献   

6.
Force fluctuations recorded in an atomic force spectroscopy experiment, during the approach of a tip functionalized with biotin towards a substrate charged with avidin, have been analyzed by a wavelet transform. The observation of strong transient changes only when a specific biorecognition process between the partners takes place suggests a drastic modulation of the force fluctuations when biomolecules recognize each other. Such an analysis allows to investigate the peculiar features of a biorecognition process. These results are discussed in connection with the possible role of energy minima explored by biomolecules during the biorecognition process.  相似文献   

7.
We describe a PCA-based genome scan approach to analyze genome-wide admixture structure, and introduce wavelet transform analysis as a method for estimating the time of admixture. We test the wavelet transform method with simulations and apply it to genome-wide SNP data from eight admixed human populations. The wavelet transform method offers better resolution than existing methods for dating admixture, and can be applied to either SNP or sequence data from humans or other species.  相似文献   

8.
Assessment of neuromuscular fatigue is essential for early detection and prevention of risks associated with work-related musculoskeletal disorders. In recent years, discrete wavelet transform (DWT) of surface electromyography (SEMG) has been used to evaluate muscle fatigue, especially during dynamic contractions when the SEMG signal is non-stationary. However, its application to the assessment of work-related neck and shoulder muscle fatigue is not well established. Therefore, the purpose of this study was to establish DWT analysis as a suitable method to conduct quantitative assessment of neck and shoulder muscle fatigue under dynamic repetitive conditions. Ten human participants performed 40 min of fatiguing repetitive arm and neck exertions while SEMG data from the upper trapezius and sternocleidomastoid muscles were recorded. The ten of the most commonly used wavelet functions were used to conduct the DWT analysis. Spectral changes estimated using power of wavelet coefficients in the 12–23 Hz frequency band showed the highest sensitivity to fatigue induced by the dynamic repetitive exertions. Although most of the wavelet functions tested in this study reasonably demonstrated the expected power trend with fatigue development and recovery, the overall performance of the “Rbio3.1” wavelet in terms of power estimation and statistical significance was better than the remaining nine wavelets.  相似文献   

9.
Analysis of phonocardiogram (PCG) signals provides a non-invasive means to determine the abnormalities caused by cardiovascular system pathology. In general, time-frequency representation (TFR) methods are used to study the PCG signal because it is one of the non-stationary bio-signals. The continuous wavelet transform (CWT) is especially suitable for the analysis of non-stationary signals and to obtain the TFR, due to its high resolution, both in time and in frequency and has recently become a favourite tool. It decomposes a signal in terms of elementary contributions called wavelets, which are shifted and dilated copies of a fixed mother wavelet function, and yields a joint TFR. Although the basic characteristics of the wavelets are similar, each type of the wavelets produces a different TFR. In this study, eight real types of the most known wavelets are examined on typical PCG signals indicating heart abnormalities in order to determine the best wavelet to obtain a reliable TFR. For this purpose, the wavelet energy and frequency spectrum estimations based on the CWT and the spectra of the chosen wavelets were compared with the energy distribution and the autoregressive frequency spectra in order to determine the most suitable wavelet. The results show that Morlet wavelet is the most reliable wavelet for the time-frequency analysis of PCG signals.  相似文献   

10.
A database independent search algorithm for the detection of phosphopeptides is described. The program interrogates the tandem mass spectra of LC-MS/MS data sets regarding the presence of phosphorylation specific signatures. To achieve maximum informational content, the complementary fragmentation techniques electron capture dissociation (ECD) and collisionally activated dissociation (CAD) are used independently for peptide fragmentation. Several criteria characteristic for peptides phosphorylated on either serine or threonine residues were evaluated. The final algorithm searches for product ions generated by either the neutral loss of phosphoric acid or the combined neutral loss of phosphoric acid and water. Various peptide mixtures were used to evaluate the program. False positive results were not observed because the program utilizes the parts-per-million mass accuracy of Fourier transform ion cyclotron resonance mass spectrometry. Additionally, false negative results were not generated owing to the high sensitivity of the chosen criteria. The limitations of database dependent data interpretation tools are discussed and the potential of the novel algorithm to overcome these limitations is illustrated.  相似文献   

11.
The B-value: a tool for monitoring data   总被引:2,自引:0,他引:2  
K K Lan  J Wittes 《Biometrics》1988,44(2):579-585
This paper considers the problem of monitoring slowly accruing data from a nonsequentially designed experiment. We describe the use of the B-value, which is a transformed Z-value, for the calculation of conditional power. In data monitoring, interim Z-values do not allow simple projections to the end of the study. Moreover, because of their popular association with P-values, Z-values are often misinterpreted. If observed trends are viewed as the realization of a Brownian motion process, the B-value and its decomposition allow simple extrapolations to the end of the study under a variety of hypotheses. Applications are presented to one- and two-sample Z-tests, the two-sample Wilcoxon rank sum test, and the log-rank test.  相似文献   

12.
Pedro is a Java application that dynamically generates data entry forms for data models expressed in XML Schema, producing XML data files that validate against this schema. The software uses an intuitive tree-based navigation system, can supply context-sensitive help to users and features a sophisticated interface for populating data fields with terms from controlled vocabularies. The software also has the ability to import records from tab delimited text files and features various validation routines. AVAILABILITY: The application, source code, example models from several domains and tutorials can be downloaded from http://pedro.man.ac.uk/.  相似文献   

13.
One of the central problems in bioinformatics is data retrieval and integration. The existing biological databases are geographically distributed across the Internet, complex and heterogeneous in data types and data structures, and constantly changing. With the current rapid growth of biomedical data, the challenge is how large volumes of data retrieved from multiple databases can be transformed and integrated automatically and flexibly. This article describes a powerful new tool, the Kleisli system, for complex queries across multiple databases and data integration.  相似文献   

14.
15.
The goal of this article is to model multisubject task‐induced functional magnetic resonance imaging (fMRI) response among predefined regions of interest (ROIs) of the human brain. Conventional approaches to fMRI analysis only take into account temporal correlations, but do not rigorously model the underlying spatial correlation due to the complexity of estimating and inverting the high dimensional spatio‐temporal covariance matrix. Other spatio‐temporal model approaches estimate the covariance matrix with the assumption of stationary time series, which is not always feasible. To address these limitations, we propose a double‐wavelet approach for modeling the spatio‐temporal brain process. Working with wavelet coefficients simplifies temporal and spatial covariance structure because under regularity conditions, wavelet coefficients are approximately uncorrelated. Different wavelet functions were used to capture different correlation structures in the spatio‐temporal model. The main advantages of the wavelet approach are that it is scalable and that it deals with nonstationarity in brain signals. Simulation studies showed that our method could reduce false‐positive and false‐negative rates by taking into account spatial and temporal correlations simultaneously. We also applied our method to fMRI data to study activation in prespecified ROIs in the prefontal cortex. Data analysis showed that the result using the double‐wavelet approach was more consistent than the conventional approach when sample size decreased.  相似文献   

16.
We describe the software, STEM (STrategic Extractor for Mascot's results), which efficiently processes large-scale mass spectrometry-based proteomics data. V (View)-mode evaluates the Mascot peptide identification dataset, removes unreliable candidates and redundant assignments, and integrates the results with key information in the experiment. C (Comparison)-mode compares peptide coverage among multiple datasets and displays proteins commonly/specifically found therein, and processes data for quantitative studies that utilize conventional isotope tags or tags having a smaller mass difference. STEM significantly improves throughput of proteomics study.  相似文献   

17.

Background

The study of nuclear architecture using Chromosome Conformation Capture (3C) technologies is a novel frontier in biology. With further reduction in sequencing costs, the potential of Hi-C in describing nuclear architecture as a phenotype is only about to unfold. To use Hi-C for phenotypic comparisons among different cell types, conditions, or genetic backgrounds, Hi-C data processing needs to be more accessible to biologists.

Results

HiCdat provides a simple graphical user interface for data pre-processing and a collection of higher-level data analysis tools implemented in R. Data pre-processing also supports a wide range of additional data types required for in-depth analysis of the Hi-C data (e.g. RNA-Seq, ChIP-Seq, and BS-Seq).

Conclusions

HiCdat is easy-to-use and provides solutions starting from aligned reads up to in-depth analyses. Importantly, HiCdat is focussed on the analysis of larger structural features of chromosomes, their correlation to genomic and epigenomic features, and on comparative studies. It uses simple input and output formats and can therefore easily be integrated into existing workflows or combined with alternative tools.

Electronic supplementary material

The online version of this article (doi:10.1186/s12859-015-0678-x) contains supplementary material, which is available to authorized users.  相似文献   

18.

Background  

Analysis of High Throughput (HTP) Data such as microarray and proteomics data has provided a powerful methodology to study patterns of gene regulation at genome scale. A major unresolved problem in the post-genomic era is to assemble the large amounts of data generated into a meaningful biological context. We have developed a comprehensive software tool, WholePathwayScope (WPS), for deriving biological insights from analysis of HTP data.  相似文献   

19.
CLIP-seq is widely used to study genome-wide interactions between RNA-binding proteins and RNAs. However, there are few tools available to analyze CLIP-seq data, thus creating a bottleneck to the implementation of this methodology. Here, we present PIPE-CLIP, a Galaxy framework-based comprehensive online pipeline for reliable analysis of data generated by three types of CLIP-seq protocol: HITS-CLIP, PAR-CLIP and iCLIP. PIPE-CLIP provides both data processing and statistical analysis to determine candidate cross-linking regions, which are comparable to those regions identified from the original studies or using existing computational tools. PIPE-CLIP is available at http://pipeclip.qbrc.org/.  相似文献   

20.

Background  

Trace or chromatogram files (raw data) are produced by automatic nucleic acid sequencing equipment or sequencers. Each file contains information which can be interpreted by specialised software to reveal the sequence (base calling). This is done by the sequencer proprietary software or publicly available programs. Depending on the size of a sequencing project the number of trace files can vary from just a few to thousands of files. Sequencing quality assessment on various criteria is important at the stage preceding clustering and contig assembly. Two major publicly available packages – Phred and Staden are used by preAssemble to perform sequence quality processing.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号