首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
Dryden IL  Walker G 《Biometrics》1999,55(3):820-825
In many disciplines, it is of great importance to match objects. Procrustes analysis is a popular method for comparing labeled point configurations based on a least squares criterion. We consider alternative procedures that are highly resistant to outlier points, and we describe an application in electrophoretic gel matching. We consider procedures based on S estimators, least median of squares, and least quartile difference estimators. Practical implementation issues are discussed, including random subset selection and intelligent subset selection (where subsets with small size or near collinear subsets are ignored). The relative performances of the resistant and Procrustes methods are examined in a simulation study.  相似文献   

2.
Abstract

The problems of evaluating results based on the analysis of complex binding models are considered and new methods are proposed to provide independent confirmation of the existence of multiple sites. A new plotting format for the results from experiments involving two ligands is introduced, and its utility is demonstrated for a) finding initial estimates for nonlinear least squares curve fitting; b) presenting the results of multiple experiments; and c) giving a new means for evaluating the significance of a third site. The general problem of finding initial estimates for models involving three classes of sites, and strategies for using nonlinear least squares curve fitting algorithms to optimize the fit are considered.  相似文献   

3.
In quantitative biology, observed data are fitted to a model that captures the essence of the system under investigation in order to obtain estimates of the parameters of the model, as well as their standard errors and interactions. The fitting is best done by the method of maximum likelihood, though least-squares fits are often used as an approximation because the calculations are perceived to be simpler. Here Brian Williams and Chris Dye argue that the method of maximum likelihood is generally preferable to least squares giving the best estimates of the parameters for data with any given error distribution, and the calculations are no more difficult than for least-squares fitting. They offer a relatively simple explanation of the methods and describe its implementation using examples from leishmaniasis epidemiology.  相似文献   

4.
A method for resolving an overlapped polypeptide pattern of sodium dodecyl sulfate-polyacrylamide gel electrophoresis was described. The procedure was essentially a Gaussian fitting using the least squares method, and could resolve more than 20 overlapped components simultaneously. The applicability to overlapped and shouldered patterns was evaluated using practical electrophoretic data with varying amounts of mitochondrial samples. The relative contents of respective polypeptide components gave a good agreement regardless of the loaded amounts.  相似文献   

5.
A program was written to perform a linear least squares curve fitting on data. It includes facilities to report the usual statistics and digital plotter output. Seven types of curves are available for fitting the data. Other features of LILLY include provision of facilities for the selection of subsets in different symbols and separate curve fitting for these subsets. The program also provides a confidence region about the fitted line and the prediction interval for data points. Examples of the use of the program are described.  相似文献   

6.
Making sound proteomic inferences using ELISA microarray assay requires both an accurate prediction of protein concentration and a credible estimate of its error. We present a method using monotonic spline statistical models (MS), penalized constrained least squares fitting (PCLS) and Monte Carlo simulation (MC) to predict ELISA microarray protein concentrations and estimate their prediction errors. We contrast the MSMC (monotone spline Monte Carlo) method with a LNLS (logistic nonlinear least squares) method using simulated and real ELISA microarray data sets.MSMC rendered good fits in almost all tests, including those with left and/or right clipped standard curves. MS predictions were nominally more accurate; especially at the extremes of the prediction curve. MC provided credible asymmetric prediction intervals for both MS and LN fits that were superior to LNLS propagation-of-error intervals in achieving the target statistical confidence. MSMC was more reliable when automated prediction across simultaneous assays was applied routinely with minimal user guidance.  相似文献   

7.
Relaxation dispersion spectroscopy is one of the most widely used techniques for the analysis of protein dynamics. To obtain a detailed understanding of the protein function from the view point of dynamics, it is essential to fit relaxation dispersion data accurately. The grid search method is commonly used for relaxation dispersion curve fits, but it does not always find the global minimum that provides the best-fit parameter set. Also, the fitting quality does not always improve with increase of the grid size although the computational time becomes longer. This is because relaxation dispersion curve fitting suffers from a local minimum problem, which is a general problem in non-linear least squares curve fitting. Therefore, in order to fit relaxation dispersion data rapidly and accurately, we developed a new fitting program called GLOVE that minimizes global and local parameters alternately, and incorporates a Monte-Carlo minimization method that enables fitting parameters to pass through local minima with low computational cost. GLOVE also implements a random search method, which sets up initial parameter values randomly within user-defined ranges. We demonstrate here that the combined use of the three methods can find the global minimum more rapidly and more accurately than grid search alone.  相似文献   

8.
Hou T  Zhang W  Wang J  Wang W 《Proteins》2009,74(4):837-846
Drug resistance significantly impairs the efficacy of AIDS therapy. Therefore, precise prediction of resistant viral mutants is particularly useful for developing effective drugs and designing therapeutic regimen. In this study, we applied a structure-based computational approach to predict mutants of the HIV-1 protease resistant to the seven FDA approved drugs. We analyzed the energetic pattern of the protease-drug interaction by calculating the molecular interaction energy components (MIECs) between the drug and the protease residues. Support vector machines (SVMs) were trained on MIECs to classify protease mutants into resistant and nonresistant categories. The high prediction accuracies for the test sets of cross-validations suggested that the MIECs successfully characterized the interaction interface between drugs and the HIV-1 protease. We conducted a proof-of-concept study on a newly approved drug, darunavir (TMC114), on which no drug resistance data were available in the public domain. Compared with amprenavir, our analysis suggested that darunavir might be more potent to combat drug resistance. To quantitatively estimate binding affinities of drugs and study the contributions of protease residues to causing resistance, linear regression models were trained on MIECs using partial least squares (PLS). The MIEC-PLS models also achieved satisfactory prediction accuracy. Analysis of the fitting coefficients of MIECs in the regression model revealed the important resistance mutations and shed light into understanding the mechanisms of these mutations to cause resistance. Our study demonstrated the advantages of characterizing the protease-drug interaction using MIECs. We believe that MIEC-SVM and MIEC-PLS can help design new agents or combination of therapeutic regimens to counter HIV-1 protease resistant strains.  相似文献   

9.
Two methods are described for fitting the Michaelis-Menten equation to sets of data with a common michaelis constant but different maximum velocities. One of them uses the method of least squares, and the other is based on the direct plot of Eisenthal & Cornish-Bowden [Biochem. J. (1974) 139, 715-720].  相似文献   

10.
A method is described for fitting a 'fraction labelled mitoses'curve to a set of data points and for estimating the values of the best fitting parameters of the cell cycle. Estimates of the SE of the parameters are obtained. The method depends on the fact that when gamma distributions are used to describe the durations of the phases of the cell cycle, the Laplace transform of a FLM curve can be described by simple analytic functions enabling a least squares fit to be made to a set of Laplace transforms of the experimental data. The method is easy to program and quick to execute.  相似文献   

11.
A hybrid analysis that combines the maximum entropy method (MEM) with nonlinear least squares (NLS) fitting has been developed to interpret a general time-dependent signal. Data that include processes of opposite sign and a slow baseline drift can be inverted to obtain both a continuous distribution of lifetimes and a sum of discrete exponentials. Fits by discrete exponentials are performed with initial parameters determined from the distribution of lifetimes obtained with the MEM. The regularization of the parameter space achieved by the MEM stabilizes the introduction of each successive exponential in the NLS fits. This hybrid approach is particularly useful when fitting by a large number of exponentials. Revision of the MEM "prior" based on features in the data can improve the lifetime distribution obtained. Standard errors in the mean are estimated automatically for raw data. The results presented for simulated data and for fluorescence measurements of protein folding illustrate the utility and accuracy of the hybrid algorithm. Analysis of the folding of dihydrofolate reductase reveals six kinetic processes, one more than previously reported.  相似文献   

12.
A computational method is presented for minimizing the weighted sum of squares of the differences between observed and expected pairwise distances between species, where the expectations are generated by an additive tree model. The criteria of Fitch and Margoliash (1967, Science 155:279-284) and Cavalli-Sforza and Edwards (1967, Evolution 21:550-570) are both weighted least squares, with different weights. The method presented iterates lengths of adjacent branches in the tree three at a time. The weighted sum of squares never increases during the process of iteration, and the iterates approach a stationary point on the surface of the sum of squares. This iterative approach makes it particularly easy to maintain the constraint that branch lengths never become negative, although negative branch lengths can also be allowed. The method is implemented in a computer program, FITCH, which has been distributed since 1982 as part of the PHYLIP package of programs for inferring phylogenies, and is also implemented in PAUP*. The present method is compared, using some simulated data sets, with an implementation of the method of De Soete (1983, Psychometrika 48:621-626); it is slower than De Soete's method but more effective at finding the least squares tree. The relationship of this method to the neighbor-joining method is also discussed.  相似文献   

13.
The receiver operating characteristic (ROC) curve is a popular tool to evaluate and compare the accuracy of diagnostic tests to distinguish the diseased group from the nondiseased group when test results from tests are continuous or ordinal. A complicated data setting occurs when multiple tests are measured on abnormal and normal locations from the same subject and the measurements are clustered within the subject. Although least squares regression methods can be used for the estimation of ROC curve from correlated data, how to develop the least squares methods to estimate the ROC curve from the clustered data has not been studied. Also, the statistical properties of the least squares methods under the clustering setting are unknown. In this article, we develop the least squares ROC methods to allow the baseline and link functions to differ, and more importantly, to accommodate clustered data with discrete covariates. The methods can generate smooth ROC curves that satisfy the inherent continuous property of the true underlying curve. The least squares methods are shown to be more efficient than the existing nonparametric ROC methods under appropriate model assumptions in simulation studies. We apply the methods to a real example in the detection of glaucomatous deterioration. We also derive the asymptotic properties of the proposed methods.  相似文献   

14.
Procedures for comparing samples with multiple endpoints   总被引:18,自引:0,他引:18  
P C O'Brien 《Biometrics》1984,40(4):1079-1087
Five procedures are considered for the comparison of two or more multivariate samples. These procedures include a newly proposed nonparametric rank-sum test and a generalized least squares test. Also considered are the following tests: ordinary least squares, Hotelling's T2, and a Bonferroni per-experiment error-rate approach. Applications are envisaged in which each variable represents a qualitatively different measure of response to treatment. The null hypothesis of no treatment difference is tested with power directed towards alternatives in which at least one treatment is uniformly better than the others. In all simulations the nonparametric procedure provided relatively good power and accurate control over the size of the test, and is recommended for general use. Alternatively, the generalized least squares procedure may also be useful with normally distributed data in moderate or large samples. A convenient expression for this procedure is obtained and its asymptotic relative efficiency with respect to the ordinary least squares test is evaluated.  相似文献   

15.
R K Misra  M D Easton 《Cytometry》1999,36(2):112-116
BACKGROUND: The coefficient of variation (CV) is often used to characterize and summarize the flow cytometry analysis of nuclear DNA of the Go/G1 peak in a cell population within an individual organism. CV values are frequently used in subsequent statistical analysis to compare experimental groups of individuals. METHODS: We explain why the conventional analysis of variance, linear comparisons and regressions that employ the F and t-tests are not appropriate for analyzing CV data sets. The weighted least squares procedure which utilizes the chi-square test is presented as an adequate method. We further explain why this type of data needs to be analyzed by this procedure. RESULTS: To illustrate the application of the weighted least squares procedure, we analyzed a real data set that had been previously analyzed by conventional methods. We found that a non-significant result (p = 1) using the latter was significant when re-analyzed with the weighted least squares procedure (p = 0.032). CONCLUSIONS: Significant differences between treatments established by the weighted least squares often go unidentified by the conventional analysis. Use of the weighted least squares procedure is recommended for analyzing CV data sets.  相似文献   

16.
Estimating the position of the bones from optical motion capture data is a challenge associated with human movement analysis. Bone pose estimation techniques such as the Point Cluster Technique (PCT) and simulations of movement through software packages such as OpenSim are used to minimize soft tissue artifact and estimate skeletal position; however, using different methods for analysis may produce differing kinematic results which could lead to differences in clinical interpretation such as a misclassification of normal or pathological gait. This study evaluated the differences present in knee joint kinematics as a result of calculating joint angles using various techniques. We calculated knee joint kinematics from experimental gait data using the standard PCT, the least squares approach in OpenSim applied to experimental marker data, and the least squares approach in OpenSim applied to the results of the PCT algorithm. Maximum and resultant RMS differences in knee angles were calculated between all techniques. We observed differences in flexion/extension, varus/valgus, and internal/external rotation angles between all approaches. The largest differences were between the PCT results and all results calculated using OpenSim. The RMS differences averaged nearly 5° for flexion/extension angles with maximum differences exceeding 15°. Average RMS differences were relatively small (< 1.08°) between results calculated within OpenSim, suggesting that the choice of marker weighting is not critical to the results of the least squares inverse kinematics calculations. The largest difference between techniques appeared to be a constant offset between the PCT and all OpenSim results, which may be due to differences in the definition of anatomical reference frames, scaling of musculoskeletal models, and/or placement of virtual markers within OpenSim. Different methods for data analysis can produce largely different kinematic results, which could lead to the misclassification of normal or pathological gait. Improved techniques to allow non-uniform scaling of generic models to more accurately reflect subject-specific bone geometries and anatomical reference frames may reduce differences between bone pose estimation techniques and allow for comparison across gait analysis platforms.  相似文献   

17.
Analysis of sedimentation velocity data for indefinite self-associating systems is often achieved by fitting of weight average sedimentation coefficients (s(20,w)) However, this method discriminates poorly between alternative models of association and is biased by the presence of inactive monomers and irreversible aggregates. Therefore, a more robust method for extracting the binding constants for indefinite self-associating systems has been developed. This approach utilizes a set of fitting routines (SedAnal) that perform global non-linear least squares fits of up to 10 sedimentation velocity experiments, corresponding to different loading concentrations, by a combination of finite element simulations and a fitting algorithm that uses a simplex convergence routine to search parameter space. Indefinite self-association is analyzed with the software program isodesfitter, which incorporates user provided functions for sedimentation coefficients as a function of the degree of polymerization for spherical, linear and helical polymer models. The computer program hydro was used to generate the sedimentation coefficient values for the linear and helical polymer assembly mechanisms. Since this curve fitting method directly fits the shape of the sedimenting boundary, it is in principle very sensitive to alternative models and the presence of species not participating in the reaction. This approach is compared with traditional fitting of weight average data and applied to the initial stages of Mg(2+)-induced tubulin self-associating into small curved polymers, and vinblastine-induced tubulin spiral formation. The appropriate use and limitations of the methods are discussed.  相似文献   

18.
Using data from the human mortality database (HMD), and five different modeling approaches, we estimate Gompertz mortality parameters for 7,704 life tables. To gauge model fit, we predict life expectancy at age 40 from these parameters, and compare predicted to empirical values. Across a diversity of human populations, and both sexes, the overall best way to estimate Gompertz parameters is weighted least squares, although Poisson regression performs better in 996 cases for males and 1,027 cases for females, out of 3,852 populations per sex. We recommend against using unweighted least squares unless death counts (to use as weights or to allow Poisson estimation) are unavailable. We also recommend fitting to logged death rates. Over time in human populations, the Gompertz slope parameter has increased, indicating a more severe increase in mortality rates as age goes up. However, it is well-known that the two parameters of the Gompertz model are very tightly (and negatively) correlated. When the slope goes up, the level goes down, and, overall, mortality rates are decreasing over time. An analysis of Gompertz parameters for all of the HMD countries shows a distinct pattern for males in the formerly socialist economies of Europe.  相似文献   

19.
Cardiac muscle tissue during relaxation is commonly modeled as a hyperelastic material with strongly nonlinear and anisotropic stress response. Adapting the behavior of such a model to experimental or patient data gives rise to a parameter estimation problem which involves a significant number of parameters. Gradient-based optimization algorithms provide a way to solve such nonlinear parameter estimation problems with relatively few iterations, but require the gradient of the objective functional with respect to the model parameters. This gradient has traditionally been obtained using finite differences, the calculation of which scales linearly with the number of model parameters, and introduces a differencing error. By using an automatically derived adjoint equation, we are able to calculate this gradient more efficiently, and with minimal implementation effort. We test this adjoint framework on a least squares fitting problem involving data from simple shear tests on cardiac tissue samples. A second challenge which arises in gradient-based optimization is the dependency of the algorithm on a suitable initial guess. We show how a multi-start procedure can alleviate this dependency. Finally, we provide estimates for the material parameters of the Holzapfel and Ogden strain energy law using finite element models together with experimental shear data.  相似文献   

20.
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号