首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 671 毫秒
1.
Nahhas RW  Wolfe DA  Chen H 《Biometrics》2002,58(4):964-971
McIntyre (1952, Australian Journal of Agricultural Research 3, 385-390) introduced ranked set sampling (RSS) as a method for improving estimation of a population mean in settings where sampling and ranking of units from the population are inexpensive when compared with actual measurement of the units. Two of the major factors in the usefulness of RSS are the set size and the relative costs of the various operations of sampling, ranking, and measurement. In this article, we consider ranking error models and cost models that enable us to assess the effect of different cost structures on the optimal set size for RSS. For reasonable cost structures, we find that the optimal RSS set sizes are generally larger than had been anticipated previously. These results will provide a useful tool for determining whether RSS is likely to lead to an improvement over simple random sampling in a given setting and, if so, what RSS set size is best to use in this case.  相似文献   

2.
Wang X  Lim J  Stokes L 《Biometrics》2008,64(2):355-363
Summary .   MacEachern, Stasny, and Wolfe (2004, Biometrics 60 , 207–215) introduced a data collection method, called judgment poststratification (JPS), based on ideas similar to those in ranked set sampling, and proposed methods for mean estimation from JPS samples. In this article, we propose an improvement to their methods, which exploits the fact that the distributions of the judgment poststrata are often stochastically ordered, so as to form a mean estimator using isotonized sample means of the poststrata. This new estimator is strongly consistent with similar asymptotic properties to those in MacEachern et al. (2004) . It is shown to be more efficient for small sample sizes, which appears to be attractive in applications requiring cost efficiency. Further, we extend our method to JPS samples with imprecise ranking or multiple rankers. The performance of the proposed estimators is examined on three data examples through simulation.  相似文献   

3.
Chen H  Stasny EA  Wolfe DA 《Biometrics》2006,62(1):150-158
The application of ranked set sampling (RSS) techniques to data from a dichotomous population is currently an active research topic, and it has been shown that balanced RSS leads to improvement in precision over simple random sampling (SRS) for estimation of a population proportion. Balanced RSS, however, is not in general optimal in terms of variance reduction for this setting. The objective of this article is to investigate the application of unbalanced RSS in estimation of a population proportion under perfect ranking, where the probabilities of success for the order statistics are functions of the underlying population proportion. In particular, the Neyman allocation, which assigns sample units for each order statistic proportionally to its standard deviation, is shown to be optimal in the sense that it leads to minimum variance within the class of RSS estimators that are simple averages of the means of the order statistics. We also use a substantial data set, the National Health and Nutrition Examination Survey III (NHANES III) data, to demonstrate the feasibility and benefits of Neyman allocation in RSS for binary variables.  相似文献   

4.
A nonparametric selected ranked set sampling is suggested. The estimator of population mean based on the new approach is compared with that using the simple random sampling (SRS), the ranked set sampling (RSS) and the median ranked set sampling (MRSS) methods. The estimator of population mean using the new approach is found to be more efficient than its counter‐parts for almost all the cases considered.  相似文献   

5.
Precision of the estimate of the population mean using ranked set sample (RSS) relative to using simple random sample (SRS), with the same number of quantified units, depends upon the population and success in ranking. In practice, even ranking a sample of moderate size and observing the ith ranked unit (other than the extremes) is a difficult task. Therefore, in this paper we introduce a variety of extreme ranked set sample (ERSSs) to estimate the population mean. ERSSs is more practical than the ordinary ranked set sampling, since in case of even sample size we need to identify successfully only the first and/or the last ordered unit or in case of odd sample size the median unit. We show that ERSSs gives an unbiased estimate of the population mean in case of symmetric populations and it is more efficient than SRS, using the same number of quantified units. Example using real data is given. Also, parametric examples are given.  相似文献   

6.
Ranked set sampling (RSS) is a sampling procedure that can be considerably more efficient than simple random sampling (SRS). When the variable of interest is binary, ranking of the sample observations can be implemented using the estimated probabilities of success obtained from a logistic regression model developed for the binary variable. The main objective of this study is to use substantial data sets to investigate the application of RSS to estimation of a proportion for a population that is different from the one that provides the logistic regression. Our results indicate that precision in estimation of a population proportion is improved through the use of logistic regression to carry out the RSS ranking and, hence, the sample size required to achieve a desired precision is reduced. Further, the choice and the distribution of covariates in the logistic regression model are not overly crucial for the performance of a balanced RSS procedure.  相似文献   

7.
Chen Z  Wang YG 《Biometrics》2004,60(4):997-1004
This article is motivated by a lung cancer study where a regression model is involved and the response variable is too expensive to measure but the predictor variable can be measured easily with relatively negligible cost. This situation occurs quite often in medical studies, quantitative genetics, and ecological and environmental studies. In this article, by using the idea of ranked-set sampling (RSS), we develop sampling strategies that can reduce cost and increase efficiency of the regression analysis for the above-mentioned situation. The developed method is applied retrospectively to a lung cancer study. In the lung cancer study, the interest is to investigate the association between smoking status and three biomarkers: polyphenol DNA adducts, micronuclei, and sister chromatic exchanges. Optimal sampling schemes with different optimality criteria such as A-, D-, and integrated mean square error (IMSE)-optimality are considered in the application. With set size 10 in RSS, the improvement of the optimal schemes over simple random sampling (SRS) is great. For instance, by using the optimal scheme with IMSE-optimality, the IMSEs of the estimated regression functions for the three biomarkers are reduced to about half of those incurred by using SRS.  相似文献   

8.
Ranked set sampling with unequal samples   总被引:3,自引:0,他引:3  
Bhoj DS 《Biometrics》2001,57(3):957-962
A ranked set sampling procedure with unequal samples (RSSU) is proposed and used to estimate the population mean. This estimator is then compared with the estimators based on the ranked set sampling (RSS) and median ranked set sampling (MRSS) procedures. It is shown that the relative precisions of the estimator based on RSSU are higher than those of the estimators based on RSS and MRSS. An example of estimating the mean diameter at breast height of longleaf-pine trees on the Wade Tract in Thomas County, Georgia, is presented.  相似文献   

9.
Sample size determination for case-control studies of chronic disease are often based on the simple 2 X 2 tabular cross-classification of exposure and disease, thereby ignoring stratification which may be considered in the analysis. One consequence of this approach is that the sample size may be inadequate to attain a specified power and size when performing a statistical analysis on J 2 X 2 tables using Cochran's (1954, Biometrics 10, 417-451) statistic or the Mantel-Haenszel (1959, Journal of the National Cancer Institute 22, 719-748) statistic. A sample size formula is derived from Cochran's statistic and it is compared with the corresponding one derived when the data are treated as unstratified, and also with two other formulas proposed for stratified data analysis. The formula developed yields values slightly higher than one recently proposed by Mu?oz and Rosner (1984, Biometrics 40, 995-1004), which assumes that both margins of each 2 X 2 table are fixed, while the present study considers only the case-control margin to be fixed.  相似文献   

10.
Chen YI 《Biometrics》1999,55(4):1258-1262
Lim and Wolfe (1997, Biometrics 53, 410-418) proposed rank-based multiple test procedures for identifying the dose levels that are more effective than the zero-dose control in randomized complete block designs when it can be assumed that the efficacy of the increasing dose levels is monotonically increasing up to a point, followed by a monotonic decrease. Modifications of the Lim-Wolfe tests are suggested that provide more practical and powerful alternatives. Two numerical examples are illustrated and the results of a Monte Carlo power study are presented.  相似文献   

11.
Wahed AS  Tsiatis AA 《Biometrics》2004,60(1):124-133
Two-stage designs, where patients are initially randomized to an induction therapy and then depending upon their response and consent, are randomized to a maintenance therapy, are common in cancer and other clinical trials. The goal is to compare different combinations of primary and maintenance therapies to find the combination that is most beneficial. In practice, the analysis is usually conducted in two separate stages which does not directly address the major objective of finding the best combination. Recently Lunceford, Davidian, and Tsiatis (2002, Biometrics58, 48-57) introduced ad hoc estimators for the survival distribution and mean restricted survival time under different treatment policies. These estimators are consistent but not efficient, and do not include information from auxiliary covariates. In this article we derive estimators that are easy to compute and are more efficient than previous estimators. We also show how to improve efficiency further by taking into account additional information from auxiliary variables. Large sample properties of these estimators are derived and comparisons with other estimators are made using simulation. We apply our estimators to a leukemia clinical trial data set that motivated this study.  相似文献   

12.
In longitudinal studies and in clustered situations often binary and continuous response variables are observed and need to be modeled together. In a recent publication Dunson, Chen, and Harry (2003, Biometrics 59, 521-530) (DCH) propose a Bayesian approach for joint modeling of cluster size and binary and continuous subunit-specific outcomes and illustrate this approach with a developmental toxicity data example. In this note we demonstrate how standard software (PROC NLMIXED in SAS) can be used to obtain maximum likelihood estimates in an alternative parameterization of the model with a single cluster-level factor considered by DCH for that example. We also suggest that a more general model with additional cluster-level random effects provides a better fit to the data set. An apparent discrepancy between the estimates obtained by DCH and the estimates obtained earlier by Catalano and Ryan (1992, Journal of the American Statistical Association 87, 651-658) is also resolved. The issue of bias in inferences concerning the dose effect when cluster size is ignored is discussed. The maximum-likelihood approach considered herein is applicable to general situations with multiple clustered or longitudinally measured outcomes of different type and does not require prior specification and extensive programming.  相似文献   

13.
Optimal sampling in retrospective logistic regression via two-stage method   总被引:1,自引:0,他引:1  
Case-control sampling is popular in epidemiological research because of its cost and time saving. In a logistic regression model, with limited knowledge on the covariance matrix of the point estimator of the regression coefficients a priori, there exists no fixed sample size analysis. In this study, we propose a two-stage sequential analysis, in which the optimal sample fraction and the required sample size to achieve a predetermined volume of a joint confidence set are estimated in an interim analysis. Additionally required observations are collected in the second stage according to the estimated optimal sample fraction. At the end of the experiment, data from these two stages are combined and analyzed for statistical inference. Simulation studies are conducted to justify the proposed two-stage procedure and an example is presented for illustration. It is found that the proposed two-stage procedure performs adequately in the sense that the resultant joint confidence set has a well-controlled volume and achieves the required coverage probability. Furthermore, the optimal sample fractions among all the selected scenarios are close to one. Hence, the proposed procedure can be simplified by always considering a balance design.  相似文献   

14.
Median ranked set sampling may be combined with size biased probability of selection. A two-phase sample is assumed. In the first phase, units are selected with probability proportional to their size. In the second phase, units are selected using median ranked set sampling to increase the efficiency of the estimators relative to simple random sampling. There is also an increase in the efficiency relative to ranked set sampling (for some probability distribution functions). There will be a loss in efficiency depending on the amount of errors in ranking the units, the median ranked set sampling can be used to reduce the errors in ranking the units selected from the population. Estimators of the population mean and the population size are considered. The median ranked set sampling with probability proportion to size and with errors in ranking is considered and compared with ranked set sampling with errors in ranking. Computer simulation results for some probability distributions are also given.  相似文献   

15.
If one or few individuals are enough to perform an action that produces a collective good and if this action has a cost, living in group can be beneficial because the cost can be shared with other individuals. Without coordination, however, the production of a collective good by the contribution of one or few individuals is inefficient and can be modelled as a volunteer's dilemma. In the volunteer's dilemma the individuals that pay the cost for the production of the collective good benefit from their action if nobody else volunteers, but the cost is wasted if too many individuals volunteer. Increasing group size reduces the need of volunteering for each member of the group; the overall benefit for the group, however, decreases too because the larger the group is, the less likely it is that the collective good is produced. This problem persists even with a high degree of relatedness between group members; an optimal, intermediate group size exists that maximizes the probability to produce the collective good.  相似文献   

16.
Gerard PD  Schucany WR 《Biometrics》1999,55(3):769-773
Seber (1986, Biometrics 42, 267-292) suggested an approach to biological population density estimation using kernel estimates of the probability density of detection distances in line transect sampling. Chen (1996a, Applied Statistics 45, 135-150) and others have employed cross validation to choose a global bandwidth for the kernel estimator or have suggested adaptive kernel estimation (Chen, 1996b, Biometrics 52, 1283-1294). Because estimation of the density is required at only a single point, we investigate a local bandwidth selection procedure that is a modification of the method of Schucany (1995, Journal of the American Statistical Association 90, 535-540) for nonparametric regression. We report on simulation results comparing the proposed method and a local normal scale rule with cross validation and adaptive estimation. The local bandwidths and normal scale rule produce estimates with mean squares that are half the size of the others in most cases. Consistency results are also provided.  相似文献   

17.
Lee K  Daniels MJ 《Biometrics》2007,63(4):1060-1067
Generalized linear models with serial dependence are often used for short longitudinal series. Heagerty (2002, Biometrics58, 342-351) has proposed marginalized transition models for the analysis of longitudinal binary data. In this article, we extend this work to accommodate longitudinal ordinal data. Fisher-scoring algorithms are developed for estimation. Methods are illustrated on quality-of-life data from a recent colorectal cancer clinical trial.  相似文献   

18.
Bhoj (1997c) proposed a new ranked set sampling (NRSS) procedure for a specific two‐parameter family of distributions when the sample size is even. This NRSS procedure can be applied to one‐parameter family of distributions when the sample size is even. However, this procedure cannot be used if the sample size is odd. Therefore, in this paper, we propose a modified version of the NRSS procedure which can be used for one‐parameter distributions when the sample size is odd. Simple estimator for the parameter based on proposed NRSS is derived. The relative precisions of this estimator are higher than those of other estimators which are based on other ranked set sampling procedures and the best linear unbiased estimator using all order statistics.  相似文献   

19.
Mattei A  Mealli F 《Biometrics》2007,63(2):437-446
In this article we present an extended framework based on the principal stratification approach (Frangakis and Rubin, 2002, Biometrics 58, 21-29), for the analysis of data from randomized experiments which suffer from treatment noncompliance, missing outcomes following treatment noncompliance, and "truncation by death." We are not aware of any previous work that addresses all these complications jointly. This framework is illustrated in the context of a randomized trial of breast self-examination.  相似文献   

20.
Two-stage randomization designs (TSRD) are becoming increasingly common in oncology and AIDS clinical trials as they make more efficient use of study participants to examine therapeutic regimens. In these designs patients are initially randomized to an induction treatment, followed by randomization to a maintenance treatment conditional on their induction response and consent to further study treatment. Broader acceptance of TSRDs in drug development may hinge on the ability to make appropriate intent-to-treat type inference within this design framework as to whether an experimental induction regimen is better than a standard induction regimen when maintenance treatment is fixed. Recently Lunceford, Davidian, and Tsiatis (2002, Biometrics 58, 48-57) introduced an inverse probability weighting based analytical framework for estimating survival distributions and mean restricted survival times, as well as for comparing treatment policies at landmarks in the TSRD setting. In practice Cox regression is widely used and in this article we extend the analytical framework of Lunceford et al. (2002) to derive a consistent estimator for the log hazard in the Cox model and a robust score test to compare treatment policies. Large sample properties of these methods are derived, illustrated via a simulation study, and applied to a TSRD clinical trial.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号