首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 320 毫秒
1.
The combined mark-recapture and line transect sampling methodology proposed by Alpizar-Jara and Pollock [Journal of Environmental and Ecological Statistics, 3(4), 311–327, 1996; In Marine Mammal Survey and Assessment Methods Symposium. G.W. Garner, S.C. Amstrup, J.L. Laake, B.F.J. Manly, L.L. McDonald, and D.C. Robertson (Eds.), A.A. Balkema, Rotterdam, Netherlands, pp. 99–114, 1999] is used to illustrate the estimation of population size for populations with prominent nesting structures (i.e., bald eagle nests). In the context of a bald eagle population, the number of nests in a list frame corresponds to a pre-marked sample of nests, and an area frame corresponds to a set of transect strips that could be regularly monitored. Unlike previous methods based on dual frame methodology using the screening estimator [Haines and Pollock (Journal of Environmental and Ecological Statistics, 5, 245–256, 1998a; Survey Methodology, 24(1), 79–88, 1998b)], we no longer need to assume that the area frame is complete (i.e., all the nests in the sampled sites do not need to be seen). One may use line transect sampling to estimate the probability of detection in a sampled area. Combining information from list and area frames provides more efficient estimators than those obtained by using data from only one frame. We derive an estimator for detection probability and generalize the screening estimator. A simulation study is carried out to compare the performance of the Chapman modification of the Lincoln–Petersen estimator to the screening estimator. Simulation results show that although the Chapman estimator is generally less precise than the screening estimator, the latter can be severely biased in presence of uncertain detection. The screening estimator outperforms the Chapman estimator in terms of mean squared error when detection probability is near 1 wheareas the Chapman estimator outperforms the screening estimator when detection probability is lower than a certain threshold value depending on particular scenarios.  相似文献   

2.
Consider a survey of a plant or animal species in which abundance or presence/absence will be recorded. Further assume that the presence of the plant or animal is rare and tends to cluster. A sampling design will be implemented to determine which units to sample within the study region. Adaptive cluster sampling designs Thompson (1990) are sampling designs that are implemented by first selecting a sample of units according to some conventional probability sampling design. Then, whenever a specified criterion is satisfied upon measuring the variable of interest, additional units are adaptively sampled in neighborhoods of those units satisfying the criterion. The success of these adaptive designs depends on the probabilities of finding the rare clustered events, called networks. This research uses combinatorial generating functions to calculate network inclusion probabilities associated with a simple Latin square sample. It will be shown that, in general, adaptive simple Latin square sampling when compared to adaptive simple random sampling will (i) yield higher network inclusion probabilities and (ii) provide Horvitz-Thompson estimators with smaller variability.  相似文献   

3.
Thompson (1990) introduced the adaptive cluster sampling design and developed two unbiased estimators, the modified Horvitz-Thompson (HT) and Hansen-Hurwitz (HH) estimators, for this sampling design and noticed that these estimators are not a function of the minimal sufficient statistics. He applied the Rao-Blackwell theorem to improve them. Despite having smaller variances, these latter estimators have not received attention because a suitable method or algorithm for computing them was not available. In this paper we obtain closed forms of the Rao-Blackwell versions which can easily be computed. We also show that the variance reduction for the HH estimator is greater than that for the HT estimator using Rao-Blackwell versions. When the condition for extra samples is 0$$ " align="middle" border="0"> , one can expect some Rao-Blackwell improvement in the HH estimator but not in the HT estimator. Two examples are given.  相似文献   

4.
The United States Environmental Protection Agency's Environmental Monitoring and Assessment Program (EMAP) is designed to describe status, trends and spatial pattern of indicators of condition of the nation's ecological resources. The proposed sampling design for EMAP is based on a triangular systematic grid and employs both variable probability and double sampling. The Horvitz-Thompson estimator provides the foundation of the design-based estimation strategy used in EMAP. However, special features of EMAP designed to accommodate the complexity of sampling environmental resources on a national scale require modifications of standard variance estimation procedures as well as development of new techniques. An overview of variance estimation methods proposed for application to EMAP's sampling strategy for discrete resources is presented.  相似文献   

5.
In this article we consider asymptotic properties of the Horvitz-Thompson and Hansen-Hurwitz types of estimators under the adaptive cluster sampling variants obtained by selecting the initial sample by simple random sampling without replacement and by unequal probability sampling with replacement. We develop an asymptotic framework, which basically assumes that the number of units in the initial sample, as well as the number of units and networks in the population tend to infinity, but that the network sizes are bounded. Using this framework we prove that under each of the two variants of adaptive sampling above mentioned, both the Horvitz-Thompson and Hansen-Hurwitz types of estimators are design-consistent and asymptotically normally distributed. In addition we show that the ordinary estimators of their variances are also design-consistent estimators.  相似文献   

6.
Adaptive two-stage one-per-stratum sampling   总被引:1,自引:0,他引:1  
We briefly describe adaptive cluster sampling designs in which the initial sample is taken according to a Markov chain one-per-stratum design (Breidt, 1995) and one or more secondary samples are taken within strata if units in the initial sample satisfy a given condition C. An empirical study of the behavior of the estimation procedure is conducted for three small artificial populations for which adaptive sampling is appropriate. The specific sampling strategy used in the empirical study was a single random-start systematic sample with predefined systematic samples within strata when the initially sampled unit in that stratum satisfies C. The bias of the Horvitz-Thompson estimator for this design is usually very small when adaptive sampling is conducted in a population for which it is suited. In addition, we compare the behavior of several alternative estimators of the standard error of the Horvitz-Thompson estimator of the population total. The best estimator of the standard error is population-dependent but it is not unreasonable to use the Horvitz-Thompson estimator of the variance. Unfortunately, the distribution of the estimator is highly skewed hence the usual approach of constructing confidence intervals assuming normality cannot be used here.  相似文献   

7.
A ranked set sampling protocol is proposed when an auxiliary variable is available in addition to the target variable in sample surveys. The protocol may be practically carried out without additional sampling effort or costs. Under the suggested sampling scheme, the estimators usually adopted in surveys with auxiliary information - such as the ratio estimator or the regression estimator - display surprising theoretical properties as well as high performance in practice.  相似文献   

8.
Thompson (1990) introduced the adaptive cluster sampling design. This sampling design has been shown to be a useful sampling method for parameter estimation of a clustered and scattered population (Roesch, 1993; Smith et al., 1995; Thompson and Seber, 1996). Two estimators, the modified Hansen-Hurwitz (HH) and Horvitz-Thompson (HT) estimators, are available to estimate the mean or total of a population. Empirical results from previous researches indicate that the modified HT estimator has smaller variance than the modified HH estimator. We analytically compare the properties of these two estimators. Some results are obtained in favor of the modified HT estimator so that practitioners are strongly recommended to use the HT estimator despite easiness of computations for the HH estimator.  相似文献   

9.
Estimators for the population sizes of animal species are similar to Horvitz-Thompson estimators — they involve dividing counts of detected animals by the probabilities of detection. Knowing detection probabilities for different subpopulations allows one to estimate each sub-population size with such an estimator, and to add the results for an estimator of the total population. In the case where the proportions of animals belonging to the different sub-populations are also known, this paper shows that using those proportions to arrive at a common average detection probability will, when used in conjunction with the total number of animals detected, result in a better estimator. We provide two examples where the inferior estimator may seem sensible.  相似文献   

10.
A biological community usually has a large number of species with relatively small abundances. When a random sample of individuals is selected and each individual is classified according to species identity, some rare species may not be discovered. This paper is concerned with the estimation of Shannons index of diversity when the number of species and the species abundances are unknown. The traditional estimator that ignores the missing species underestimates when there is a non-negligible number of unseen species. We provide a different approach based on unequal probability sampling theory because species have different probabilities of being discovered in the sample. No parametric forms are assumed for the species abundances. The proposed estimation procedure combines the Horvitz–Thompson (1952) adjustment for missing species and the concept of sample coverage, which is used to properly estimate the relative abundances of species discovered in the sample. Simulation results show that the proposed estimator works well under various abundance models even when a relatively large fraction of the species is missing. Three real data sets, two from biology and the other one from numismatics, are given for illustration.  相似文献   

11.
In this paper, we consider design-based estimation using ranked set sampling (RSS) in finite populations. We first derive the first and second-order inclusion probabilities for an RSS design and present two Horvitz–Thompson type estimators using these inclusion probabilities. We also develop an alternate Hansen–Hurwitz type estimator and investigate its properties. In particular, we show that this alternate estimator always outperforms the usual Hansen–Hurwitz type estimator in the simple random sampling with replacement design with comparable sample size. We also develop formulae for ratio estimator for all three developed estimators. The theoretical results are augmented by numerical and simulation studies as well as a case study using a well known data set. These show that RSS design can yield a substantial improvement in efficiency over the usual simple random sampling design in finite populations.  相似文献   

12.
Randomized graph sampling (RGS) is an approach for sampling populations associated with or describable as graphs, when the structure of the graph is known and the parameter of interest is the total weight of the graph. RGS is related to, but distinct from, other graph-based approaches such as snowball and network sampling. Graph elements are clustered into walks that reflect the structure of the graph, as well as operational constraints on sampling. The basic estimator in RGS can be constructed as a Horvitz-Thompson estimator. I prove it to be design-unbiased, and also show design-unbiasedness of an estimator of the sample variance when walks are sampled with replacement. Covariates can be employed for variance reduction either through improved assignment of selection probabilities to walks in the design step, or through the use of alternative estimators during analysis. The approach is illustrated with a trail maintenance example, which demonstrates that complicated approaches to assignment of selection probabilities can be counterproductive. I describe conditions under which RGS may be efficient in practice, and suggest possible applications.  相似文献   

13.
Practical problems facing adaptive cluster sampling with order statistics (acsord) are explored using Monte Carlo simulation for three simulated fish populations and two known waterfowl populations. First, properties of an unbiased Hansen-Hurwitz (HH) estimator and a biased alternative Horvitz-Thompson (HT) estimator are evaluated. An increase in the level of population aggregation or the initial sample size increases the efficiencies of the two acsord estimators. For less aggregated fish populations, the efficiencies decrease as the order statistic parameter r (the number of units about which adaptive sampling is carried out) increases; for the highly aggregated fish and waterfowl populations, they increase with r. Acsord is almost always more efficient than simple random sampling for the highly aggregated populations. Positive bias is observed for the HT estimator, with the maximum bias usually occurring at small values of r. Secondly, a stopping rule at the Sth iteration of adaptive sampling beyond the initial sampling unit was applied to the acsord design to limit the otherwise open-ended sampling effort. The stopping rule induces relatively high positive bias to the HH estimator if the level of the population aggregation is high, the stopping level S is small, and r is large. The bias of HT is not very sensitive to the stopping rule and its bias is often reduced by the stopping rule at smaller values of r. For more aggregated populations, the stopping rule often reduces the efficiencies of the estimators compared to the non-stopping-rule scheme, but acsord still remains more efficient than simple random sampling. Despite its bias and lack of theoretical grounding, the HT estimator is usually more efficient than the HH estimator. In the stopping rule case, the HT estimator is preferable, because its bias is less sensitive to the stopping level.  相似文献   

14.
Coverage, i.e., the area covered by the target attribute in the study region, is a key parameter in many surveys. Coverage estimation is usually performed by adopting a replicated protocol based on line-intercept sampling coupled with a suitable linear homogeneous estimator. Since coverage is a parameter which may be interestingly represented as the integral of a suitable function, improved Monte Carlo strategies for implementing the replicated protocol are introduced in order to achieve estimators with small variance rates. In addition, new specific theoretical results on Monte Carlo integration methods are given to deal with the integrand functions arising in the special coverage estimation setting.
Lucio BarabesiEmail:
  相似文献   

15.
Adaptive cluster sampling (ACS) is an efficient sampling design for estimating parameters of rare and clustered populations. It is widely used in ecological research. The modified Hansen-Hurwitz (HH) and Horvitz-Thompson (HT) estimators based on small samples under ACS have often highly skewed distributions. In such situations, confidence intervals based on traditional normal approximation can lead to unsatisfactory results, with poor coverage properties. Christman and Pontius (Biometrics 56:503–510, 2000) showed that bootstrap percentile methods are appropriate for constructing confidence intervals from the HH estimator. But Perez and Pontius (J Stat Comput Simul 76:755–764, 2006) showed that bootstrap confidence intervals from the HT estimator are even worse than the normal approximation confidence intervals. In this article, we consider two pseudo empirical likelihood functions under the ACS design. One leads to the HH estimator and the other leads to a HT type estimator known as the Hájek estimator. Based on these two empirical likelihood functions, we derive confidence intervals for the population mean. Using a simulation study, we show that the confidence intervals obtained from the first EL function perform as good as the bootstrap confidence intervals from the HH estimator but the confidence intervals obtained from the second EL function perform much better than the bootstrap confidence intervals from the HT estimator, in terms of coverage rate.  相似文献   

16.
The initial use of composite sampling involved the analysis of many negative samples with relatively high laboratory cost (Dorfman sampling). We propose a method of double compositing and compare its efficiency with Dorfman sampling. The variability of composite measurement samples has environmental interest (hot spots). The precision of these estimates depends on the kurtosis of the distribution; leptokurtic distributions (2 > 0) have increased precision as the number of field samples is increased. The opposite effect is obtained for platykurtic distributions. In the lognormal case, coverage probabilities are reasonable for < 0.5. The Poisson distribution can be associated with temporal compositing, of particular interest where radioactive measurements are taken. Sample size considerations indicate that the total sampling effort is directly proportional to the length of time sampled. If there is background radiation then increasing levels of this radiation require larger sample sizes to detect the same difference in radiation.  相似文献   

17.
In many applications of line intersect sampling, transects consist of multiple, connected segments in a prescribed configuration. The relationship between the transect configuration and the selection probability of a population element is illustrated and a consistent sampling protocol, applicable to populations composed of arbitrarily shaped elements, is proposed. It is shown that this protocol obviates the arbitrary practice of treating multiple intersections of a single particle as independent probabilistic events and preserves the design-unbiasedness of Kaisers (1983, Biometrics 39, 965–976) conditional and unconditional estimators, suitably generalized to segmented transect designs. The relative efficiency and utility of segmented transect designs are also discussed from a fixed population perspective.  相似文献   

18.
Surface soils (0–15 cm) were sampled at 10–20 km intervals along two transects in Venezuela. One (1162 km, 70 samples) ran west to east parallel with the Caribbean coastline, the other (920 km, 92 samples) ran south to north from the frontier with Brazil to the Caribbean shore. Sampling took place in both a wet and a dry season. Trace metals were extracted from dried, sieved (<2 mm) soil with boiling aqua regia followed by analysis by ICP or flame AAS. Metal values did not differ significantly between the two seasons and dates were averaged. Geometric mean values for the west–east transect were: Cr=41.5, Cu 17.9, Cs 3.6, Li=13.9, Mn=294, Ni=21.3, Pb=17.4, Sr=39.4, V=60.4 and Zn = 83.7g g–1, respectively. Similarly, for the south–north transect Cr=21.3, Cu=4.3, Cs=1.1, Li=2.0, Mn=55.7, Ni=4.4, Pb=6.1, Sr=13.3, V=28.2 and Zn=16.7g g–1, respectively. A classification of samples by lithology showed surface soil composition to be related to rock composition. Metal values were low in the soils in the south of the country, in the Guyana highlands (Gran Sabana). Low Zn contents were prevalent. Lead contents were affected by roadside fallout from vehicles using leaded petrol except that high Pb contents of soils in the Gran Sabana were of more complex origin.  相似文献   

19.
Ranked set sampling can provide an efficient basis for estimating parameters of environmental variables, particularly when sampling costs are intrinsically high. Various ranked set estimators are considered for the population mean and contrasted in terms of their efficiencies and useful- ness, with special concern for sample design considerations. Specifically, we consider the effects of the form of the underlying random variable, optimisation of efficiency and how to allocate sampling effort for best effect (e.g. one large sample or several smaller ones of the same total size). The various prospects are explored for two important positively skew random variables (lognormal and extreme value) and explicit results are given for these cases. Whilst it turns out that the best approach is to use the largest possible single sample and the optimal ranked set best linear estimator (ranked set BLUE), we find some interesting qualitatively different conclusions for the two skew distributions  相似文献   

20.
We compare the performance of a number of estimators of the cumulative distribution function (CDF) for the following scenario: imperfect measurements are taken on an initial sample from afinite population and perfect measurements are obtained on a small calibration subset of the initial sample. The estimators we considered include two naive estimators using perfect and imperfect measurements; the ratio, difference and regression estimators for a two-phasesample; a minimum MSE estimator; Stefanski and Bay's SIMEX estimator (1996); and two proposed estimators. The proposed estimators take the form of a weighted average of perfect and imperfect measurements. They are constructed by minimizing variance among the class of weighted averages subject to an unbiasedness constraint. They differ in the manner of estimating the weight parameters. The first one uses direct sample estimates. The second one tunes the unknown parameters to an underlying normal distribution. We compare the root mean square error (RMSE) of the proposed estimator against other potential competitors through computer simulations. Our simulations show that our second estimator has the smallest RMSE among thenine compared and that the reduction in RMSE is substantial when the calibration sample is small and the error is medium or large.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号