Sumber : http://www.epa.gov/nheerl/arm/surdesignfaqs.htm A sample is any subset of the target population, i.e., any collection of its elements. Sampling methods may be classified into either probability-based sample methods or non-probabilitybased sampling methods. Probability-based methods are discussed in subsequent sections. Non-probability methods include chunk samples, expert choice samples, and quota samples. Chunk Samples. Scientists often draw conclusions using an arbitrary or fortuitous collection of sites. The sites are gathered haphazardly or "happen to be handy." Often the scientist implicitly assumes that the sites are typical for a larger universe of sites about which conclusions are desired. Such an assumption has only the individual's judgment as a basis and can not be easily defended. The sites are an unknown "chunks" of the target population and consequently no basis exists to make a scientific inference to the target population without invoking assumptions that can not be verified. Expert Choice Samples. Expert choice sampling is a form of judgment sampling that is a more developed form of non-random selection. An expert, or experts, may define a set of criteria to be met for a site to be included in the sample. Not all sites that meet the criteria are included. Criteria usually result in the designation of "typical" sites for the study. A fairly good sample may result given that the expert was skillful in defining the criteria and locating typical sites that met the criteria. However there is no way to be sure. A different expert would probably use different criteria or pick different sites that et the criteria. Without invoking additional assumptions, no basis exists to make inferences to the target population and know the uncertainty associated with the inference. Quota Samples. Quota sampling is commonly used in market research. The target population is divided according to one or more characteristics, e.g., age, sex, and geographic area. For two age groups, two sexes, and three geographic areas, a total of 12 population cells are defined. The cells are similar to strata in stratified random sampling. A quota sample then contains a pre-determined number of individuals in each of the 12 cells. The interviewer then simply "fills the quota" for each cell. The individuals may be the first individuals encountered or the interviewer may have the option of using judgment in selecting the individuals. The sample of individuals in a cell is either a chunk sample or a judgment sample. Individuals may refuse or be unavailable, but new individuals are contacted until the quota of individuals is achieved. However, the problem of selection bias due to non-response still remains. Hence as before there is no basis for an inference to the target population. How many sample sites to use? The most commonly asked question is: How many sample sites do I need? This is an important question as it directly determines the precision of any statement derived from the sample data. An answer requires detailed information on all the estimates that will be produced from the survey, the precision desired for each estimate, and knowledge of the variability expected. A reality faced in most studies is that the number of objectives creates a need for many more sample sites than budget and operational constraints allow. Consequently, the total number of sites in many situations is known from these constraints and the question is which objectives are the most important. It is usual to have some subobjectives dropped due to sample size limitations. Sample size calculations are available in most survey sampling textbook and will not be discussed here. One specific situation of interest is when the objectives call for the estimation of a proportion, e.g., proportion of stream length that meets a designated use. In this case, sample size calculations depend only on the proportion, precision required, and confidence required. Approximate precision estimates for proportions can be obtained by assuming the survey designs are simple random samples. Under this condition the estimated precision can be estimated using procedures given by Cochran (1987) for proportions. Go to Top Precision, as a percent, is determined from precision = Z1- * 100 * Sqrt[ p(1-p)/n] To calculate precision requires knowledge of p, the proportion to be estimated. However, a conservative estimate of precision can be obtained by assuming p to be 0.5, which gives the maximum variance. Z1- is related to the level of confidence required for the estimate. If desire 90% confidence, then use 1.645. If desire 95% confidence, then use 2. Table 1. Precision to achieve 90% confidence in estimates of selected proportions. Assumed Precision with 90% Confidence Precision with 95% Confidence for alternative sample sizes for alternative sample sizes Proportion (percent) 25 50 100 400 1000 25 50 100 400 1000 20% ±13 ±9 ±7 ±3 ±2 ±16 ±11 ±8 ±4 ±3 50% ±17 ±12 ±8 ±4 ±3 ±20 ±13 ±10 ±5 ±3 If the survey designs are actually based on the spatially-restricted survey designs, the actual precision estimates are expected to be lower (better) than those stated. Go to Top How do Sites get selected? Several processes lead to the selection of sites. The first process identifies the resource characteristics and target population and results in a sample frame that contains all sites within the target population. The second process establishes a spatial grid and hierarchical structure that result in cells containing single, or a small number/area of sites. These two results are then combined resulting in each site, or small number/area of sites assigned a hierarchical cell address. Randomization and statistical weighting processes produce a sequence of all sites from which a systematic random sample is selected. Sumber : http://home.ubalt.edu/ntsbarsh/stat-data/Surveys.htm Pertemuan :19,20,21,22,23,24 Introduction The main idea of statistical inference is to take a random sample from a population and then to use the information from the sample to make inferences about particular population characteristics such as the mean (measure of central tendency), the standard deviation (measure of spread) or the proportion of units in the population that have a certain characteristic. Sampling saves money, time, and effort. Additionally, a sample can, in some cases, provide as much information as a corresponding study that would attempt to investigate an entire population-careful collection of data from a sample will often provide better information than a less careful study that tries to look at everything. We must study the behavior of the mean of sample values from different specified populations. Because a sample examines only part of a population, the sample mean will not exactly equal the corresponding mean of the population. Thus, an important consideration for those planning and interpreting sampling results, is the degree to which sample estimates, such as the sample mean, will agree with the corresponding population characteristic. In practice, only one sample is usually taken (in some cases such as "survey data analysis" a small "pilot sample" is used to test the data-gathering mechanisms and to get preliminary information for planning the main sampling scheme). However, for purposes of understanding the degree to which sample means will agree with the corresponding population mean, it is useful to consider what would happen if 10, or 50, or 100 separate sampling studies, of the same type, were conducted. How consistent would the results be across these different studies? If we could see that the results from each of the samples would be nearly the same (and nearly correct!), then we would have confidence in the single sample that will actually be used. On the other hand, seeing that answers from the repeated samples were too variable for the needed accuracy would suggest that a different sampling plan (perhaps with a larger sample size) should be used. Quota Sampling: Quota sampling is availability sampling, but with the constraint that proportionality by strata be preserved. Thus the interviewer will be told to interview so many white male smokers, so many black female nonsmokers, and so on, to improve the representatives of the sample. Maximum variation sampling is a variant of quota sampling, in which the researcher purposively and non-randomly tries to select a set of cases, which exhibit maximal differences on variables of interest. Further variations include extreme or deviant case sampling or typical case sampling. What Is the Margin of Error Estimation is the process by which sample data are used to indicate the value of an unknown quantity in a population. Results of estimation can be expressed as a single value; known as a point estimate, or a range of values, referred to as a confidence interval. Whenever we use point estimation, we calculate the margin of error associated with that point estimation. For example; for the estimation of the population proportion, by the means of sample proportion (P), the margin of errors calculated often as follows: ±1.96 [P(1-P)/n]1/2 In newspapers and television reports on public opinion pools, the margin of error is often appears in small font at the bottom of a table or screen, respectively. However, reporting the amount of error only, is not informative enough by itself, what is missing is the degree of the confidence in the findings. The more important missing piece of information is the sample size n. that is, how many people participated in the survey, 100 or 100000? By now, you know it well that the larger the sample size the more accurate is the finding, right? The reported margin of error is the margin of "sampling error". There are many nonsampling errors that can and do affect the accuracy of polls. Here we talk about sampling error. The fact that subgroups have larger sampling error than one must include the following statement: "Other sources of error include but are not limited to, individuals refusing to participate in the interview and inability to connect with the selected number. Every feasible effort is made to obtain a response and reduce the error, but the reader (or the viewer) should be aware that some error is inherent in all research." If you have a yes/no question in a survey, you probably want to calculate a proportion P of Yes's (or No's). Under simple random sampling survey, the variance of P is P(1-P)/n, ignoring the finite population correction, for large n, say over 30. Now a 95% confidence interval is P - 1.96 [P(1-P)/n]1/2, P + 1.96 [P(1-P)/n]1/2. A conservative interval can be calculated, since P(1-P) takes its maximum value when P = 1/2. Replace 1.96 by 2, put P = 1/2 and you have a 95% consevative confidence interval of 1/n1/2. This approximation works well as long as P is not too close to 0 or 1. This useful approximation allows you to calculate approximate 95% confidence intervals. References and Further Readings: Casella G., and R. Berger, Statistical Inference, Wadsworth Pub. Co., 2001. Kish L., Survey Sampling, Wiley, 1995. Lehmann E., and G. Casella, Theory of Point Estimation, Springer Verlag, New York, 1998. Levy P., and S. Lemeshow, Sampling of Populations: Methods and Applications, Wiley, 1999. Sample Size Determination The question of how large a sample to take arises early in the planning of any survey. This is an important question that should be treated lightly. To take a large sample than is needed to achieve the desired results is wasteful of resources whereas very small samples often lead to that are no practical use of making good decision. The main objective is to obtain both a desirable accuracy and a desirable confidence level with minimum cost. Pilot Sample: A pilot or preliminary sample must be drawn from the population and the statistics computed from this sample are used in determination of the sample size. Observations used in the pilot sample may be counted as part of the final sample, so that the computed sample size minus the pilot sample size is the number of observations needed to satisfy the total sample size requirement. People sometimes ask me, what fraction of the population do you need? I answer, "It's irrelevant; accuracy is determined by sample size alone" This answer has to be modified if the sample is a sizable fraction of the population. For an item scored 0/1 for no/yes, the standard deviation of the item scores is given by SD = [p(1-p)/N] 1/2 where p is the proportion obtaining a score of 1, and N is the sample size. The standard error of estimate SE (the standard deviation of the range of possible p values based on the pilot sample estimate) is given by SE= SD/N½. Thus, SE is at a maximum when p = 0.5. Thus the worst case scenario occurs when 50% agree, 50% disagree. The sample size, N, can then be expressed as largest integer less than or equal to 0.25/SE2 Thus, for SE to be 0.01 (i.e. 1%), a sample size of 2500 would be needed; 2%, 625; 3%, 278; 4%, 156, 5%, 100. Note, incidentally, that as long as the sample is a small fraction of the total population, the actual size of the population is entirely irrelevant for the purposes of this calculation. Sample sizes with regard to binary data: n = [t2 N p(1-p)] / [t2 p(1-p) + 2 (N-1)] with N being the size of the total number of cases, n being the sample size, the expected error, t being the value taken from the t distribution corresponding to a certain confidence interval, and p being the probability of an event. For a finite population of size N, the standard error of the sample mean of size n, is: [(N -n)/(nN)]½ There are several formulas for the sample size needed for a t-test. The simplest one is n = 2(Z+Z)22/D2 which underestimates the sample size, but is reasonable for large sample sizes. A less inaccurate formula replaces the Z values with t values, and requires iteration, since the df for the t distribution depends on the sample size. The accurate formula uses a non-central t distribution and it also requires iteration. The simplest approximation is to replace the first Z value in the above formula with the value from the studentized range statistic that is used to derive Tukey's follow-up test. If you don't have sufficiently detailed tables of the studentized range, you can approximate the Tukey follow-up test using a Bonferroni correction. That is, change the first Z value to Z where k is the number of comparisons. Neither of these solutions is exact and the exact solution is a bit messy. But either of the above approaches is probably close enough, especially if the resulting sample size is larger than (say) 30. A better stopping rule for conventional statistical tests is as follows: Test some minimum (pre-determined) number of subjects. Stop if p-value is equal to or less than .01, or p-value equal to or greater than .36; otherwise, run more subjects. Obviously, another option is to stop if/when the number of subjects becomes too great for the effect to be of practical interest. This procedure maintains about 0.05. We may categorized probability proportion to size (PPS) sampling, stratification, and ratio estimation (or any other form of model assisted estimation) as tools that protect one from the results of a very unlucky sample. The first two (PPS sampling and stratification) do this by manipulation of the sampling plan (with PPS sampling conceptually a limiting case of stratification). Model assisted estimation methods such as ratio estimation serve the same purpose by introduction of ancillary information into the estimation procedure. Which tools are preferable depends, as others have said, on costs, availability of information that allows use of these tools, and the potential payoffs (none of these will help much if the stratification/PPS/ratio estimation variable is not well correlated with the response variable of interest). There are also heuristic methods for determination of sample size. For example, in healthcare behavior and process measurement sampling criteria are designed for a 95% CI of 10 percentage points around a population mean of 0.50; There is a heuristic rule: "If the number of individuals in the target population is smaller than 50 per month, systems do not use sampling procedures but, attempt to collect data from all individuals in the target population." Sumber : http://home.ubalt.edu/ntsbarsh/stat-data/Surveys.htm Introduction The main idea of statistical inference is to take a random sample from a population and then to use the information from the sample to make inferences about particular population characteristics such as the mean (measure of central tendency), the standard deviation (measure of spread) or the proportion of units in the population that have a certain characteristic. Sampling saves money, time, and effort. Additionally, a sample can, in some cases, provide as much information as a corresponding study that would attempt to investigate an entire population-careful collection of data from a sample will often provide better information than a less careful study that tries to look at everything. We must study the behavior of the mean of sample values from different specified populations. Because a sample examines only part of a population, the sample mean will not exactly equal the corresponding mean of the population. Thus, an important consideration for those planning and interpreting sampling results, is the degree to which sample estimates, such as the sample mean, will agree with the corresponding population characteristic. In practice, only one sample is usually taken (in some cases such as "survey data analysis" a small "pilot sample" is used to test the data-gathering mechanisms and to get preliminary information for planning the main sampling scheme). However, for purposes of understanding the degree to which sample means will agree with the corresponding population mean, it is useful to consider what would happen if 10, or 50, or 100 separate sampling studies, of the same type, were conducted. How consistent would the results be across these different studies? If we could see that the results from each of the samples would be nearly the same (and nearly correct!), then we would have confidence in the single sample that will actually be used. On the other hand, seeing that answers from the repeated samples were too variable for the needed accuracy would suggest that a different sampling plan (perhaps with a larger sample size) should be used. Quota Sampling: Quota sampling is availability sampling, but with the constraint that proportionality by strata be preserved. Thus the interviewer will be told to interview so many white male smokers, so many black female nonsmokers, and so on, to improve the representatives of the sample. Maximum variation sampling is a variant of quota sampling, in which the researcher purposively and non-randomly tries to select a set of cases, which exhibit maximal differences on variables of interest. Further variations include extreme or deviant case sampling or typical case sampling. What Is the Margin of Error Estimation is the process by which sample data are used to indicate the value of an unknown quantity in a population. Results of estimation can be expressed as a single value; known as a point estimate, or a range of values, referred to as a confidence interval. Whenever we use point estimation, we calculate the margin of error associated with that point estimation. For example; for the estimation of the population proportion, by the means of sample proportion (P), the margin of errors calculated often as follows: ±1.96 [P(1-P)/n]1/2 In newspapers and television reports on public opinion pools, the margin of error is often appears in small font at the bottom of a table or screen, respectively. However, reporting the amount of error only, is not informative enough by itself, what is missing is the degree of the confidence in the findings. The more important missing piece of information is the sample size n. that is, how many people participated in the survey, 100 or 100000? By now, you know it well that the larger the sample size the more accurate is the finding, right? The reported margin of error is the margin of "sampling error". There are many nonsampling errors that can and do affect the accuracy of polls. Here we talk about sampling error. The fact that subgroups have larger sampling error than one must include the following statement: "Other sources of error include but are not limited to, individuals refusing to participate in the interview and inability to connect with the selected number. Every feasible effort is made to obtain a response and reduce the error, but the reader (or the viewer) should be aware that some error is inherent in all research." If you have a yes/no question in a survey, you probably want to calculate a proportion P of Yes's (or No's). Under simple random sampling survey, the variance of P is P(1-P)/n, ignoring the finite population correction, for large n, say over 30. Now a 95% confidence interval is P - 1.96 [P(1-P)/n]1/2, P + 1.96 [P(1-P)/n]1/2. A conservative interval can be calculated, since P(1-P) takes its maximum value when P = 1/2. Replace 1.96 by 2, put P = 1/2 and you have a 95% consevative confidence interval of 1/n1/2. This approximation works well as long as P is not too close to 0 or 1. This useful approximation allows you to calculate approximate 95% confidence intervals. References and Further Readings: Casella G., and R. Berger, Statistical Inference, Wadsworth Pub. Co., 2001. Kish L., Survey Sampling, Wiley, 1995. Lehmann E., and G. Casella, Theory of Point Estimation, Springer Verlag, New York, 1998. Levy P., and S. Lemeshow, Sampling of Populations: Methods and Applications, Wiley, 1999. Sample Size Determination The question of how large a sample to take arises early in the planning of any survey. This is an important question that should be treated lightly. To take a large sample than is needed to achieve the desired results is wasteful of resources whereas very small samples often lead to that are no practical use of making good decision. The main objective is to obtain both a desirable accuracy and a desirable confidence level with minimum cost. Pilot Sample: A pilot or preliminary sample must be drawn from the population and the statistics computed from this sample are used in determination of the sample size. Observations used in the pilot sample may be counted as part of the final sample, so that the computed sample size minus the pilot sample size is the number of observations needed to satisfy the total sample size requirement. People sometimes ask me, what fraction of the population do you need? I answer, "It's irrelevant; accuracy is determined by sample size alone" This answer has to be modified if the sample is a sizable fraction of the population. For an item scored 0/1 for no/yes, the standard deviation of the item scores is given by SD = [p(1-p)/N] 1/2 where p is the proportion obtaining a score of 1, and N is the sample size. The standard error of estimate SE (the standard deviation of the range of possible p values based on the pilot sample estimate) is given by SE= SD/N½. Thus, SE is at a maximum when p = 0.5. Thus the worst case scenario occurs when 50% agree, 50% disagree. The sample size, N, can then be expressed as largest integer less than or equal to 0.25/SE2 Thus, for SE to be 0.01 (i.e. 1%), a sample size of 2500 would be needed; 2%, 625; 3%, 278; 4%, 156, 5%, 100. Note, incidentally, that as long as the sample is a small fraction of the total population, the actual size of the population is entirely irrelevant for the purposes of this calculation. Sample sizes with regard to binary data: n = [t2 N p(1-p)] / [t2 p(1-p) + 2 (N-1)] with N being the size of the total number of cases, n being the sample size, the expected error, t being the value taken from the t distribution corresponding to a certain confidence interval, and p being the probability of an event. For a finite population of size N, the standard error of the sample mean of size n, is: [(N -n)/(nN)]½ There are several formulas for the sample size needed for a t-test. The simplest one is n = 2(Z+Z)22/D2 which underestimates the sample size, but is reasonable for large sample sizes. A less inaccurate formula replaces the Z values with t values, and requires iteration, since the df for the t distribution depends on the sample size. The accurate formula uses a non-central t distribution and it also requires iteration. The simplest approximation is to replace the first Z value in the above formula with the value from the studentized range statistic that is used to derive Tukey's follow-up test. If you don't have sufficiently detailed tables of the studentized range, you can approximate the Tukey follow-up test using a Bonferroni correction. That is, change the first Z value to Z where k is the number of comparisons. Neither of these solutions is exact and the exact solution is a bit messy. But either of the above approaches is probably close enough, especially if the resulting sample size is larger than (say) 30. A better stopping rule for conventional statistical tests is as follows: Test some minimum (pre-determined) number of subjects. Stop if p-value is equal to or less than .01, or p-value equal to or greater than .36; otherwise, run more subjects. Obviously, another option is to stop if/when the number of subjects becomes too great for the effect to be of practical interest. This procedure maintains about 0.05. We may categorized probability proportion to size (PPS) sampling, stratification, and ratio estimation (or any other form of model assisted estimation) as tools that protect one from the results of a very unlucky sample. The first two (PPS sampling and stratification) do this by manipulation of the sampling plan (with PPS sampling conceptually a limiting case of stratification). Model assisted estimation methods such as ratio estimation serve the same purpose by introduction of ancillary information into the estimation procedure. Which tools are preferable depends, as others have said, on costs, availability of information that allows use of these tools, and the potential payoffs (none of these will help much if the stratification/PPS/ratio estimation variable is not well correlated with the response variable of interest). There are also heuristic methods for determination of sample size. For example, in healthcare behavior and process measurement sampling criteria are designed for a 95% CI of 10 percentage points around a population mean of 0.50; There is a heuristic rule: "If the number of individuals in the target population is smaller than 50 per month, systems do not use sampling procedures but, attempt to collect data from all individuals in the target population."