Which sampling provides a known non zero chance of selection for each population element?

Non-Probability Sampling

Alison Galloway, in Encyclopedia of Social Measurement, 2005

Introduction

Definition

In a non-probability sample, some members of the population, compared to other members, have a greater but unknown chance of selection. There are five main types of non-probability sample: convenience, purposive, quota, snowball, and self-selection.

Comparison with Probability Samples

The main feature present in a probability sample, but generally absent in a non-probability sample, is a sampling frame. Probability samples are possible only when there is a complete and up-to-date list of the members [names and/or addresses] of the population under investigation. There are, of course, many situations in which it is not possible to obtain a sampling frame. For example, when researching football supporters, it is not feasible to obtain a list of names or addresses of people who attend games. It should be possible to obtain a list of people who are official members of the team, but this is unlikely to comprise all supporters. Similarly, there is not a sampling frame easily available to researchers of all the elderly in a population. Non-probability techniques therefore have to be used whenever there is no readily available and complete list of the population under investigation.

Advantages and Disadvantages of Non-Probability Sampling

The most obvious advantage in non-probability sampling is clearly the ability to target particular groups of the population. Non-probability methods are often dismissed or criticized because they do not have the statistical foundations of probability methods. However, a survey using, for example, random, systematic, or stratified sampling may adopt methods such as postal delivery, which characteristically has extremely poor response rates. It could certainly be argued that as many valid conclusions can be drawn from a well- constructed study using non-probability methods, compared with a probability survey to which only 10% of the sample responded. Researchers would need to be confident that those 10% were truly representative of the population as a whole.

Non-probability methods also have the advantage in typically being less expensive to conduct. Savings, in terms of both money and time, can be achieved not so much by the sampling method per se, but rather by the forms of delivery that are available for these methods. For example, face-to-face delivery can be cheaper than postal approaches, particularly where oversampling has had to be used to compensate for the typically poor response rates of a mailed survey.

Read full chapter

URL: //www.sciencedirect.com/science/article/pii/B0123693985003820

Planning the Study

Bill Albert, ... Donna Tedesco, in Beyond the Usability Lab, 2010

2.7.2 Sampling techniques

There are two main types of samples: probability and nonprobability samples. Nonprobability samples are cases where you do not know of every unique member of the population in question [i.e., the entire user group in our case]. Another way to describe it is when every member of the population does not have an equal chance of being invited to participate. Probability samples are when you do know of every unique member of the population and therefore each has a probabilistic chance of being invited for the sample [e.g., 100 users of a product, each has a 1/100 chance of being invited]. Here's a taste of a couple of common nonprobability sampling techniques.

Convenience sampling. This is the most common nonprobability sample. You might send invitations to people in your company, students from a school you're affiliated with, the city you live in, and so on. It's referred to as “convenience” sampling because unless the targeted user group is truly limited to those people, it is likely introducing some bias to recruit just a particular slice of the population.

Snowball sampling. This is a type of convenience sampling in which those participants invited invite other participants and so on to create a pyramid effect.

With probability sampling, you can choose a more scientific way to sample because you know the number and characteristics of the true population. For instance, a particular product has a small contingent of users. Some of the common techniques include the following.

Simple random sampling. This is a known population of users from which you take a random sample via some means, such as a program or application, an Excel formula, or a simple “pick out of a hat” lottery.

Stratified sampling. You assign everyone in the population to a specific [but meaningful] group and then take another probability sample within each category. The users are therefore chosen randomly, but there is a representation from each “strata.” Note that the nonprobability sampling method that correlates to stratified sampling is called quota sampling. It means that despite not knowing the true population, you divide the users you know about into groups and still try to get some representation from that group [usually via convenience sampling].

Systematic random sampling [also known as just systematic sampling]. The idea here is that you list all of the users in no particular order [in fact, it should not be in any logical order] and then pick every Nth user. You define what N is; for instance, if there are 1000 users total and the goal is to invite 500 to participate, you'd simply take every second person from the list and invite them.

Multistage sampling. This takes different samples in different ways to eliminate bias. For example, you may do a random sample, then stratified, and so on.

Chances are that the number and characteristics of users for a product you're testing may not be entirely known, especially for Web sites, and still further for Web sites that don't require users to register or create an account [and thus can't be tracked]. If using a generic participant panel, it's likely a convenience sample and is not necessary to worry about probability sampling techniques. However, in cases where you're providing a recruiting service with part of a customer or user list from which to invite people, you may want to use one of the probability sampling techniques discussed. For example, if there are 10,000 customers on a customer list and it needs to be whittled down to a 1000 person sample for study invitations, you might use a simple random or stratified sampling method to get a representative sample of customers.

If recruiting by posting on the Web via message boards, forums, or in newspaper ads, rather than using email invitations, it's likely a snowball sample as people might forward on the information to others that they know. Just be aware that this comes with a self-selection bias. This type of bias is where the participants who choose to participate may have certain outstanding and overrepresentative characteristics, such as being the ones who particularly love and/or hate the Web site enough to participate. One way to minimize this bias is to use a phased launching strategy [breaking up the study launch into multiple groups].

TIP

Check out //stattrek.com/Reading/Sampling.aspx for references to some good books on sampling methods and the biases associated with them.

Read full chapter

URL: //www.sciencedirect.com/science/article/pii/B9780123748928000028

Sample Surveys: Nonprobability Sampling

J.J. Forster, in International Encyclopedia of the Social & Behavioral Sciences, 2001

Nonprobability sampling describes any method for collecting survey data which does not utilize a full probability sampling design. Nonprobability samples are usually cheaper and easier to collect than probability samples. However, there are a number of drawbacks. Such methods can be prone to selection bias, and standard design-based methods of inference cannot be used to ensure approximately unbiased estimators of population quantities or to provide associated measures of precision. In this article, some of the more common methods of nonprobability sampling, quota sampling in particular, are introduced. Their advantages and disadvantages are discussed, and a formal framework for assessing the validity of inferences from nonprobability samples is described.

Read full chapter

URL: //www.sciencedirect.com/science/article/pii/B008043076700499X

Socialization: Political

M.K. Jennings, in International Encyclopedia of the Social & Behavioral Sciences, 2001

2 Methodology

The majority of the literature rests on a foundation of empirical methodology. Attention has focused more on individuals than on institutions and systems as units of analysis.

There are various modes of data gathering. Surveys of probability and nonprobability samples compose the major mode. Three variants have been used. One consists of self-administered questionnaires of elementary and secondary school students. The use of such instruments for young children has been questioned. A second variant consists of personal interviews, either face to face or by telephone. Some of these surveys employ a structured format while others are more semi-structured. Investigators using either mode typically employ statistical analysis. Other modes of data collection include intensive interviews, sometimes combined with participant observation and projective techniques, with a small number of subjects [Coles 1987]. Analysis in this instance is usually qualitative in nature. Although not typically classified under the heading of political socialization, analytic biographies of political figures—employing a variety of data sources—also constitute a mode of data gathering. Studies based on experiments are relatively rare.

Most investigations are cross-sectional, that is, the subjects were studied only once. However, a few over-time studies of the same individuals exist, thereby enabling more insight into developmental patterns. Some studies stretch from pre to full adulthood and involve either special populations or more general ones. Other longitudinal projects, especially those assessing the impact of educational institutions and political events, cover a shorter period of time.

Replicated surveys of the same populations provide historical trend data. Longitudinal studies of pre-adult populations are now available in a number of countries via national and international archives. Similarly, replicated surveys of adult birth cohorts are widely available. These surveys facilitate analyses devoted to assessing the probable impact of preadult experiences as cohorts age.

Read full chapter

URL: //www.sciencedirect.com/science/article/pii/B0080430767012365

General Interviewing Issues

Chauncey Wilson, in Interview Techniques for UX Practitioners, 2014

Sampling Methods

Sampling, the process of choosing the subset of people who will represent your population of users, is a complex topic that is described only briefly here. The two major types of sampling are probability and nonprobability [Bailey, 1994; Levy & Lemeshow, 1999; Robson, 2002]. In probability sampling, the probability of selection of each participant is known. In nonprobability sampling, the interviewer does not know the probability that a person will be chosen from the population. Probability sampling is expensive and time-consuming and may not even be possible because there is no complete list of everyone in a population. For many interview studies, you are likely to be dealing with nonprobability samples where you can use one or a combination of the following approaches [Bailey, 1994; Robson, 2002]:

Quota sampling. You try to obtain participants in relative proportion to their presence in the population. You might, for example, try to get participants in proportion to a distribution of age ranges.

Dimensional sampling. You try to include participants who fit the critical dimensions of your study [time spent as an architect or engineer, time using a particular product, experience with a set of software tools].

Convenience sampling. You choose anyone who meets some basic screening criteria. Many samples in UCD are convenience samples that can be biased in subtle ways. For example, the easiest people to find might be users from favorite companies that are generally evangelists of your product. You might end up with a “positivity bias” if you use participants from your favorite companies.

Purposive sampling. You choose people by interest, qualifications, or typicality [they fit a general profile of the types of participants who would be typical users of a product]. Samples that meet the specific goals of the study are sought out. For example, if you are trying to understand how experts in a particular field work on complex projects, you might seek out the “best of the best” and use them for your interviews.

Snowball sampling. You identify one good participant [based on your user profile or persona] who is then asked to name other potential participants, and so on. Snowball sample is useful when there is some difficulty in identifying members of a population. For example, if you are looking for cosmologists who use complex visualization tools, you might find one and then ask him or her about any friends or colleagues in the field who might want to be interviewed.

Extreme samples. You want people who are nontraditional or who have some exceptional knowledge that will provide an extreme or out-of-the-box perspective.

Extreme Input Can Be Useful

The use of “extremes” in user research can provide inspiration [Jansen et al., 2013] and help you understand the limits of a system. In addition to extreme samples of users, you can also explore extreme data sets that are large and dirty [something that usability research often ignores in small-scale testing] and extreme scenarios that highlight risks and rare, but critical, usage patterns.

Heterogeneous samples. You select the widest range of people possible on the dimensions of greatest interest [e.g., you might choose people from many industries, countries, genders, and experience ranges].

For any type of user research, it is important to be explicit about your sampling method and its limitations and biases.

Read full chapter

URL: //www.sciencedirect.com/science/article/pii/B9780124103931000065

Web-Based Survey

R. Michael Alvarez, Carla VanBeselaere, in Encyclopedia of Social Measurement, 2005

Developing Web Surveys

Respondents

Web-based surveys are only useful if they actually generate data, thus recruiting respondents is a priority. As discussed in the section on Web-survey typology, there are many different ways to recruit subjects. The choice of recruitment method will of course depend on the objectives of the survey. If the target population can be identified and easily contacted then producing probability-based Web surveys should be feasible. If, however, the intended target population is not well defined or readily contactable over the Internet, nonprobability respondent recruitment methods may be necessary. Inferences made about population parameters from nonprobability surveys are potentially problematic although several techniques have been proposed to improve the representativeness of Internet surveys.

Alvarez et al. discuss two prominent methods for recruiting respondents over the Internet. The first involves Web advertisements. Advertisements on various Web sites or newsgroups encouraging people to complete the Web survey is a fairly effective way of obtaining a large nonprobability sample of respondents. Another method to recruit respondents is through subscription or coregistration procedures. This involves asking individuals registering for another service whether they would like to provide their e-mail address and participate in Internet surveys. Once respondents provide their e-mail addresses, they can be contacted by e-mail to participate in Web-based surveys.

Web-Based Survey Panels

Because recruiting respondents over the Internet can be somewhat complicated, survey panels are popular. Rather than asking respondents to complete a single survey, Web-based survey panels recruit subjects to participate in a series of surveys. In order to obtain probability-based subjects, potential respondents are often initially contacted by telephone. Once respondents agree to participate in a survey panel, they are contacted by e-mail when they are required to complete a new survey. Using panels, researchers can draw samples from the registered respondents in order to undertake studies of specific subpopulations. Knowledge Networks claims that their Web-based survey panels is particularly effective for market research. Panels also offer the opportunity to examine temporal changes in respondent behavior and beliefs.

Researchers are currently studying the long-term effectiveness of Web-survey panels. Although the concept is relatively new, studies to date do not indicate that extended participation in Internet panels affects respondent behavior. However, preliminary data indicates that response rates do tend to decline with panel tenure. A significant problem for many Internet panels is that participants are frequently unreachable by e-mail because they have changed e-mail addresses or there are technical problems. These issues need to be continually examined especially for long-standing panels.

Creating the Survey

An advantage of Web-based surveys is that they are relatively easy to conduct. All that is needed is a Web site and some basic Web programming skills. Many surveys are created simply using Hypertext Markup Language [HTML]; there are dozens of HTML editors available and they are becoming increasingly sophisticated and easy to use. Data from surveys can be captured either by programming the form to e-mail the data to a specified address or through a common gateway interface [CGI] script. Several HTML development packages automate the process of developing CGI scripts necessary to capture data from HTML forms. Internet survey companies have even developed computer programs that automatically create surveys.

Despite the fact that Web-based surveys are easy to implement, their effective use requires an understanding of the methodological issues presented above. While Web surveys have many potential uses, making general statements about large populations based on Internet survey results is currently problematic. The Web opens up a whole new realm of survey possibilities, but it is important to evaluate surveys based on the fundamental criteria outlined in this article.

Read full chapter

URL: //www.sciencedirect.com/science/article/pii/B012369398500390X

Sample Surveys: Model-based Approaches

S.L. Lohr, in International Encyclopedia of the Social & Behavioral Sciences, 2001

2 Models in Descriptive and Analytic Uses of Surveys

Estimating a finite population mean or total is an example of a descriptive use of a survey: the characteristics of a particular finite population are of interest. In much social science research, survey data are used for analytic purposes: investigating relationships between factors and testing sociological theories. Data from the US National Crime Victimization Survey may be used to estimate the robbery rate in 1999 [descriptive], or they may be used to investigate a hypothesized relationship between routine activities and likelihood of victimization [analytic]. In the former case, the population of inference is definite and conceivably measurable through a census. In the latter, the population of inference is conceptual; the investigator may well be interested in predicting the likelihood of victimization of a future person with given demographic and routine activity variables.

Smith [1994] argued that design-based inference is the appropriate paradigm for official descriptive statistics based on probability samples. Part of his justification for this position was the work of Hansen et al. [1983], who provided an example in which small deviations from an assumed model led to large biases in inference. Brewer [1999] summarized work on design-based and model-based estimation for estimating population totals and concluded that a model-assisted generalized regression estimator [see Särndal et al. 1992], used with design-based inference, captures the best features of both approaches. Models must of course always be used for inference in nonprobability samples [see Sample Surveys: Nonprobability Sampling]; they may also be desirable in probability samples that are too small to allow the central limit theorem to be applied for inference.

Lohr [1999 Chap. 11] distinguished between obtaining descriptive official statistics and uncovering a ‘universal truth’ in an analytic use of a survey. Returning to the depression example, the investigator might be interested in the relationship between depression score [y] and variables such as marital status, financial resources, number of chronic health problems, and ability to care for oneself. In this case, the investigator would be interested in testing a theory that would be assumed to hold not just for the particular population of 10,000 women but for other populations as well, and should be making inferential statements about the βs in model [2]. The quantity for inference in the design-based setting is bp, the least squares estimate of β that would be obtained if the xis and yi were known for all 10,000 persons in the finite population. The quantity bp would rarely be of primary interest to the investigator, though, since it is merely a summary statistic for this particular finite population. In social research, models are generally motivated by theories, and a model-based analysis allows these theories to be tested empirically.

The generalized least squares estimator of β, β̭LS, would be the estimator of choice under a pure model-based approach because of its optimality properties under the proposed model. This estimator is, however, sensitive to model misspecification. An alternative, which achieves a degree of robustness to the model at the expense of a possibly higher variance, is to use the design-based estimator of bp. If the proposed stochastic model is indeed generating the finite population and if certain regularity conditions are met, an estimator that is consistent for estimating bp will also be consistent for estimating β. Under this scenario, a design-based estimate for bp also estimates the quantity of primary interest β and has the advantage of being less sensitive to model misspecification.

Regardless of philosophical differences on other matters of inference, it is generally agreed that two aspects of descriptive statistics require the use of models. All methods currently used to adjust for nonresponse [see Nonsampling Errors] employ models to relate nonrespondents to respondents, although the models are not necessarily testable. In small area estimation, sample sizes in some subpopulations of interest are too small to allow estimates of sufficient precision; models are used to relate such subpopulations to similar subpopulations and to useful covariates.

Read full chapter

URL: //www.sciencedirect.com/science/article/pii/B0080430767005283

Sample Design

W. Penn Handwerker, in Encyclopedia of Social Measurement, 2005

Selection Criteria Provide the Ingredients for Sample Designs

Cases can be selected on the basis of one or more of six criteria:

1.

Availability

2.

Fulfilling a size quota

3.

Random [or known probability] selection

4.

Case characteristics

5.

Presence in specific enumeration units

6.

Presence along transects or at specific map coordinates

All samples that utilize random [or known probability] selection are called probability samples. If one does not employ random selection, one produces one of four different forms of nonprobability samples.

Nonprobability Samples

If you select a predetermined number or proportion of cases with specific case characteristics, or from specific enumeration units, transects, or sets of map coordinates, you produce a quota sample. If you select cases on the basis of case characteristics to acquire specific forms of information, you produce a purposive [judgment] sample. If you select cases simply because they will participate in your study, you produce an availability [convenience] sample. If cases become available because one case puts you in contact with another, or other cases, you produce a snowball sample.

Probability Samples

Probability samples are distinguished from nonprobability samples because the former exhibit known sampling distributions that warrant parameter estimation with classical statistical tests [e.g., chi-squared, t test, and F ratio]. By convention, we identify parameters with Greek letters, such as β [beta], α [alpha], ɛ [epsilon], ρ [rho], and σ [sigma]. Samples, in contrast, yield statistics. By convention, we identify statistics with Latin letters and words [e.g., b, median, percentage, and mean]. Each statistic constitutes a point estimate of a parameter, which is one's single best guess about the value of the parameter.

Statistics constitute point estimates of parameters because samples of populations cannot perfectly replicate the properties of the populations from which they derive. Every sample yields different findings, and every statistic contains three sources of error [construct, measurement, and sampling]. Construct error derives from trying to measure a construct that imperfectly fits the culture or cultures found in the population studied. Measurement error derives from imperfections in the means by which a value is assigned to an observation from a set of possible outcomes. To the extent to which significant construct and measurement errors can be ruled out, the difference between a specific statistic and the population parameter constitutes sampling error in that specific sample. Measurements of the same variable made on a large number of samples of the same size drawn from the same population exhibit a characteristic sampling distribution of errors around the parameter. Some statistics underestimate the parameter, whereas others overestimate the parameter.

Sampling errors may reflect chance or bias. Sampling errors that derive from chance exhibit characteristic distributions. Many such sampling distributions [the family of t distributions and the normal distribution] are symmetrical and are summarized by a mean of 0 and a standard deviation of 1. The average amount of error in a sampling distribution is called the standard error rather than standard deviation to distinguish sampling distributions from the frequency distributions of the variables studied in social science research.

Although some statistics underestimate the parameter and others overestimate it, when cases are selected independently and have the same probability of inclusion in any one sample, sampling errors come solely from chance. When this condition applies, the sampling distribution of all possible statistics reveals that most statistics come very close to the parameter, and the average amount of sampling error is 0. With statistics that exhibit a normal sampling distribution, for example, 68% of all sample statistics fall within ±1.00 standard errors of the parameter, and 95% of all sample statistics fall within ±1.96 standard errors of the parameter.

Small samples contain large amounts of sampling error because randomly selected extreme values exert great effects. Large samples contain small amounts of sampling error and thus estimate parameters very precisely. Sample precision is measured by the size of confidence intervals. Accurate samples yield confidence intervals that contain the parameter a given proportion [usually 95%] of the time. Statistical test findings apply to samples of all sizes because they incorporate into their results the degree of sampling error contained in samples of different sizes. Confidence intervals for small samples are wider than confidence intervals for large samples, but statistics from both large and small samples estimate parameters equally accurately.

This generalization holds only for statistics from samples that are reasonably unbiased. Unbiased samples are those in which all members of the population have an equal chance of selection. The only way to reliably obtain a reasonably unbiased sample is to employ the random selection criterion.

Simple Random Samples

Simple random samples [SRSs] constitute the reference standard against which all other samples are judged. The procedure for selecting a random sample requires two steps. First, make a list of all members of the population. Second, randomly select a specific number of cases from the total list. Random selection may rely on tables of pseudo-random numbers or the algorithms that generate uniform pseudo-random number distributions in statistical analysis software such as SYSTAT. One may sample with or without replacing cases selected for the sample back into the population. Sampling without replacement produces unequal probabilities of case selection, but these are inconsequential except with very small populations. More important, even SRSs overestimate the true standard error by the factor, N/N−n. Application of the finite population multiplier, [N−n]/N, will produce correct standard errors. The importance of this correction increases as the ratio of sample size [n] to population size [N] increases.

Random Systematic Samples

Random systematic samples [RSSs] constitute a variation on SRSs in which random selection of a starting point is substituted for random selection of all cases. For example, to select an RSS of 20% of a population, randomly select a number between 1 and 5, make your first case the one with the randomly selected number, and select every fifth case thereafter. To select an RSS of 5% of a population, randomly select a number between 1 and 20, make your first case the one with the randomly selected number, and select every 20th case thereafter.

Periodicity in a list of population members introduces significant bias into RSSs. In the absence of periodicity, and with a known population size, to determine a sampling interval [k], divide the size of the population [N] by a desired sample size [n]. RSSs produce unbiased samples when k is an integer. The bias introduced when k is not an integer is inconsequential with large populations. However, if you know the size of the population, the following procedure always yields unbiased estimates:

1.

Randomly select a number [j] between 1 and N.

2.

Express the ratio [j/k] as an integer and a remainder [m].

3.

When m equals 0, select the case numbered k as your first sample element; when m does not equal 0, select the case numbered m as your first sample element.

Stratified, Cluster, Transect, and Case-Control Samples

All other probability samples incorporate SRSs or RSSs into the selection process. Stratified samples, for example, consist of a series of simple random or random systematic samples of population sectors identified by case characteristics [e.g., age, class, gender, and ethnicity] or combinations of characteristics [e.g., old and young women, and old and young men]. Disproportionally stratified samples employ a quota criterion to oversample population sectors that might otherwise be insufficiently represented in the final sample. Cluster samples consist of samples in which cases are selected from SRSs or RSSs of enumeration units that contain sets of cases, such as households, hospitals, city blocks, buildings, files, file drawers, or census enumeration districts. Probability proportional to size samples are cluster samples in which the number of cases selected from specific enumeration units matches a quota proportional to the size of unit relative to the entire population. Transect samples consist of samples in which cases or enumeration units are selected from SRSs or RSSs of units that lie along randomly drawn transects or randomly selected map coordinates. Case-control samples consist of a set of purposefully [judgmentally] identified cases, a small set of which may be selected randomly, plus a set of randomly selected controls. This sampling procedure originated in epidemiology, in which cases are characterized by specific health conditions not experienced by controls. However, the procedure is readily generalizable by defining cases and controls by reference to a binary variable that distinguishes cases with a specific experience from controls without that experience.

Read full chapter

URL: //www.sciencedirect.com/science/article/pii/B0123693985000761

Sociology

Joseph W. Elder, in Encyclopedia of Social Measurement, 2005

Quantitative Research

For sociologists, one of the main purposes of quantitative research is to make statements with some degree of confidence about sizable human groups. Such statements typically accept the hypothetico-deductive premises of regularity and recurrence seen to be present in the natural and physical sciences. Quantitative research often begins with the researcher selecting the population [or universe] to be studied, based on the nature of the questions to be answered. The selected universe could be all the citizens of a designated nation, laborers in a sweatshop, or the homeless in a large urban center. It could be women chief executive officers of Fortune 500 corporations, or children raised in homes with same-gender parents. If the universe is large, the researcher might wish to select a manageable-sized sample from which to gather data. The sample could be selected in such a way that every unit in the universe has an equal likelihood of being selected [a probability sample]. Or the sample could be structured in a way that guarantees the inclusion of certain subsectors [e.g., genders, ages, locations, ethnic groups] of the universe. An initial sample could be selected on the basis of convenience [a nonprobability sample] and then expanded by “snowball” referrals to additional units of the sample. The manner in which the sample is selected directly affects the confidence with which findings from the sample can be generalized statistically to the universe. A researcher can introduce additional degrees of sophistication by engaging in time-series sampling, paired- comparative sampling, and longitudinal panel studies.

The hypothetico-deductive approach calls for examining statements of purported invariant relationships [hypotheses] between antecedent [independent] variables and consequent [dependent] variables, presented so that the relationship is capable of falsification. In order to be capable of falsification, such statements must be nontautological and without empirically continuous antecedents or consequents, and they must specify the relationships between the antecedents and consequents [e.g., as necessary conditions or sufficient conditions]. If, after the data have been collected and examined, the hypotheses have not been falsified, it can be stated that the data demonstrate the hypotheses. But it cannot be stated that the hypotheses have been “proved.” “Proving” invariance requires evidence that has yet to be gathered.

In order for a statement of invariance to be examined, the variables in the statement must be operationalized [i.e., made capable of observation and measurement]. If, for example, the statement refers to domestic violence, democracy, or intelligence, instructions must be provided for how data are to be gathered as evidence of domestic violence, democracy, or intelligence. Furthermore, the case must be made that the data obtained by following the instructions are valid indicators [i.e., actually provide evidence] of domestic violence, democracy, etc. This case must be made rhetorically, because it cannot be made empirically.

Comparisons are at the heart of sociological research: comparisons of sociological phenomena at two points in time, comparisons of dependent variables following changes in independent variables, and quasi-experiments in which treatment effects on dependent variables are compared with control effects on comparable dependent variables while accounting for experimenter effects. Quantitative research tends to focus on structural agency [rather than human agency] and hence recurrent comparability. The goal of such research is to demonstrate empirically supported sociological relationships, often in sizable human groups. In order to gather systematic information from large numbers of people, sociological researchers frequently conduct surveys for which they prepare data-gathering instruments, such as questionnaires [for written responses] and interview schedules [for oral responses]. The questions asked in these instruments may be close ended [with limited fixed choices] or open ended. The major advantage of close-ended questions is the ease of final tabulation. The major advantage of open-ended questions is the possible acquisition of useful unanticipated information. If replies to open-ended questions are to be used systematically, they must be coded. Raters who code the open-ended answers must be trained so as to minimize differences in their coding patterns and to maximize their interrater reliability.

Prior to using questionnaires or interview-schedules in the field, it is essential to pretest them with a sample of respondents who will not be included in the final survey. Pretests may uncover problems of question clarity and accuracy. Because the phrasing of questions can alter respondents' answers, efforts are usually made to ask each respondent identically worded questions. At times, however, wording equivalency may be more important than wording identity. In a survey of sexual behavior, for example, alternative words were substituted for certain sexual practices so that respondents from different social backgrounds could understand what sexual practices were being referenced. Designers of questionnaires and interview schedules can prepare scales [such as Guttman scales, Likert scales, or Thurston scales] to measure relative degrees of attitudes or practices. Interview schedules require trained interviewers to gather data in face-to-face conversations or over the telephone. One advantage of using interview schedules rather than questionnaires is the possibility of asking respondents to answer follow-up questions for clarification or additional details. Advantages of using questionnaires generally include greater speed and lower cost of data collection.

Reliability [the extent to which questionnaires or interviews give the same results in repeated trials] is desirable in survey instruments. Asking the same question in different ways can test a procedure's reliability. In a survey of sexual behavior, for example, differently worded questions were asked at different points in the interview about the number and gender of the respondent's recent sexual partners. Afterward, the respondents' replies to each of the different questions were compared for consistency, and hence for procedure reliability. Another technique for testing a survey's reliability is to reinterview a random selection of already interviewed subjects. The reinterview can achieve two ends: it can establish the fact that the subjects were indeed interviewed, and it can identify differences between the ways subjects originally answered the questions and the ways they answered the questions the second time. The narrowness of differences between the first and second answers would be a measure of the survey instrument's reliability. Arranging for questions to be asked in multiple languages requires special care. A useful technique is to have one translator convert the English questions into Spanish [for example] and a different translator convert the Spanish questions back into English. Comparing the original English with the “back-translated” English could identify words or phrases needing to be retranslated.

The heart of the sociological enterprise involves concepts and the measurement of concepts. Over the years, increasingly sophisticated statistical procedures have refined researchers' abilities to analyze data. But sociology's fundamental philosophical problems remain the same: How are social collectivities accurately defined and measured, and how are their causal relationships established?

Read full chapter

URL: //www.sciencedirect.com/science/article/pii/B0123693985002875

What type of probability sampling is non

Probability sampling uses random sampling techniques to create a sample. For each element in the sample, the probability is known and non-zero. In principal, every element of the population has the same chance at being included in the sample. This is a achieved with a sampling frame.

In which sampling method every element of the population does not get a chance of being selected?

In non-probability sampling, the sample is selected based on non-random criteria, and not every member of the population has a chance of being included. Common non-probability sampling methods include convenience sampling, voluntary response sampling, purposive sampling, snowball sampling, and quota sampling.

What type of sampling is non

Non-probability sampling is a method of selecting units from a population using a subjective [i.e. non-random] method. Since non-probability sampling does not require a complete survey frame, it is a fast, easy and inexpensive way of obtaining data.

Which sampling is also known as the method of chance selection?

Probability sampling refers to the selection of a sample from a population, when this selection is based on the principle of randomization, that is, random selection or chance.

Bài Viết Liên Quan

Chủ Đề