Skip to main content

Section 6.2 Sampling Methods and Sampling Bias

As we mentioned in a previous section, the first thing we should do before conducting a survey is to identify the population that we want to study. Suppose we are hired by a politician to determine the amount of support he has among the electorate should he decide to run for another term. What population should we study? Every person in the district? Not every person is eligible to vote, and regardless of how strongly someone likes or dislikes the candidate, they don't have much to do with him being re-elected if they are not able to vote.

What about eligible voters in the district? That might be better, but if someone is eligible to vote but does not register by the deadline, they won't have any say in the election either. What about registered voters? Many people are registered but choose not to vote. What about "likely voters?"

This is the criteria used in much political polling, but it is sometimes difficult to define a "likely voter." Is it someone who voted in the last election? In the last general election? In the last presidential election? Should we consider someone who just turned 18 a "likely voter?" They weren't eligible to vote in the past, so how do we judge the likelihood that they will vote in the next election?

In November 1998, former professional wrestler Jesse "The Body" Ventura was elected governor of Minnesota. Up until right before the election, most polls showed he had little chance of winning. There were several contributing factors to the polls not reflecting the actual intent of the electorate:

  • Ventura was running on a third-party ticket and most polling methods are better suited to a two-candidate race.

  • Many respondents to polls may have been embarrassed to tell pollsters that they were planning to vote for a professional wrestler.

  • The mere fact that the polls showed Ventura had little chance of winning might have prompted some people to vote for him in protest to send a message to the major-party candidates.

But one of the major contributing factors was that Ventura recruited a substantial amount of support from young people, particularly college students, who had never voted before and who registered specifically to vote in the gubernatorial election. The polls did not deem these young people likely voters (since in most cases young people have a lower rate of voter registration and a turnout rate for elections) and so the polling samples were subject to sampling bias: they omitted a portion of the electorate that was weighted in favor of the winning candidate.

Definition 6.2.1. Sampling Bias.

A sampling method is biased if every member of the population doesn’t have equal likelihood of being in the sample.

So even identifying the population can be a difficult job, but once we have identified the population, how do we choose an appropriate sample? Remember, although we would prefer to survey all members of the population, this is usually impractical unless the population is very small, so we choose a sample. There are many ways to sample a population, but there is one goal we need to keep in mind: we would like the sample to be representative of the population.

Returning to our hypothetical job as a political pollster, we would not anticipate very accurate results if we drew all of our samples from among the customers at a Starbucks, nor would we expect that a sample drawn entirely from the membership list of the local Elks club would provide a useful picture of district-wide support for our candidate.

One way to ensure that the sample has a reasonable chance of mirroring the population is to employ randomness. The most basic random method is simple random sampling.

Definition 6.2.2. Simple Random Sample.

A random sample is one in which each member of the population has an equal probability of being chosen. A simple random sample is one in which every member of the population and any group of members has an equal probability of being chosen.

Example 6.2.3.

If we could somehow identify all likely voters in the state, put each of their names on a piece of paper, toss the slips into a (very large) hat and draw 1000 slips out of the hat, we would have a simple random sample.

In practice, computers are better suited for this sort of endeavor than millions of slips of paper and extremely large headgear.

It is always possible, however, that even a random sample might end up not being totally representative of the population. If we repeatedly take samples of 1000 people from among the population of likely voters in the state of Washington, some of these samples might tend to have a slightly higher percentage of Democrats (or Republicans) than does the general population; some samples might include more older people and some samples might include more younger people; etc. In most cases, this sampling variability is not significant.

Definition 6.2.4. Sampling Variability.

The natural variation of samples is called sampling variability.

This is unavoidable and expected in random sampling, and in most cases is not an issue.

To help account for variability, pollsters might instead use a stratified sample.

Definition 6.2.5. Stratified Sampling.

In stratified sampling, a population is divided into a number of subgroups (or strata). Random samples are then taken from each subgroup with sample sizes proportional to the size of the subgroup in the population.

Example 6.2.6.

Suppose in a particular state that previous data indicated that the electorate was comprised of 39% Democrats, 37% Republicans and 24% independents. In a sample of 1000 people, they would then expect to get about 390 Democrats, 370 Republicans and 240 independents. To accomplish this, they could randomly select 390 people from among those voters known to be Democrats, 370 from those known to be Republicans, and 240 from those with no party affiliation.

Stratified sampling can also be used to select a sample with people in desired age groups, a specified mix ratio of males and females, etc. A variation on this technique is called quota sampling.

Definition 6.2.7. Quota Sampling.

Quota sampling is a variation on stratified sampling, wherein samples are collected in each subgroup until the desired quota is met.

Example 6.2.8.

Suppose the pollsters call people at random, but once they have met their quota of 390 Democrats, they only gather people who do not identify themselves as a Democrat.

You may have had the experience of being called by a telephone pollster who started by asking you your age, income, etc. and then thanked you for your time and hung up before asking any "real" questions. Most likely, they already had contacted enough people in your demographic group and were looking for people who were older or younger, richer or poorer, etc. Quota sampling is usually a bit easier than stratified sampling, but also does not ensure the same level of randomness.

Another sampling method is cluster sampling, in which the population is divided into groups, and one or more groups are randomly selected to be in the sample.

Definition 6.2.9. Cluster Sampling.

In cluster sampling, the population is divided into subgroups (clusters), and a set of subgroups are selected to be in the sample.

Example 6.2.10.

If the college wanted to survey students, since students are already divided into classes, they could randomly select 10 classes and give the survey to all the students in those classes. This would be cluster sampling.

Definition 6.2.11. Systematic Sampling.

In systematic sampling, every nth member of the population is selected to be in the sample.

Example 6.2.12.

To select a sample using systematic sampling, a pollster calls every 100th name in the phone book.

Systematic sampling is not as random as a simple random sample (if your name is Albert Aardvark and your sister Alexis Aardvark is right after you in the phone book, there is no way you could both end up in the sample) but it can yield acceptable samples.

Perhaps the worst types of sampling methods are convenience samples and voluntary response samples.

Definition 6.2.13. Convenience Sampling and Voluntary Response Sampling.

Convenience sampling is samples chosen by selecting whoever is convenient.

Voluntary response sampling is allowing the sample to volunteer.

Example 6.2.14.

A pollster stands on a street corner and interviews the first 100 people who agree to speak to him. This is a convenience sample.

Example 6.2.15.

A website has a survey asking readers to give their opinion on a tax proposal. This is a self-selected sample, or voluntary response sample, in which respondents volunteer to participate.

Usually voluntary response samples are skewed towards people who have a particularly strong opinion about the subject of the survey or who just have way too much time on their hands and enjoy taking surveys.

Problem 6.2.16. Try It Now.

In each case, indicate what sampling method was used:

  1. Every 4th person in the class was selected

  2. A sample was selected to contain 25 men and 35 women

  3. Viewers of a new show are asked to vote on the show’s website

  4. A website randomly selects 50 of their customers to send a satisfaction survey to

  5. To survey voters in a town, a polling company randomly selects 10 city blocks, and interviews everyone who lives on those blocks.

Answer.
  1. Systematic

  2. Stratified or Quota

  3. Voluntary Response

  4. Simple Random

  5. Cluster

There are number of ways that a study can be ruined before you even start collecting data. The first we have already explored - sampling or selection bias, which is when the sample is not representative of the population. One example of this is voluntary response bias, which is bias introduced by only collecting data from those who volunteer to participate. This is not the only potential source of bias.

Note 6.2.17. Sources of bias.

Some sources of bias in studies are:

  • Sampling bias - when the sample is not representative of the population

  • Voluntary response bias - the sampling bias that often occurs when the sample is volunteers

  • Self-interest study - bias that can occur when the researchers have an interest in the outcome

  • Response bias - when the responder gives inaccurate responses for any reason

  • Perceived lack of anonymity - when the responder fears giving an honest answer might negatively affect them

  • Loaded questions - when the question wording influences the responses

  • Non-response bias - when people refusing to participate in the study can influence the validity of the outcome

Example 6.2.18.

Consider a recent study which found that chewing gum may raise math grades in teenagers[1]. This study was conducted by the Wrigley Science Institute, a branch of the Wrigley chewing gum company. This is an example of a self-interest study; one in which the researches have a vested interest in the outcome of the study. While this does not necessarily ensure that the study was biased, it certainly suggests that we should subject the study to extra scrutiny.

Example 6.2.19.

A survey asks people “when was the last time you visited your doctor?” This might suffer from response bias, since many people might not remember exactly when they last saw a doctor and give inaccurate responses.

Sources of response bias may be innocent, such as bad memory, or as intentional as pressuring by the pollster. Consider, for example, how many voting initiative petitions people sign without even reading them.

Example 6.2.20.

A survey asks participants a question about their interactions with members of other races. Here, a perceived lack of anonymity could influence the outcome. The respondent might not want to be perceived as racist even if they are, and give an untruthful answer.

Example 6.2.21.

An employer puts out a survey asking their employees if they have a drug abuse problem and need treatment help. Here, answering truthfully might have consequences; responses might not be accurate if the employees do not feel their responses are anonymous or fear retribution from their employer.

Example 6.2.22.

A survey asks “do you support funding research of alternative energy sources to reduce our reliance on high- polluting fossil fuels?” This is an example of a loaded or leading question - questions whose wording leads the respondent towards an answer.

Loaded questions can occur intentionally by pollsters with an agenda, or accidentally through poor question wording. Also a concern is question order, where the order of questions changes the results. A psychology researcher provides an example:

My favorite finding is this: we did a study where we asked students, 'How satisfied are you with your life? How often do you have a date?' The two answers were not statistically related - you would conclude that there is no relationship between dating frequency and life satisfaction. But when we reversed the order and asked, 'How often do you have a date? How satisfied are you with your life?' the statistical relationship was a strong one. You would now conclude that there is nothing as important in a student's life as dating frequency.
Example 6.2.23.

A telephone poll asks the question “Do you often have time to relax and read a book?”, and 50% of the people called refused to answer the survey. It is unlikely that the results will be representative of the entire population. This is an example of non-response bias, introduced by people refusing to participate in a study or dropping out of an experiment. When people refuse to participate, we can no longer be so certain that our sample is representative of the population.

Problem 6.2.24. Try It Now.

In each situation, identify a potential source of bias:

  1. A survey asks how many sexual partners a person has had in the last year

  2. A radio station asks readers to phone in their choice in a daily poll.

  3. A substitute teacher wants to know how students in the class did on their last test. The teacher asks the 10 students sitting in the front row to state their latest test score.

  4. High school students are asked if they have consumed alcohol in the last two weeks.

  5. The Beef Council releases a study stating that consuming red meat poses little cardiovascular risk.

  6. A poll asks “Do you support a new transportation tax, or would you prefer to see our public transportation system fall apart?”

Answer.
  1. Response bias - historically, men are likely to over-report, and women are likely to under-report to this question.

  2. Voluntary response bias - the sample is self-selected

  3. Sampling bias - the sample may not be representative of the whole class

  4. Lack of anonymity

  5. Self-interest study

  6. Loaded question