POLITICAL POLLING TECHNIQUES (Weeks 2-3)

(Note: these are actual class notes, valuable to those having an excused class absence, or those wishing to review their class notes for the test. Double spaced notes reflect subjects that are so important that they are likely to be asked about on a test.)

A poll is a sample of the population that is designed to be representative of the entire population. As such, you may read about a national poll of only 1,000 people, but if it is done with the correct scientific method, it should be a pretty accurate representation of the entire population. Historically, though, we have had at least four problems with polls:

1) Biased samples- the sample does not accurately represent important aspects of the population, because it has an over-representation or under-representation of a particular group. For example, the Literary Digest poll in 1936 was pretty accurate in previous presidential elections, but it sampled its readers and car owners, and those groups were of a higher socioeconomic status (SES, high incomes and education levels) than the general population. FDR was running for re-election with liberal economic policies that were popular with the lower but not the upper SES, so the poll had the Republican Landon beating him in a landslide. Actually, FDR ended up winning in a landslide, as the poll had a biased sample. As such, today we check our polling samples with census data and then weight the results by demographic characteristics to ensure that we have representative samples.

2) Time-bound polls. Polls are only accurate for the exact time that they were conducted, since people can change their opinions over time. Some inaccurate polls were in 1948, when pollsters stopped polling weeks before the election, as Republican challenger Dewey was easily leading. Democratic incumbent Truman conducted an aggressive whistle-stop campaign where he blasted Republicans as only caring for the rich and big business and reminded voters that FDR and the Democrats had given them Social Security and protected their right to join labor unions. Despite a bad economy, Truman won an upset election victory, and the next day he laughingly held up a Republican newspaper’s (Chicago Tribune) erroneous headline of “Dewey Defeats Truman.” (The textbook on page 32 also uses this election as an example of a biased sample, as the pollsters’ quota sampling method underrepresented the lower SES.) Another example of a time-bound poll was in 1980, when Republican Reagan had only a slight lead in the polls over incumbent Jimmy Carter (who faced a bad economy and international disasters). Over the weekend, the one debate sank into the thoughts of the undecideds, as Reagan rebutted Carter’s claims that he was a conservative extremist by saying, “There you go again, Mr. President. Golly shucks. I didn’t support Medicare in the 1960s because I supported the alternate free market plan called Eldercare.” Reagan’s closing statement was: “Are you better off today than you were four years ago? Can your paycheck buy as many groceries as it could four years ago? If so, vote for the incumbent. If not, vote for a change, give my program a chance.” Also, the day before the election was the one-year anniversary of the Iranian hostage crisis, and the network news programs proclaimed: “Day 365. America held hostage.” So the national polls that had stopped polling the Friday before the election were wrong, as Reagan won in a landslide, and even brought in a Republican controlled Senate. Two pollsters were correct, however. They were the candidates’ own pollsters, who kept polling until the night before the election. And so when President Carter was flying down to Plains Georgia to vote in his hometown, his pollster walked up to him and said, “Mr. President. We’ve just finished our polling, and you have lost. You have lost big. Your party is going to lose the Senate, maybe even the House.” And so poor Jimmy Carter got off the plane and tearfully told the crowd, “I hope I haven’t let you down.” So the moral is, keep polling right up to election day.

3) Hard to estimate likely voters. Some people will lie and say they plan to be good citizens and to vote, when they will not really vote. So pollsters ask a series of questions to determine who is likely to vote- likely voters. They can ask about interest in the campaign, knowledge of where your polling place is, and similar political interest and knowledge questions. The Mississippi Poll asked three questions to determine likely voters: reported likelihood of voting with only the Definitely Will Vote category considered as likely voters (not the Somewhat Likely category); interest in the political campaigns; ability to recall their U.S. House incumbent’s name (only about half can recall the name without prompting). A further problem is to determine who is likely to vote in a party’s primary, if you are asking about their party primary vote intention. In the 1991 Mississippi gubernatorial contests, one pollster estimated that about one-third of those voting on primary day would vote in the Republican primary, and they predicted that a moderate conservative (party switcher state auditor Pete Johnson) would beat a strong conservative (life-long GOP activist and construction company owner Kirk Fordice). Consistent with history at that time, only about 10% of votes were cast in the GOP primary, and they were the more conservative and partisan Republicans, so Fordice won. The poll probably relied on a party identification question to determine the party primary voted in, and since the real contest at that time was in the Democratic primary, many “weak” and “independent” Republicans voted in the Democratic primary (which was a slugfest between incumbent governor Ray Mabus and his challenger, Congressman Wayne Dowdy). If the poll had just considered strong Republicans as GOP primary voters, they would have gotten down to about the 10% level, it would have been a more conservative group, and Fordice would have been favored. Fordice went on to upset Mabus in the general election, becoming the first Republican governor of Mississippi since Reconstruction.

4) Social desirability response bias. Many polls, particularly in the Rust Belt, underestimated Trump’s vote in 2016. It is likely that Trump was so controversial that some of his voters weren’t willing to admit that they planned to vote for him, so they said they were undecided or even said they planned to vote for Clinton. The press and Democratic politicians were labeling him as a bigot (and continued to do so throughout his presidency). One of my students said that her father had told a pollster that he planned to vote for Hillary Clinton. “Why did you say that, Dad? You know you’re going to vote for Trump?” “Uh, I don’t want anyone to get the wrong impression.” This problem continued into 2020, as again some polls underestimated Trump’s support. Our Mississippi Poll had a similar problem in 2014, as we greatly underestimated conservative activist Chris McDaniel’s support in the GOP primary against incumbent Senator Thad Cochran. In that case, we may have had a biased response rate as some conservatives just refused to even answer a poll from a perceived “liberal university,” so McDaniel’s supporters may have been underrepresented in the sample. After McDaniel led by 1% in the first GOP primary, Cochran had to fight for his life to win the runoff, and he then went on to easily win the general election against a Democrat (our poll correctly called the general election outcome). This is a hard question to deal with, and media and politicians who keep calling candidates names just make the lives of pollsters harder.

 

Sampling Error. An Economist/YouGov national poll in July 22-25, 2023 had Biden receiving 44% to Trump's 40%. With a sample of 1306 registered voters, the sample error was +- 3%. That means that in the entire population of registered voters, Biden's vote would be 95% likely to range from 41% to 47%; Trump's vote in the population would range from 37% to 43%. As such, the results are too close to call, since it is possible that Trump is ahead in the population of likely voters. (see website RealClearPolitics.com) Three things affect the level of sample error:

1) Sample size- the larger the sample size, the less the error. So, it is better to interview 1,100 people (yielding only 3% error) than 400 people (5% error). It’s like flipping a coin, the more tosses, the closer you get to the 50-50 split of a two-sided coin, heads and tails.

2) Homogeneity of population- the more united the population is on an issue, the smaller the sample size. A sample size of 400 with a population evenly divided on an issue (half want to vote Republican and half Democratic) yields a 5% sample error. A sample of 400 people who are overwhelmingly for an issue by a 90% value yields a 3% sample error. (Such a question might be: are you proud of your nation?) See the sample error chart in my Political Analysis class notes.

3) A cluster sample produces higher sample error, about 20% higher. A cluster sample is when your sample members are not independently selected from each other. For example, in-person surveys (and even two-stage random digit dialing phone surveys) may randomly choose up to 5 people on the same city block, and since those people may share similar SES and race characteristics, they are not independently selected. As such, your sample size is not as great as you think, so you have more sample error. When our Mississippi Polls used two-stage random digit dialing, we began by sampling 600 people. The chart at a 50-50 worst case split gave a 4.1% sample error. We added 20% to that, which is .2. So, 4.1 X .2 yields .82 additional error. So our total sample error for our cluster sample was about 5%, not 4.1% (4.1 + .82), which is what we reported.

 

Advantages and disadvantages of the three major types of surveys historically used:

In-person surveys (where pollsters go door-to-door to survey people). They use a multi-stage cluster design, discussed later.

Advantages:

1) Observe the respondent, and clear up confusion about the question that shows on their face. You can slowly ask the question a second time, but you can’t add any more information to the question.

2) Obtain objective information about the respondent (example, if they say they have a low income, but they have a new Cadillac in their driveway, you can report their likely untruthfulness).

3) You can use visual aids. When ranking candidates on a 100-point feeling thermometer in terms of how hot (like them) or cold (dislike) you are to them, you can show the respondent an actual thermometer. You can do card sort for five ideological groups, having five boxes with labels ranging from very liberal to very conservative, give them cards with the pictures of the candidates, and then have them put each card in the appropriate box to measure their perception of the candidate’s ideology.

Disadvantages:

1) Expensive. You have to pay the cost of the interviewers traveling across the country (for national surveys), then the cost of their staying at a motel, and their own individual time doing all of that. Plus, this process can take up to two months for all of the interviewers. Such national in-person surveys can cost hundreds of thousands of dollars, or even millions.

2) Safety of the interviewer may be endangered. Your interviewers are going into all kinds of neighborhoods that may have high crime, neighborhoods where even a pizza company may refuse to go into. You may face legal liability if they are hurt.

3) Interviewer fraud. It may be hard to monitor your interviewers, as they work on their own. They may falsify some of their surveys. If someone is interviewing in New Orleans, they may be in a bar on Bourbon Street and filling out the forms themselves (Uh, this is a liberal, Democrat, voted for Biden. Uh, this next survey is a conservative Republican, loves Trump. Etc.). You have to have some way of verifying that they actually interviewed a real person, perhaps by having them report the phone number or address of each respondent, then contacting that person to ensure that the survey was actually done (You then have to remove such identifying information from each survey to ensure that the results are anonymous.).

Telephone surveys (where pollsters sit in a room, and call people). The methods used are discussed later.

Advantages:

1) Quick. You can complete your entire survey in a few days or a few weeks. You can use as many phones as the room will contain and for as many interviewers as you can hire. Typically, you phone from 5:30-9:30 PM during the weeknights, and most of the day on Saturdays and Sundays. You could technically even complete a survey in one day, but that is not advisable, since you need to call back people whom you can’t get so that you have a representative sample (don’t leave out people who are busy with work or with their social life).

2) Cost effective. You do not have to pay travel expenses; you are just paying for the interviewers’ time when they are on the phone or dialing the phone. You also have the cost of the computers and their upkeep, plus the cost of buying the phone numbers from a marketing firm. When we did the Mississippi Poll, our cost was as little as $2,000 (for the marketing firm phone numbers); students in Political Analysis did the calls as their lab requirement; two professors donated their time supervising and upkeeping the computer equipment.

3) You can eliminate fraud, due to the centralized interviewing. As supervisor, I would walk around the room and listen to the student interviewers, so it was clear that they were actually doing the interviews with real people. If you hire a supervisor, make sure that they do their job, and don’t study for their graduate courses or socialize with a best friend.

4) Interviewer safety. Interviewers work in this one (or two) room, so there is no danger of their going into unsafe neighborhoods. Our Mississippi Poll polling was done at the SSRC (Social Science Research Center) in the Research Park at MSU, and interviewers could park literally 20 feet from their polling room, so even the walk from their car to the building was safe (watch out turning into the Research Park, though, as it is a high traffic area).

Disadvantages:

1) Historically, we worried about excluding people without telephones or those with only cell phones. When we started polling in 1981, only about 80% of Mississippi households had telephones, so we worried about under sampling the lower SES. By the turn of the century, households were up to 98% phone coverage, even in Mississippi. But our methodology sampled only landlines, so the rise of cell phones led us to eventually greatly under sample young adults. Our current methodology includes cell phones as well as land lines, so this is not a big disadvantage anymore.

2) You can’t use visual aids, so you are entirely dependent on the voice of the interviewer. This hasn’t been a big problem for the Mississippi poll, as the interviewer just slowly repeats what the response categories of complex questions are, such as very liberal, somewhat liberal, moderate or middle of the road, somewhat conservative, very conservative. Race and gender of the interviewer has not been a problem. I thought we’d have a problem with a real country white guy who yelled into the phone when asking questions, but he ended up having a high response rate (I guess he sounded like a regular guy, hunter, fisherman, whatever.). So this is also not a big disadvantage. As you can see, telephone interviewing has many advantages and few disadvantages.

Mail surveys (use bulk mail, use three waves or mailings if people don’t return them).

Advantages:

1) Cheap. No cost for interviewers. You only pay for paper, printing, postage and return envelopes and their postage. Requiring that respondents use a number 2 pencil means that the returned forms can be read into a machine that then produces an electronic database, so no cost of typing in all of the survey responses. You can add the cost of a graduate student to process the mailings, which we did for our NSF grants studying grassroots party activists in Mississippi.

2) You can use mail surveys for some specialized populations. A specialized population is more interested in the subject of the survey, so they are more likely to complete the surveys. Thus, we used the mail surveys for the NSF grants and sent them to each county party’s chair and its committee members. It is not generally advisable to use mail surveys for the general population, for the following reasons cited under disadvantages.

Disadvantages:

1) They exclude illiterates. Unless someone in the home reads the surveys to them, which may seldom happen. So you can end up with a biased sample.

2) You can’t control who answers the survey. You may want to randomly select men and women, but a man may get the survey, and he may just give it to the woman of the house to answer. Or, the adult may give the survey to their teenage kid (Hey, you’re taking civics. Why don’t you fill out this thing?). So, the person answering the survey may not even be eligible, age-wise.

3) You can’t control the order of the questions asked. We may prefer that respondents answer general life questions first, then a vote choice item, then specific issues and candidate trait items, and lastly sensitive personal items such as their income and race. We want them to answer the issues and candidate trait items after the vote choice, so that their vote is not influenced by what the researcher thinks are the important issues. But unlike the other two methods of surveying, in mail surveys the respondent can look through the entire questionnaire before answering, so they can answer out of order, so to speak. That can influence the results.

4) Mail surveys are slow, as they take months to complete. With a 3-wave method to increase your response rate, you may send a second mailing out to those who don’t response to the first mailing, and that second mailing may go out a month after the first. The third wave may be sent a month after the second wave. So, it is a slow process. Indeed, how valid is the process, given that public opinion can change over time, so different people are responding at different time points?

5) Incomplete forms. What do you do if they leave out a whole page, or skip individual items? If they skipped items, we assumed that they just had no opinions on those items, or refused to answer. If they left out a whole page, we’d xerox that page, and send it back to them, and politely ask them to complete it and send it back to us. Only half at best sent it back to us. A general low response rate was not a problem for us, since we used a 3-wave mailing system; on the first wave we might get a 35% response rate; the second wave might yield an additional 10%; the third wave might yield 5%; so the total response rate might be 50%, which was pretty good back at the turn of the century (when we did the NSF study).

 

Sampling Techniques:

1) Multi-stage Cluster Sampling. It is usually used in in-person national surveys, where a national list of the population is typically not available. The following stages are: PSUs (Primary Sampling Units), such as U.S. House districts, maybe randomly select 100 of them; choose a city/cities or a rural area; for each city, randomly select 3 city blocks; for each city block, send the interviewer to draw up each housing unit, and randomly select 5 of them; the interviewer visits each household, gets the first name of each adult in that household, and randomly selects 1 to interview, and comes back if that person is not home. In this example, the total sample size is 1500 (100 X 3 X 5 X 1 = 1500).

2) Telephone Survey Sampling. Telephone directory sampling historically, you may have 50 pages, 4 columns on each page, if you want a sample size of 400, then randomly select 2 names in each column. Problem with phone directory sampling is that unlisted numbers are not included, nor are people who just moved into the community, so you can have some sampling bias. Correct for this problem by using random digit dialing, based on a table or computer-generated random numbers, which would include those left out by directory sampling. The Mississippi Poll in the 1980s and 1990s used Two Stage Random Digit Dialing, whereby eligible telephone exchanges (the first 3 digits of a phone number, about 350 were used in the state) and the next two digits were randomly generated and dialed, and 80% of the time it would be a non-existent number. But when you got a real household, you would then repeat the first five digits up to five times with new random last two digits attached to each; this method was used since the phone company tended to assign adjacent numbers, so non-existent numbers would drop to 25%. So this was more efficient in terms of interviewer time, but it is a form of cluster sampling as those same first five digits would be from the same community. After the turn of the century, we just purchased these “working household numbers” from a marketing firm, which guaranteed that 90% of the phone numbers would indeed be real households.

3) Sampling Within the Household. The Carter-Trodahl method ensures a random selection of each adult in each household by asking two questions- how many adults live in the household, and how many of them are men. It has at least 4 versions of a selection table (in a 2-adult, one male and one female house, 2 of the versions would tell the interviewer to interview the man, and 2 of the versions would say interview the woman. The Mississippi Poll used 12 versions of the selection table, thereby permitting up to a 4-adult household to be accurately represented. The biggest problem with this method was interviewee resistance to answering such personal information at the outset of the survey, so that took a few minutes to explain why we needed that info. Sociologists used a method that we then turned to, which was the Last Birthday Method; it asks to talk to the adult of the household who had the most recent birthday. The problem with that method is that most people would not take the time to figure out who had the most recent birthday, so they’d just say It’s Me about 90% of the time. So we got an oversampling of women. We then corrected for that problem by Weighting the Sample, which comes later.

 

Demographic Groups Historically Under sampled by Telephone Polls:

1) High school dropouts. They are very busy at blue-collar jobs so they often lack time, and they may lack interest in politics and government. They may also feel that their opinions aren’t important or valued by college-educated pollsters. As such, we oversample the college educated by nearly double. In the 2014 poll, 38% of our sample was college grads or higher, compared to only 21% of the adult population (census data). Our sample consisted of 12% high school dropouts and 23% high school educated, compared to a census of 18% dropouts and 30% high school grads. In 2010, though, we accurately sampled high school grads, but not dropouts.

2) Lower income. Again, very busy at work, often lacking political interest. Years ago, they were less likely to have the money for a phone. They tend to have lower education levels.

3) African Americans and other racial minorities. These groups tend to have a lower SES level. I was initially concerned that in Mississippi there might be a fear of answering sensitive political questions given the state’s troubled racial history, but by the 1981 poll this did not seem to be a problem. In the 2014 poll, 29% of our sample was African American, compared to 35% of the adult population, and this level of under sampling minorities was about the same in previous polls.

4) Men are under sampled repeatedly. They are hard to get, less likely to be home, often less verbal, more assertive in being unwilling to answer surveys. Plus, the last birthday method gets fewer men. In the 2014 poll, 39% of the sample were men, compared to 47.5% of the adult population.

5) The young adults, as they are socially active, busy, mobile in residency, and less interested in politics than older groups. In the 2014 poll, 16.5% of the sample were under 30 age, compared to 23% of the adult population. The last survey where we relied solely on land lines saw a huge problem, as only 6% of the sample were under 30, compared to 23.5% of the population. So today we combine both land lines and cell phones, as do other pollsters.

6) Old adults historically were under sampled, as they were more likely to be deaf, had health problems preventing long phone conversations, or were institutionalized in nursing homes. Today this is not a problem, as it is easier to represent those over 60 than it is to represent those under 30. In 2014, 32% of our sample was over 60, compared to 26% of the adult population.

 

Weighting the Sample to ensure a representative sample. Multiple stages.

1) You should determine during the survey how many phones the respondent could have been reached from, and how many adults had access to those phones. Historically, you would create a Weight variable using SPSS computer package that would be the number of adults in the household, divided by the number of different telephone numbers. Thus, in a two-person household, the numerator would be a 2, which would be twice that of a one-person household; that would compensate for each person in the 2-person household having only half of the chance of being included in the survey compared to the person in the 1-person household, since we only interview one person in each household. If a person’s household could be reached by dialing two different phone numbers, that number 2 would go into the denominator; thus, the person had double the chance of being called than a person who had only one telephone number (95% of the population at least), so their response would be cut in half. So that is the first weight variable.

2) You now compare the weighted sample with census data to see how representative your sample is. You then create a Weight2 Variable that is equal to Weight1 * (a different number based on the category of the variable that you are correcting for). In 2010 we so under sampled young adults that we corrected for Age first. We used the SPSS, Transform, Compute menu. Target variable was Weight2. Numeric Expression was Weight1 * (the population % of the age category divided by the sample % of the age category). At bottom left, use If, then Include If whatever age group you are correcting for. Such as age < 30, then the value is 23.5/5.9, the population percent divided by the sample percent. That value now gives each young adult in the sample close to 4 votes, so the sample percent of young adults will be equal to the population percent of young adults. You have to do that for each age category you are using. You then compare the Weight2 demographic frequencies of your sample with the census data, and you will find that age is corrected for. You now have to correct for other problem demographics. We did sex next, so Weight3 = Weight2 X (male or female correction). Weight4 corrected for education, and Weight5 corrected for race. Weight6 just took Weight5 and divided by a constant, whatever was necessary to get the sample size back to what it originally was. For instance, if you originally had 600 people in your sample, but your weighted sample now had 1200 (if you had all 2-adult households), Weight6 would be Weight5 X .5. That ensures that you don’t mislead a reader or get statistically significant results that don’t really exist.

 

How accurate are polls? Despite the problems with recent Trump polls, polls in this century have generally been pretty accurate. We gauge accuracy often from presidential election results, since it is a measurable event that we can compare poll results with. In examining state polls of the presidential race in 2004, for example, polls were most accurate if they were conducted as recently as possible, such as within five days of the election. Another way of increasing accuracy was to combine the results from multiple polls, since you are basically increasing your sample size and thereby reducing your sample error. That is why in my Political Analysis class, student papers will often combine the three most recent state polls.

Check out my website for a link to this subject. A more recent analysis of polling accuracy came to similar conclusions. See G. Elliott Morris' book Strength in Numbers: How Polls Work and Why We Need Them, page 111.

 

Given the increasing problems we have had with polling, check out the most recent results from the Mississippi Poll, which was last done in 2014.

 

Check out the Mississippi Poll website. One interesting link on that page is a summary of methods used in the polls, which extend from 1981 thru 2014. It shows yet another problem with polling, and that is declining response rate. From 1981 thru 1990, the response rates were over 70%. From 1992 thru 1999, they were in the 60’s. From 2000-2006, they were about 50% (meaning that half of the people we wanted to interview refused to even talk to us). In 2008-10, it was about 41%. By 2012-14 it had fallen to 26-31%. A low response rate can cause validity problems, since one wonders if the sample is indeed representative of the population. Of course, you can weight the sample by demographic characteristics, but is a weighted male who participates the same in attitudes to a male who refuses to be interviewed, for example. Page 35 of the text talks about this problem in depth, as they encountered response rates becoming as low as 9%.

 

Wording of survey questions considerations:

1) Avoid loaded or leading questions. Ensure that you have reasonable opposing categories.

2) Avoid double barreled questions, which ask about two separate but related issues, such as do you support higher taxes for teacher pay and a new stadium. Break them up into two questions.

3) Avoid complex questions, often having double negatives. In 1981 Miss. Poll, we asked an agree-disagree item, do you favor or oppose a constitutional amendment stopping all abortions except those needed to protect the health of the mother. Many people said I oppose that item, I oppose abortion. We had to explain, then you would favor an amendment stopping all abortions. Uh, I guess. But it would not stop abortions needed to protect the mother’s health. I’m even lost at this point. We soon switched to a simpler agree-disagree item, do you agree or disagree that a woman should be able to have an abortion as a matter of personal choice.

4) Avoid unbalanced alternatives. A national poll in 1981 asked respondents whether they thought that blacks were treated the same or worse by public officials. Our poll asked: How well do you think blacks are treated by public officials in Mississippi? Are blacks treated worse than whites, better than whites, or about the same? Interesting results, as the races were split on this issue with some whites actually saying that African Americans were treated better than whites.

5) Acquiescence bias. This is an agreement bias, often found on agree-disagree items. An example is the state spending items in the Mississippi Poll. Now I'm going to ask you about some issues facing state and local government in Mississippi. As you know, most of the money government spends comes from the taxes you and others pay. For each of the following, please tell me whether you think state and local government in Mississippi should be spending more, less, or about the same as now. How about. . .? There are about 10 different state programs, each very short, such as Public Grade School and High School, Streets and Highways, Health Care and Hospitals. It is very easy for people to just say Spend More on all or most of them. Fortunately, you can still compare across the programs to identify which of the programs is most and least popular, such as public education’s popularity. An advantage of such a long list of items with the same response categories is that you can get through the interview very fast, so that makes up for the other 50 or so items that we ask about.

6) Sensitive personal items- people may refuse to answer them. So for age, don’t ask their age, ask in what year they were born. Don’t ask their income, list eight categories that are in $10,000 units, and ask them to choose the right intervals. I initially thought that asking their race might be a problem, so we asked: are you white, black, or what, which students modified by replacing what with other; it wasn’t a problem.

7) Social desirability- people may lie when asked about sensitive social issues, such as their sex practices, and what their views are on race or sex equality issues. We use pretty well-established indicators, so we do get fairly honest responses.