19  Survey Research Literature

This page provides references in readings for survey research. This page is continually being revised and updated to reflect new resources. Do you have suggestions for readings we should add to this page? Send us an email with your suggestions and a link to this page!

Quicklinks:

20 Introduction to Survey Research Literature

Krosnick, Jon A. 1999. “Survey Research.” Annual Review of Psychology. 50:537-567.
AUTHOR ABSTRACT For the first time in decades, conventional wisdom about survey methodology is being challenged on many fronts. The insights gained can not only help psychologists do their research better but also provide useful insights into the basics of social interaction and cognition. This chapter reviews some of the many recent advances in the literature, including the following: New findings challenge a long-standing prejudice against studies with low response rates; innovative techniques for pretesting questionnaires offer opportunities for improving measurement validity; surprising effects of the verbal labels put on rating scale points have been identified, suggesting optimal approaches to scale labeling; respondents interpret questions on the basis of the norms of everyday conversation, so violations of those conventions introduce error; some measurement error thought to have been attributable to social desirability response bias now appears to be due to other factors instead, thus encouraging different approaches to fixing such problems; and a new theory of satisficing in questionnaire responding offers parsimonious explanations for a range of response patterns long recognized by psychologists and survey researchers but previously not well understood.
Robert Groves et al. 2009. Survey Methodology, 2nd Edition. John Wiley and Sons.
BOOK DESCRIPTION

This new edition of Survey Methodology continues to provide a state-of-the-science presentation of essential survey methodology topics and techniques. The volume’s six world-renowned authors have updated this Second Edition to present newly emerging approaches to survey research and provide more comprehensive coverage of the major considerations in designing and conducting a sample survey.

Key topics in survey methodology are clearly explained in the book’s chapters, with coverage including sampling frame evaluation, sample design, development of questionnaires, evaluation of questions, alternative modes of data collection, interviewing, nonresponse, post-collection processing of survey data, and practices for maintaining scientific integrity. Acknowledging the growing advances in research and technology, the Second Edition features:

  • Updated explanations of sampling frame issues for mobile telephone and web surveys New scientific insight on the relationship between nonresponse rates and nonresponse errors

  • Restructured discussion of ethical issues in survey research, emphasizing the growing research results on privacy, informed consent, and confidentiality issues

  • The latest research findings on effective questionnaire development techniques

  • The addition of 50% more exercises at the end of each chapter, illustrating basic principles of survey design

  • An expanded FAQ chapter that addresses the concerns that accompany newly established methods

Providing valuable and informative perspectives on the most modern methods in the field, Survey Methodology, Second Edition is an ideal book for survey research courses at the upper-undergraduate and graduate levels. It is also an indispensable reference for practicing survey methodologists and any professional who employs survey research methods.
Lior Gideon. 2012. Handbook of Survey Methodology for the Social Sciences. New York: Springer
BOOK DESCRIPTION Surveys enjoy great ubiquity among data collection methods in social research: they are flexible in questioning techniques, in the amount of questions asked, in the topics covered, and in the various ways of interactions with respondents. Surveys are also the preferred method by many researchers in the social sciences due to their ability to provide quick profiles and results. Because they are so commonly used and fairly easy to administer, surveys are often thought to be easily thrown together. But designing an effective survey that yields reliable and valid results takes more than merely asking questions and waiting for the answers to arrive. Geared to the non-statistician, the Handbook of Survey Methodology in Social Sciences addresses issues throughout all phases of survey design and implementation. Chapters examine the major survey methods of data collection, providing expert guidelines for asking targeted questions, improving accuracy and quality of responses, while reducing sampling and non-sampling bias. Relying on the Total Survey Error theory, various issues of both sampling and non-sampling sources of error are explored and discussed. By covering all aspects of the topic, the Handbook is suited to readers taking their first steps in survey methodology, as well as to those already involved in survey design and execution, and to those currently in training. Featured in the Handbook: • The Total Survey Error: sampling and non-sampling errors. • Survey sampling techniques. • The art of question phrasing. • Techniques for increasing response rates • A question of ethics: what is allowed in survey research? • Survey design: face-to-face, phone, mail, e-mail, online, computer-assisted.? • Dealing with sensitive issues in surveys. • Demographics of respondents: implications for future survey research. • Dealing with nonresponse, and nonresponse bias The Handbook of Survey Methodology in Social Sciences offers how-to clarity forresearchers in the social and behavioral sciences and related disciplines, including sociology, criminology, criminal justice, social psychology, education, public health, political science, management, and many other disciplines relying on survey methodology as one of their main data collection tools.

21 Data Collection Literature

Bauman, Adrian, P. Phongsavan, A. Cowle, E. Banks, L. Jorm, K. Rogers, B. Jalaludin, and A. Grunseit. 2016. “Maximising Follow-up Participation Rates in a Large Scale 45 and Up Study in Australia.” Emerging Themes in Epidemiology 13:6.
AUTHOR ABSTRACT The issue of poor response rates to population surveys has existed for some decades, but few studies have explored methods to improve the response rate in follow-up population cohort studies. A sample of 100,000 adults from the 45 and Up Study, a large population cohort in Australia, were followed up 3.5 years after the baseline cohort was assembled. A pilot mail-out of 5000 surveys produced a response rate of only 41.7 %. This study tested methods of enhancing response rate, with three groups of 1000 each allocated to (1) receiving an advance notice postcard followed by a questionnaire, (2) receiving a questionnaire and then followup reminder letter, and (3) both these strategies.The enhanced strategies all produced an improved response rate compared to the pilot, with a resulting mean response rate of 53.7 %. Highest response was found when both the postcard and questionnaire reminder were used (56.4 %) but this was only significantly higher when compared to postcard alone (50.5 %) but not reminder alone (54.1 %). The combined approach was used for recruitment among the remaining 92,000 participants, with a resultant further increased response rate of 61.6 %. Survey prompting with a postcard and a reminder follow-up questionnaire, applied separately or combined can enhance follow-up rates in large scale survey-based epidemiological studies.
Dutwin, David, Scott Keeter, and Courtney Kennedy. “Bias from wireless substitution in surveys of Hispanics.” Hispanic Journal of Behavioral Sciences 32.2 (2010): 309-328.
AUTHOR ABSTRACT Increasingly, American households are choosing to forgo ownership of landline telephones in favor of cell phones. Presently, more than 25% of Hispanics now only own a cell phone. Concern about potential bias from non-coverage of this “cell-only” population in traditional general population RDD (random digit dial) telephone interviewing has been a particular focus among the survey research community. Because Hispanics lead all ethnic and racial groups in the percentage who are now cell-only, it is critical to understand the impact this has on research of Hispanics, which to date is still largely accomplished using traditional landline telephone survey methods. This article explores the extent to which non-coverage of cell-only Hispanics leads to raw bias in landline telephone surveys of Hispanics and the degree to which survey weights may correct for such bias.
Kim, ChangHwan and Christopher R. Tamborini. 2014. “Response Error in Earnings: An Analysis of the Survey of Income and Program Participation Matched With Administrative Data.” Sociological Methods & Research 43(1):39-72.
AUTHOR ABSTRACT This article examines the problem of response error in survey earnings data. Comparing workers’ earnings reports in the U.S. Census Bureau’s Survey of Income and Program Participation (SIPP) to their detailed W-2 earnings records from the Social Security Administration, we employ ordinary least squares (OLS) and quantile regression models to assess the effects of earnings determinants and demographic variables on measurement errors in 2004 SIPP earnings in terms of bias and variance. Results show that measurement errors in earnings are not classical, but mean-reverting. The directions of bias for subpopulations are not constant, but varying across levels of earnings. Highly educated workers more correctly report their earnings than less educated workers at higher earnings levels, but they tend to overreport at lower earnings levels. Black workers with high earnings underreport to a greater degree than comparable whites, while black workers with low earnings overreport to a greater degree. Some subpopulations exhibit higher variances of measurement errors than others. Blacks, Hispanics, high school dropouts, part-year employed workers, and occupation ‘‘switchers’’ tend to misreport—both over- and underreport—their earnings rather than unilaterally in one direction. The implications of our findings are discussed.
Kleiner, Brian, Oliver Lipps, and Eliane Ferrez. 2015. “Language Ability and Motivating among Foreigners in Survey Responding. Journal of Survey Statistics and Methodology 3:339-360.
AUTHOR ABSTRACT With increasing migration and linguistic diversification in many countries, survey researchers and methodologists should consider whether data provided by individuals with variable levels of command of the survey language are of the same quality. This paper examines the question of whether answers from resident foreign respondents who do not master available survey languages may suffer from problems of comprehension of survey items, especially items that are more complicated in terms of content and/or form. In addition, it addresses the extent to which motivation may affect the response quality of resident foreigners. We analyzed data from two large-scale surveys conducted in Switzerland, a country with three national languages and a burgeoning foreign population, employing a set of dependent measures of response quality, including don’t know responses, extreme responding, mid-5 responding, recency effects, and straight-lining. Results show overall poorer response quality among foreigners, and indicate that both reduced language mastery and motivation among foreigners are relevant factors. This is especially true for foreign groups from countries that do not share a common language with those spoken in Switzerland. A general conclusion is that the more distant respondents are culturally and linguistically from the majority mainstream within a country, the more their data may be negatively affected. We found that more complex types of questions do generally lead to poorer response quality, but to a much lesser extent than respondent characteristics, such as nationality, command of the survey language, level of education, and age.
Olsen, Frank, Birgit Abelsen, and Jan Abel Olsen. 2012. “Improving Response Rate and Quality of Survey Data with a Scratch Lottery Ticket Incentive.” Medical Research Methodology 12:52.
AUTHOR ABSTRACT The quality of data collected in survey research is usually indicated by the response rate; the representativeness of the sample, and; the rate of completed questions (item-response). In attempting to improve a generally declining response rate in surveys considerable efforts are being made through follow-up mailings and various types of incentives. This study examines effects of including a scratch lottery ticket in the invitation letter to a survey. Questionnaires concerning oral health were mailed to a random sample of 2,400 adults. A systematically selected half of the sample (1,200 adults) received a questionnaire including a scratch lottery ticket. One reminder without the incentive was sent. The incentive increased the response rate and improved representativeness by reaching more respondents with lower education. Furthermore, it reduced item nonresponse. The initial incentive had no effect on the propensity to respond after the reminder. When attempting to improve survey data, three issues become important: response rate, representativeness, and item-response. This study shows that including a scratch lottery ticket in the invitation letter performs well on all the three.

22 Sample Design and Selection Literature

Alwin, Duane F. 2014. “Investigating Response Errors in Survey Data.” Sociological Methods & Research, 43(1), 3-14.
AUTHOR ABSTRACT This special issue of Sociological Methods & Research contributes to recent trends in studies that exploit the availability of multiple measures in sample surveys in order to detect the level and patterning to measurement errors. Articles in this volume focus on topics in one of (or some combination of) the three areas: (1) those that develop and test theoretical hypotheses regarding the behavior of measurement errors under specific conditions of measurement, (2) those that focus on the methodological problems encountered in the design of data collection permitting the estimation of measurement models, and (3) those that focus on the evaluation of existing models for detecting and quantifying the nature of measurement errors. The designs included in these investigations include those that incorporate follow-up probes, record-check studies, multitrait-multimethod designs, longitudinal designs, and latent class models for assessing measurement errors for categorical variables.
Bauman, Adrian, Philayrath Phongsavan, Alison Cowle, Emily Banks, Louisa Jorm, Kris Rogers, Bin Jalaludin, and Anne Grunseit. 2016. “Maximising follow-up participation rates in a large scale 45 and Up Study in Australia.” Emerging Themes in Epidemiology, 13(6), 1-7.
AUTHOR ABSTRACT

Background: The issue of poor response rates to population surveys has existed for some decades, but few studies have explored methods to improve the response rate in follow-up population cohort studies.

Methods: A sample of 100,000 adults from the 45 and Up Study, a large population cohort in Australia, were followed up 3.5 years after the baseline cohort was assembled. A pilot mail-out of 5000 surveys produced a response rate of only 41.7 %. This study tested methods of enhancing response rate, with three groups of 1000 each allocated to (1) receiving an advance notice postcard followed by a questionnaire, (2) receiving a questionnaire and then followup reminder letter, and (3) both these strategies.

Results: The enhanced strategies all produced an improved response rate compared to the pilot, with a resulting mean response rate of 53.7 %. Highest response was found when both the postcard and questionnaire reminder were used (56.4 %) but this was only significantly higher when compared to postcard alone (50.5 %) but not reminder alone (54.1 %). The combined approach was used for recruitment among the remaining 92,000 participants, with a resultant further increased response rate of 61.6 %.

Conclusions: Survey prompting with a postcard and a reminder follow-up questionnaire, applied separately or combined can enhance follow-up rates in large scale survey-based epidemiological studies.

Keywords: Epidemiological studies, Follow-up, Response rates, Recruitment strategies
Yeager, D. S., Krosnick, J. A., Chang, L., Javitz, H. S., Levendusky, M. S., Simpser, A., et al. 2011. “Comparing the Accuracy of RDD Telephone Surveys and Internet Surveys Conducted with Probability and Non-probability Samples.” Public Opinion Quarterly, 75(4), 709-747.
AUTHOR ABSTRACT This study assessed the accuracy of telephone and Internet surveys of probability samples and Internet surveys of non-probability samples of American adults by comparing aggregate survey results against benchmarks. The probability sample surveys were consistently more accurate than the non-probability sample surveys, even after post-stratification with demographics. The non-probability sample survey measurements were much more variable in their accuracy, both across measures within a single survey and across surveys with a single measure. Post-stratification improved the overall accuracy of some of the non-probability sample surveys but decreased the overall accuracy of others. Higher completion and response rates of the surveys were associated with less accuracy. Accuracy did not appear to change from 2004/2005 to 2009 for any of the methods, and these conclusions are reinforced by data collected in 2008 as well. These results are consistent with the conclusion that non-probability samples yield data that are neither as accurate as nor more accurate than data obtained from probability samples.

23 Questionnaire Development and Layout Literature

Bradley, Kelly D., Michael R. Peabody, Kathryn S. Akers, and Nichole M Knutson. 2015. “Rating Scales in Survey Research: Using the Rasch Model to Illustrate the Neutral Middle Category Measurement Flaw.” Survey Practice 8(2).
AUTHOR ABSTRACT The quality of the instrument used in the measurement process of survey data is fundamental to successful outcomes. Issues regarding content and structure are primary during instrument development, but the rating scale is just as important. Specifically for Likert-type questionnaires, the words used to describe rating categories and the placement of a neutral or not sure category is at the core of this measurement issue. This study utilizes the Rasch model to assess the quality of an instrument and structure of the rating scale for a typical data set collected at an institution of higher education. The importance of category placement and an evaluation of the use of a neutral middle category for Likert-type survey data is highlighted in this study.
Krosnick, Jon A. and Stanley Presser. 2010 ‘Question and Questionnaire Design’ in: Handbook of Survey Research. 2nd Ed. Emerald. pp. 263-313.
AUTHOR INTRODUCTION

The heart of a survey is its questionnaire. Drawing a sample, hiring, and training interviewers and supervisors, programming computers, and other preparatory work is all in service of the conversation that takes place between researchers and respondents. Survey results depend crucially on the questionnaire that scripts this conversation (irrespective of how the conversation is mediated, e.g., by an interviewer or a computer). To minimize response errors, questionnaires should be crafted in accordance with best practices.

Recommendations about best practices stem from experience and common lore, on the one hand, and methodological research, on the other. In this chapter, we first offer recommendations about optimal questionnaire design based on conventional wisdom (focusing mainly on the words used in questions), and then make further recommendations based on a review of the methodological research (focusing mainly on the structural features of questions).

We begin our examination of the methodological literature by considering open versus closed questions, a difference especially relevant to three types of measurement: (1) asking for choices among nominal categories (e.g., ‘‘What is the most important problem facing the country?’’), (2) ascertaining numeric quantities (e.g., ‘‘How many hours did you watch television last week?’’), and (3) testing factual knowledge (e.g., ‘‘Who is Joseph Biden?’’).

Next, we discuss the design of rating scales. We review the literature on the optimal number of scale points, consider whether some or all scale points should be labeled with words and/or numbers, and examine the problem of acquiescence response bias and methods for avoiding it. We then turn to the impact of response option order, outlining how it varies depending on whether categories are nominal or ordinal and whether they are presented visually or orally.

After that, we assess whether to offer ‘‘don’t know’’ or no-opinion among a question’s explicit response options. Next we discuss social desirability response bias (a form of motivated misreporting) and recall bias (a form of unmotivated misreporting), and recommend ways to minimize each. Following that, we consider the ordering of questions within a questionnaire and then discuss methods for testing and evaluating questions and questionnaires. Finally, we offer two more general recommendations to guide questionnaire development.
Liu, Mingnan, Sunghee Lee, Frederick G. Conrad. 2015. “Comparing Extreme Response Styles between Agree-Disagree and Item-Specific Scales.” Public Opinion Quarterly 79(4):952-975.
AUTHOR ABSTRACT Abstract Although Likert scales in agree-disagree (A/D) format are popular in surveys, the data quality yielded by them is controversial among researchers. Recognizing the measurement issues involved with the A/D format, researchers have developed other question formats to measure attitudes. In this study, we focused on an alternative question type, the item-specific (IS) question, which asks the respondent to choose an option that best describes his or her attitude. Using political efficacy items from the American National Election Studies (ANES), we compare extreme response style (ERS) between A/D and IS scales. Latent class factor analysis showed that ERS exists in both A/D and IS scale formats, but differs slightly across the two. Also, when analyzing ERS within subjects across two waves, there is only a single ERS for both question formats, after controlling for the correlation within respondents. The last finding suggests that ERS is a stable characteristic.
Pasek, Josh, and Jon A. Krosnick. 2010. “Optimizing Survey Questionnaire Design in Political Science: Insights from Psychology.” The Oxford Handbook of American Elections and Political Behavior, edited by Jan E. Leighley.
AUTHOR ABSTRACT This article provides a summary of the literature’s suggestions on survey design research. In doing so, it points researchers toward question formats that appear to yield the highest measurement reliability and validity. Using the American National Election Studies as a starting point, it shows the general principles of good questionnaire design, desirable choices to make when designing new questions, biases in some question formats and ways to avoid them, and strategies for reporting survey results. Finally, it offers a discussion of strategies for measuring voter turnout in particular, as a case study that poses special challenges. Scholars designing their own surveys should not presume that previously written questions are the best ones to use. Applying best practices in questionnaire design will yield more accurate data and more accurate substantive findings about the nature and origins of mass political behavior.
Presser, Stanley. 1990. “Measurement Issues in the Study of Social Change.” Social Forces 68(3):856-868.
AUTHOR ABSTRACT The assumption that relationships between variables are unaffected by formal features of the measurement indicators has been shown to hold for most associations between attitudes and other variables, where the other variables are elf/ier background characteristics or other attitudes. This article tests the assumption of “form-resistant correlations” in the context of studying social change, where the other variable is time. It examines whether alternate forms of the same attitude item behave in similar fashion across time.
Reja, Ursa, Katja Lozar Manfreda, Valentina Hlebec, and Vasja Vehovar. 2003. “Open-ended vs. Close-ended Questions in Web Questionnaires. pp 159-177 in Developments in Applied Statistics, edited by Anuska Ferligoj and Andrej Mrvar. Ljubljana:FDV.
AUTHOR ABSTRACT Two quite different reasons for using open-ended as opposed to close-ended questions can be distinguished. One is to discover the responses that individuals give spontaneously; the other is to avoid the bias that may result from suggesting responses to individuals. However, open-ended questions also have disadvantages in comparison to close-ended, such as the need for extensive coding and larger item non-response. While this issue has already been well researched for traditional survey questionnaires, not much research has been devoted to it in recently used Web questionnaires. We therefore examine the differences between the open-ended and the close-ended question form in Web questionnaires by means of experiments within the large-scale RIS 2001 Web survey. The question “What is the most important, critical problem the Internet is facing today?” was asked in an open-ended and two close-ended question forms in a split-ballot experiment. The results show that there were differences between question forms in univariate distributions, though no significant differences were found in the ranking of values. Close-ended questions in general yield higher percentages than open-ended question for answers that are identical in both question forms. It seems that respondents restricted themselves with apparent ease to the alternatives offered on the close-ended forms, whereas on the open-ended question they produced a much more diverse set of answers. In addition, our results suggest that open-ended questions produce more missing data than close-ended. Moreover, there were more inadequate answers for open-ended question. This suggests that open-ended questions should be more explicit in their wording (at least for Web surveys, as a self administered mode of data collection) than close-ended questions, which are more specified with given response alternatives.
Revilla, Melanie A., Willem E. Saris, and Jon A. Krosnick. 2014. “Choosing the Number of Categories in Agree-Disagree Scales.” Sociological Methods & Research 43(1):73-97.
AUTHOR ABSTRACT Although agree–disagree (AD) rating scales suffer from acquiescence response bias, entail enhanced cognitive burden, and yield data of lower quality, these scales remain popular with researchers due to practical considerations (e.g., ease of item preparation, speed of administration, and reduced administration costs). This article shows that ifresearchers want to use AD scales, they should offer 5 answer categories rather than 7 or 11, because the latter yield data of lower quality. This is shown using data from four multitrait-multimethod experiments implemented in the third round of the European Social Survey. The quality of items with different rating scale lengths were computed and compared.
Saris, Willem E., Melanie Revilla, Jon A. Krosnick, and Eric M Shaeffer. 2010.”Comparing Questions with Agree/Disagree Response Options to Questions with Item-specific Response Options.” Survey Research Methods 4 (1), pp. 61-79.
AUTHOR ABSTRACT Although Agree/Disagree (A/D) rating scales are hugely popular in the social sciences, a large body of research conducted during more than five decades has documented the bias that results from acquiescence in responses to these items. This may be a reason to prefer questions with Item Specific (IS) response options, but remarkably little research has explored whether responses to A/D rating scale questions are indeed of lower quality than responses to questions with IS response options. Using a research design that combines the advantages of a random assignment between-subjects experiment and the multitrait-multimethod approach in the context of representative sample surveys, we found that responses to A/D rating scale questions indeed had much lower quality than responses to comparable questions offering IS response options. These results attest to the superiority of questions with IS response options.
Schuman, Howard, Jacob Ludwig, & Jon A. Krosnick. 1986. “The Perceived Threat of Nuclear War, Salience, and Open Questions. Public Opinion Quarterly 50(4):519-536.
AUTHOR ABSTRACT The most serious problem facing the United States, according to many scientific and political leaders, is the threat of nuclear war. Yet the standard survey question on the most important problem facing the country has often shown little public concurrence with this assumption. Our article uses experimentation in national samples to test whether this difference can be traced to limitations in either the form or the wording of the standard question. The results indicate that there are some important systematic differences between open and closed versions of the question, and also differences that result from reference to the nation as distinct from the world, but neither type of difference accounts for the infrequent mention of nuclear war on the standard question. Instead, other evidence indicates that most Americans believe that nuclear war is not going to happen at all, or that if it does happen it will be too far in the distant future to be of pressing concern to them personally.
Schwarz, Norbert. 2010. “If it’s Easy to Read, it’s Easy to Do, Pretty, Good, and True.” The Psychologist 23(2).
AUTHOR INTRODUCTION

Thinking can feel easy or difficult. But what effect does the ease or difficulty of reading a text have on information processing? Can something as seemingly irrelevant as the print font in which information is presented influence how information is evaluated, or even whether it is accepted as true or false? What are the practical implications for everyday life?

KEY QUESTIONS: What is the likely role of metacognitive feelings of ease and difficulty in your own field of psychology? What are the implications for teaching, counselling, advertising, health education, and political communication? What do these influences imply for the rationality of human judgement?
Schwarz, Norbert. 1999. “Self-Reports: How the Questions Shape the Answers.” American Psychologist 54(2): 93-105 American Psychological Association.
AUTHOR ABSTRACT Self-reports of behaviors and attitudes are strongly influenced by features of the research instrument, including question wording, format, and context. Recent research has addressed the underlying cognitive and communicative processes, which are systematic and increasingly well-understood. I review what has been learned, focusing on issues of question comprehension, behavioral frequency reports, and the emergence of context effects in attitude measurement. The accumulating knowledge about the processes underlying self-reports promises to improve questionnaire design and data quality.
Schwarz, Norbert. 1995. “What Respondents Learn from Questionnaires: The Survey Interview and the Logic of Conversation.” International Statistical Institute 63(2):153-177.
AUTHOR ABSTRACT Drawing on psychological theories of language comprehension and conversational conduct, I review what respondents learn from various features of questionnaires, including the response alternatives and the context in which a question is asked. This review indicates that many apparent “artifacts” in survey measurement reflect a systematic and thoughtful use of contextual information that is licensed by the tacit rules that govern the conduct of conversation in daily life. Implications for questionnaire construction are discussed.
Schwarz, Norbert, Eryn Newman, and William Leach. 2016. “Making the Truth Stick and The Myths Fade: Lessons from Cognitive Psychology. Behavioral Science & Policy.
AUTHOR ABSTRACT Erroneous beliefs are difficult to correct. Worse, popular correction strategies may backfire and further increase the spread and acceptance of misinformation. People evaluate the truth of a statement by assessing its compatibility with other things they believe, its internal consistency, amount of supporting evidence, acceptance by others, and the credibility of the source. To do so, they can draw on relevant details (an effortful analytic strategy) or attend to the subjective experience of processing fluency (a less effortful intuitive strategy). Throughout, fluent processing facilitates acceptance of the statement – when thoughts flow smoothly, people nod along. Correction strategies that make false information more fluent (e.g., through repetition or pictures) can therefore increase its later acceptance. We review recent research and offer recommendations for more effective correction strategies.
Schwarz, N., & Oyserman, D. (2001). Asking Questions About Behavior: Cognition, Communication, and Questionnaire Construction. American Journal Of Evaluation, 22(2), 127.
AUTHOR ABSTRACT Evaluation researchers frequently obtain self-reports of behaviors, asking program participants to report on process and outcome-relevant behaviors. Unfortunately, reporting on one’s behavior poses a difficult cognitive task, and participants’ reports can be profoundly influenced by question wording, format, and context. We review the steps involved in answering a question about one’s behavior and highlight the underlying cognitive and communicative processes. We alert researchers to what can go wrong and provide theoretically grounded recommendations for pilot testing and questionnaire construction.
Sturgis, Patrick, Caroline Roberts, and Patten Smith. 2014. “Middle Alternatives Revisited: How the neither/nor Response Acts as a Way of Saying ‘I Don’t Know’?” Sociological Methods & Research 43(1): 15-38.
AUTHOR ABSTRACT A persistent problem in the design of bipolar attitude questions is whether or not to include a middle response alternative. On the one hand, it is reasonable to assume that people might hold opinions which are ‘neutral’ with regard to issues of public controversy. On the other, question designers suspect that offering a mid-point may attract respondents with no opinion, or those who lean to one side of an issue but do not wish to incur the cognitive costs required to determine a directional response. Existing research into the effects of offering a middle response alternative has predominantly used a split-ballot design, in which respondents are assigned to conditions which offer or omit a midpoint. While this body of work has been useful in demonstrating that offering or excluding a mid-point substantially influences the answers respondents provide, it does not offer any clear resolution to the question of which format yields more accurate data. In this paper, we use a different approach. We use follow-up probes administered to respondents who initially select the mid-point to determine whether they selected this alternative in order to indicate opinion neutrality, or to indicate that they do not have an opinion on the issue. We find the vast majority of responses turn out to be what we term ‘face-saving don’t knows’ and that reallocating these responses from the mid-point to the don’t know category significantly alters descriptive and multivariate inferences. Counter to the survey-satisficing perspective, we find that those with this tendency is greatest amongst those who express more interest in the topic area.
Tourangeau, Roger, and Rom W. Smith. 1996. “Asking Sensitive Questions: The Impact of Data Collection Mode, Question Format, and Question Context.”Public Opinion Quarterly 60(2): 275-304.
AUTHOR ABSTRACT This study compared three methods of collecting survey data about sexual behaviors and other sensitive topics: computer-assisted personal interviewing (CAPI), computer-assisted self-administered interviewing (CASI), and audio computer-assisted self-administered interviewing (ACASI). Interviews were conducted with an area probability sample of more than 300 adults in Cook County, Illinois. The experiment also compared open and closed questions about the number of sex partners and varied the context in which the sex partner items were embedded. The three mode groups did not differ in response rates, but the mode of data collection did affect the level of reporting of sensitive behaviors: both forms of self-administration tended to reduce the disparity between men and women in the number of sex partners reported. Self-admimstration, especially via ACASI, also increased the proportion of respondents admitting that they had used illicit drugs. In addition, when the closed answer options emphasized the low end of the distribution, fewer sex partners were reported than when the options emphasized the high end of the distribution; responses to the open-ended versions of the sex partner items generally fell between responses to the two closed versions.

24 Data Collection and Processing Literature

Alwin, Duane F., Kristina Zeiser, and Don Gensimore. 2014. “Reliability of Self-reports of Financial Data in Surveys: Results From the Health and Retirement Study.” Sociological Methods & Research 43(1):98-136.
AUTHOR ABSTRACT This article reports an investigation of errors of measurement in self-reports of financial data in the Health and Retirement Study (HRS), one of the major social science data resources available to those who study the demography and economics of aging. Results indicate significantly lower levels of reporting reliability of the composite variables in the HRS relative to those found for ‘‘summary’’ income approaches used in other surveys. Levels of reliability vary by type of income source—reports of monthly benefit levels from sources such as Social Security or the Veterans Administration achieve near-perfect levels of reliability, whereas somewhat less regular sources of household income that vary across time in their amounts are measured less reliably. One major area of concern resulting from this research, which may be beneficial to users of the HRS surveys, involves the use of imputation in the handling of missing data. We found that imputation of values for top-end open income brackets can produce a substantial number of outliers that affect sample estimates of relationships and levels of reliability. Imputed income values in the HRS should be used with great care.
Berzofsky, Marcus E., Paul P. Biemer, and William D. Kalsbeek. 2014. “Local Dependence in Latent Class Analysis of Rare and Sensitive Events.” Sociological Methods & Research 43(1):137-170.
AUTHOR ABSTRACT For survey methodologists, latent class analysis (LCA) is a powerful tool for assessing the measurement error in survey questions, evaluating survey methods, and estimating the bias in estimates of population prevalence. LCA can be used when gold standard measurements are not available and applied to essentially any set of indicators that meet certain criteria for identifiability. LCA offers quality inference, provided the key threat to model validity— namely, local dependence—can be appropriately addressed either in the study design or in the model-building process. Three potential causes threaten local independence: bivocality, behaviorally correlated error, and latent heterogeneity. In this article, these threats are examined separately to obtain insights regarding (a) questionnaire designs that reduce local dependence, (b) the effects of local dependence on parameter estimation, and (c) modeling strategies to mitigate these effects in statistical inference. The article focuses primarily on the analysis of rare and sensitivity outcomes and proposes a practical approach for diagnosing and mitigating model failures. The proposed approach is empirically tested using real data from a national survey of inmate sexual abuse where measurement errors are a serious concern. Our findings suggest that the proposed modeling strategy was successful in reducing local dependence bias in the estimates, but its success varied by the quality of the indicators available for analysis. With only three indicators, the biasing effects of local dependence can usually be reduced but not always to acceptable levels.
Millar, M. M., & Dillman, D. A. 2011. “Improving Response to Web and Mixed-mode Surveys.” Public Opinion Quarterly.
AUTHOR ABSTRACT We conducted two experiments designed to evaluate several strategies for improving response to Web and Web/mail mixed-mode surveys. Our goal was to determine the best ways to maximize Web response rates in a highly Internet-literate population with full Internet access. We find that providing a simultaneous choice of response modes does not improve response rates (compared to only providing a mail response option). However, offering the different response modes sequentially, in which Web is offered first and a mail follow-up option is used in the final contact, improves Web response rates and is overall equivalent to using only mail. We also show that utilizing a combination of both postal and email contacts and delivering a token cash incentive in advance are both useful methods for improving Web response rates. These experiments illustrate that although different implementation strategies are viable, the most effective strategy is the combined use of multiple response-inducing techniques.
Krosnick, J. A. (1999). Survey research. Annual Review Of Psychology, 50(1), 537.
AUTHOR ABSTRACT Comments on the challenges on the methodologies used for surveys conducted for psychological research. Challenges to prejudice against studies with low response rates; Techniques for pretesting questionnaires offering opportunities to improve measurement validity; Effects of the verbal labels on rating scale points; Violations of survey conventions that cause errors.

25 Data Analysis Literature

Kristensen, Kai & Jacob Eskildsen. 2011. “Is the Net Promoter Score a Reliable Performance Measure?” Proceedings on the 2011 IEEE ICQR.
AUTHOR ABSTRACT Customer satisfaction and customer loyalty have become very important concepts in modern management and quality models. A number of measurement frameworks have been proposed, and especially the American Customer Satisfaction Index and its European counterpart, EPSI Rating, have been accepted as good solutions. In 2003, however, Reichheld published an article in HBR, in which he claims that the Net Promoter Score (introduced by himself), is the only number you need to grow, and the only number you need to manage customer loyalty. We claim that the NPS is an inefficient and unreliable measure of customer loyalty and this paper clearly demonstrate that the NPS is by far inferior to the standard measures of loyalty used by the ACSI and EPSI Rating.
Morgan, Neal A. and Lopo Leotte Rego. 2006. “The Value of Different Customer Satisfaction and Loyalty Metrics in Predicting Business Performance.” Marketing Science 25(5):426-439.
AUTHOR ABSTRACT Managers commonly use customer feedback data to set goals and monitor performance on metrics such as “Top 2 Box” customer satisfaction scores and “intention-to-repurchase” loyalty scores. However, analysts have advocated a number of different customer feedback metrics including average customer satisfaction scores and the number of “net promoters” among a firm’s customers. We empirically examine which commonly used and widely advocated customer feedback metrics are most valuable in predicting future business performance. Using American Customer Satisfaction Index data, we assess the linkages between six different satisfaction and loyalty metrics and COMPUSTAT and CRSP data-based measures of different dimensions of firms’ business performance over the period 1994–2000. Our results indicate that average satisfaction scores have the greatest value in predicting future business performance and that Top 2 Box satisfaction scores also have good predictive value. We also find that while repurchase likelihood and proportion of customers complaining have some predictive value depending on the specific dimension of business performance, metrics based on recommendation intentions (net promoters) and behavior (average number of recommendations) have little or no predictive value. Our results clearly indicate that recent prescriptions to focus customer feedback systems and metrics solely on customers’ recommendation intentions and behaviors are misguided.
Sharp, Bryon. 2008. “Net Promoter Score Fails the Test.” Marketing Research, American Marketing Association.
AUTHOR ABSTRACT Beginning in this issue, Byron Sharp, director of Ehrenberg-Bass Institution, University of South Australia, will be writing a column dealing with issues in marketing research. In this issue, he tackles the Net Promoter Score.