642
Views
0
CrossRef citations to date
0
Altmetric
Research Article

Plain language in web questionnaires: effects on data quality and questionnaire evaluation

ORCID Icon, ORCID Icon & ORCID Icon
Received 11 Apr 2023, Accepted 08 Dec 2023, Published online: 25 Dec 2023

ABSTRACT

In web surveys, no interviewer is present to clarify question comprehension problems, which can be particularly prevalent among respondents with low literacy skills. Although plain language is used in various contexts to improve text comprehensibility, its use in social science questionnaires has not been investigated to date. Using a between-subjects design, we compared the use of plain language and standard language in a web questionnaire in terms of data quality and respondents’ evaluation of the questionnaire. Plain language did not generally have a positive effect on data quality and questionnaire evaluation. However, in the plain-language condition, we found less item nonresponse and more response differentiation among respondents who spoke a language other than German at home, which suggests that especially respondents who likely have low literacy skills benefit from plain language. Based on our findings, we outline opportunities for future research and give practical recommendations.

Introduction

Survey researchers face various challenges with web surveys, including not having an interviewer present to help respondents to answer the questions. Respondents generally go through four cognitive steps when answering survey questions (e.g. Tourangeau et al., Citation2000). First, they have to comprehend the question by reading all its components, which implies processing the structure of the question and understanding the syntax and the words used. Second, they have to retrieve the relevant information from memory, and third, they have to use that information to make the required judgments. And finally, fourth, they have to provide an appropriate answer by selecting a response option or giving a response in their own words (Sudman et al., Citation1996; Tourangeau et al., Citation2000). The first step – question comprehension – is essential for the subsequent steps. If respondents are not conscientious about reading and understanding all components of a survey question, they will not know what it is about and what information is needed for an appropriate answer. Thus, comprehension problems in the first step of the cognitive response process may negatively affect the subsequent steps (Wenz et al., Citation2021).

To perform the four steps of the cognitive response process conscientiously and comprehensively – which Krosnick (Citation1991, Citation1999) called ‘optimizing’ – respondents must have certain cognitive abilities and be motivated to make the necessary effort. Otherwise, they may employ a ‘satisficing response strategy,’ which means they may execute all four steps, but less conscientiously and comprehensively (weak satisficing); or they may skip some or all the steps (strong satisficing; Krosnick, Citation1991, Citation1999). In addition to respondent ability and motivation, task difficulty also influences the likelihood of satisficing (Krosnick, Citation1991, Citation1999). Satisficing can take various forms, all of which affect data quality. For example, respondents may speed through the questionnaire, answering the questions very quickly without investing the time necessary to properly read and process them (e.g. Conrad et al., Citation2017; Zhang & Conrad, Citation2014); they may straightline by selecting the same or almost the same answer options for all items in a battery (e.g. Kim et al., Citation2019; Roßmann et al., Citation2018); they may choose the ‘don’t know’ option (e.g. Deutskens et al., Citation2004; Krosnick et al., Citation2002); they may leave one or more questions unanswered (e.g. Wenz et al., Citation2021; Yan & Curtin, Citation2010); or they may break off the survey (e.g. Mavletova & Couper, Citation2015; Steinbrecher et al., Citation2015).

In contrast to respondent characteristics, task difficulty is a satisficing-fostering aspect that survey researchers can directly influence. One way to make the response task easier, thereby counteracting comprehension problems and reducing the effort necessary to process the questions, may be to use plain language – that is, a simplified version of standard language, which is intended to be easy to read and understand.

In the present study, we aimed to fill a research gap by examining the use of plain language in web questionnaires. We conducted a web survey among respondents from an online access panel to explore the following research questions: (1) Does the use of plain language in a web questionnaire improve data quality compared with using standard language? (2) Does the use of plain language in a web questionnaire improve respondents’ evaluation of the questionnaire compared with using standard language?

Background

Plain language

Plain language is intended to be easy to read and understand; for weak readers or non-experts on a particular topic, it offers an alternative to standard-language texts (Maaß, Citation2020). Accessible and inclusive communication is a high-priority issue in many countries. The United Nations Convention on the Rights of People with Disabilities (CRPD) identifies plain language as a means of providing information in an accessible format (United Nations, Citation2017). Plain language has been the subject of extensive discussion, research, and legislation in the United States, the United Kingdom, Canada, and Australia since the 1960s (Petelin, Citation2010). In recent years, other countries around the world have become involved in the plain language movement, and many of them are now members of the Plain Language Association International (PLAIN), a nonprofit organization that aims inter alia to increase public awareness of plain language and to promote the use of and research and development in plain language (Plain Language Association International PLAIN, Citationn.d.). While there are several handbooks and practical guidelines on the use of plain English (e.g. Cutts, Citation2020), ‘plain language’ (Einfache Sprache) is a relatively new concept in Germany (Maaß, Citation2020), where it is not defined by strict rules but rather by a set of recommendations (Pottmann, Citation2019).

Plain language is characterized by short and concise sentences using the active rather than the passive voice. Using simple syntax and providing explanations or examples for rare terms or abstract matters is recommended; so, too, is avoiding foreign words, metaphors, and abstract terms. Beyond that, plain language aims not only to simplify the wording but also to adapt the structure and design of the written text. Thus, plain language can be considered a ‘complete package’ characterized by clarity of wording (e.g. word choice, length of sentences), structure (e.g. thematic sequence), and design (e.g. visual layout), so that ‘intended readers can easily find what they need, understand what they find, and use the information’ (International Plain Language Federation, Citationn.d..)

Survey questions in plain language

The idea of simplifying the language of survey questions to improve data quality and respondent experience is not new. Previous studies have examined the impact of different question wording – often using cognitive pretests or experimental designs. Cognitive pretesting is a well-established method in survey research. To assess the comprehensibility of individual questions and to identify problems in answering the questionnaire before it is fielded, respondents are asked to express their associations with the survey questions and words used (Lenzner et al., Citation2016; Presser et al., Citation2004). However, one shortcoming of cognitive pretesting is that due to time and resource constraints, usually only a small number of questions can be tested with a small group of respondents. In addition, as Eisner et al. (Citation2019) noted, cognitive pretesting does not offer systematic ways of finding alternatives for problematic words and phrases.

Previous experimental studies on survey data quality have also focused mainly on question wording. For example, Lenzner (Citation2012) found that survey questions with low comprehensibility in terms of different psycholinguistic text features (e.g. low-frequency words, vague terms, or complex syntax) decreased data quality. Blasius and Friedrichs (Citation2009) tested the use of high-brow (or elaborate) versus low-brow (or everyday) question wording and its effects on data quality. They showed that using low-brow language resulted in more diverse responses; however, the distribution of responses differed between the two language versions. Slavec and Vehovar (Citation2015) used either low-frequency (unfamiliar) or high-frequency (familiar) words in survey questions. They found evidence of lower break-off rates and lower perception of difficulty when the high-frequency question wording was used. These three studies show that the choice of words in survey questions can affect data quality and the respondents’ perception of the survey questions. However, to our knowledge, the use of plain language as a complete package that goes beyond the variation of individual words and terms has not yet been systematically studied in survey research.

Plain language can be used in web questionnaires to help respondents understand the questions as quickly, simply, and completely as possible. It has the potential to counteract comprehension problems that may occur in the first step of the cognitive response process. Because reading and understanding a question requires a certain level of language proficiency, comprehension problems when answering survey questions are more likely among respondents with low literacy skills (Lenzner, Citation2012; Wenz et al., Citation2021). According to the results of the second LEO – Living with Low Literacy survey, a large-scale survey of adult reading and writing skills in Germany, 12% of the adult population are considered to have low literacy and are able to read and write only simple sentences or even single words (Grotlüschen et al., Citation2020). Previous research in Germany has found that individuals with low formal education or a low family educational background are more likely to have lower literacy levels, and that a first language other than German is also a predictor of low literacy skills (Grotlüschen et al., Citation2020). Plain language can be of benefit not only to respondents with low literacy skills; it can also make it easier for all other respondents to complete web questionnaires. We assume that making questions easier to understand and reducing cognitive effort through plain language can reduce satisficing response behavior and improve data quality and respondents’ survey experience in general.

Data and methods

Sample and survey

The web survey was conducted in Germany from November 26 to December 4, 2020. The questionnaire included questions about family, relationships, and values. Respondents were sampled from a German online access panel. We used quotas for gender, age, and education. Of the 5,332 panelists who started answering the questionnaire, 30 were screened out because they were not members of the target population (age 18–69 years), and 1,179 were excluded due to filled quotas. Ninety-eight panelists broke off the survey, resulting in a break-off rate of 2.4% (Callegaro & DiSogra, Citation2008). Following the American Association for Public Opinion Research (The American Association for Public Opinion Research AAPOR, Citation2016) definition, we classified as ‘break-offs’ those cases that answered less than 50% of all applicable questions. A total of 4,025 respondents completed the questionnaire, with an average completion time of 10.7 min (median = 8.8 min).

Experimental design

Our questionnaire comprised three parts: the mandatory quota questions, the main section with 47 questions, and the evaluation questions at the end of the questionnaire. The between-subjects design was part of the main section of the questionnaire, with respondents being randomly assigned either to the control group with questions in standard language (n = 1,024) or the experimental group with questions in plain language (n = 981). Random assignment to the two language versions was successful, as indicated by nonsignificant differences between the two groups in terms of gender (χ2(2) = 0.29, p =.865), age (χ2(2) = 0.175, p =.916), and education (χ2(2) = 4.844, p =.089). The questionnaire included another experiment that is not reported in the current study, which is why the total number of cases was greater than the sum of the numbers reported above.

The 47 questions in the main section of the questionnaire were adapted from the German version of the international Generations and Gender Survey (GGS; Gauthier et al., Citation2018), which is implemented in Germany as part of the Family Research and Demographic Analysis (FReDA) panel study (Schneider et al., Citation2021). These questions and the evaluation questions at the end of the questionnaire were translated into plain language by Lebenshilfe Bremen e.V., an organization that provides professional translations of texts into plain language (see https://leichte-sprache.de/uebersetzungen/leichte-sprache/). The translation included changes in the wording (e.g. avoiding complex words), sentence structure (e.g. converting one sentence into two sentences), and visual design (e.g. adding line breaks) of the questions. Some changes resulted in longer question texts (especially introductions) and response option labels (see Figure A1 in the Appendix). Due to longer response option labels in the plain-language version, we presented all item batteries in both language versions in a vertical item-by-item format instead of a grid format. In doing so, we also followed the recommendation that the visual presentation of questions in plain language should make the response task easier for respondents. To ensure comparability between the two language versions, the visual design of the questions in the main section was the same for all respondents. For the same reason, respondents in both the experimental and the control condition received the evaluation questions in plain language (for the English translations of the questions in both language versions, see Table A1 in the Appendix).

Dependent variables

We computed various data quality indicators based on the 47 questions in the main section of the questionnaire, counting each item in an item battery as a single question. In addition, we calculated an evaluation index based on seven items asked at the end of the questionnaire (see Table A2 in the Appendix).

Break-offs

We created a binary variable that indicates whether a respondent broke off the survey during the main section of the questionnaire (0 = no break-off; 1 = break-off).

Item nonresponse

Respondents could skip questions without being forced to answer. For each respondent, we calculated the proportion of questions with missing answers. This indicator could take values from 0 to 1, where 0.5 means a respondent left 50% of the questions unanswered.

Completion time

Completion times are a proxy measure for the cognitive effort a respondent spent on answering the questions. Using the open-source Embedded Client Side Paradata (ECSP) script developed by Schlosser and Höhne (Citation2020), we collected page-wise timestamps in milliseconds for each survey page. We calculated the absolute completion time for the main section of the questionnaire for each respondent by adding up the page-wise response times. To account for differences in question length between the two language versions, we additionally calculated the relative completion time as the quotient of the page-wise response times divided by the number of characters in the respective question. We excluded outliers page-wise at two standard deviations above the group mean. Completion times are reported in seconds.

Time to first click

The time to first click captures the time respondents initially spent reading and comprehending the question before they selected an answer. Unlike the completion time measures, this indicator does not consider subsequent revisions of an initially answered question. And in the case of multiple items on a web page, this indicator is limited to the first response to one of the items. Using the ECSP script (Schlosser & Höhne, Citation2020), we measured the time until the first click was made on a web page and averaged these times across all pages in the main section. Again, we excluded outliers at two standard deviations above the group mean and reported the time to first click in seconds.

Speeding

Speeding means that respondents answer a question too quickly to be able to fully read and comprehend it (Conrad et al., Citation2017). To identify speeding respondents, we used the ‘scanning’ threshold method proposed by Andreadis (Citation2014), which assumes that 39.375 characters can be read per second. In addition to reading a question, a respondent needs additional time to process it. Thus, a constant is added that takes the value of 1.4 seconds for simple attitude questions and 2.0 seconds for more complex attitude questions. For each respondent, we calculated the proportion of questions in which speeding was present. The indicator can take values from 0 to 1, with higher values indicating more speeding.

Nondifferentiation

Nondifferentiation means that respondents use the same or nearly the same response option for all items in a battery. As the items cover very different aspects, a low degree of differentiation is undesirable, or even implausible. We used three indices to measure different aspects of nondifferentiating response behavior (Kim et al., Citation2019). For all three indices, we calculated the average value at respondent level across the six item batteries in the main section of the questionnaire.

First, we computed straightlining (Zhang & Conrad, Citation2014) as the proportion of item batteries in which respondents used the same scale point for each item in the battery. This indicator can range from 0 to 1, where a value of 0.5 means a respondent chose the same scale point in 50% of the item batteries.

Second, we calculated the coefficient of variation as CV=sxˉ (McCarty & Shrum, Citation2000), where s represents the standard deviation and xˉ the mean of a respondent’s answers across the items in a battery. This indicator is a measure of the distance between the scale points used by the respondents. It can take values from 0 to greater than 0, where 0 means that each item was answered with the same scale point, and higher values indicate a greater distance between the scale points used by a respondent.

Third, we calculated the probability of differentiation as a measure of variation in the scale points that respondents used to answer item batteries. According to McCarty and Shrum (Citation2000), the probability of differentiation is defined as Pd=1i=1nPi, where n is the number of scale points and Pi is the proportion of the answers at each point of the scale. As the item batteries differed in terms of the number of items as well as the number of scale points, the maximum value of Pd differed between the item batteries. Thus, we standardized the values for Pd by scaling the values of the individual batteries to a range between 0 and 1, with larger values indicating that a respondent used a greater number of different scale points to answer an item battery. Respondents who answered less than three items in an item battery were excluded from analyses of Pd.

Questionnaire evaluation

We calculated an evaluation index based on (a) six items covering different aspects of the questionnaire (i.e. ‘interesting,’ ‘varied,’ ‘important for science,’ ‘long,’ ‘difficult,’ and ‘too personal’), which respondents rated on a 5-point scale from 1 (no, not at all) to 5 (yes, always; Kaczmirek et al., Citation2014); and (b) an additional item asking respondents how they found the questionnaire overall, which they rated on a 5-point scale from 1 (very good) to 5 (very bad; see Table A2 in the Appendix). The items ‘long,’ ‘difficult,’ and ‘too personal’ were reverse coded before analysis, so that higher index values indicate a positive evaluation across all seven items. Respondents who left one or more questionnaire evaluation items unanswered were excluded from analyses of this dependent variable.

Methods

We employed logistic regressions for ‘break-off’ as a binary dependent variable and the language version offered as independent variable. For the continuous dependent variables (i.e. item nonresponse, completion time, time to first click, speeding, nondifferentiation, and questionnaire evaluation), we used t tests for independent samples.

In addition to the full-sample analyses, we conducted analyses for subgroups with low literacy skills. As the assessment of literacy skills was not part of our study, we drew on previous research on individuals in Germany who tend to have low literacy skills (Grotlüschen et al., Citation2020). According to Grotlüschen et al. (Citation2020), low formal education, first language other than German, and low family educational background are the strongest predictors of low literacy skills in Germany. As we did not collect data on family educational background, we were limited to the respondents’ level of education. For the first language spoken, we relied on information about the language that respondents predominantly speak at home and on information about whether respondents were born in Germany and had German citizenship. We distinguished subgroups by level of education (low, medium, and high), German citizenship (yes or no), Germany as country of birth (yes or no), and language predominantly spoken at home (German vs. other language). We re-ran the analyses for each of these subgroups.

Results

Full-sample analyses

shows our results for data quality and questionnaire evaluation depending on whether the questions were provided in plain language or standard language.

Table 1. Results* on data quality and questionnaire evaluation.

We found significant effects of language version on absolute completion time. Respondents who received the questions in plain language took on average about 20 seconds longer to answer than respondents who answered the questions in standard language. Regarding the relative completion time, which takes into consideration that the plain language questions were generally longer (i.e. the question texts and response options had more characters), the differences between the two language versions were no longer significant. Regarding nondifferentiation, we found significant differences in the coefficient of variation (CV); providing questions in plain language rather than standard language resulted in more differentiated responses in terms of a larger distance between the scale points used to answer the items in a battery. We did not find significant effects of language version on straightlining, probability of differentiation, break-off, item nonresponse, and time to first click. Further, we found no significant effects of language version on questionnaire evaluation, irrespective of whether the evaluation items were considered together or individually.

Subsample analyses

shows the results for the subgroups that are generally at higher risk of lower literacy.

Table 2. Results* for subgroups on data quality and questionnaire evaluation.

For respondents who speak a language other than German at home, we found significantly less item nonresponse when the questions were presented in plain language compared with standard language. For this subgroup, we also found significantly less straightlining and a higher probability of differentiation for questions in plain language compared with standard language.

We found significantly longer absolute completion times for plain language compared with standard language only among respondents who presumably had higher literacy skills (i.e. higher formal education, German citizenship, born in Germany, speaking German at home), but not for respondents who were more likely to have had low literacy skills (i.e. low formal education, not born in Germany, and speaking a language other than German at home). For the remaining indicators of data quality, as well as for questionnaire evaluation, we found no additional or differing results in the subsample analyses compared with the full sample.

Conclusion

Summary of results and discussion

In the present study, we focused on the use of plain language in web questionnaires and its impact on data quality and questionnaire evaluation. Although plain language is used in various contexts to improve text comprehensibility and ensure inclusion and accessibility, to our knowledge its use in population-wide social science surveys has not been investigated to date. Using plain language involves more than just explaining or replacing individual words; it is a complete package that includes simplifying the wording and sentence structure and clarifying the visual design of written text. In general, plain language is intended to make texts easier to read and understand. In self-administered (web) surveys, no interviewer is present to help respondents if they have problems comprehending the questions. Using plain language might help to prevent comprehension problems and thus ensure high-quality responses – especially among respondents with lower literacy skills, who are more likely to have comprehension problems. We further reasoned that making the first step of the cognitive response process easier for all respondents might help to reduce the effort required to process the questions, which might lead to a more positive evaluation of the questionnaire.

The results of our full-sample analyses were mixed. We found that absolute completion times were slightly longer when plain language was used compared with standard language. As we know from previous research, longer response times may indicate more conscientious response behavior, because comprehensive cognitive question processing takes more time (e.g. Callegaro et al., Citation2009; Lenzner et al., Citation2010; Toepoel et al., Citation2008). This result can therefore be interpreted as an indication that data quality was higher because respondents took more time to thoroughly comprehend the question and think about an answer. An alternative explanation for this finding may be the increased length of the question stems and response options in plain language, which required respondents to spend more time reading them. This interpretation is supported by the fact that when the number of characters was considered, there was no longer a difference in relative completion times between the two language versions. Regarding response differentiation in item batteries, we found a slightly higher coefficient of variation for plain language, but no statistically significant differences for straightlining and probability of differentiation between the two language versions. Thus, simplifying the language may help respondents take clearer positions on various issues, as evidenced by greater distances between the scale points used to answer the items in a battery. However, it does not necessarily result in respondents using a greater number of different scale points to answer item batteries. Break-off, item nonresponse, and time to first click, as further data quality indicators, were not affected by the different language versions, nor was questionnaire evaluation. Thus, using plain language in web questionnaires does not appear to be a panacea for improving data quality or respondent experience.

Our analyses for subgroups that were more likely to have low literacy skills (e.g. respondents with low formal education or with a first language other than German) yielded some promising findings. Our results suggest that plain language can potentially reduce item nonresponse and straightlining among respondents who are at risk of lower literacy skills. Thus, plain language may make it easier for some respondents to understand the questions and thereby reduce their likelihood to engage in satisficing behavior.

Implications for questionnaire design

Our findings have several practical implications for questionnaire design. First, we found no indication that plain language triggered detrimental response behavior in the full sample for any of the data quality indicators. We interpret this finding as a positive sign for those who fear that respondents with higher literacy skills might perceive more simplified language as unsophisticated or even taxing. Although we stress that more research is needed on the use of plain language in social science questionnaires, our findings may be a first indication that using plain language is feasible and does not hamper data quality.

Second, we found that developing a plain-language version of our standard questions was in itself helpful and beneficial for question development. In trying to simplify the question wording, the complexity of our questions became apparent, which was something we had taken for granted before their translation into plain language. It also helped us to shed a different light on question development and explicitly link it to the cognitive response process. Moreover, taking the perspective that plain language is a complete package led us to acknowledge that easing question comprehension involves not only the wording (e.g. word choice, length of sentences) but also the structure (e.g. thematic sequence) and design (e.g. visual layout) of written text. We see a benefit in routinely incorporating the exercise of creating an alternative question version in plain language into question development – if time and resource constraints allow.

Third, using plain language appears to make the cognitive response process easier for respondents with lower literacy skills. Thus, providing an alternative questionnaire for those respondents might be a helpful tool for researchers who have detected data quality issues in these respondent groups.

Limitations and future research opportunities

We designed the present exploratory study to initiate a discussion on the use of plain language in questionnaires to make the response process easier for respondents with low literacy and to make surveys more inclusive. Thus, our study has several limitations that are at the same time opportunities for further research on using plain language in questionnaires. First, we tested our questionnaire based on a sample from a non-probability online access panel. Online access panelists are likely to be experienced respondents, which may have masked the observed differences between the two language versions and limited the generalizability of our findings. However, we see such samples as a conservative test case, because the detection of effects that depend on literacy is less likely in a sample composed of individuals who have a basic level of literacy (i.e. the ability to sign up for an access panel) compared with a general population sample. While we acknowledge that non-probability samples can be used for exploratory studies such as ours to pave the way for more resource-intensive testing, we strongly recommend testing the use of plain language in probability-based samples of the general population.

Second, the relevant subgroups of respondents with low literacy skills were comparatively small, thus limiting our possibilities to conduct more in-depth analyses. Anticipating this issue when relying on respondents from an online access panel, we used quota sampling with education as one of our quota variables to ensure at least some variation across subgroups. In addition, we did not assess respondents’ actual literacy skills, but rather created subgroups based on assumptions derived from previous findings about predictors of low literacy skills. Consequently, we strongly recommend that future studies investigate plain-language versions in a sample that includes more individuals with (potentially) low literacy skills. More profound insights could be gained by assessing respondents’ literacy skills prior to investigating the effect of plain language on their response behavior.

Third, the process of translating survey questions into plain language holds potential for further research. There are no translation processes and standards specifically designed and used for adapting survey questions to plain language. We relied on professionals who specialize in translating texts into plain language for general purposes, but not specifically for survey questions. Thus, we see two major avenues for future research. First, measurement equivalence between the translated and the original language versions needs to be assessed. Our study was aimed at testing the feasibility of using plain language in web questionnaires and did not attempt to assess measurement equivalence. Assessing measurement equivalence is especially relevant when adapting well-established scales or questionnaires to plain language in ongoing longitudinal surveys. Second, we strongly encourage future research on standardized ‘checklists’ that provide information on which elements of a survey question need to be translated into plain language and how to create a version of the question that is easier for respondents to understand and answer.

Building on our exploratory study, we see potential to further investigate the use of plain language in questionnaires. Using plain language not only has the potential to improve the quality of responses by making the first step of the cognitive response process – question comprehension – easier; it can also make surveys more inclusive. Questionnaires that are more accessible because of plain language might motivate and empower individuals to participate who could not or would not respond to questionnaires in standard language.

Ethics declaration

Because our study was non-interventional (i.e. a survey), ethical approval was not required. However, we obtained informed consent from our respondents.

Supplemental material

Supplemental Material

Download MS Word (90 KB)

Disclosure statement

No potential conflict of interest was reported by the author(s).

Supplemental data

Supplemental data for this article can be accessed online at https://doi.org/10.1080/13645579.2023.2294880.

Additional information

Notes on contributors

Irina Bauer

Irina Bauer is a PhD student at GESIS – Leibniz Institute for the Social Sciences in Mannheim, Germany. Her research interests include (visual) questionnaire design and data quality in web surveys. Email: [email protected]. Postal address: GESIS – Leibniz Institute for the Social Sciences, B6.4-5, 68159 Mannheim, Germany.

Tanja Kunz

Tanja Kunz is a senior researcher at GESIS – Leibniz Institute for the Social Sciences in Mannheim, Germany. Her current research interests include the design and implementation of mixed-mode survey data collection, with a focus on questionnaire design and data quality. Email: [email protected]. Postal address: GESIS – Leibniz Institute for the Social Sciences, B6.4-5, 68159 Mannheim, Germany.

Tobias Gummer

Tobias Gummer is a senior researcher and team leader at GESIS – Leibniz Institute for the Social Sciences in Mannheim, Germany, and apl. Professor at the University of Mannheim. His methodological research interests include survey design, data quality, nonresponse, and correction methods for biases. Email: [email protected]. Postal address: GESIS – Leibniz Institute for the Social Sciences, B6.4-5, 68159 Mannheim, Germany

References

  • Andreadis, I. (2014). Data quality and data cleaning. In D. Garzia & S. Marschall (Eds.), Matching voters with parties and candidates: Voting advice applications in a comparative perspective (pp. 79–91). ECPR Press.
  • Blasius, J., & Friedrichs, J. (2009). The effect of phrasing scale items in low-brow or high-brow language on responses. International Journal of Public Opinion Research, 21(2), 235–247. https://doi.org/10.1093/ijpor/edp018
  • Callegaro, M., & DiSogra, C. (2008). Computing response metrics for online panels. Public Opinion Quarterly, 72(5), 1008–1032. https://doi.org/10.1093/poq/nfn065
  • Callegaro, M., Yang, Y., Bhola, D. S., Dillman, D. A., & Chin, T.‑Y. (2009). Response latency as an indicator of optimizing in online questionnaires. Bulletin of Sociological Methodology/Bulletin de Méthodologie Sociologique, 103(1), 5–25. https://doi.org/10.1177/075910630910300103
  • Conrad, F. G., Couper, M. P., Tourangeau, R., & Zhang, C. (2017). Reducing speeding in web surveys by providing immediate feedback. Survey Research Methods, 11(1), 45–61. https://doi.org/10.18148/srm/2017.v11i1.6304
  • Cutts, M. (2020). Oxford guide to plain English (5th ed.). Oxford University Press.
  • Deutskens, E., Ruyter, K. de, Wetzels, M., & Oosterveld, P. (2004). Response rate and response quality of Internet-based surveys: An experimental study. Marketing Letters, 15(1), 21–36. https://doi.org/10.1023/B:MARK.0000021968.86465.00
  • Eisner, L., Clémence, A., Roberts, C., Joost, S., & Theler, J.‑M. (2019). Developing attitude measures based on respondents’ representations of unfamiliar objects: An application to attitudes toward biodiversity. Field Methods, 31(1), 56–75. https://doi.org/10.1177/1525822X18797280
  • Gauthier, A. H., Cabaço, S. L. F., & Emery, T. (2018). Generations and Gender Survey study profile. Longitudinal and Life Course Studies, 9(4), 456–465. https://doi.org/10.14301/llcs.v9i4.500
  • Grotlüschen, A., Buddeberg, K., Dutz, G., Heilmann, L., & Stammer, C. (2020). Low literacy in Germany: Results from the second German literacy survey. European Journal for Research on the Education andLearning of Adults, 11(1), 127–143. https://doi.org/10.3384/rela.2000-7426.rela9147
  • International Plain Language Federation (n.d.). Plain Language: Definitions. https://www.iplfederation.org/plain-language/
  • Kaczmirek, L., Bandilla, W., Schaurer, I., & Struminskaya, B. (2014). GESIS Online Panel Pilot: Multitopic introductory wave (survey 1) (ZA5582 Data file Version 1.0.0). GESIS Data Archive. https://doi.org/10.4232/1.11570
  • Kim, Y., Dykema, J., Stevenson, J., Black, P., & Moberg, D. P. (2019). Straightlining: Overview of measurement, comparison of indicators, and effects in mail–web mixed-mode surveys. Social Science Computer Review, 37(2), 214–233. https://doi.org/10.1177/0894439317752406
  • Krosnick, J. A. (1991). Response strategies for coping with the cognitive demands of attitude measures in surveys. Applied Cognitive Psychology, 5(3), 213–236. https://doi.org/10.1002/acp.2350050305
  • Krosnick, J. A. (1999). Survey research. Annual Review of Psychology, 50, 537–567. https://doi.org/10.1146/annurev.psych.50.1.537
  • Krosnick, J. A., Holbrook, A. L., Berent, M. K., Carson, R. T., Hanemann, W. M., Kopp, R. J., Mitchell, C., Presser, S., Ruud, P. A., Smith, V. K., Moody, W. R., Green, M. C., & Conaway, M. (2002). The impact of “no opinion” response options on data quality: Non-attitude reduction or an invitation to satisfice? Public Opinion Quarterly, 66(3), 371–403. https://doi.org/10.1086/341394
  • Lenzner, T. (2012). Effects of survey question comprehensibility on response quality. Field Methods, 24(4), 409–428. https://doi.org/10.1177/1525822X12448166
  • Lenzner, T., Kaczmirek, L., & Lenzner, A. (2010). Cognitive burden of survey questions and response times: A psycholinguistic experiment. Applied Cognitive Psychology, 24(7), 1003–1020. https://doi.org/10.1002/acp.1602
  • Lenzner, T., Neuert, C., & Otto, W. (2016). Cognitive pretesting. In GESIS Survey Guidelines. GESIS – Leibniz Institute for the Social Sciences. https://doi.org/10.15465/gesis-sg_en_010
  • Maaß, C. (2020). Easy language – Plain language – Easy language plus: Balancing comprehensibility and acceptability. Frank & Timme.
  • Mavletova, A., & Couper, M. P. (2015). A meta-analysis of breakoff rates in mobile web surveys. In D. Toninelli, R. Pinter, & P. de Pedraza (Eds.), Mobile research methods: Opportunities and challenges of mobile research methodologies (pp. 81–98). Ubiquity Press. https://doi.org/10.5334/bar.f
  • McCarty, J. A., & Shrum, L. J. (2000). The measurement of personal values in survey research: A test of alternative rating procedures. Public Opinion Quarterly, 64(3), 271–298. https://doi.org/10.1086/317989
  • Petelin, R. (2010). Considering plain language: Issues and initiatives. Corporate Communications: An International Journal, 15(2), 205–216. https://doi.org/10.1108/13563281011037964
  • Plain Language Association International (PLAIN) (n.d.). Who we are. https://plainlanguagenetwork.org/plain/who-we-are/
  • Pottmann, D. M. (2019). Leichte Sprache and Einfache Sprache – German plain language and teaching DaF German as a foreign language. Studia Linguistica, 38, 81–94. https://doi.org/10.19195/0137-1169.38.6
  • Presser, S., Couper, M. P., Lessler, J. T., Martin, E., Martin, J., Rothgeb, J. M., & Singer, E. (2004). Methods for testing and evaluating survey questions. Public Opinion Quarterly, 68(1), 109–130. https://doi.org/10.1093/poq/nfh008
  • Roßmann, J., Gummer, T., & Silber, H. (2018). Mitigating satisficing in cognitively demanding grid questions: Evidence from two web-based experiments. Journal of Survey Statistics and Methodology, 6(3), 376–400. https://doi.org/10.1093/jssam/smx020
  • Schlosser, S., & Höhne, J. K. (2020). ECSP – Embedded Client Side Paradata (Version v1) [Computer software]. Zenodo. https://doi.org/10.5281/zenodo.3782592
  • Schneider, N. F., Bujard, M., Wolf, C., Gummer, T., Hank, K., & Neyer, F. J. (2021). Family Research and Demographic Analysis (FReDA): Evolution, framework, objectives, and design of ”The German Family-Demographic Panel Study”. Comparative Population Studies, 46. https://doi.org/10.12765/CPoS-2021-06
  • Slavec, A., & Vehovar, V. (2015). The role of word frequencies in detecting unfamiliar terms and their effect on response quality. Psihologija, 48(4), 327–344. https://doi.org/10.2298/PSI1504327S
  • Steinbrecher, M., Roßmann, J., & Blumenstiel, J. E. (2015). Why do respondents break off web surveys and does it matter? Results from four follow-up surveys. International Journal of Public Opinion Research, 27(2), 289–302. https://doi.org/10.1093/ijpor/edu025
  • Sudman, S., Bradburn, N. M., & Schwarz, N. (1996). Thinking about answers: The application of cognitive processes to survey methodology (1st ed.). Jossey-Bass Publishers.
  • The American Association for Public Opinion Research (AAPOR). (2016). Standard definitions: Final dispositions of case codes and outcome rates for surveys (9th ed.). AAPOR. https://aapor.org/wp-content/uploads/2022/11/Standard-Definitions20169theditionfinal.pdf
  • Toepoel, V., Das, M., & van Soest, A. (2008). Effects of design in web surveys. Public Opinion Quarterly, 72(5), 985–1007. https://doi.org/10.1093/poq/nfn060
  • Tourangeau, R., Rips, L. J., & Rasinski, K. A. (2000). The psychology of survey response. Cambridge University Press.
  • United Nations. (2017, January 20). Convention on the Rights of Persons with Disabilities (CRPD): Article 2 – Definitions. https://www.un.org/development/desa/disabilities/convention-on-the-rights-of-persons-with-disabilities/article-2-definitions.html
  • Wenz, A., Al Baghal, T., & Gaia, A. (2021). Language proficiency among respondents: Implications for data quality in a longitudinal face-to-face survey. Journal of Survey Statistics and Methodology, 9(1), 73–93. https://doi.org/10.1093/jssam/smz045
  • Yan, T., & Curtin, R. (2010). The relation between unit nonresponse and item nonresponse: A response continuum perspective. International Journal of Public Opinion Research, 22(4), 535–551. https://doi.org/10.1093/ijpor/edq037
  • Zhang, C., & Conrad, F. G. (2014). Speeding in web surveys: The tendency to answer very fast and its association with straightlining. Survey Research Methods, 8(2), 127–135. https://doi.org/10.18148/srm/2014.v8i2.5453