Download Using Questionnaire Design to Fight Nonresponse Bias in Web

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Ease of doing business index wikipedia , lookup

Web analytics wikipedia , lookup

Transcript
Using Questionnaire Design
to Fight Nonresponse
Bias in Web Surveys
Social Science Computer Review
Volume 28 Number 2
May 2010 251-267
# 2010 SAGE Publications
10.1177/0894439309340751
http://ssc.sagepub.com
hosted at
http://online.sagepub.com
Paula Vicente
Elizabeth Reis
ISCTE-Lisbon University Institute
The technical potential of the Internet offers survey researchers a wide range of possibilities for
web surveys in terms of questionnaire design; however, the abuse of technical facilities can
detract respondents from cooperating rather than motivating them. Within the web survey
methodology literature, many contributions can be found on how to write a ‘‘good’’
questionnaire. The outcomes are however scattered and researchers and practitioners may
find it difficult to obtain an overall picture. The article reviews the latest empirical research
on how questionnaire characteristics affect response rates. The article is divided into three
main sections: an introduction where the various forms of nonresponse in web surveys are
described; a second section presenting questionnaire features affecting nonresponse—
general structure, length, disclosure of survey progress, visual presentation, interactivity,
and question/response format—and a final section that summarizes the options in terms of
questionnaire design and its implications for nonresponse rate.
Keywords:
rate
web surveys; questionnaires; nonresponse rate; dropout rate; item nonresponse
Introduction
With the growing number of Internet users and the increasing popularity of broader
bandwidth, it is clear that web survey research is becoming more and more prevalent especially in organizational environments, in access panels, or generally in contexts where the
target population can be well identified.
Like mail questionnaires, the web questionnaires are self-administered but the Internet
offers researchers a number of powerful advantages over mail surveys. In web surveys,
there is no printing or mailing of the questionnaires, data already come in an electronic format, and there can be permanent, 24-hr a day data collection. All this permits significant
cost and time savings. Additionally, the Internet allows interaction with the respondent that
can work in favor of better completed questionnaires, without missing or commission
errors.
Authors’ Note: Please address correspondence to Paula Vicente, ISCTE-Lisbon University Institute, Lisbon,
1649-029, Portugal; e-mail: [email protected].
251
Downloaded from ssc.sagepub.com at Aalto University on March 29, 2016
252 Social Science Computer Review
But along with the positive aspects offered by Internet as a survey research tool, there are
important methodological concerns that cannot be ignored if survey quality is to be
guaranteed. One of those concerns relates to response rates. Web surveys are generally connected with low response rates (Couper, 2000; Lozar Manfreda, Berzelak, Vehovar, Bosnjak, & Haas, 2008; Lozar Manfreda, Bosnjak, Haas, & Vehovar, 2005; Umbach, 2004) and
this can be a cause for bias in survey estimates. Other problems include coverage and sampling error (Couper, 2000). Because not everybody has access or uses the Internet, webbased survey estimates may suffer from coverage bias because they exclude some segments
of the population; moreover Internet penetration is not evenly distributed across segments
of the population: Internet users tend to be younger, with higher educational and income
levels than the general population (European Commission, 2008; Hudson, Seah, Hite, &
Haab, 2004). Sampling error occurs during the process of selecting the sample of
respondents, for example, a nonprobability selection scheme is a cause of sampling error
as not every element of the sampling frame has a nonzero chance of being selected for the
survey.
The designing and implementation of a web survey therefore involves devoting
much attention to procedures aimed at controlling the risk of nonresponse, coverage, and
sampling error.
When developing and implementing a strategy aimed at reducing nonresponse rates in a
survey, several design features can be manipulated. Basically, the surveyor can act in two
ways: (a) proactively, either by offering the individuals some kind of incentive that indeed
seduces them into cooperating or by implementing a contact strategy with the respondents
that convinces them of the importance of cooperating (e.g., conducting multiple contacts) or
(b) alternatively (or additionally) defensively, by manipulating certain aspects of the survey
in order not to put the individuals off cooperating; designing a ‘‘good’’ questionnaire is one
way of doing this.
Differences in questionnaire design have long been acknowledged as affecting response
rates in mail surveys (Carroll, 1994; Dillman, Sinclair, & Clark, 1993; Heberlein and Baumgartner, 1975; Jansen, 1985). Although some of the design principles established for mail
surveys may be translated to web surveys because both methods are self-administered, others require specific treatment in the Internet context because the technical features of the
Internet allow a respondent-survey interaction that is distinct from that of paper questionnaires (Couper, Traugott, & Lamias, 2001).
Within the web survey methodology literature, many contributions provide tips and suggestions on how to write a ‘‘good’’ questionnaire in order not to detract respondents from
cooperating. The outcomes are however scattered and researchers and practitioners may
find it difficult to obtain an overall picture. The objective of this article is to use the latest
empirical research to propose a categorization of questionnaire characteristics which
influence response rates and to summarize a set of questionnaire design principles for web
surveys that help researchers considering conducting a web-based survey to (a) know the set
of possibilities of questionnaire design features, (b) be aware of the impact each design
choice may have on response rates, and (c) choose the ‘‘best’’ design for the questionnaire.
Nonresponse rate is defined in survey methodology as the percentage of people contacted
that did not respond to the survey (Groves et al., 2004). For web surveys, there are a number
of steps in this whole process of contacting and trying to get information from the
252
Downloaded from ssc.sagepub.com at Aalto University on March 29, 2016
Using Questionnaire Design to Fight Nonresponse Bias / Vicente, Reis
253
respondents and several types of (non)response rates can be distinguished during this process. However, because we are dealing with questionnaire design, we restrict our analysis to
the web survey process from the moment the respondent starts to answer the web questionnaire. We shall concentrate on questionnaire design features that can influence respondents’
participation during the filling out of the questionnaire but have limited impact on the initial
decision to participate in a web survey. Specifically we will refer to
1.
2.
3.
Dropout rate: the percentage of those who prematurely abandoned the questionnaire, that
is represents the percentage of respondents who started the questionnaire but did not end it.
A dropout can occur immediately after the start, in the middle, or near the end of the
questionnaire.
Item nonresponse rate: the unanswered questions (‘‘no opinions,’’ ‘‘don’t knows,’’ or ‘‘no
answer’’) as a percentage of the total number of questions in the questionnaire.
Overall completion rate: the percentage of respondents who completed the questionnaire
of all those who started the questionnaire.
Questionnaire Features Affecting Nonresponse
Web survey methodology offers much advice on how to write a questionnaire in order
not to detract individuals from cooperating. Although a lot of experimental studies are
available about the effects of questionnaire features on nonresponse, there are few categorizations structuring that knowledge.
The work by Couper, Tourangeau, and Kenyon (2004) distinguishes two types of web
questionnaire features that are likely to influence participants’ motivation to start and complete a survey till the end. The first is task elements, which includes factors that influence
the way the respondents deal with the questionnaire (e.g., question wording, response
options, instructions, navigational clues); the second is style elements, comprising a set
of elements related to the look of the questionnaire (e.g., typefaces, background color, logotypes, and other elements). Similarly, Jackob and Zerback (2006) arrange web questionnaires features into two groups: visual appearance and structural appearance. The first
category comprises the stylistic elements of the questionnaire and the second includes the
structural/task features of the questionnaire. However, these classifications are quite generalist, especially in relation to structural features, because most experimental research
reported in the literature allows several factors with distinct effects on response rates to
be distinguished within this group, that are worth separating into specific categories of web
questionnaires features. It is our conviction that a more refined classification can be
achieved.
Ganassali (2008) has proposed a framework to study the effects of questionnaire
characteristics on response quality in web surveys based on the response process of a web
questionnaire. According to the author, this process begins by the respondent giving a quick
look through the questionnaire to apprehend its length and make a preliminary assessment
of the level of effort required to answer it. Second, the respondents will get an impression of
the friendliness of the form according to the balance between text and illustrations. The
third step will be the reading of the questions and the assessment of the wording. Then the
respondent is supposed to apprehend the interactive components of the survey interview,
253
Downloaded from ssc.sagepub.com at Aalto University on March 29, 2016
254 Social Science Computer Review
and finally, the respondent will look at the response format to know exactly what kind of
task and what kind of data is expected: ticks, numbers, texts, and so on. Based on these
successive stages, the author identified five groups of questionnaire factors with influence
on response quality.
Alongside the process that determines the quality with which the respondent answers the
questions, there is another process that determines whether the respondent gives an answer
or leaves a particular question unanswered, whether the respondent continues till the questionnaire is completed or gives up before finishing it. When at the outset, the respondent
gives a quick look through the questionnaire to apprehend its general layout and its length,
he or she will come to the decision whether to proceed with the questionnaire or to immediately give up. On deciding to advance, the respondent will read the questions and find out
how they are supposed to be answered; the questions and response format may determine
whether he or she leaves a particular question unanswered or even whether he or she drops
out or continues to the next question. As the respondent advances through the questionnaire,
he or she gets a clearer picture of its visual presentation, how to navigate across questions,
and how much still remains till the end of the questionnaire; all these factors can act as
motivators or detractors to continue.
Starting from the two-group categorization of questionnaire features with influence on
response rates already present in the literature and joining the successive stages of the contact process between the respondents and the questionnaire itself, we suggest a six-category
classification for web questionnaires features capable of affecting nonresponse: general
structure, length, disclosure of survey progress, visual presentation, interactivity, and
question/response format. When describing the effect of each feature-type category on
nonresponse, we will distinguish, whenever possible, the effect on dropout rates, item
nonresponse rates, and overall completion rates.
General Structure
One of the features that define the general structure of a questionnaire is how the questions are presented. Basically the options can balance between a scroll design and a screen
design. A scroll design is like the electronic version of a paper questionnaire. The questions
are displayed on a single web page, the respondent can scroll from question to question to
read and select the responses. A button is placed at the end of the page to submit the
responses to the web site of the survey and works much like the act of mailing a paper
questionnaire. At the other extreme, the screen design is very similar to a Computer
Assisted Telephone Interviewing (CATI) survey conducted in the Internet environment.
Questions appear on several screens/pages with a button to submit the responses on each
screen; each screen can contain just one or a number of questions. Responses on each screen
are submitted prior to receiving the next screen. Hybrid surveys fall between these two
extremes and may provide a mixture of screen and scroll, for example, multiple scrolling
sections.
Nonresponse due to the choice of one design or the other affects mainly the item nonresponse rate. The research by Lozar Manfreda, Batagelj, and Vehovar (2002) in which a
comparison between single, long scrolling page, and multiple page design was made taking
a national web survey revealed higher item nonresponse in the scroll design. The two
254
Downloaded from ssc.sagepub.com at Aalto University on March 29, 2016
Using Questionnaire Design to Fight Nonresponse Bias / Vicente, Reis
255
designs were also compared in terms of dropout rates but the difference between them was not
statistically significant. Because no forced response procedure (an aspect that will be discussed in Section Interactivity) was used in either design, respondents might have skipped
individual survey questions, either by mistake or on purpose, owing to the scrolling between
questions required in the one-page design, thus causing a higher item nonresponse rate.
A comparison of the scroll and screen versions of the same questionnaire was also made
by Tourangeau, Couper, Galesic, and Givens (2004) using a national survey on physicians.
Among other issues, the study evaluated completion rates and item nonresponse. Results
revealed no significant differences between the two designs in terms of completion rates.
However, item omission rates were higher in the scroll version and the most significant
differences were found in items involving skipping patterns.
Peytchev, Couper, McCabe, and Crawford (2006) compared the scroll version with the
screen version of the same questionnaire in a survey of university students. The two designs
were equivalent in terms of completion rates and dropout rates, and in terms of item
nonresponse, the scroll version produced only a slightly higher rate than the screen version.
The authors justified the similarities between the two designs by the fact that the sample
was composed of university students who are typically very experienced in using the
Internet and therefore naturally comfortable with any design.
A second issue in the structure of the questionnaire relates to how it is sent: either
embedded or in an attached format. An embedded questionnaire means the researcher
embeds the questions in the text of the e-mail. Alternatively, the questionnaire can be sent
as an attachment to the e-mail, either as a link or as a document. To respond to an attached
e-mail survey, the respondent must have the hardware and the software that enables him or
her to download, read, and upload the file and he or she must know how to execute the various response steps. While scroll designs can be either embedded or attached, screen
designs require the attached format. The embedded format was more common in the early
days of web surveys when computer hardware and software were not so sophisticated.
Nowadays, the most usual procedure is to attach the questionnaire namely by providing
links to the survey in the e-mail. We have found just one study dealing with this topic; the
research by Dommeyer and Moriarty (2000) compared the embedded version (i.e., the
questions embedded in the body of the e-mail) with the attached version (i.e., the survey
was accessed by means of a link included within the e-mail) of the same questionnaire in
a survey on students’ attitudes toward drinking. The authors found significant differences
in completion rates between the modes with the embedded version producing an 8% higher
response rate than the attached version. No differences were found for item nonresponse.
The authors concluded that the embedded version was more comfortable and secure
because it did not require the downloading and uploading of a foreign file and there was
no threat of virus.
Length
The participation in a survey always implies a burden for the respondent translated in the
time he or she has to spend answering the questions. Therefore, the decision to carry on till
the end of a survey is to a great extent influenced by the effort required of the respondent.
Common sense tells us that there is likely to be a lower response rate in longer
255
Downloaded from ssc.sagepub.com at Aalto University on March 29, 2016
256 Social Science Computer Review
questionnaires than in shorter ones. The length is not always measured in the same way in
the studies: this can be done by using number of pages, number of questions, or the completion time. Some studies show that questionnaire length has a modest effect on response
rates in mail surveys (Dillman et al., 1993, Haraldsen, Stalnacke, & Blackwell, 1999). Some
meta-analyses of surveys also indicate that the impact of length on response rates is modest
(Yammarino, Skinner, & Childers, 1991; Yu & Cooper, 1983).
Several studies have examined the effect of questionnaire dimension on response rates in
web surveys. The meta-analysis conducted by Cook, Heath, and Thompson (2000) to investigate the determinants of response rates in web surveys found no significant correlation
between questionnaire length and response rates. However, questionnaire length was found
to affect response rates in subsequent studies. In their research on design features
influencing response rates, Lozar Manfreda and Vehovar (2002) questioned web survey
researchers on their experience with web surveys. The respondents selected questionnaire
length, from among the possible reasons, as the problem that most accounted for dropouts in
web surveys. The effect of questionnaire length on web surveys nonresponse has been studied by Deutskens, Ruyter, Wetzels, and Oosterveld (2004) in a multiclient attitude and
usage study where a shorter version of the questionnaire (15 to 30 min to complete) was
compared to the long version (30 to 45 min to complete). The short version yielded a higher
response rate (24.5% vs. 17.1%). The dropout rate was higher in the long version and the
respondents stopped sooner in the long version than in the short version. Item nonresponse
(measured by the percentage of ‘‘don’t know’’ answers) was also higher in the long version.
In a more recent survey among university students on young peoples’ consumption
patterns, Ganassali (2008) evaluated the effect of questionnaire size by comparing a short
questionnaire (20 questions) with a longer one (42 questions). The author proved that
questionnaire length had a significant effect on dropout rates and found higher dropout rates
in the longer version of the questionnaire.
Another issue related to questionnaire length is the announcement a priori of questionnaire length. Hogg and Miller (2003) used two survey experiments—a concept test survey
regarding interactive television and a dining chain customer loyalty survey—to investigate
how revealing questionnaire length in advance affected dropout rates. Respondents were
chosen from a panel of web users. In both surveys, the shorter version of the questionnaire
was assigned to individuals who previously indicated they had no more than 15 min to
complete the survey, while the longer version was assigned for individuals who previously
indicated they had up to 30 min to complete the survey. The authors found higher dropout
rates among those assigned to the longer version of the questionnaire in both surveys.
Crawford, Couper, and Lamias (2001) also conducted an experiment to evaluate whether
the previous announcement of questionnaire length would affect the percentage of people
who begin the survey and whether dropouts would be higher when the survey took longer
than the promised completion time. The author found that respondents who were informed
that the survey would take 8 to 10 min had a lower nonresponse rate than those who were
told it would take 20 min. However, the longer version group had a lower dropout rate once
they started the survey. The two groups were not significantly different in terms of the point
at which the respondents stopped filling in the questionnaire. Galesic (2005, 2006) replicated to some extent this experiment and investigated the effect of announced length on
response rates, specifically the effect of announcing a shorter (10 min) versus a longer
256
Downloaded from ssc.sagepub.com at Aalto University on March 29, 2016
Using Questionnaire Design to Fight Nonresponse Bias / Vicente, Reis
257
(30 min) completion time and the effect if the announced length did not correspond to the
actual length. A web survey on unemployment in Croatia was used for the experiment. The
author found that 76% of the respondents of the 10-min group started the questionnaire
compared with 64% of the 30-min group. Moreover, the dropout rate was higher in the long
version (50% vs. 28%). When the actual length was much longer than the announced length,
the dropout rates were higher than when the actual length matched the announced length but
dropouts occurred later in the questionnaire.
Disclosure of Survey Progress
Besides questionnaire length, there is the issue of the respondent’s perception of
progress. While it is easy for a paper-questionnaire respondent to run over the pages of the
questionnaire and see how much longer it will take to reach the end, in web surveys this may
not be so easy. In the case of scroll design questionnaires, the respondents can scroll up and
down to find out how many questions there are in much the same way as he or she would do
in a paper questionnaire. In the case of screen design questionnaires, it is not usually
possible to scroll between screens so a progress indicator can be included in the survey
when considered necessary to give respondents some feedback on their progress through
the questionnaire. The progress information can be displayed in a graphical format (e.g.,
as a bar chart or as a pie chart) or as a text message (e.g., x% of survey complete). The
progress indicator can be permanently visible, that is, appears in every screen or it can
appear only at the respondent’s request (e.g., display a show progress link on each
screen where the respondent can click whether he or she wants to know his or her progress).
Alternatively, or additionally, progress information can be displayed on some screens at key
(transition) points of the survey (e.g., display a screen with the message you are about one
third through the survey).
The objective of progress indicators in web surveys is to reduce dropouts and thus
increase the completion rate. Several studies have sought to investigate whether the inclusion of a progress indicator is beneficial to the response rate. Couper et al. (2001) designed
an experiment using a survey with university students where a random half of the sample
was given a questionnaire with a progress indicator while the other half was not. Contrary to
what the authors expected, the difference between groups in terms of completion rates was
not statistically significant; moreover, the group with the progress indicator took longer on
average to complete the survey which the authors explained by the longer download times
of the progress indicator version of the questionnaire. The conclusion was that longer download times associated with the progress indicator dampened the potential positive effect of
including a progress indicator.
In the study by Crawford et al. (2001), the problem of slow downloading was controlled
and the progress indicator was found to have a negative effect when the with-progress indicator and the no-progress indicator groups were compared, namely the completion rate was
lower for those receiving the progress indicator version. A possible explanation for this
rather unexpected outcome came from the fact that the progress indicator reported the progress through the questionnaire inaccurately, that is, the progress indicator (that detected the
number of questions completed out of the total number of questions in the survey) underestimated the amount of time (or effort) that had been completed. This difference was
257
Downloaded from ssc.sagepub.com at Aalto University on March 29, 2016
258 Social Science Computer Review
caused mainly by the open-ended questions sections of the questionnaire where it is more
difficult than in the closed-ended questions section to predict the completion time.
Conrad, Couper, Tourangeau, and Galesic (2005) studied two additional issues regarding
the effectiveness of progress indicators in reducing dropouts in web surveys: the impact of
progress speed (constant, fast-to-slow, or slow-to-fast) of the progress indicator and the
impact of frequency with which the progress information was displayed (always-on, intermittent, or on-demand). The survey was on lifestyle attitudes and respondents were
recruited from two commercial opt-in panels. The higher dropout rates were found in the
slow-to-fast version (slower progress at the start increasing to fast near the end of the questionnaire) of the progress indicator and most of the dropouts occurred earlier in this version
than in the other two versions. As for the frequency of the progress information display, the
results in the sample that could request progress information on-demand revealed that only
37% of the respondents requested that information and progress information was displayed
on only 1.1% of the screens. The authors also found an interaction effect between speed of
progress and the display of feedback information: there are more dropouts where there is
slow early progress and more frequent feedback (it seems that frequent display is seen
by the respondents as something that delays their progress); the dropout rate is reduced
where there is faster early progress and more frequent feedback (either permanently on
or intermittent). In short, in long questionnaires, the inclusion of a permanently-on progress
indicator may not be very effective in reducing dropouts because it conveys that much of
the questionnaire remains unanswered; and respondents are very sensitive to early
information when presented on every page (whether this information is good news or bad),
therefore intermittent feedback may be a good solution for minimizing costs and
maximizing benefits.
Visual Presentation
The possibility of using the advanced graphics technology supported by current web
browsers is one of the most often cited advantages of web surveys. Graphics can be used
to illustrate survey questions, decreasing the burden on respondents if they are designed for
easy navigation through the questionnaire and/or improving respondents’ motivation and
satisfaction. The latter can be achieved by generating a valuable feeling of having a ‘‘good
time’’ or ‘‘fun’’ while answering the web questionnaire (Lozar Manfreda et al., 2002).
Web survey questionnaires are based largely on visual language. Redline and Dillman
(1999) distinguishes between three different types of visual language: graphic language
(fonts, font sizes, borders, and tables); symbolic language (e.g., arrows that guide the
respondents through the questionnaire); and numerical language (used to number the
questions and response items). Couper (2001) includes font size, font type, color, layout,
symbols, images, animation, and other graphics as components of visual language.
Ganassali (2008) refers that the visual information in a questionnaire can be displayed
on three different levels: questions in which images play a major part (e.g., logotype recognition), images as supplements to the text (e.g., illustrations), and incidental images (e.g.,
the screen background).
Dillman, Tortora, Conradt, and Bowker (1998) studied the effect of using a ‘‘fancy’’
design or a ‘‘plain’’ design in the same questionnaire on the response rates for a web survey.
258
Downloaded from ssc.sagepub.com at Aalto University on March 29, 2016
Using Questionnaire Design to Fight Nonresponse Bias / Vicente, Reis
259
Both questionnaires used individual screen formats, with scrolling reserved for individual
questions that required more than one screen for viewing. No progress indicator was used in
either format. Both questionnaires had the same wording and question order but they
differed in some design specific features. More specifically, the fancy version used bright
colors (purple and pink) with a constantly changing figure/ground format (vs. black and
white colors in the plain version), html tables for constructing tables (vs. no use of html
tables in the plain version), and placement of answer categories on the extreme right of each
screen (vs. placing answers on the left side of the screen in the plain version). The plain
version proved to be better in terms of completion rate (41% in the plain vs. 36% in the
fancy); additionally, dropouts occurred on a later page in the plain version than in the fancy
version which meant that the plain version was more likely to be fully completed. The
authors explained their findings by the fact that the plain version required less transmission
time and less time to be completed than the fancy version.
The influence of advanced graphics for illustrating survey questions, that is, logotypes,
on nonresponse was evaluated in the experiment by Lozar Manfreda et al. (2002). In a
questionnaire on web site recognition, a random group of respondents was administered
a version of the questionnaire with a page presenting site logotypes and their names; the
other group was administered the version where only the names of the sites were written.
All the other features of the questionnaire were identical. The use of logotypes proved to
have a positive effect on item nonresponse because a lower item nonresponse rate was
found for the group with the logotype version. However, the use of logotypes caused a
higher dropout rate: the percentage of respondents abandoning the survey at the page
containing the question with the logotypes was higher than in the version of the questionnaire without logotypes; this was more marked among respondents using the slowest
Internet connection indicating that the use of advanced graphics can cause some downloading problems for those with the slowest types of Internet connection to the point that they
quit the survey.
Deutskens et al. (2004) also evaluate the impact in a multiclient attitude and usage study
on response rate of visual versus text version of the same questionnaire. The visual version
meant that respondents could see the brand name and the picture or logo in a list of products
while in the text version they only saw the name of the products. The visual presentation of
the questionnaire had a lower response rate than the text version although item nonresponse
was higher in the text version than in the visual version; respondents in the text group
indicated ‘‘don’t know’’ more often than respondents in the visual group. As in the studies
previously cited, the authors justify the lower overall response rate in questionnaires with
visual enhancements by the longer downloading times of visual presentations.
Interactivity
In every self-administered survey, the respondents tend to seek information from the
questionnaire that allows them to dissipate any doubt about the meaning of the questions,
how the answers should be given, or how to navigate through the questionnaire. One of the
advantages of web questionnaires over paper questionnaires is that it can ‘‘interact’’ with
the respondents, thus helping them to know what to do during questionnaire completion.
259
Downloaded from ssc.sagepub.com at Aalto University on March 29, 2016
260 Social Science Computer Review
Screen designs have several interactive options that can be implemented in web questionnaires: the automatic jumps between questions which allow the omission of inapplicable
questions based on previous responses, the display of missing data messages alerting the
respondent to unanswered questions, the presentation of instructions or explanations based
on answers given by the respondent to previous questions, or the adaptation of question text
according to answers given to previous questions.
Most studies on interactivity focus on its effect on data quality (e.g., Conrad, Couper,
Tourangeau, & Peytchev, 2006; Conrad et al., 2005); little research has been conducted
on the impact of automatic interactive questionnaires on response rates. We have found
three studies on the effect of interactivity on response rates, namely the using of forcedresponse procedures. These procedures consist of building the questionnaire in such a way
that respondents are forced to provide an answer to each question before being allowed to
answer any subsequent ones. In these cases, a ‘‘next’’ button is placed on each screen that
prevents the respondent proceeding if a valid answer has not been given to the questions
presented on the screen. A warning message is displayed to alert the respondents to
unanswered questions.
DeRouvray and Couper (2002) examined the effect of alerting respondents to unanswered questions in a comparison of various strategies aimed at reducing ‘‘no-opinion’’
responses. They found that item nonresponse went down when the alerting message was
displayed and was even lower if the options ‘‘don’t know’’/’’no opinion’’ were not provided. More recently, Stieger, Reips, and Voracek (2007) conducted an experiment with
university students to assess the effect of forced response on dropout rates. The same questionnaire was compared between two design mode groups—the forced-response group, that
is, those who tried to skip a question without answering it and therefore saw an error page
and the nonforced-response group, that is, those who never saw the error page because they
answered all the items without trying to skip any item. The authors found a dropout rate of
64% in the forced-response group as opposed to only 19% in the nonforced-response group;
moreover, when analyzing the moment/screen of the questionnaire where dropouts
occurred, it was clear that the dropout rate was higher for the forced-response group in all
the screens of the survey.
The research by Ganassali (2008) compared the interactive with the noninteractive
version of the same questionnaire on a survey about young people’s consumption patterns
and proved that the interactive version (in which an alerting message was displayed for the
open-ended question in case of no answer) produced a lower item nonresponse rate than the
noninteractive version.
Question/Response Format
Questionnaires can contain closed-ended questions or open-ended questions.
Closed-ended questions require respondents to select the answers that best fit their ideas
or characteristics from a set of response options. Open-ended questions allow respondents
to formulate the answers in their own words rather than selecting them from a list of answer
categories. Closed-ended questions are usually easier to respond and are likely to reduce the
completion time of the questionnaire. Open-ended questions require more thinking and
260
Downloaded from ssc.sagepub.com at Aalto University on March 29, 2016
Using Questionnaire Design to Fight Nonresponse Bias / Vicente, Reis
261
writing than closed-ended questions and are therefore more burdensome for the respondent
(Groves et al., 2004).
In web questionnaires, open-ended questions require that a text field or a text area be supplied on the screen/page for respondents to insert his or her answer, while a decision has to
be made with closed-ended questions on how to present the answer categories, for example,
radio buttons (all answering options are shown; the respondent selects his or her options by
a mouse click on the corresponding ‘‘button’’) or drop-down boxes (the list of answer
options is hidden and it is only shown after a mouse click; the respondent selects his or her
options by pointing and clicking the mouse on the intended option).
The meta-analysis by Lozar Manfreda and Vehovar (2002) on determinants of nonresponse in web surveys proved question type has a significant effect on dropout rates; higher
dropout rates were likely to be found in surveys that included open-ended and/or difficultto-answer questions. Jackob and Zerback (2006) report the results of a survey on media trust
conducted with students where a first sample of students was presented with an open-ended
question format and a second sample in which the same question was presented as closedended with a list of response options. The results showed an item nonresponse rate of 11%
in the open-ended sample against 16% in the closed-ended sample. Although this evidence
is not striking, it indicates that open-ended questions are not necessarily worse.
Turning to the impact of response format on response rates, Couper et al. (2001) studied
the impact of radio buttons versus entry boxes on response rates in a survey of university
students. In the radio buttons version, a ‘‘don’t know’’ and ‘‘no answer’’ response had to
be explicitly selected, whereas respondents had the option of simply leaving an entry box
blank in the entry box version. The outcomes revealed higher item nonresponse in the entry
box version.
Couper, Tourangeau, Conrad, and Singer (2006) conducted an experiment to explore the
utility of a particular input device in web surveys—visual analog scales—comparing it with
radio button input and numeric entry in a text box on a series of bipolar questions eliciting
views on genetic versus environmental causes of various behaviors. A visual analog scale is
a kind of graphic rating scale on which the respondent directly marks his or her position on a
‘‘continuous’’ line; this line represents the scale.
The input type was found to have a significant effect on the overall completion rate, with
the radio button version showing the higher completion rate. Regarding missing data, none
of the versions had an explicit ‘‘don’t know’’ option and respondents could skip items without answering them; the radio button version had the lowest item nonresponse rate followed
by the text box version and then by the visual scale analog version. The item nonresponse
rate in the latter version was more than double that of the other two versions. Although this
experiment was designed to evaluate the effect of the visual scale analog, it also allowed
comparisons to be made of the other two formats, leading to the conclusion that the performance of the radio buttons version was better in terms of response rates.
The impact of response format was also evaluated in the experiment by Healey (2007).
Using a survey on the general population of New Zealand, the author compared two
response formats, namely radio buttons versus drop-downs boxes, by randomly assigning
one half of the sample to one treatment and the second half of the sample to the other. Three
aspects of response rates were evaluated: completion rates, item nonresponse at survey
level, and item nonresponse at question level. Although no significant difference was found
261
Downloaded from ssc.sagepub.com at Aalto University on March 29, 2016
262 Social Science Computer Review
between the two treatments regarding completion rates and item nonresponse at survey
level, at question level there was evidence that drop-down boxes led to slightly higher item
nonresponse than radio buttons.
Table 1 outlines the major findings from the literature revision. It presents the design features that can be considered when designing the questionnaire, where the major impact of
each design feature is likely to be felt—dropout rate, item nonresponse rate, or overall completion rate—and the literature sources cited in this review on each design feature.
Web Questionnaires: Which Design Is Best?
The purpose of this article was to classify web questionnaire features and discuss the
impact of questionnaire design on nonresponse by gathering the most relevant knowledge
available about web survey methodology and web survey nonresponse. The results of the
empirical research are not always consensual about the effect of a particular design practice
on response rates. Several factors can account for this. First, the target population is not the
same across experiments and the level of expertise of the target population dealing with the
Internet influences the contact process between the respondent and the questionnaire and
thus the outcomes in terms of response rates. Second, the lack of a standard definition for
some concepts makes it difficult to compare the various experiments, for example, there is
no consensus on how the length of a questionnaire should be measured and therefore it is
hard to establish the length of the questionnaire which detracts respondents from cooperating. Finally, there are factors that are difficult for the researcher to control but that are capable of influencing the contact process between respondent and questionnaires, for example,
the speed of the Internet connection of the respondents’ computer.
This review of the literature on web questionnaire design and survey nonresponse presents a picture of what is known about the effects of questionnaire design on web surveys
response rates. In short:
Screen designs tend to yield a lower item nonresponse rate than scroll designs. Moreover,
screen designs allow for a better control of the response process, for example, in the case of
dropouts, the question where the respondent abandoned the questionnaire can be located
thus enabling the analysis of causes behind this.
Surveys with lengthy questionnaires tend to have lower overall completion rates mainly
because the increased burden of long questionnaires tends to increase dropout rates.
The respondents’ perception of the burden is more important than the burden itself. The
inclusion of a progress indicator can give the respondent some feedback as to the point
of the questionnaire where he or she stands. If the respondent does not see his or her progress moving significantly forward, the likelihood of dropout increases. This is more likely
to happen in longer questionnaires. Therefore, an intermittent progress indicator instead of
one that is permanently on can work better to keep respondents’ motivation higher and thus
decrease dropouts.
Illustration enhancements of the questionnaires can make the questionnaire more attractive
and pleasant thus contributing to higher completion rates. But visual enhancements can
also cause premature abandon of the survey if the questionnaire becomes difficult to navigate or download due to software or hardware compatibility problems. Although these
262
Downloaded from ssc.sagepub.com at Aalto University on March 29, 2016
263
Downloaded from ssc.sagepub.com at Aalto University on March 29, 2016
f. Include versus not include forced
response procedures
g. Closed-ended versus open-ended
questions
h. Radio buttons versus other response
format
d. Include versus not include progress
information
e. ‘‘Fancy’’ versus ‘‘plain’’ design
b. Embedded versus attached
questionnaire
c. Long versus short questionnaires
a. Scroll versus screens designs
Design Feature
Overall
Item
Nonresponse Completion
Rate
Rate
Dropout
Rate
Couper et al. (2001); Couper, Tourangeau, and Peytchev (2006); Healey
(2007)
Lozar Manfreda and Vehovar (2002); Deutskens, Ruyter, Wetzels, and
Oosterveld (2004); Ganassali (2008)
Couper, Traugott, and Lamias (2001); Crawford, Couper, and Lamias
(2001); Conrad, Couper, Tourangeau, and Galesic (2005)
Dillman, Tortora, Conradt, and Bowker (1998); Lozar Manfreda et al.
(2002); Deutskens et al. (2004)
DeRouvray and Couper (2002); Stieger, Reips, and Voracek (2007);
Ganassali (2008)
Lozar Manfreda and Vehovar (2002); Jackob and Zerback (2006)
Lozar Manfreda, Batagelj, and Vehovar (2002); Tourangeau, Couper,
Galesic, and Givens (2004); Peytchev, Couper, McCabe, and Crawford
(2006)
Dommeyer and Moriarty (2000)
Literature Sources
Table 1
Design Features for Web Survey Questionnaires, (Non)Response Rates, and Literature Sources
264 Social Science Computer Review
technical problems can be expected to decline in the future as computer and Internet connection speed increases, for now the empirical evidence on the use of graphics in web questionnaires points to a moderate use of graphically complex designs in order not to detract
respondents from completing the questionnaire.
The radio button format instead of drop-down boxes or text entry for closed-ended questions seems to work in favor of lower item nonresponse.
When designing a web questionnaire, decisions must be taken on the six features described
here—general structure, length, disclosure of survey progress, visual presentation, interactivity, and question/response format. Ideally, these features are combined to simultaneously
guarantee low dropouts, low item nonresponse, and high overall completion rate. The literature review clearly demonstrates that it is not easy to build a set of methodological rules
that grants good results in all types of (non)response rates—dropouts, item nonresponse,
and overall completion rate—in all web surveys. To fulfill the objective on one aspect, it
is not uncommon for other aspects to be put at risk. In such cases, researchers must try
to find the ‘‘best’’ mix of design features to counterbalance the different aspects.
For example, the combination ‘‘screen design þ forced response procedure þ radio
buttons’’ may have a favorable effect on reducing item nonresponse because according
to the literature all these features affect item response. Respondents are less likely to unintentionally skip a question in screen designs than in scroll designs because the respondent
sees few questions in each screen in screen designs; in contrast, the whole questionnaire can
be seen by using the scroll bar in scroll designs. As a result, item nonresponse rate is likely
to be reduced in screen designs. However, if a scroll design is necessary for some reason
(e.g., cost or software constraints), it is prudent to implement a procedure to alert for unanswered questions to control the risk of high item nonresponse. The inclusion of a procedure
to force response to all questions may also be pertinent in surveys where a complete data set
is crucial (e.g., when required by the data analysis procedures that are to be used) to
guarantee a zero item nonresponse rate, whether a screen or a scroll design is chosen.
A combination of the type ‘‘moderate length þ progress information þ moderately fancy
design’’ may act favorably to control for dropouts because the literature review presents the
length, the disclosure of survey progress, and visual presentation as factors with a
significant impact on dropout rate. A ‘‘moderately long’’ questionnaire is more likely to
be completed till the end than a longer one. However, if a long questionnaire is unavoidable
(e.g., due to the objectives of the survey), something must be done to counterbalance the
negative effect that size is expected to have on the completion rate—the insertion of a
‘‘motivating’’ progress indicator, namely one that shows real progress every time the
respondent looks at it, is likely to help keep the respondent till the end of the questionnaire;
more recent research suggests that an intermittent display of progress information seems to
be the best option.
The literature review does not provide a direct answer for the best combination of
‘‘general structure, length, disclosure of survey progress, visual presentation, interactivity,
and question/response format’’ to enhance response rates. On one hand, the effect of some
design features on response rates is not yet conclusive. On the other, further research is
required for a full understanding of the effect of some specific combinations of design
features on response rates as most experimental research designs on questionnaire features
264
Downloaded from ssc.sagepub.com at Aalto University on March 29, 2016
Using Questionnaire Design to Fight Nonresponse Bias / Vicente, Reis
265
focus on the effect of one of the key factors—general structure, length, disclosure of survey
progress, visual presentation, interactivity, or question/response format—and not on the
entire sequence of the six factors. Nevertheless, taking into consideration the specific
context in which web surveys are made—an electronic environment, with some degree
of technical complexity, and where there is no interviewer to help the respondent to fill
in the questionnaire—it can be emphasized that the web surveyor should note that it is more
important for a questionnaire to be easy and motivating for respondents to understand and
complete than to have one that it is easy to administer.
Web surveys research is a rapidly changing field so future updates of best practices to
lower nonresponse will undoubtedly accommodate ongoing developments. Moreover, the
shortcomings found in the literature and referred above certainly make web surveys
research a rich area for future investigation.
References
Carroll, S. (1994). Questionnaire design affects response rate. Marketing News, 28, 14.
Conrad, F., Couper, M., Tourangeau, T., & Galesic, M. (2005). Interactive feedback can improve quality of
responses in web surveys. Paper presented at the 60th American Association for Public Opinion Research
Annual Conference, Phoenix, AZ.
Conrad, F., Couper, M., Tourangeau, T., & Peytchev, A. (2006). Use and non-use of clarification features in web
surveys. Journal of Official Statistics, 22, 245-269.
Cook, C., Heath, F., & Thompson, R. L. (2000). A meta-analysis of response rates in web- or Internet-based
surveys. Educational and Psychological Measurement, 60, 821-836.
Couper, M. (2000). Web surveys: A review of issues and approaches. Public Opinion Quarterly, 64, 464-494.
Couper, M. (2001). Web surveys: The questionnaire design challenge. Proceedings of the 53rd Session of the
International Statistical Institute, Seoul, Korea.
Couper, M., Tourangeau, R., Conrad, F., & Singer, E. (2006). Evaluating the effectiveness of visual analog
scales: A web experiment. Social Science Computer Review, 24, 227-245.
Couper, M., Tourangeau, R., & Kenyon, K. (2004). Picture this! Exploring visual effects in web surveys. Public
Opinion Quarterly, 68, 255-266.
Couper, M., Traugott, M., & Lamias, M. (2001). Web survey design and administration. Public Opinion
Quarterly, 65, 230-253.
Crawford, S., Couper, M., & Lamias, M. (2001). Web surveys: Perceptions of burden. Social Science Computer
Review, 19, 146-162.
DeRouvray, C., & Couper, M. (2002). Designing a strategy for reducing ‘‘no opinion’’ responses in web-based
surveys. Social Science Computer Review, 20, 3-9.
Deutskens, E., Ruyter, K., Wetzels, M., & Oosterveld, P. (2004). Response rate and response quality of Internetbased surveys: An experimental study. Marketing Letters, 15, 21-36.
Dillman, D., Sinclair, M., & Clark, J. (1993). Effects of questionnaire length, respondent-friendly design, and a
difficult question on response rates for occupant-addressed census mail surveys. Public Opinion Quarterly,
57, 289-304.
Dillman, D., Tortora, R., Conradt, J., & Bowker, D. (1998). Influence of the plain vs fancy design on response
rates for web surveys. Proceedings of the Survey Research Methods Section, American Statistical
Association.
Dommeyer, C., & Moriarty, E. (2000). Comparing two forms of an e-mail survey: Embedded vs. attached.
Journal of the Market Research Society, 42, 39-53.
European Commission. (2008). Eurobarometer 293, European Commission, Brussels.
Galesic, M. (2005). Effects of questionnaire length on quality of responses in web surveys. Paper presented at
the ESF Workshop, Dubrovnik, Croatia.
265
Downloaded from ssc.sagepub.com at Aalto University on March 29, 2016
266 Social Science Computer Review
Galesic, M. (2006). Dropouts on the web: Effects of interest and burden experienced during an online survey.
Journal of Official Statistics, 22, 313-328.
Ganassali, S. (2008). The influence of the design of web survey questionnaires on the quality of responses.
Survey Research Methods, 2, 21-32.
Groves, R., Fowler, F., Jr., Couper, M., Lepkowski, J., Singer, E., & Tourangeau, R. (2004). Survey methodology. New York: John Wiley and Sons.
Haraldsen, G., Stalnacke, M., & Blackwell, K. (1999). Empirical contributions to a general survey response
model. Paper presented at the International Conference on Survey Nonresponse, Portland, OR.
Healey, B. (2007). Drop down and scroll mice: The effect of response option format and input mechanism
employed on data quality in web surveys. Social Science Computer Review, 25, 111-128.
Heberlein, T., & Baumgartner, R. (1975). Factors affecting response rates to mailed questionnaires: A quantitative analysis of the published literature. American Sociological Review, 43, 447-462.
Hogg, A., & Miller, J. (2003). Watch out for dropouts. Quirk’s Marketing Research Review, 1137. Retrieved from
http://www.websm.org/uploadi/editor/Watch%20out%20for%20dropouts.doc.
Retrieved date 19th March 2010.
Hudson, D., Seah, L. -H., Hite, D., & Haab, T. (2004). Telephone presurveys, self-selection and non-response
bias to mail and Internet surveys in economic research. Applied Economic Letters, 11, 237-240.
Jackob, N., & Zerback, T. (2006). Improving quality by lowering non-response: A guideline for online surveys.
Paper presented at the WAPOR Seminar ‘‘Quality Criteria in Survey Research VI,’’ Cadenabbia, Italy.
Jansen, J. (1985). Effect of questionnaire layout and size and issue-involvement on response rates in mail
surveys. Perceptual and Motor Skills, 61, 139-142.
Lozar Manfreda, K., Batagelj, Z., & Vehovar, V. (2002). Design of web survey questionnaires: Three basic
experiments. Journal of Computer-Mediated Communication, 7, 3. Retrieved from http://www.
websm.org/uploadi/editor/Lozar_2002_Design.doc. Retrieved date 19th March 2010.
Lozar Manfreda, K., Berzelak, J., Vehovar, V., Bosnjak, M., & Haas, I. (2008). Web surveys versus other survey
modes: A meta-analysis comparing response rates. International Journal of Market Research, 50, 79-104.
Lozar Manfreda, K., Bosnjak, M., Haas, I., & Vehovar, V. (2005). Web survey response rates compared to other
modes—A meta analysis. Paper presented at the 60th American Association for Public Opinion Research
Annual Conference, Miami Beach, FL.
Lozar Manfreda, K., & Vehovar, V. (2002). Survey design features influencing response rates in web surveys.
Paper presented at the International Conference on Improving Surveys, Copenhagen, Denmark.
Peytchev, A., Couper, M., McCabe, S., & Crawford, S. (2006). Web survey design: Paging versus scrolling.
Public Opinion Quarterly, 70, 596-607.
Redline, C., & Dillman, D. (1999). The influence of auxiliary, symbolic, numeric and verbal language on
navigational compliance in self-administered questionnaires. U.S. Census Bureau. Retrieved from
http://survey.sesrc.wsu.edu/dillman/papers/Auxiliary,Symbolic,Numeric%
20paper--with%20Cleo.pdf. Retrieved date 19th March 2010.
Stieger, S., Reips, U.-D., & Voracek, M. (2007). Forced-response in online surveys: Bias from reactance and an
increase in sex-specific dropout. Journal of the American Society for Information Science and Technology,
58, 1653-1660.
Tourangeau, R., Couper, M., Galesic, M., & Givens, J. (2004). A comparison of two web-based surveys:
Static versus dynamic versions of the NAMCS questionnaire. RC33 6th International Conference on
Social Science Methodology: Recent Developments and Applications in Social Research Methodology,
Amsterdam, Holland.
Umbach, P. (2004). Web surveys: Best practices. New Directions for Institutional Research, 121, 23-38.
Yammarino, F., Skinner, S., & Childers, T. (1991). Understanding mail survey response behavior: A meta-analysis.
Public Opinion Quarterly, 55, 613-639.
Yu, J., & Cooper, H. (1983). A Quantitative review of research design effects on response rates to questionnaires. Journal of Marketing Research, 20, 36-44.
266
Downloaded from ssc.sagepub.com at Aalto University on March 29, 2016
Using Questionnaire Design to Fight Nonresponse Bias / Vicente, Reis
267
Paula Vicente is an assistant professor in the quantitative methods department, ISCTE-Lisbon University Institute, Portugal. Her research interests are focused on survey methodology, Internet surveys, and mixed-mode
effects. She may be reached at [email protected].
Elizabeth Reis is a full professor in the quantitative methods department, ISCTE-Lisbon University Institute,
Portugal. She is currently vice president of GIESTA, a multidisciplinary centre at ISCTE that combines
methodological and substantive research in survey methodology and statistics. She may be reached at
[email protected].
267
Downloaded from ssc.sagepub.com at Aalto University on March 29, 2016