Professional Documents
Culture Documents
Humanizing Self-Administered Surveys: Experiments On Social Presence in Web and IVR Surveys
Humanizing Self-Administered Surveys: Experiments On Social Presence in Web and IVR Surveys
Humanizing Self-Administered Surveys: Experiments On Social Presence in Web and IVR Surveys
www.elsevier.com/locate/comphumbeh
Abstract
Social interface theory has had widespread influence within the field of human–computer
interaction. The basic thesis is that humanizing cues in a computer interface can engender
responses from users similar to those produced by interactions between humans. These
humanizing cues often confer human characteristics on the interface (such as gender) or sug-
gest that the interface is an agent actively interacting with the respondent. In contrast, the
survey interviewing literature suggests that computer administration of surveys on highly
sensitive topics reduces or eliminates social desirability effects, even when such humanizing
features as recorded human voices are used. In attempting to reconcile these apparently con-
tradictory findings, we varied features of the interface in two Web surveys and a telephone
survey. In the first Web experiment, we presented an image of (1) a male researcher, (2) a
female researcher, or (3) the study logo at several points throughout the questionnaire. This
experiment also varied the extent of personal feedback provided to the respondent. The sec-
ond Web study compared three versions of the survey: (1) One that included a photograph of
a female researcher and text messages from her; (2) another version that included only the text
messages; and (3) a final version that included neither the picture nor the personalizing mes-
sages. Finally, we carried out a telephone study using a method—interactive voice response
(IVR)—in which the computer plays a recording of the questions over the telephone and
respondents indicate their answers by pressing keys on the telephone handset. The IVR study
varied the voice that administered the questions. All three surveys used questionnaires that
included sensitive questions about sexual behavior and illicit drug use and questions on gen-
der-related attitudes.
§
An earlier version of this paper was originally presented at the CHI ’01 Conference on Human Fac-
tors in Computing Systems in Seattle, Washington in April, 2001.
* Corresponding author.
E-mail address: rtourangeau@survey.umd.edu (R. Tourangeau).
0747-5632/03/$ - see front matter # 2002 Elsevier Science Ltd. All rights reserved.
PII: S0747-5632(02)00032-8
2 R. Tourangeau et al. / Computers in Human Behavior 19 (2003) 1–24
We find limited support for the social interface hypothesis. There are consistent, though
small, effects of the ‘‘gender’’ of the interface on reported gender attitudes, but few effects
on socially desirable responding. We propose several possible reasons for the contradictory
evidence on social interfaces.
# 2002 Elsevier Science Ltd. All rights reserved.
Keywords: Social interfaces; Web surveys; IVR surveys; Social desirability
1. Introduction
Two distinct research traditions have examined how features of the interfaces of
computerized systems can alter the character of the interaction and thereby affect
the responses people give to the system. One tradition—social interface theory
(Dryer, 1999; Kiesler & Sproull, 1997; Reeves & Nass, 1997)—is attracting con-
siderable interest within the field of human–computer interaction. Much of the
support for social interface theory comes from laboratory-based studies. The other
tradition involves the systematic comparison of different modes of survey data col-
lection. Recent mode comparisons have focused on two of the newest modes of data
collection—Web surveys and a method of telephone data collection variously refer-
red to as telephone audio computer-assisted self-interviewing (T-ACASI), touchtone
data entry (TDE), and interactive voice response (IVR). This second tradition relies
on field studies in which the same questions are administered under the different
methods of data collection. Unfortunately, the two traditions seem to come to dif-
ferent conclusions about the impact of incidental features of interfaces.
A growing body of laboratory research suggests that relatively subtle cues (such as
‘‘gendered’’ text or simple inanimate line drawings of a face) in a computer interface
can evoke reactions similar to those produced by a human, including social desir-
ability effects. Nass, Moon, and Green (1997), for example, conclude that the ten-
dency to stereotype by gender can be triggered by such minimal cues as the voice on
a computer. Based on the results of a series of experiments that varied a number of
cues in computer tutoring and other tasks, Nass and his colleagues (Fogg & Nass,
1997; Nass, Fogg, & Moon, 1996; Nass, Moon, & Carney, 1999; Nass et al., 1997)
argue that people treat computer interfaces as though they were people—that is, as
social actors rather than inanimate tools (see also Couper, 1998).
Additional support for the hypothesis that a computer interface can function as a
virtual human presence comes from a study by Walker, Sproull, and Subramani
(1994). They administered questionnaires to people using either a text display or one
of two talking-face displays to ask the questions. Those interacting with a talking-
face display spent more time, made fewer mistakes, and wrote more comments than
did people interacting with the text display. However, people who interacted with an
expressive face liked the face and the experience less than those who interacted with
an inexpressive face. In a subsequent experiment, Sproull, Subramani, Kiesler,
R. Tourangeau et al. / Computers in Human Behavior 19 (2003) 1–24 3
Walker, and Waters (1996) varied the expression of a talking face on a computer-
administered career counseling interview; one face was stern, the other pleasant. The
faces were computer-generated images with animated mouths. They found that:
Social interface researchers have used a range of humanizing cues to foster a sense
of virtual presence. Some studies use cues that confer human-like qualities on the
interface, such as gender or specific personality traits (e.g. pleasantness). Other
studies have tried to create the impression that the interface is actively interacting
with the user. Our studies examined both types of humanizing cues, varying both the
gender associated with the interfaces and the degree of apparent agency.
If the social interface theory is correct, it has important implications for survey
practice. Within the survey industry, there is an increasing trend toward the use of
computer-assisted interviewing, and especially the use of the World Wide Web, for
administration of surveys (Couper, 2001; Couper and Nicholls, 1998). In addition,
more and more surveys include sensitive questions (on sexual behavior, drug use,
etc.), raising concerns about social desirability effects and interviewer influences on
the answers. Partly in response to these developments, more surveys have come to
rely on computer-assisted self-interviewing (CASI) methods, in which the respon-
dent interacts directly with the computer to answer questions. The most recent
manifestation of this trend is the development of audio-CASI, in which the respon-
dent listens to a recording of an interviewer reading the questions and enters the
responses into the computer. A number of studies have compared CASI and audio-
CASI to alternative methods of data collection in field-based experiments. The gen-
eral finding in the survey literature is that self-administration increases reports about
sensitive behaviors and that computerized forms of self-administration—that is,
CASI methods (including audio-CASI)—maintain or increase these advantages over
interviewer-administered methods (Tourangeau & Smith, 1998). Some have gone so
far as to argue that voice does not matter when asking questions about sexual
behavior (e.g. Turner, Forsyth, O’Reilly, Cooley, Smith, Rogers, & Miller, 1998;
Turner, Ku, Rogers, Lindberg, Pleck, & Sonenstein, 1998), although these claims
have yet to be tested empirically.
The survey results appear to contradict the findings of the social interface
researchers. If subtle humanizing cues do indeed influence the behavior of computer
users, we would fully expect the gender of the voice to affect the answers given to
survey questions on topics such as gender attitudes and sexual behavior. Further,
the increasing use of multimedia tools on the Web make it possible to add a variety
of humanizing visual and aural cues to Web surveys; these cues could reduce or even
4 R. Tourangeau et al. / Computers in Human Behavior 19 (2003) 1–24
completely eliminate the gains from self-administration for items eliciting sensitive
information. It is thus important to explore the apparent contradiction between the
social interface and survey methods work and to try to bring these two strands of
research together.
There are numerous differences between the two research traditions that could
account for the discrepant results. For one, virtually all of the social interface
research has been conducted with students as volunteer subjects. In contrast, the
survey-based findings are from probability samples of broader populations (e.g.
teenage males, women between 15 and 44 years old, the adult US population). In the
former, the number of subjects is typically measured in tens or scores while, in
the latter, sample sizes go up to the thousands.
The measurement settings also differ sharply. The social interface work is typically
done in a laboratory setting, free from distractions and from threats to privacy. Most
of the CASI surveys are conducted in the respondent’s home with an interviewer
present—and sometimes with other family members present as well. The more con-
trolled environment of the laboratory may well focus subjects’ attention on the
experimental manipulation more than would be true in an uncontrolled real-world
setting where the respondents are not aware they are in an experiment and where
outside distractions may draw much of their attention. Furthermore, the measure-
ment devices differ considerably between the two approaches. The social interface
experiments often use subjects’ performance on a computer task as the dependent
measure. When questionnaire measures are used in social interface studies, they are
typically measures designed to assess social desirability or impression management.
The findings from the survey world are based on measures of highly sensitive beha-
viors (e.g. abortions, number of sex partners, engagement in high risk sexual beha-
viors, illicit drug use, etc.); the items are often drawn from ongoing national surveys.
We are engaged in a program of research to explore the issue of the effect of
interface design and social interface features on survey responses. Here we report on
the results of two Web experiments and an experiment using IVR to investigate the
effect of interface features on social desirability effects and other response problems
in surveys. In all three of the studies we describe here, some versions of the survey
questionnaire included gender cues (conveyed by photographs or recorded voices),
allowing us to compare ‘‘male’’ and ‘‘female’’ versions of the questionnaire or gen-
dered with neutral versions. In addition, our first and third studies included cues
designed to create a sense of agency. In the first study, we varied whether or not the
interface gave tailored feedback to the respondents, based on their answers to earlier
questions; in the final study, we varied whether the recorded voices administering the
questions used first or third person language.
Our first study was designed to explore the effect of characteristics of the interface
on the responses obtained to survey questions administered via the World
Wide Web. We varied two dimensions in an attempt to increase the sense of social
R. Tourangeau et al. / Computers in Human Behavior 19 (2003) 1–24 5
2.1. Methods
We created six versions of the Web survey instrument that differed on the dimen-
sions of personalization and interaction. At several points in the questionnaire,
the personalized versions of the questionnaire displayed pictures of one of the
researchers (either Tourangeau or Steiger). A third version of the questionnaire
presented the logo for the study (‘‘Study of Attitudes and Lifestyles’’) instead of
either investigator’s picture. Along with the pictures, the program displayed relevant
statements from the investigator (Hi! My name is Roger Tourangeau. I’m one of the
investigators on this project. Thanks for taking part in my study.) There were, thus,
three levels of the personalization variable, involving the presentation of pictures
and messages from the male investigator, the female investigator, or only the pre-
sentation of the study logo.
We also varied the level of interaction in the Web survey. The high interaction
versions of the questionnaire addressed the respondent by name and used the first
person in introductions and transitional phrases (e.g. Thanks, [NAME]. Now I’d
like to ask you a few questions about the roles of men and women) and occasionally
echoed back to the respondents their earlier answers (According to your responses,
you exercise once daily). The low interaction versions used more impersonal lan-
guage (The next series of questions is about the roles of men and women) and gave less
tailored feedback (Thank you for this information). Examples of screens from each
condition designs are shown in Figs. 1–3.
We crossed the two experimental variables, resulting in a 32 experiment. We
randomly assigned respondents to one of the six cells in the design, with final sample
sizes shown in Table 1.
2.1.1. Questionnaire
The survey questionnaire contained a variety of items, most of them from existing
instruments:
Gender attitudes: eight items from Kane and Macauley’s (1993) study regarding
the roles of men and women (e.g. Thinking about men as a group, do you think men
have too much influence, about the right amount of influence, or too little influence in
society?);
Socially undesirable behaviors: three items on drinking and illicit drug use and
three on diet and exercise;
Socially desirable behaviors: two items on voting and church attendance;
Self-reported social desirability: sixteen items from the Marlowe–Crowne Social
Desirability Scale (Crowne & Marlowe, 1964) and the 20-item Impression Man-
agement scale from the Balanced Inventory of Desirable Responding (BIDR;
Paulhus, 1984), all of them using a true/false answer format;
6 R. Tourangeau et al. / Computers in Human Behavior 19 (2003) 1–24
Fig. 1.
Trust: three items on trust from the General Social Survey and National Election
Studies (e.g. Most people can be trusted);
Debriefing questions: nine items to assess social presence and evaluate the inter-
view experience (e.g. How much was this interview like an ordinary conversation?
How much was it like dealing with a machine?);
Demographic questions: eight items eliciting the respondent’s age, race, sex and so on.
Many of the items are known to be susceptible to various sorts of response effects.
For example, Kane and Macauley showed that responses to the gender items partly
reflects the sex of the interviewer who administers them; on the average, women
interviewers elicit more pro-feminist responses than men interviewers do. We inclu-
ded the gender attitude items to see whether our attempt to personalize the interface
produced similar effects—that is, more pro-feminist responses with the ‘‘female’’
than with the ‘‘male’’ interface. We included the items on diet, exercise, drinking,
drug use, voting, and attendance at church to test the hypothesis that humanizing
the interface (both by personalizing it and by making it more interactive) would
increase the number of socially desirable responses and decrease the number of
socially undesirable responses. The items we chose are typical of sensitive items that
appear in surveys, and they show systematic reporting problems.
The social desirability and impression management items have been used for
similar purposes to measure socially desirable responding in several studies on social
presence (e.g. Moon, 1998; Sproull et al., 1996), and we included them in our studies
for the sake of comparability. We included the trust items to see whether the
experimental variables had the most impact on those who are low on trust (as
Aquilino & LoSciuto, 1990, found). We included the demographic items to assess
R. Tourangeau et al. / Computers in Human Behavior 19 (2003) 1–24 7
Fig. 2.
Fig. 3.
A total of 3047 sample members completed the questionnaire, for a response rate
of approximately 20%. Less than 1% of the initial e-mail invitations bounced back
as invalid addresses. Another 434 persons (3% of the sample) began the survey but
broke off without finishing it. We focus here on the respondents who completed the
survey. The number of completed cases per cell is shown in Table 1.
The number of cases we obtained far exceeds that for most of the experimental
studies on social interfaces (which typically include 10–20 subjects per cell). Thus,
the study should have ample statistical power to detect the effects of the manipu-
lations. Furthermore, the respondents to our survey represent a much more diverse
group than is typically found in laboratory-based experiments. However, we
acknowledge that neither the initial invitees nor the final respondents are repre-
sentative of any larger population. Our focus is on detecting differences across
experimental manipulations rather than on making inference to some broader
population.
2.2. Results
Table 1
Number of subjects per cell
Low High
Table 2
Scale means, by condition (standard errors in parentheses)
Table 3
Percentages reported selected behaviors, by condition
Our second Web study was designed as a partial replication of the first. Given
the general lack of effects in the first study, we were concerned that the feedback
12 R. Tourangeau et al. / Computers in Human Behavior 19 (2003) 1–24
in the low interaction version may have been too mechanical; for instance,
when the respondents’ answers were fed back to them, the answers were in par-
entheses. In addition, the BIDR items were administered as dichotomous true/
false items (to be consistent with the Marlowe–Crowne items) rather than as
seven-point scales scored as dichotomies after the fact (following the procedure
outlined by Paulhus, 1984). We refined the design to correct these potential prob-
lems and administered the revised experimental treatments to a new sample of
respondents.
3.1. Method
We collapsed the study into a one-way design with three versions of the Web
questionnaire: (1) One version displayed the female researcher’s picture and pro-
vided feedback to the respondent; (2) the second version provided the same feedback
as the first version but displayed the study logo; and (3) the final version displayed
the logo but provided no feedback. In addition, we changed the format of the feed-
back text so it would appear more natural. We also experimentally compared the
two-point version of the BIDR items with the seven-point version. This resulted in a
32 experiment with three levels of the social presence manipulation crossed with
two versions of the BIDR items.
3.1.2. Questionnaire
Aside from the change in the experimental variables noted above, the ques-
tionnaire we used was very similar to the one used in the first Web study. The
two-point version of the impression management scale was presented as a series of
true/false items, with one point given for each response indicating an attempt to
create a positive impression. The items in the seven-point version were presented
as scales ranging from 1 (not true) to 7 (very true), with the items dichotomized
after the fact. One point was added for every 6 or 7 response, after reverse scoring
negatively-scaled items (following Paulhus, 1991). Thus, both versions of the
impression management scale have a minimum of 0 and a maximum of 20. To
shorten the questionnaire, we dropped the Marlowe-Crowne items from the sec-
ond study.
R. Tourangeau et al. / Computers in Human Behavior 19 (2003) 1–24 13
3.2. Results
Table 4
Scale means, by condition (standard errors in parentheses)
Table 5
Percentages reported selected behaviors, by condition
4. IVR experiment
1
Turner and his colleagues refer to this method as telephone audio computer assisted self-interviewing,
or T-ACASI (see Cooley, Miller, Gribble, & Turner, 2000; Gribble, Miller, Cooley, Catania, Pollack, &
Turner, 2000).
R. Tourangeau et al. / Computers in Human Behavior 19 (2003) 1–24 15
Table 6
Mean ratings of Web survey, by condition (ns in parentheses)
4.1. Methods
We compared six versions of an IVR interview. The different versions used dif-
ferent voices to record the questionnaires (a male voice, a female voice, or the male
voice reading the questions themselves and the female voice reading the answer
options) and also varied whether the questionnaires used personalizing language
(e.g. the first person). The respondents were selected via random digit dialing (RDD)
and were contacted by a live telephone interviewer, who switched those agreeing to
take part into the automated IVR interview. The IVR questionnaire included a
subset of the items from the Web studies, including the gender attitude items and
questions on a number of sensitive behaviors.
The IVR portion of the interview lasted an average of about 8 min for those who
made it to the end of the questionnaire.
4.1.3. Questionnaire
The questionnaire was a streamlined version of the ones used in our two Web
surveys. For approximately half the respondents, the IVR questionnaire included
the innocuous opening items (asking the respondent’s age, and educational attain-
ment). Those items were then followed by 12 items on gender roles, seven items on
sensitive behaviors (covering exercise, diet, illicit, drug use, voting, and church
attendance), four debriefing questions, and three demographic items.2 The remain-
ing questionnaires included the same questions, except that they omitted the intro-
ductory questions (which the live interviewer had already administered before the
switch to IVR).
4.2. Results
The most important findings from the experiment involve responses to the sensi-
tive items and to the gender role questions; our presentation focuses on these. We
also briefly discuss our findings regarding the breakoffs and responses to the
debriefing items.
2
The gender role items used in the Web studies were expanded and modified to suit the IVR mode.
The cocaine item was dropped from the IVR study because of concerns about the sensitivity of this item.
R. Tourangeau et al. / Computers in Human Behavior 19 (2003) 1–24 17
4.2.3. Breakoffs
The results presented so far exclude data from the cases who broke off without
completing the IVR questionnaire (though the results do not change appreciably if
we include the responses from those who quit part way through the interview).
Here we focus on the effect of the experimental variables on the breakoff rates.
Because a substantial proportion of the breakoffs occurred before the respondent
had answered a single question (225 out of 576 cases), we analyzed three related
18 R. Tourangeau et al. / Computers in Human Behavior 19 (2003) 1–24
Table 7
Mean sensitive admissions by condition (ns in parentheses)
Table 8
Proportion completing IVR questionnaire, by condition (ns in parentheses)
Across our three studies, our attempts to humanize the interfaces produced much
weaker effects than those reported by Nass, Sproull, or their colleagues, though our
efforts to give our questionnaires a human feel seem much more blatant than theirs
and our studies feature much larger samples. Table 10 summarizes the main features
of our studies and the major findings.
Table 9
Sample Size 3045 completes; 436 partials 2987 completes; 268 partials 1022 completes; 351 partials
Sensitive Admissions
Scale No effects No effects No effects
Individual items Effect of picture on reported voting Effects of version: Use of marijuana in past No effects
year, Lifetime cocaine use
Ratings of Interview
Chat No effect Effect of version Effects of voice, voice-personalization
Form No effect – –
Machine Effect of picture Effect of version Effect of voice
In the IVR study, an additional 225 cases were switched to the IVR system but hung up without answering a single question.
R. Tourangeau et al. / Computers in Human Behavior 19 (2003) 1–24 21
reports about sensitive behaviors, we find little reason for concern. In three tries
across two types of interface, we find only one significant effect of the interface, for
one version of the BIDR. Though significant, the effect is quite small, accounting
for less than 1% of the variance in the BIDR scores from that version; the gender of
the respondent accounts for approximately the same proportion of the variation in
BIDR scores in that study.
We find even less evidence for interface effects with scales constructed from sensi-
tive questions taken from real surveys. None of our three studies found an overall
effect for any of our experimental variables. Looking at five individual items in the
three studies, we find a total of three significant effects (out of a total of 35 possible
main effects and interactions) and none of these replicates across studies.
The results for reports about gender role attitudes are more consistent with the
social interface theory. We used the same items shown to be affected by the gender
of human interviewers (Kane & Macauley, 1993) and found that our attempt to
manipulate the gender of the interface had analogous results—when the researcher
displayed to the respondent was a woman, respondents gave more pro-feminist
answers. Only our final study, which varied the gender of the interviewer who
recorded the questions, failed to show the expected differences. Again, the size of
these gender-of-interface effects is quite small. In our first Web study, for example,
the correlation between scores on the gender attitude scale and the gender of the
researcher displayed was 0.050; again, this is far smaller than the impact of the sex of
the respondent (r=0.338).
It seems clear that respondents noticed our variations of the interface. In all three
studies, the experimental variables had a significant impact on the debriefing items.
Our efforts to humanize the interface did succeed in making the experience of com-
pleting the surveys seem more like talking to a friend and less like interacting with a
machine (see the bottom panel of Table 10). But this change in the feel of the inter-
view did not seem to have much impact on respondent’s answers to the substantive
questions. Nor did the experimental variables have any impact on the likelihood
that respondents would complete the questionnaire once they started it. The only
variable that affected breakoff rates involved giving a foretaste of the questions
before they began the automated questionnaire.
In short, we find at best limited support for the social interface theory. We are puz-
zled by the discrepancy between our results and those of the social interface research-
ers. We included some of the same measures used in the past work (e.g. the BIDR),
used strong humanizing cues, and fielded much larger samples than the earlier studies.
22 R. Tourangeau et al. / Computers in Human Behavior 19 (2003) 1–24
Several things could account for the discrepant findings. First, our experimental
manipulations may not have been blatant enough to produce the hypothesized
effects. We believe our manipulations to be at least as obvious as many of those in
the social interface research studies; these studies often use very subtle cues, such as
a label on a computer monitor (Moon, 1998) or the shape of a mouth on a compu-
ter-displayed face (Sproull et al., 1996; see also Kiesler & Sproull, 1997; Parise,
Kiesler, Sproull, & Waters, 1999). A related possibility is that our manipulations did
not foster the sense of the computer as a social agent. In the Web studies, the inves-
tigators were clearly third parties, who were not actively participating in the inter-
action between the respondents and Web questionnaire. In the social interface
studies, by contrast, the computer itself is presented as the active agent and the
language the computer uses and the line drawings and facial expressions it presents
are used to convey the characteristics of that agent. Our manipulations may thus be
examples of computer-mediated communication (CMC) rather than human–
computer interaction (HCI), and the consequences of the two may differ. We are
currently conducting a follow-up study to examine this possibility. However, Nass,
Isbister, and Lee (2001) argue that the same effects should occur under both CMC
and HCI conditions.
Another explanation of the discrepant results could be the differences in the sam-
ples. The social interface studies mostly use college students; ours encompassed a
broader range of the population. In our Web studies, we had big enough samples to
examine several variables that might interact with the experimental variables (whe-
ther the respondent was currently a student, his or her age, prior survey experience,
and level of trust). For example, we tested the hypotheses that students are more
sensitive to the characteristics of the interface and that respondents with prior
experience with Web surveys are less sensitive to them. None of these hypotheses
received much support—we did not find any significant interactions between these
individual differences variables and the experimental variables on the reporting of
sensitive information or gender attitudes.
A final possible explanation, which we could not test, is that the demand char-
acteristics of laboratory-based experiments yield results that are not replicated in
our field studies. In the lab-based studies, the subjects are typically students and they
are aware of being in an experiment. As a result, they may be especially sensitive to
small variations in interfaces. The same variations may pass unnoticed by most
survey respondents, who are typically unaware that they are in an experiment
(believing instead that the purpose of the study is to elicit their views on particular
topics) and who may be distracted as they answer. Any of these differences
may account for the failure of the social interface theory to replicate beyond the
laboratory.
Given the influence of the social interface perspective in human–computer inter-
action research and interface design, it is important to understand whether and
how the findings from this work translate to the real-world experiences of those
who interact with computers. In two such applications—Web surveys and an
automated telephone interview—we find little support for the social interface
hypothesis.
R. Tourangeau et al. / Computers in Human Behavior 19 (2003) 1–24 23
Acknowledgements
This research was supported by the National Science Foundation, Award No.
SES-9910882. We thank Robert Tortora and Margrethe Montgomery of the Gallup
Organization for their assistance in planning and carrying out this project.
References
Aquilino, W., & LoSciuto, L. (1990). Effect of interview mode on self-reported drug use. Public Opinion
Quarterly, 54, 362–395.
Athey, K. R., Coleman, J. E., Reitman, A. P., & Tang, J. (1960). Two experiments showing the effects of
the interviewer’s racial background on responses to questionnaires concerning racial issues. Journal of
Applied Psychology, 44, 562–566.
Cooley, P. C., Miller, H. G., Gribble, J. N., & Turner, C. F. (2000). Automating telephone surveys: Using
T-ACASI to obtain data on sensitive topics. Computers in Human Behavior, 16, 1–11.
Couper, M. P. (1998). Review of Byron Reeves and Clifford Nass, ‘‘The media equation: how people
treat computers, television, and new media like real people and places’’. Journal of Official Statistics, 13,
441–443.
Couper, M. P. (2001). Web surveys: a review of issues and approaches. Public Opinion Quarterly, 64,
464–494.
Couper, M. P., & Nicholls, W. L. II. (1998). The history and development of computer assisted survey
information collection. In M. P. Couper, R. P. Baker, J. Bethlehem, C. Z. F. Clark, J. Martin, et al.
(Eds.), Computer assisted survey information collection. New York: John Wiley.
Crowne, D., & Marlowe, D. (1964). The approval motive. New York: John Wiley.
Devine, P. G. (1989). Stereotypes and prejudice: their automatic and controlled components. Journal of
Personality and Social Psychology, 56, 5–18.
Dillman, D. (2000). Mail and Internet surveys: the tailored design method (2nd ed.). New York: John Wiley.
Dommeyer, C. J., & Ruggiero, L. A. (1996). The effects of a photograph on mail survey response. Mar-
keting Bulletin, 7, 51–57.
Dryer, D. C. (1999). Getting personal with computers: how to design personalities for agents. Applied
Artificial Intelligence, 13, 273–295.
Fogg, B. J., & Nass, C. (1997). Silicon sycophants: the effects of computers that flatter. International
Journal of Human–Computer Studies, 46, 551–561.
Gribble, J. N., Miller, H. G., Cooley, P. C., Catania, J. A., Pollack, L., & Turner, C. F. (2000). The
impact of T-ACASI interviewing on reported drug use among men who have sex with men. Substance
Use & Misuse, 35(6–8), 869–890.
Hutchinson, K. L., & Wegge, D. G. (1991). The effects of interviewer gender upon response in telephone
survey research. Journal of Social Behavior and Personality, 6, 573–584.
Kane, E. W., & Macauley, L. J. (1993). Interviewer gender and gender attitudes. Public Opinion Quarterly,
57, 1–28.
Kiesler, S., & Sproull, L. (1997). ‘‘Social’’ human–computer interaction. In B. Friedman (Ed.), Human
values and the design of computer technology. Stanford, CA: CSLI Press.
Moon, Y. (1998). Impression management in computer-based interviews: the effects of input modality,
output modality, and distance. Public Opinion Quarterly, 62, 610–622.
Moon, Y. (2000). Intimate exchanges: using computers to elicit self-disclosure from consumers. Journal of
Consumer Research, 26, 323–339.
Nass, C., Fogg, B. J., & Moon, Y. (1996). Can computers be teammates?. International Journal of
Human–Computer Studies, 45, 669–678.
Nass, C., Isbister, K., & Lee, E.-J. (2001). Truth is beauty: researching embodied conversational agents. In
J. Cassell, J. Sullivan, S. Prevost, & E. Churchill (Eds.), Embodied conversational agents (pp. 374–402).
Cambridge, MA: MIT Press.
24 R. Tourangeau et al. / Computers in Human Behavior 19 (2003) 1–24
Nass, C., Moon, Y., & Carney, P. (1999). Are people polite to computers? Responses to computer-based
interviewing systems. Journal of Applied Social Psychology, 29, 1093–1110.
Nass, C., Moon, Y., & Green, N. (1997). Are machines gender neutral? Gender-stereotypic responses to
computers with voices. Journal of Applied Social Psychology, 27, 864–876.
Parise, S., Kiesler, S., Sproull, L., & Waters, K. (1999). Cooperating with life-like interface agents. Com-
puters in Human Behavior, 15, 123–142.
Paulhus, D. L. (1984). Two-component models of socially desirable responding. Journal of Personality and
Social Psychology, 46, 598–609.
Paulhus, D. L. (1991). Measurement and control of response bias. In J. P. Robinson, P. R. Shaver, &
L. S. Wrightsman (Eds.), Measures of personality and social psychological attitudes (pp. 17–59). San
Diego: Academic Press.
Reeves, B., & Nass, C. (1997). The media equation: how people treat computers, television, and new media
like real people and places. Cambridge: CSLI and Cambridge University Press.
Schuman, H., & Converse, J. (1971). The effects of black and white interviewers on black responses in
1968. Public Opinion Quarterly, 35, 44–68.
Sproull, L., Subramani, M., Kiesler, S., Walker, J. H., & Waters, K. (1996). When the interface is a face.
Human–Computer Interaction, 11, 97–124.
Tourangeau, R., & Smith, T. (1998). Collecting sensitive information with different modes of data col-
lection. In M. P. Couper, R. P. Baker, J. Bethlehem, C. Z. F. Clark, J. Martin, et al. (Eds.), Computer
assisted survey information collection. New York: John Wiley.
Tourangeau, R., & Steiger, D. M. (1999). The impact of mode of data collection on response rates. Paper
presented at the International Conference on Nonresponse, Portland, Oregon, 28 October.
Turner, C. F., Forsyth, B. H., O’Reilly, J. M., Cooley, P. C., Smith, T. K., Rogers, S. M., & Miller, H. G.
(1998). Automated self-interviewing and the survey measurement of sensitive behaviors. In M. P. Couper
(Ed.), Computer assisted survey information collection. New York: John Wiley.
Turner, C. F., Ku, L., Rogers, S. M., Lindberg, L. D., Pleck, J. H., & Sonenstein, F. L. (1998). Adolescent
sexual behavior, drug use and violence: Increased reporting with computer survey technology. Science,
280, 867–873.
Walker, J., Sproull, L., & Subramani, R. (1994). Using a human face in an interface. In Proceedings of the
Conference on Human Factors in Computers (pp. 85–99). Boston: ACM Press.