Computers as Persuasive Social Actors
Computers as Persuasive Social Actors
Shortly after midnight, a resident of a small town in southern California called the police to report hearing a man inside a house nearby screaming Im going to kill you! Im going to kill you! Officers arrived on the scene and ordered the screaming man to come out of the house. The man stepped outside, wearing shorts and a Polo shirt. The officers found no victim inside the house. The man had been yelling at his computer. 1
No studies have shown exactly how computing products trigger social responses in humans, but as the opening anecdote demonstrates, at times people do respond to computers as though they were living beings. The most likely explanation is that social responses to certain types of computing systems are automatic and natural; human beings are hardwired to respond to cues in the environment, especially to things that seem alive in some way.2 At some level we cant control these social responses; they are instinctive rather than rational. When people perceive social presence, they naturally respond in social waysfeeling empathy or anger, or following social rules such as taking turns. Social cues from computing products are important to understand because they trigger such automatic responses in people. This chapter will explore the role of computing products as persuasive social actorsthe third corner in the functional triad (Figure 5.1). These products persuade by giving a variety of social cues that elicit social responses from their human users.
89
90
Persuasive Technology
Figure 5.1
Computers as social actors.
Tool
p .
Medium
Social actor
Creates relationship
A social actor can be persuasive by
Rewarding people with positive feedback Modeling a target behavior or attitude Providing social support
Figure 5.2
Nintendos Pocket Pikachu was designed to motivate users to be physically active.
The Tamagotchi craze in the late 1990s was perhaps the first dramatic demonstration of how interacting directly with a computer could be a social experience.3 People interacted with these virtual pets as though they were alive: They played with them, fed them, bathed them, and mourned them when they died. Tamagotchi was soon followed by Nintendos Pocket Pikachu (Figure 5.2) a digital pet designed to persuade. Like other digital pets, Pikachu required care and feeding, but with a twist: the device contained a pedometer that could register and record the owners movements. For the digital creature to thrive, its owner had to be physically active on a consistent basis. The owner had to walk, run, or jumpanything to activate the pedometer. Pocket Pikachu is a simple example of a computing device functioning as a persuasive social actor.
91
As shown in Table 5.1, I propose that five primary types of social cues cause people to make inferences about social presence in a computing product: physical, psychological, language, social dynamics, and social roles. The rest of this chapter will address these categories of social cues and explore their implications for persuasive technology.
92
Persuasive Technology
Figure 5.3
The Banana-Rama slot machine interface has onscreen characters that influence gamblers to keep playing.
93
Principle of Attractiveness
If someone is physically attractive, people tend to assume they also have a host of admirable qualities, such as intelligence and honesty.10 Similarly, physically attractive computing products are potentially A computing technology more persuasive than unattractive products. If an interface, device, or that is visually attractive to onscreen character is physically attractive (or cute, as the Banana-Rama target users is likely to be characters are), it may benefit from the halo effect; users may assume the more persuasive as well. product is also intelligent, capable, reliable, and credible.11 Attractiveness issues are prominent in one of the most ambitious and frustratingefforts in computing: creating human-like faces that interact with people in real time. Over the past few decades, researchers have taken important steps forward in making these faces more technically competent, with better facial expressions, voices, and lip movements. However, because of the programming challenges involved, many of these interactive faces are not very attractive, as shown in Figure 5.4. If interactive faces are to be used for persuasive purposes, such as counseling, training, or advertising, they need to be visually pleasing to be most effective.12 Two studies performed at the School of Management at Boston University reinforce the power of attractiveness. In the initial study, in 1996, researchers had participants play a two-person social dilemma game, in which participants could cooperate with an onscreen character or could choose to serve their own selfish purposes. In this initial study, the onscreen character (representative of the technology at the time) looked unattractive, even creepy, in my view. And it received rather low cooperation rates: just 32%. A couple of years later, the researchers repeated the study. Thanks to technology developments, in this second study, the onscreen character looked less artificial and, I would argue, more attractive and less creepy. This new and improved character garnered cooperation rates of a whopping 92%a figure
Figure 5.4
These interactive faces, created in the mid- to late 1990s, show that researchers have a long way to go in developing attractive, human-like interactive onscreen characters.
94
Persuasive Technology
that in this study was statistically indistinguishable from cooperation rates for interacting with real human beings. The researchers concluded that the mere appearance of a computer character is sufficient to change its social influence.13 Of course, people have different opinions about what is attractive. Evaluations vary from culture to culture, generation to generation, and individual to individual. (However, judging attractiveness is not entirely subjective; some elements of attractiveness, such as symmetry, are predictable).14 Because people have different views of whats attractive, designers need to understand the aesthetics of their target audiences when creating a persuasive technology product. The more visually attractive the product is to its target audience, the more likely it is to be persuasive. The designer might review the magazines the audience reads and music they listen to, observe the clothes they wear, determine what trends are popular with them, and search for other clues to what they might find attractive. With this information, the designer can create a product and test it with the target group. Computing technology also can convey social presence without using physical characters. We confirmed this in laboratory experiments at Stanford. We designed interface elements that were simple dialog boxesno onscreen characters, no computer voices, no artificial intelligence. Yet participants in the experiments responded to these simple computer interfaces as though they were responding to a human being. Among other things, they reported feeling better about themselves when they were praised by a computer and reciprocated favors from a computer. (These experiments are discussed in detail later in this chapter.)
95
Its not only computer novices who infer these psychological qualities; my research with experienced engineers showed that even the tech savvy treat computing products as though the products had preferences and personalities.16
96
Persuasive Technology
How do you create a dominant or submissive computer? For our study, we created a dominant computer interface by using bold, assertive typefaces for the text. Perhaps more important, we programmed the dominant computer to go first during the interaction and to make confident statements about what the user should do next. Finally, to really make sure people in the study would differentiate between the dominant and submissive computers, we added a confidence scale below each of these messages, indicating on a scale of 1 to 10 how confident the computer was about the suggestion it was offering. The dominant computer usually gave confidence ratings of 7, 8, and 9, while the submissive computer offered lower confidence ratings. For example, if a participant was randomly assigned to interact with the dominant computer, he or she would read the following on the screen while working with the computer on the Desert Survival Task: In the desert, the intense sunlight will clearly cause blindness by the second day. The sunglasses are absolutely important. This assertion from the computer was in a bold font, with a high confidence rating. In contrast, if a person was assigned to the submissive computer, he or she would read a similar statement, but it would be presented in this way: In the desert, it seems that the intense sunlight could possibly cause blindness by the second day. Without adequate vision, dont you think that survival might become more difficult? The sunglasses might be important. This statement was made in an italicized font, along with a low ranking on the confidence meter. To further reinforce the concept of submissiveness, the computer let the user make the first move in the survival task. Another step in preparing for this study was to find dominant and submissive people to serve as study participants. We asked potential participants to fill out personality assessments. Based on the completed assessments, we selected 48 people who were on the extreme ends of the continuumthe most dominant and the most submissive personalities. We found these participants by having almost 200 students take personality tests. Some, but not all, were engineers, but all participants had experience using computers. In the study, half of the participants were dominant types and half were submissive.
97
In conducting the study, we mixed and matched the dominant and submissive people with the dominant and submissive computers. In half the cases, participants worked with a computer that shared their personality type. In the other half, participants worked with a computer having the opposite personality. The information provided by all the computers was essentially the same. Only the computers style of interacting differed, as conveyed through text in dialog boxes: either the computer was dominant (The intense sunlight will clearly cause blindness), or it was submissive (It seems that the intense sunlight could possibly cause blindness). After we ran the experiment and analyzed the data, we found a clear result: participants preferred working with a computer they perceived to be similar to themselves in personality style. Dominant people preferred the dominant computer. Submissive people preferred the submissive computer. Specifically, when working with a computer perceived to be similar in personality, users judged the computer to be more competent and the interaction to be more satisfying and beneficial. In this study we didnt measure persuasion directly, but we did measure key predictors of persuasion, including likability and credibility.
Created dominant and submissive computer personalities Chose as participants people who were at extremes of dominant or submissive Mixed and matched computer personalities with user personalities Result: Participants preferred computers whose personalities matched their own.
The evidence from this study suggests that computers can motivate and persuade people more effectively when they share personality traits with themat least in terms of dominance and submission. For designers of persuasive technology, the findings suggest that products may be more persuasive if they match the personality of target users or are similar in other ways.
98
Persuasive Technology
Participants were given a problem to solve and assigned to work on the problem either with a computer they were told was a teammate or a computer that was given no label. For all participants, the interaction with the computer was identical; the only difference was whether or not the participant believed the computer was a teammate. The results compared to responses of other participants: people who worked with a computer labeled as their teammate reported that the computer was more similar to them, that it was smarter, and that it offered better information. These participants also were more likely to choose the problem solutions recommended by the computers.
After completing the study, we examined the data and found significant differences between the conditions. When compared with other participants,
99
people who worked with a computer labeled as their teammate reported that the computer was more similar to them, in terms of approach to the task, suggestions offered, interaction style, and similarity of rankings of items needed for survival. Even more interesting, participants who worked with a computer labeled as a teammate thought the computer was smarter and offered better information. In addition, participants who worked on the task with a computer labeled as a teammate reported that the computer was friendlier and that it gave higherquality information. Furthermore, people who perceived the computer to be similar to themselves reported that the computer performed better on the task.24 During the study we also measured peoples behavior. We found that computers labeled as teammates were more effective in influencing people to choose problem solutions that the computer advocated. In other words, teammate computers were more effective in changing peoples behavior. All in all, the study showed that the perception of shared affiliation (in this case, being on the same team) made computers seem smarter, more credible, and more likableall attributes that are correlated with the ability to persuade. Among people, similarity emerges in opinions and attitudes, personal traits, lifestyle, background, and membership.25 Designers of persuasive technology should be aware of these forms of similarity and strive to build them into their products. One company that has done a good job of this is Ripple Effects, Inc., Principle of Similarity which helps schools, youth-serving organizations, and businesses change social behavior in ways that improve performance.26 The comPeople are more readily panys Relate for Teens CD-ROM leverages the principle of similarity to persuaded by computing make its product more persuasive to its target audiencetroubled teens. technology products that It conveys similarity through the language it uses, the style of its art are similar to themselves (which includes graffiti and dark colors), audio (all instructions are given in some way. by teen voices), and photos and video clips that feature other, similar teens. Researchers at Columbia University and New York University have shown that the product produces positive effects on teen behavior, including significant reductions in aggressive acts, increases in prosocial acts, and improvements in educational outcomes.27 As this example and the Stanford research suggests, designers can make their technology products more persuasive by making them similar to the target audience. The more that users can identify with the product, the more likely they will be persuaded to change their attitudes or behavior in ways the product suggests.
100
Persuasive Technology
101
They were wrong. My research showed that Oscillotechs scopes made a much less favorable impression on users than did a competitors scopes. This competitor had been gaining market share at the expense of Oscillotech. While many factors led to the competitors success, one clear difference was the personality its scopes projected: the messages from the competitors scopes were invariably warm, helpful, and friendly, but not obsequious or annoying. What was more convincing was a controlled study I performed. To test the effects of simply changing the error messages in Oscillotechs scopes, I had a new set of messagesdesigned to portray the personality of a helpful senior engineerburned into the Oscillotech scopes and tested users reactions in a controlled experiment. The result? On nearly every measure, people who used the new scope rated the device more favorably than people who used the previous version of the scope, with the unfriendly messages. Among other things, users reported that the new scope gave better information, was more accurate, and was more knowledgeable. In reality, the only difference between the two scopes was the personality of the message. This was the first time Oscillotech addressed the issue of the personality of the devices it produced. This example illustrates the potential impact of psychological cues in computing products. While it is a benign example, the broader issue of using computer technology to convey a human-like psychologyespecially as a means to persuade peopleis a controversial area that has yet to be fully explored and that is the subject of much debate. (Chapter 9 will address some of the ethical issues that are part of the debate.)
102
Persuasive Technology
Figure 5.5
Quicken rewards users with a celebratory message each time they reconcile their accounts.
offers recommendations, and lists a host of separate stores tailored to my preferences. Each page I click on addresses me by name and lists more recommendations. To keep me online, Im asked if the sites recommendations are on target and to supply more information if they are not. The designers goal, its safe to say, is to persuade me and other users to maximize our online purchases. Iwin.com, a leading site for online gaming and lotteries, uses language in a very different way. The site conveys a young, brash attitude in an attempt to persuade users to log in, amass iCoins (a type of online currency) by performing various activities on the site, and use that currency to play lotteries. When you arrive at the homepage, you can answer a Daily Poll and gain 25 iCoins. One sample question was Whos eaten more? You can choose from two answers: Pac Man or Dom DeLuise. Obviously this question is not serious, more of a teaser to get people to start clicking and playing. The submit button for the survey doesnt even say submit. Instead, it reads Hey big daddy or something similarly hip (the message changes each time you visit the site). If you keep playing games on the main page without logging in, youll see this message in large type: Whats the deal? You dont call, you dont log in . . . Is it me? And if you continue to play games without logging in, again youll get a prompt with attitude:
103
Well, you could play without logging in, but you wont win anything. Its up to you. Later, as you log out of your gaming session, the Web site says: Youre outta here! Thanks for playing! The use of language on this Web site is very different from Amazon. Note how the creators of Iwin.com had the option to choose standard language to move people through the transactional elements (register, log in, log out, enter lotteries for prizes) of the online experience. Instead, they crafted the language to convey a strong online personality, one that has succeeded in acquiring and keeping users logging in and playing games.32
Figure 5.6
Dialog boxes were a key element in our research on the impact of positive feedback from computers.
104
Persuasive Technology
In all, each study participant received 12 messages from the computer during the game. They saw text messages such as: Very smart move. Your new addition will enhance this game in a variety of ways. Your question makes an interesting and useful distinction. Great job! Great! Your suggestions show both thoroughness and creativity. Ten of the messages were pure praise. The other two were less upbeat: All the players received the following warning message after their fourth contribution to the game: Be careful. Your last question may steer the game in the wrong direction. After their eighth contribution, players received this somewhat negative message: Okay, but your question will have a negligible effect on overall search efficiency. Previous studies had shown that adding the non-praise messages to the mix increased the credibility of the 10 praise messages. After participants finished playing the 20 Questions game with the computer, they responded to a questionnaire about their experience. The questionnaire had a few dozen questions about how they felt, their view of the computer, their view of the interaction, and their view of the computers evaluations of their work. In analyzing the data, we compared the two praise conditions (sincere and insincere) along with a third condition that offered no evaluation, just the text, Begin next round. The findings were clear. Except for two questions that focused on the sincerity of the computer, participants responded to true praise and flattery identicallyand these responses were positive. In essence, after people received computer praisesincere or notthey responded significantly more positively than did people who received no evaluation. Specifically, compared to the generic, no evaluation condition, the data show that people in both conditions who received praise
Felt better about themselves Were in a better mood Felt more powerful Felt they had performed well
105
Principle of Praise
By offering praise, via words, images, symbols, or sounds, computing technology can lead users to be more open to persuasion.
Found the interaction engaging Were more willing to work with the computer again Liked the computer more Thought the computer had performed better
Although these arent direct measures of persuasion, these positive reactions open the door to influence. These findings illustrate the importance of using language in ways that will set the stage for persuasive outcomes, rather than in ways that build barriers to influence. Languageeven language used by a computing systemis never neutral. It can promote or hinder a designers persuasion goals.
Social Dynamics
Most cultures have set patterns for how people interact with each otherrituals for meeting people, taking turns, forming lines, and many others. These rituals are social dynamicsunwritten rules for interacting with others. Those who dont follow the rules pay a social price; they risk being alienated. Computing technology can apply social dynamics to convey social presence and to persuade. One example is Microsofts Actimates characters, a line of interactive toys introduced in the late 1990s. The Microsoft team did a great deal of research into creating toys that would model social interactions.35 The goal of the toys, of course, is to entertain kids, but it also seems that the toys are designed to use social rituals to persuade kids to interact with the characters. Consider the Actimates character named DW (Figure 5.7). This interactive plush toy says things like I love playing with you and Come closer. I want to tell you a secret. These messages cue common social dynamics and protocols. By cueing social dynamics, DW affects how children feel and what they do. DWs expressions of friendship may lead children to respond with similar expressions or feelings. DWs invitation to hear her secret sets up a relationship of trust and support. E-commerce sites also use social dynamics to help interactions succeed. They greet users, guide people to products they may like, confirm whats being
106
Persuasive Technology
Figure 5.7
DW, an interactive plush toy developed by Microsoft, puts social dynamics into motion.
purchased, ask for any needed information, and thank people for making the transaction.36 In short, they apply the same social dynamics users might encounter when shopping in a brick-and-mortar store. An application of social dynamics can be found by users of Eudora, an email program. If you dont register the product immediately, every week or so the program will bring up a dialogue box, inviting you to register. The registration screen (also shown in the Introduction, as an example of how computers can be persistent) has some funny, informal text (As a registered user of Eudora we wont nag you as often as we do. Well also erect a giant statue in your image on the front lawn of our corporate headquarterswith a note below: Giant statue offer void on the planet Earth.) All of this text is designed to persuade you to choose one of two buttons in the dialogue box (Figure 5.8): Take me to the registration page! or Maybe later.
107
Figure 5.8
The dialogue box that Eudora uses to persuade users to register.
Eudora doesnt give users the option of clicking no (although you can avoid choosing either option above by simply clicking the close button). Instead, to get the dialogue box to vanish and get on with their task at hand, people most likely click on Maybe later. By clicking on this box, the user has made an implicit commitment that maybe he or she will register later. This increases the likelihood that the user will feel compelled to register at some point. The Eudora dialogue box seems simple-mindedeven goofy. But its actually quite clever. The goofy content and language serve a few purposes: elevating the mood of the user, making the request seem fun and easy, positioning the requestor as approachable and good-humored. Perhaps the main purpose of the goofiness is to serve as a distraction, just as people can use distractions effectively during negotiations. The truth is that Eudora is making a very serious request. Getting people to eventually say yes is vital to the future of the product. And the dynamic that plays out with this dialogue box is not so different from the social dynamics that play out during serious human-human exchanges, such as asking for a raise. (If you can get the boss to say maybe later rather than no in response to your request for a raise, youre in a much more powerful position when you raise the issue again later, as the boss has made a subtle commitment to considering it.) Other social dynamics can be set in motion when people interact with computing products. Users may succumb to peer pressure from computers. Or they may judge information as more accurate when it comes from multiple computing sources. These and many other social dynamics have yet to be tested, but based on early efforts by Amazon.com, Eudora, and others, the potential for using technology to leverage social dynamics appears to be strong.
108
Persuasive Technology
109
Only half of the participants worked with computers that provided this high quality of help and information. The other half also went into the lab alone and were given the same task but used a computer that provided low-quality help. The computer looked identical and had the same interface. But when these participants asked the computer to find information on the items, the information that came back was not very helpful. We again had pretested information and knew what would seem a plausible result of an information search, but because of our pilot tests we knew the information would not be useful to the participants (Small Flashlight: Easy to find yellow Lumilite flashlight is there when you need it. Batteries included.). In setting up the experiment this way, our goal was to have two sets of participants. One group would feel the computer had done a favor for them; the other group would feel the computer had not been helpful. In a subsequent, seemingly unrelated task (it was related, but we hid this fact from participants), each participant was given the opportunity to help a computer create a color palette that matched human perception. The computer would show three colors, and the participants would rank the colors, light to dark. The participants could do as many, or as few, of these comparisons as they wished for the computer. Because this was a controlled study, half of the participants worked with the same computer on the second task, the color perception task (the reciprocity condition), and half of the participants worked with a different computer (the control condition). Those who worked with the same helpful computer on the second taskthe color perception taskhad an opportunity to reciprocate the help the computer had provided earlier. (During the experiment, we never mentioned anything about reciprocity to participants.) In contrast, those who worked with a different computer served as a control group. After completing the study, we analyzed the data and found that people did indeed reciprocate to the computer that helped them. Participants who returned to work with the initially helpful computer performed more work for that computer on the second task. Specifically, participants in the reciprocity condition performed more color evaluationsalmost double the number than those who worked with a different, although identical, computer on the second task. In summary, the study showed that people observed a common social dynamic, the rule of reciprocity; they repaid a favor that a computer had done for them.
110
Persuasive Technology
Participants entered a room with two computers and were assigned a task of finding information, with the help of one of the computers. Half of the participants used the computer that was helpful in finding the information; the other half used the computer that was not helpful. In a subsequent task, participants were asked to help one of the computers to create a color palette. Half of the participants worked with the same computer theyd worked with on the initial task; half worked with the other computer. Result: Those participants who worked with the same helpful computer on both tasks performed almost twice as much work for their computers on the second task as did the other participants.
This reciprocity study included control conditions that rule out other possible explanations for the results. One such explanation is that getting good information during the first task made participants happy, so they did more work in the second task. This explanation is ruled out because half of those who received good information from a computer for the first task used a different but identical-looking computer on the second task, but only those who used the same computer for both tasks showed the reciprocity effect. The other alternative explanation is that people who remained at the Principle of Reciprocity same workstation for the second task were more comfortable and familiar with the chair or the setup, leading to an increase in work on the secPeople will feel the need ond task. This explanation can be ruled out because participants who to reciprocate when comreceived bad information from the computer in the first task and used the puting technology has same computer on the second task did less work, not more, indicating done a favor for them. that people may have retaliated against the computer that failed to help them on the previous task. (The retaliation effect may be even more provocative than the reciprocity effect, though it is not very useful for designing persuasive computer products.) With the alternative explanations ruled out, the evidence suggests that the rule of reciprocity is such a powerful social dynamic that people followed it when working with a machine. The implication for designers of persuasive technology is that the rule of reciprocityan important social dynamiccan be applied to influence users. A simple example that leverages the rule of reciprocity is a shareware program
111
that, after multiple uses, might query the user with a message such as You have enjoyed playing this game ten times. Why not pay back the favor and register?
112
Persuasive Technology
Figure 5.9
Broderbund created a fictional authority figurea teacherto persuade users of its typing program.
Thats also why Broderbund used the image of a teacher when creating its popular software program Mavis Beacon Teaches Typing (Figure 5.9). Mavis Beacon is a marketing creation, not a real person. But her physical image, including prim hairdo, and her name suggest a kindly, competent high school typing teacher. By evoking the image of a teacher, Broderbund probably hoped that its software would gain the influence associated with that role. Although the power of authority has received the most attention in forPrinciple of Authority mal persuasion studies, authority roles arent the only social roles that influence people. Sometimes influence strategies that dont leverage Computing technology power or status also can be effective. Consider the roles of friend, that assumes roles of entertainer, and opponent, each of which can cause people to change authority will have their attitudes or behavior. enhanced powers of Ask Jeeves is a search engine that takes on the role of a butler (Figure persuasion. 5.10) to distinguish its product from competing search engines.41 When you visit ask.com, you ask a simple question, and Jeeves the butler is at your service, searching his own database and the most common Web search engines. Its likely that setting up the search engine in the role of a butler was a deliberate attempt to influence. In terms of attitude, the creators likely wanted peo-
113
Figure 5.10
The Ask Jeeves Web search engine adopts a social role of butler or servant.
ple to feel the site would be easy to use, the service was helpful, and the site would treat them as special and importantall attributes associated with a butler or servant. In terms of behavior, the Ask Jeeves creators probably hoped a butler-based Web site would influence users to return and use the site frequently, developing a kind of ongoing social relationship with the character, something the other search engines dont provide. If the popularity of Web sites is any indication, the Ask Jeeves strategy is working. The site consistently ranksaccording to some accountsin the top 20 Web sites in terms of unique visitors.42 Another example, the Personal Aerobics Trainer (PAT), takes the concept of persuasiveness one step further. PAT is a virtual interactive fitness trainer created by James Davis of MIT. The system lets users choose the type of coach they will find the most motivational (including the Virtual Army Drill Sergeant shown in Figure 5.11). The virtual coach uses computer vision technology to watch how the person is performing and offers positive feedback (Good job! and Keep up the good work!) or negative feedback (Get moving!).43 For computers that play social roles to be effective in motivating or persuading, its important to choose the role model carefully or it will be counterproductive. Adult authority figures might work well for traditional business types,
114
Persuasive Technology
Figure 5.11
The Personal Aerobics Trainer (PAT) is an interactive fitness trainer that lets users choose from a variety of coaches, including a virtual army drill sergeant.
but teenagers may not respond. One user may prefer the Army Drill Sergeant, and another may find it demotivating. The implication for designers of persuasive technology that incorporates social role-playing: know your target audience. As the PAT system suggests, a target audience can have multiple user groups. Designers should provide a way for different groups of users to choose the social roles they prefer.
115
With that in mind, when is it appropriate to make the social quality of the product more explicit? In general, I believe its appropriate to enhance social cues in leisure, entertainment, and educational products (smart toys, video games, kids learning applications). Users of such applications are more likely to indulge, accept, and perhaps even embrace an explicit cyber social actor either embodied or not. The Actimates toys are a good example. In the case of Actimates, one purpose of the toys is to teach social dynamics, so designers can rightly focus on maximizing the use of social cues. When is it not appropriate to enhance social cues? When the sole purpose of the technology product is to improve efficiency. When I buy gas for my car, I choose a station with gas pumps that take credit cards directly. I dont want to deal with a cashier; Im not looking for a social experience. I believe the analogy applies to interactive technologies, such as word processing programs or spreadsheets, that people use to perform a task more efficiently. For such tasks, its best to minimize cues for social presence, as social interactions can slow things down. This is probably why Amazon.com and other e-commerce sites use social dynamics but do not have an embodied agent that chats people up. As in brick-and-mortar stores, when people buy things they are often getting work done; its a job, not a social event. Enhancing social cues for such applications could prove to be distracting, annoying, or both. The quality and repetition of the social cue should be of concern to designers as well. For example, dialogue boxes designed to motivate need to be crafted with care to avoid being annoyingly repetitious. When I created my experiment to study praise in dialogue boxes, I started out with dozens of possible ways to praise users and winnowed down the options, through user testing and other means, to 10 praise messages that would show up during the task. Users never got the same type of praise twice; they were praised many times, but the message was varied so it didnt feel repetitious.
116
Persuasive Technology
2. B. Reeves and C. Nass, The Media Equation: How People Treat Computers, Television, and New Media Like Real People and Places (New York: Cambridge University Press, 1996). 3. In many ways the Tamagotchis success was helpful for those of us working with Cliff Nass at Stanford University. Since the early 1990s we had been researching peoples social responses to computers (see Reeves and Nass (1996) for a summary). Before the Tamagotchi craze, many people didnt fully understand our researchor they didnt understand why we would study computers in this way. However, after the Tamagotchi became widely known, our research made more sense to people; they had seen how the humancomputer relationship could take on social qualities. 4. P. Zimbardo and M. Leippe, The Psychology of Attitude Change and Social Influence (New York: McGraw-Hill, 1991). 5. Group polarization is a phenomenon whereby people, when part of a group, tend to adopt a more extreme stance than they would if they were alone. Social facilitation is a phenomenon whereby the presence of others increases arousal. This helps a person to perform easy tasks but hurts performance on difficult tasks. 6. B. J. Fogg, Charismatic Computers: Creating More Likable and Persuasive Interactive Technologies by Leveraging Principles from Social Psychology, doctoral dissertation, Stanford University (1997). 7. For academic work on computers as persuasive social actors, see the following: a. B. J. Fogg, Charismatic Computers: Creating More Likable and Persuasive Interactive Technologies by Leveraging Principles from Social Psychology, doctoral dissertation, Stanford University (1997). b. B. J. Fogg and C. I. Nass, How users reciprocate to computers: An experiment that demonstrates behavior change, in Extended Abstracts of the CHI97 Conference of the ACM/SIGCHI (New York: ACM Press, 1997). c. B. J. Fogg and C. I. Nass, Silicon sycophants: The effects of computers that flatter, International Journal of Human-Computer Studies, 46: 551561 (1997). d. C. I. Nass, B. J. Fogg, and Y. Moon, Can computers be teammates? Affiliation and social identity effects in human-computer interaction, International Journal of HumanComputer Studies, 45(6): 669678 (1996). 8. For more information on the role of attractiveness in persuasion, see the following: a. E. Berscheid and E. Walster, Physical attractiveness, in L. Berkowitz (ed.), Advances in Experimental Social Psychology, Vol. 7, (New York: Academic, 1974), pp. 158216. b. S. Chaiken, Communicator physical attractiveness and persuasion, Journal of Personality and Social Psychology, 37: 13871397 (1979). 9. H. Sigall and N. Osgrove, Beautiful but dangerous: Effects of offender attractiveness and nature of crime on juridic judgement, Journal of Personality and Social Psychology, 31: 410414 (1975).
117
10. For more on the connection between attractiveness and other qualities, see the following: a. K. L. Dion, E. Bersheid, and E. Walster, What is beautiful is good, Journal of Personality and Social Psychology, 24: 285290 (1972). b. A. H. Eagly, R. D. Ashmore, M. G. Makhijani, and L. C. Longo, What is beautiful is good, but . . . : A meta-analytic review of research on the physical attractiveness stereotype, Psychological Bulleting, 110: 109128 (1991). 11. The principle of physical attractiveness probably applies to Web sites as well, although my research assistants and I could find no quantitative public studies on how attractiveness of Web sites affects persuasiona good topic for future research. 12. These interactive faces typically have not been created for persuasive purposesat least not yet. Researchers in this area usually focus on the significant challenge of simply getting the faces to look and act realistically, with an eye toward developing a new genre of human-computer interaction. In the commercial space, people working on interactive talking faces are testing the waters for various applications, ranging from using these interactive faces for delivering personalized news to providing automated customer support online. 13. S. Parise, S. Kiesler, L. Sproull, and K. Waters, Cooperating with life-like interface agents, Computers in Human Behavior, 15 (2): 123142 (1999). Available online as an IBM technical report at http://domino.watson.ibm.com/cambridge/research.nsf/ 2b4f81291401771785256976004a8d13/ce1725c578ff207d8525663c006b5401/$FILE/ decfac48.htm. 14. For more on the predictability of attractiveness, see the following: a. M. Cunningham, P. Druen, and A. Barbee, Evolutionary, social and personality variables in the evaluation of physical attractiveness, in J. Simpson and D. Kenrick (eds.), Evolutionary Social Psychology (Mahwah, NJ: Lawrence Erlbaum, 1997), 109140. b. J. H. Langlois, L. A. Roggman, and L. Musselman, What is average and what is not average about attractive faces? Psychological Science, 5: 214220 (1994). 15. When it comes to emotions and computers, Dr. Rosalind Picards Affective Computing Research Group at MIT has been blazing new trails. For more information about the groups work, see http://affect.media.mit.edu/. 16. My research involving engineers and social responses to computer devices was performed for HP Labs in 1995. 17. H. Tajfel, Social Identity and Intergroup Relations (Cambridge, England: Cambridge University Press, 1982). 18. R. B. Cialdini, Influence: Science and Practice, 3rd ed. (New York: HarperCollins, 1993). 19. Persuasion scholar Robert Cialdini writes, As trivial as . . . similarities may seem, they appear to work. . . . even small similarities can be effective in producing a positive response to another. Robert B. Cialdini, Influence: Science & Practice (Boston: Allyn and Bacon, 2000). See also H. Tajfel, Human Groups and Social Categories (Cambridge: Cambridge University Press, 1981).
118
Persuasive Technology
See also H. Tajfel, Social Identity and Intergroup Relations (Cambridge, England: Cambridge University Press, 1982). 20. C. I. Nass, Y. Moon, B. J. Fogg, B. Reeves, and D. C. Dryer, Can computer personalities be human personalities? International Journal of Human-Computer Studies, 43: 223239 (1995). 21. For each study, we adapted the desert survival task from J. C. Lafferty and P. M. Eady, The Desert Survival Problem (Plymouth, MI: Experimental Learning Methods, 1974). 22. J. M. Digman, Personality structure: An emergence of the five-factor model, The Annual Review of Psychology, 41: 417440 (1990). 23. C. I. Nass, B. J. Fogg, and Y. Moon, Can computers be teammates? Affiliation and social identity effects in human-computer interaction, International Journal of HumanComputer Studies, 45(6): 669678 (1996). 24. B. J. Fogg, Charismatic Computers: Creating More Likable and Persuasive Interactive Technologies by Leveraging Principles from Social Psychology, doctoral dissertation, Stanford University (1997). 25. Part of my list about similarity comes from S. Shavitt and T. C. Brock, Persuasion: Psychological Insights and Perspectives (Needham Heights, MA: Allyn and Bacon, 1994). 26. http://www.rippleeffects.com. 27. To read about the research, see http://www.rippleeffects.com/research/studies.html. 28. To get a sense of how computers can use emotions to motivate and persuade, see R. Picard, Affective Computing (Cambridge, MA: MIT Press, 1997). See also the readings on Dr. Picards Web site: http://affect.media.mit.edu/AC_readings.html. 29. C. Marshall and T. O. Maguire, The computer as social pressure to produce conformity in a simple perceptual task, AV Communication Review, 19: 1928 (1971). 30. For more about the advisability (pro and con) of incorporating social elements into computer systems, see the following: a. B. Shneiderman, Designing the User Interface: Strategies for Effective Human-Computer Interaction (Reading, MA: Addison Wesley Longman, 1998). b. B. Shneiderman and P. Maes, Direct manipulations vs. interface agents, Interactions, 4(6): 4261 (1997). 31. For more about the advisability of incorporating social elements into computer systems, see the following: a. B. Reeves and C. Nass, The Media Equation: How People Treat Computers, Television, and New Media Like Real People and Places (New York: Cambridge University Press, 1996). b. B. Shneiderman and P. Maes, Direct manipulations vs. interface agents, Interactions, 4(6): 4261 (1997).
119
32. As of this writing, the Web rankings and search directory top9.com ranks iwin.com as the #1 site in the sweepstakes and lottery category, and trafficranking.com ranks the site above americanexpress.com, nih.gov, hotjobs.com, and nfl.com. 33. For more on the effects of praise, see the following: a. E. Berscheid and E. H. Walster, Interpersonal Attraction, 2nd ed. (Reading, MA: AddisonWesley, 1978). b. E. E. Jones, Interpersonal Perception (New York: W. H. Freeman, 1990). c. J. Pandey and P. Singh, Effects of Machiavellianism, other-enhancement, and powerposition on affect, power feeling, and evaluation of the ingratiator, Journal of Psychology, 121: 287300 (1987). 34. For more on investigations into praise from computers, see the following: a. B. J. Fogg, Charismatic Computers: Creating More Likable and Persuasive Interactive Technologies by Leveraging Principles from Social Psychology, doctoral dissertation, Stanford University (1997). b. B. J. Fogg and C. I. Nass, Silicon sycophants: The effects of computers that flatter, International Journal of Human-Computer Studies, 46: 551561 (1997). 35. For more on designing interfaces for children, see the following: a. Erik Strommen and Kristin Alexander, Emotional interfaces for interactive aardvarks: Designing affect into social interfaces for children, Proceeding of the CHI 99 Conference on Human Factors in Computing Systems: The CHI Is the Limit, 528535 (1999). b. Erik Strommen, When the interface is a talking dinosaur: Learning across media with ActiMates Barney, Conference Proceedings on Human Factors in Computing Systems, 228295 (1998). 36. An online article in Info World (Online agents evolve for customer service, Dec. 11, 2000, http://www.itworld.com/AppDev/1248/IWD001211hnenabler/online) notes that The new breed of agent is highly customizable, allowing e-tailers to brand them as they wish. The agents also aim to assist and reassure customers by emoting: If an agent cant answer a question, it expresses disappointment and tries again or it pushes the user toward another channel of service. If the customer expresses interest in a product, the agent can suggest related items. If a transaction is completed, the agent can thank the customer and urge him or her to come back soon. 37. A. W. Gouldner, The norm of reciprocity: A preliminary statement, American Sociological Review, 25: 161178 (1960). 38. For more on investigations into computers and reciprocity, see B. J. Fogg, Charismatic Computers: Creating More Likable and Persuasive Interactive Technologies by Leveraging Principles from Social Psychology, doctoral dissertation, Stanford University 1997. 39. His book is entitled Computer Power and Human Reason (San Francisco: W. H. Freeman, 1976).
120
Persuasive Technology
40. For a good introductory discussion of computerized therapy with pointers to other sources, see the Web page of John Suler, Ph.D., at http://www.rider.edu/users/suler/ psycyber/eliza.html. 41. www.ask.com. 42. In March 2002, top9.com rated AskJeeves.com just above Google, Excite, and About.com. In May 2002, the Search Engine Watch ranked AskJeeves as the #3 search engine, behind Google and Yahoo (source: http://searchenginewatch.com/sereport/02/05-ratings.html). Although these and other reports differ, the main point remains the same: AskJeeves is a popular site. 43. For more information on PAT, see http://www-white.media.mit.edu/~jdavis/Pat/.