Virtual Assistant Personality Preference Among Urban English Speakers - An investigation into how bot personality impacts customer experience ...

Page created by Rodney Webb
 
CONTINUE READING
Virtual Assistant Personality Preference Among Urban English Speakers - An investigation into how bot personality impacts customer experience ...
Virtual Assistant Personality Preference
    Among Urban English Speakers
 An investigation into how bot personality impacts customer experience
             during interactions with AI-powered assistants

                           By Brinda Mehra
Virtual Assistant Personality Preference Among Urban English Speakers - An investigation into how bot personality impacts customer experience ...
Report Summary
Virtual assistants, chatbots and other artificially intelligent conversational agents have been
gaining an immense amount of traction in recent times. Many market research companies pre-
dict that interactions with these AI-powered assistants will boom over the next few years. How-
ever, the current satisfaction levels with virtual assistant interactions are not as high as they
could be, as these assistants sometimes fail to deliver on the quality of the interaction.

This paper looks at the possibility of integrating personality into virtual assistants in order to
understand whether personality can have an actual positive impact on user perceptions of and
experiences with these assistants. While the concept of ‘chatbot personality’ has been
well-documented in the context of general conversational agents, there is a lack of exploration
with regards to enterprise virtual assistants and personality –– a gap which this paper attempts
to address.
Virtual Assistant Personality Preference Among Urban English Speakers - An investigation into how bot personality impacts customer experience ...
Table of Contents

SECTION I - OVERVIEW                                                                                                          1
Understanding the need to improve the quality of human-bot interactions through imbued personality.

SECTION II - THEORETICAL BACKGROUND                                                                                           2
Exploring the theoretical framework underlying the study, as well as successful implementations of bot personality
in the industry.

SECTION III - DEFINING THE BOT PERSONALITIES                                                                                  4
The selection of three distinct bot personality types for the purpose of the study - Transactional, Prosocial and Friendly.

SECTION IV - EXPERIMENTAL DESIGN                                                                                              6
The methodology for the two experiments conducted as part of the study.

SECTION V - RESULTS                                                                                                           8
A summary and analysis of the findings derived from the two experiments.

SECTION VI - CONCLUSION                                                                                                       12
The implications of the study and the road ahead.

REFERENCES                                                                                                                    13
Virtual Assistant Personality Preference Among Urban English Speakers - An investigation into how bot personality impacts customer experience ...
Section I - Overview
By 2020, the average person was predicted to have more daily interactions with an AI-powered
chatbot or virtual assistant than with their own spouse [1]. This prediction, made in 2016, has not
entirely borne out, but at the time it would have been great news for the AI industry. After all,
one of the key aims for artificial intelligence is to provide a conversational experience that is
indistinguishable from that of a human being [2]. One can argue then that people holding more
conversations with their virtual assistants than their spouses would certainly achieve that goal,
if not exceed it.

However, as we draw closer to 2020, it does not seem like the current level of satisfaction with
chatbot interactions heralds the beginning of a golden age for virtual assistants.

The general reception of virtual assistants is more positive than negative – 38% of people
globally rate their perception of virtual assistants as positive whereas only 11% rate it as neg-
ative - but the majority remain neutral [3]. This suggests that AI companies must understand
how to leverage the conversational experience in order to both gain and retain users.

Although there may be multiple ways in which companies are able to augment the conversa-
tional experience in order to achieve their goal of user engagement and retention, this paper
will attempt to make an argument for the development of personality within the virtual assistant
to achieve this goal. Research on the role of personality in humans suggest that it plays a key
role in maintaining social relationships as well as making social interactions productive and
enjoyable, which has major implications for user interactions with virtual assistants[4].

This paper explores how personality can influence customer experiences with AI assistants
across two different use-cases: food-ordering and education.

                                                                                                      1
Section II - Theoretical Background
Current research on the conversational abilities of chatbots and virtual assistants suggests that
there is a major gap between the level of the conversation that these assistants can hold and
user expectations of the interaction. Although such agents promise human-like dialogue, many
of these interactions often fall short of the mark [5].

Virtual assistants are able to embody the functional aspects of human conversation but
often fail to embody and encapsulate the major characteristics that define human conver-
sation, thus causing users to feel that interactions with such software agents are dry, stilt-
ed and uncomfortable [6].

In fact, many feel that the key attributes of a human conversation that make a social interaction
successful (such as the ability to understand intent, trust, active listening and humor) cannot be
organically developed in a conversational interaction with a virtual assistant and express their
doubts that a software agent could ever achieve the ability to hold conversations like a human
[7].

Naturally, this poses several problems for the widespread adoption of virtual assistants in daily
life, as users are unlikely to continue interacting with these assistants if all their interactions are
constrained, task-based and leave no scope for any further interaction. Moreover, when this
lack of conversational abilities is coupled with the fact that users tend to personify computer
agents, often treating them like real people and expecting empathy, humor and other human
characteristics from the interaction, there is an underlying need for these AI assistants to devel-
op more anthropomorphic qualities in order to fulfill user expectations [8].

Perhaps the easiest way to anthropomorphize chatbots and virtual assistants is to imbue
them with a personality.

Although recent reports suggest that personality is not a key requirement for a satisfactory
interaction, with an estimated 48% of users stating that they would ‘rather interact with a virtual
assistant that solves their problems’ than one with personality, one can make an argument for
the impact that personality can have [3]. Perhaps the most glowing recommendation for the
development of personality comes with the phenomenal success of Microsoft’s Xiaoice, a gen-
eral conversational chatbot deployed on the Chinese social media network WeChat. Developed
using an Empathic Computing Framework, NLG and context retention, Xiaoice mimics the per-
sonality of an eighteen year old girl who is witty, self-effacing and has a wonderful sense of
humor, often flirting with her users and engaging them in deep conversations [9]. The presence
of this personality plays a major role in Xiaoice’s success – it boasts of around 660 million users
worldwide and a Conversation Turns Per Unit metric (CPS) of 23, which is a significantly higher
number than that of many other chatbots [10].

                                                                                                          2
Naturally, this leads many to question whether personality is the main factor behind Xiaoice’s
success. Although multiple other factors, such as the presence of image recognition or senti-
ment analysis, play an equally important role in its success, the key role personality plays in
human social relationships suggests that it may also be the reason behind Xiaoice’s immense
popularity.

Research on the role of personality in human social interactions suggests that it can influence
how satisfying a social interaction can be, with an interaction between compatible personalities
being far more productive and enjoyable than any other interactions [11], [4]. Similarly, personality
is also responsible for helping classify other people’s behaviour into simplifying traits, which
then allows us to predict other people’s behaviour and guide our own responses to it, making it
a linchpin for the success of human relationships [12], [13].

Although research on the importance of personality has focused exclusively on human relation-
ships until now, it becomes imperative to understand how this information can be applied to
human-computer interaction (HCI) because of new developments in understanding how the
human brain reacts to technology. According to the Media Equation theory, humans tend to
treat technology and technologically mediated communication in the same way they would
 According
treat        to the Media Equation theory, humans tend to treat technology and technologi-
      real people.
 cally mediated communication in the same way they would treat real people.

This process happens subconsciously and automatically, with studies suggesting this reaction
may occur because human brains have not evolved to deal with twenty-first century technology
and therefore are unable to distinguish many facets of the real world from the virtual [14], [15].

Given the spate of research that supports Media Equation Theory and the importance of per-
sonality in human social interactions, there has naturally been a lot of interest garnered in how
AI companies and researchers can capitalize upon these psychological phenomena to amelio-
rate virtual assistant interactions, with chatbot personality slowly becoming a key area of focus.

Building upon this, Haptik carried out a small but in-depth research experiment on how person-
ality can influence user experiences in order to contribute to the existing literature on this topic,
as well as broach the discussion of deploying virtual assistant personality in an enterprise con-
text, which has historically been wary of bringing in any sort of personality into bot interactions
until fairly recently.

                                                                                                        3
Section III - Defining The Bot Personalities
On the basis of previous research into virtual assistant personality, Haptik’s study also utilized
the Five-Factor Model of personality (FFM) in order to develop the three key personalities it
would test. In psychological literature, FFM consists of five broad trait dimensions that are the
most common descriptors used to describe people. These five dimensions, namely Openness
to Experience, Conscientiousness, Extraversion, Agreeableness and Neuroticism represent a
range between the two extremes of the trait, where most people tend to fall (Goldberg, 1993).
The three key bot personality types were developed using IBM’s Primary and Secondary Char-
acteristics Table, which is also based on the FFM and used three different combinations of the
traits of Extraversion, Agreeableness and Conscientiousness to define the bot personalities, as
described below.

                                                                            Transactional Personality

                                                                                          I'd like to book
Transactional Personality: A transactional personality would                              a flight.

be defined as one that is serious, highly principled and restrained.
According to the IBM chart, this chatbot would score high on Con-
scientiousness and low on Agreeableness, while the interaction
with this personality would be highly goal-oriented and have little-
                                                                                Sure. Give me a
to no social niceties involved, with sharp, to-the-point text.                  minute, while I
                                                                                search for flights.

    Prosocial Personality

               I'd like to book
               a flight.
                                  Prosocial Personality: A prosocial personality, by definition, is one
                                  that is extremely helpful. It is also cooperative, polite and consider-
                                  ate, scoring high on both Conscientiousness and Agreeableness in
                                  the IBM Chart. Interactions with this personality would contain sev-
                                  eral social niceties, repeated assurances of its desire to help and
     Hi! I'd be glad to           some markers of engagement like exclamation marks and emoti-
     help you search
     for a flight.                cons.

                                                                                                             4
Friendly Personality

                                                                                     I'd like to book
Friendly Personality: Much like its name suggests, interactions                      a flight.
with this bot are meant to be reminiscent of an interaction with a
friend. Vibrant, enthusiastic and social, this personality will score
high on the dimensions of Extraversion and Agreeableness, utilis-
ing several markers of engagement like images, gifs, emoticons,
                                                                           Hey there! Just
letter reduplication and slang in order to make the interaction            sit back and relax
more casual in nature                                                      while I find you
                                                                           a flight!

Since Haptik’s Bot Builder Tool operates on a rule-based system, these personalities were
developed manually rather than generativally via a corpus database. In order to test for accura-
cy, sections of the chatbot copy were processed via IBM’s emotion analyser to check whether
the correct sentiment was being expressed. The results suggested that the personalities were
fairly well defined, with the Transactional personality receiving a dominant tone of Analytical
(0.66); Prosocial personality, the dominant tones of Joy (0.61); and Friendly personality, the
dominant tones of Joy (0.63) and Confidence (0.66). In order to test for the applicability of per-
sonality across use-cases, two use cases were tested to demonstrate how personality impacts
user experience.

                                                                                                        5
Section IV - Experimental Design
EXPERIMENT I

The first study focused on a food-ordering use-case, with users interacting with three virtual
assistants for the fictional company of WcDonalds. These bots were developed on Haptik’s Bot
Builder Tool and were in accordance with the description of the three personality types men-
tioned in the previous section. Forty users aged 18-53 were asked to participate in this experi-
ment (M = 31.25) in a repeated-measures design, which entailed all users interacting with all
three versions of the WcDonalds chatbot, namely: WcBot (the transactional bot), Roland Wc-
Donald (the prosocial bot) and Ron (the friendly bot). Following their interactions with each of
the chatbots, they were asked to rate their experience for each of the chatbots across three
key dimensions on a Likert type item questionnaire. These dimensions were:

    Likelihood of Re-interaction

    Comfort Level

    Productivity & Efficiency

          WcBot is online                                        Roland WcDonald is online                           Ron is online
          A.I Powered WcDonalds Chatbot                          Wc Donald’s very own Chatbot!                       Wc Donald’s Chatbot at your service

                                                                                                               IT’S TIME TO GET THIS PAR-TAY
                                                                                                               STARTED
                                                           Hey there, I’m the WcDonald’s
                                                           Chatbot Roland WcDonald
                                             Get Started                                                                                            YESS LET’S GO!

                                                           I would love to help you get the best
                                                           meal of your life today!
    Hello, I am the WcDonald’s Chatbot                                                                         So, what do you feel like starting
                                                                                                               with?? Pick one even tho it’s hard
                                                           Please choose how you would like to
    This is how I can assist you:
                                                           start:

     Place Orders Directly      Show Extra Value Deals                                                            HUNGRY! Let me order now!!        Deal Me Up!
                                                            Let’s Order Directly!    Check Extra Value Deals

                        Store Locator                                                                                             FMY Please!
                                                                              Locate Stores

                                                                                                                                                                     6
EXPERIMENT II

The second study was a conceptual replication of the first and tested both a different use-case
as well as a different age group in order to understand whether personality could affect user
experiences across use-cases and demographics. Twelve participants aged 11-15 were asked to
watch three .mp4 files of an interaction with a chat-based dictionary on a Facebook Messenger
interface, which was developed using BotSociety. Much like the previous study, this study also
tested all three previously defined personalities. However in this case, the names of the bots
were WordHelper (the transactional bot), WordAid (the prosocial bot) and Word!Baby (the
friendly bot). Unlike the previous study, participants only watched the video of the interaction
and did not actually interact with it. In order to understand which chatbot interaction appealed
to the participants the most, they were asked to pick which chatbot they would have liked to
interact with and explain why. All responses were collected via WhatsApp.

                                                                                                   7
Section V - Results
There was an overwhelming preference for the friendly personality across both studies, as is
evident from the findings.

EXPERIMENT I
The friendly bot Ron scored higher than the transactional and prosocial personalities across all
three dimensions tested, suggesting that the introduction of a fun, exuberent personality does
positively impact user experience. Further evidence of the overwhelming preference for the
friendly personality across all three criteria is provided below:

Likelihood of Re-interaction:

   The friendly bot personality had the highest mean rating across all three personalities,
receiving 4.05 out of a scale of 5, which was much higher than the mean ratings for the trans-
actional or the prosocial personalities.

   42.5% of participants gave the friendly bot the highest possible ranking of ‘5’ for likelihood
of reinteraction, compared to 12.5% for the prosocial bot and 15% for the transactional bot.

  Participants were almost three times as likely to state that they would definitely interact with
the friendly bot than with the transactional or prosocial bots.

                   Transactional Bot

                        Prosocial Bot

                          Friendly Bot

                                           0         1        2        3        4        5

            Fig.1: Average Ratings for Users’ Likelihood of Re-interaction Across Bot Personalities

                                                                                                      8
Comfort Level:

Participants also seemed to feel more comfortable with the friendly bot than with the transac-
tional or prosocial bots

   40% of participants said that they felt extremely comfortable interacting with the friendly
bot, in comparison to 17.5% for the transactional tbot and 7.5% for the prosocial bot.

  In terms of average ratings, the friendly bot also had the highest mean rating for users’ com-
fort levels, scoring 4.05 in comparison to the significantly lower scores for the transactional
bot (3.325) and prosocial bots (3.3).

   Average ranking of the friendly chatbot was almost 22% higher than that of the transac-
tional chatbot.

                     5

                     4

                     3

                     2

                     1

                     0

                    Transactional              Prosocial                 Friendly
                        Bot                       Bot                      Bot

                  Fig.2: Mean Ratings for Users’ Comfort Levels Across Bot Personalities

                                                                                                   9
Productivity and Efficiency:

In the context of productivity and efficiency, the friendly bot was almost four times as likely to
be considered ‘extremely productive and efficient’ than the transactional bot.

  45% of participants gave the friendly bot personality the highest ranking, compared to only
12.5% for the transactional bot and 17.5% for the prosocial bot.

  The friendly bot had the highest mean score in productivity and efficiency as well, scoring a
rating of 4.175 compared to the transactional bots rating of 3.525 or the prosocial bots rating
of 3.625.

              Transactional
              Bot

              Prosocial
              Bot

              Friendly
              Bot

                Fig.3: Average Ratings for Productivity & Efficiency Across Bot Personalities

                                                                                                     10
EXPERIMENT II

There was a similar preference expressed by participants towards the friendly bot in this experi-
ment. 66.67% of the participants expressed a preference for interacting with the friendly bot
Word!Baby over the other two, citing that they enjoyed the interaction because it felt just
like talking to another human being. Participants also commented on the fact that the use of
gifs and emojis made it easier to understand the definition of the word, as did the fact that it was
disseminated in a manner that was far easier to understand than that of the transactional or pro-
social bots. Some of the participants reviews are mentioned below:

            “I would have liked to talk to the first chatbot (Word!Baby) because I think that it
            explains the definition of the word in a simpler manner with an explanation which
            everyone can understand. Also I like the fact that it talks exactly like a human
            because it seems as if the chatbot is your friend.”
                                                                                        –– Male, 13

            “ [I like] this one [Word!Baby] because [I] felt like I was chatting with someone
            real and who knew who [I am]”
                                                                                        –– Male, 12

            “ I would like to chat with Word!Baby because it was like more like a casual
            conversation and could deliver the meaning very easily and was to the point.
            Also the added visual pictures/gif helps you remember the word better. “
                                                                                  –– Female, 15

                                                                        WordHelper
                                                                        25%

                                                                               WordAid
                                                                               8.3%
                    Word!Baby
                    66.7%

                      Fig. 4: User Preferences for Interactions Across Bot Personalities

However, unlike the previous experiment, the transactional bot personality scored higher
ratings than the prosocial personality in this study. 25% of the participants stated that they
would like to interact with the transactional bot WordHelper, stating that they found the
interaction objective, concise and good for people in a hurry. Only 8.33% of users stated that
they would like to interact with the prosocial bot WordAid.
                                                                                                       11
Section VI - Conclusion
This study strongly supports the idea that personality can positively affect customer experienc-
es, with the friendly bot personality being preferred by participants across ages, use-cases,
domains and chatbot types.

The results suggest that users may enjoy interacting with a virtual assistant that can converse
at a level that is evocative of an interaction between friends as opposed to an interaction that is
highly goal-oriented and leaves no scope for further developments.

Although a fairly small study that is focused on the Indian consumer market, the outcome of this
study does call for a renewed exploration into how AI companies can integrate personality into
their chatbots or other digital agents in an enterprise context –– especially since the incorpora-
tion of anthropomorphic qualities seems to be the ‘secret sauce’ in making human-AI interac-
tions successful.

                                                                                                      12
References

Beaver, L. (2017, May 11). Chatbots are gaining traction. Retrieved August 6, 2019, from
https://www.businessinsider.com/chatbots-are-gaining-traction-2017-5?IR=T

Churches, O., Nicholls, M., Thiessen, M., Kohler, M., & Keage, H. (2014). Emoticons in mind: An
event-related potential study. Social neuroscience, 9(2), 196-202.

Clark, L., Pantidi, N., Cooney, O., Doyle, P., Garaialde, D., Edwards, J., ... & Wade, V. (2019, April).
What Makes a Good Conversation?: Challenges in Designing Truly Conversational Agents. In
Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems (p. 475).
ACM.

Clark, L. (2018, July). Social boundaries of appropriate speech in HCI: A politeness perspective.
In The 32nd International BCS Human Computer Interaction COnference (HCI 2018), Belfast,
United Kingdom, 4-6 July 2018. BCS Learning and Development.

Dryer, D. C. (1999). Getting personal with computers: how to design personalities for agents.
Applied artificial intelligence, 13(3), 273-295.

Goldberg, L. R. (1993). The structure of phenotypic personality traits. American psychologist,
48(1), 26.

Hornigold, T. (2019, July 14). This Chatbot has Over 660 Million Users-and It Wants to Be Their
Best Friend. Retrieved August 6, 2019, from https://singularityhub.com/2019/07/14/this-chat-
bot-has-over-660-million-users-and-it-wants-to-be-their-best-friend/

Horowitz, L. M., Dryer, D. C., & Krasnoperova, E. N. (1997). The circumplex structure of inter-
personal problems. Circumplex models of personality and emotions, 347-384.

IBM. (2015, September 11). How conversation (with context) will usher in the AI future.
Retrieved August 6, 2019, from https://www.ibm.com/watson/advantage-reports/future-of-ar-
tificial-intelligence/ai-conversation.html

Levy, H. P. (2016, October 18). Gartner Predicts a Virtual World of Exponential Change.
Retrieved August 6, 2019, from https://www.gartner.com/smarterwithgartner/gartner-pre-
dicts-a-virtual-world-of-exponential-change/

Luxton, D., Anderson S. L., & Anderson, M. (2016). Chapter 11 - Ethical Issues and Artificial
Intelligence Technologies in Behavioral and Mental Health Care. In Artificial Intelligence in
Behavioural and Mental Health Care (pp 255-276). Cambridge, Massachusetts: Academic
Press.

                                                                                                           13
McTear, M., Callejas, Z., & Griol, D. (2016). The Conversational Interface: Talking to Smart Devic-
es: Springer International Publishing.

Press, G. (2017, May 17). AI By The Numbers: 33 Facts And Forecasts About Chatbots And Voice
Assistants. Retrieved August 6, 2019, from
https://www.forbes.com/sites/gilpress/2017/05/15/ai-by-the-numbers-33-facts-and-
forecasts-about-chatbots-and-voice-assistants/#5c028577731f

Reeves, B., & Nass, C. I. (1996). The media equation: How people treat computers, television, and
new media like real people and places. Cambridge university press.

Schultz, D. and Schultz, S. (1998). Theories of Personality. 6th ed. Pennsylvania: Brooks/Cole
Publishing, p.348.

Spencer, G. (2018, November 01). Much more than a chatbot: China’s Xiaoice mixes AI with ...
Retrieved July 10, 2019,from https://news.microsoft.com/apac/features/much-more-
than-a-chatbot-chinas-xiaoice-mixes-ai-with-emotions-and-wins-over-millions-of-fans/

Uleman, J. S., Hon, A., Roman, R. J., & Moskowitz, G. B. (1996). On-line evidence for spontaneous
trait inferences at encoding. Personality and Social Psychology Bulletin, 22(4), 377-394.

Zhou, L., Gao, J., Li, D., & Shum, H. Y. (2018). The design and implementation of XiaoIce, an em-
pathetic social chatbot. arXiv preprint arXiv:1812.08989.

                                                                                                      14
Thank You!

Interesting in developing a Conversational AI solution for your enterprise?

              Reach out to our team at enterprise@haptik.ai
You can also read