Abstract
Conversational agents are increasingly popular in various domains of application. Due to their ability to interact with users in human language, anthropomorphizing these agents to positively influence users’ trust perceptions seems justified. Indeed, conceptual and empirical arguments support the trust-inducing effect of anthropomorphic design. However, an opposing research stream that has widely been overlooked provides evidence that human-likeness reduces agents’ trustworthiness. Based on a thorough analysis of psychological mechanisms related to the contradicting theoretical positions, we propose that the agent substitution type acts as a situational moderator variable on the positive relationship between anthropomorphic design and agents’ trustworthiness. We argue that different agent types are related to distinct user expectations that influence the cognitive evaluation of anthropomorphic design. We further discuss how these differences translate into neurophysiological responses and propose an experimental set-up using a combination of behavioral, self-reported and eye-tracking data to empirically validate our proposed model.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Schuetzler, R.M., Grimes, M., Giboney, J.S., Buckman, J.: Facilitating natural conversational agent interactions: lessons from a deception experiment. In: Proceedings of the 35th International Conference on Information Systems, Auckland, pp. 1–16 (2014)
Nunamaker Jr., J.F., Derrick, D.C., Elkins, A.C., Burgoon, J.K., Patton, M.W.: Embodied conversational agent-based Kiosk for automated interviewing. J. Manag. Inf. Syst. 28, 17–48 (2011)
Kassner, L., Hirmer, P., Wieland, M., Steimle, F., Königsberger, J., Mitschang, B.: The social factory: connecting people, machines and data in manufacturing for context-aware exception escalation. In: Proceedings of the 50th Hawaii International Conference on System Sciences (HICSS), pp. 1673–1682 (2017)
Gartner: (2016) https://www.gartner.com/doc/3471559?ref=SiteSearch&sthkw=chatbot&fnl=search&srcId=1–3478922254
Forbes: (2016) http://www.forbes.com/sites/parmyolson/2016/02/23/chat-bots-facebook-telegram-wechat/#4c95b5762633
Braun, A.: Chatbots in der Kundenkommunikation. Springer, Berlin (2013)
Wang, W., Benbasat, I.: Attributions of trust in decision support technologies: a study of recommendation agents for e-commerce. J. Manag. Inf. Syst. 24, 249–273 (2008)
Pavlou, P.A., Gefen, D.: Building effective online marketplaces with institution-based trust. Inf. Syst. Res. 15, 37–59 (2004)
Gefen, D., Straub, D.W.: Consumer trust in B2C e-commerce and the importance of social presence: experiments in e-products and e-services. Omega 32, 407–424 (2004)
Nass, C., Steuer, J., Tauber, E.R.: Computers are social actors. In: Proceedings of the SIGCHI conference on human factors in computing systems, pp. 72–78, New York (1994)
Reeves, B., Nass, C.: The media equation: how people treat computers, television? New media like real people? Comput. Math Appl. 5, 33 (1997)
Cassell, J., Bickmore, T.: External manifestations of trustworthiness in the interface. Commun. ACM 43, 50–56 (2000)
Bickmore, T., Cassell, J.: Relational agents. In: Proceedings of the SIGCHI conference on Human Factors in Computing Systems—CHI ’01, New York (2001)
Muir, B.M.: Trust between humans and machines, and the design of decision aids. Int. J. Man Mach. Stud. 27, 527–539 (1987)
Dijkstra, J.J., Liebrand, W.B.G., Timminga, E.: Persuasiveness of expert systems. Behav. Inf. Technol. 17, 155–163 (1998)
Dijkstra, J.J.: User agreement with incorrect expert system advice. Behav. Inf. Technol. 18, 399–411 (1999)
Madhavan, P., Wiegmann, D.A.: Similarities and differences between human–human and human–automation trust: an integrative review. Theor Issues Ergon. Sci. 8, 277–301 (2007)
Mosier, K.L., Skitka, L.J.: Human decision makers and automated decision aids: made for each other. Autom. Hum. Perform. Theory Appl. pp. 201–220 (1996)
Nass, C., Moon, Y.: Machines and mindlessness: social responses to computers. J. Soc. Issues 56, 81–103 (2000)
Riedl, R., Mohr, P., Kenning, P., Davis, F., Heekeren, H.: Trusting humans and avatars: behavioral and neural evidence. In: Proceedings of the Thirty Second International Conference on Information Systems, Shanghai (2011)
Epley, N., Waytz, A., Akalis, S., Cacioppo, J.T.: When we need a human: motivational determinants of anthropomorphism. Soc. Cogn. 26, 143–155 (2008)
Waytz, A., Heafner, J., Epley, N.: The mind in the machine: anthropomorphism increases trust in an autonomous vehicle. J. Exp. Soc. Psychol. 52, 113–117 (2014)
Gong, L.: How social is social responses to computers? The function of the degree of anthropomorphism in computer representations. Comput. Hum. Behav. 24, 1494–1509 (2008)
de Visser, E.J., Monfort, S.S., McKendrick, R., Smith, M.A.B., McKnight, P.E., Krueger, F., Parasuraman, R.: Almost human: anthropomorphism increases trust resilience in cognitive agents. J. Exp. Psychol. Appl. 22, 331–349 (2016)
Dzindolet, M.T., Peterson, S.A., Pomranky, R.A., Pierce, L.G., Beck, H.P.: The role of trust in automation reliance. Int. J. Hum. Comput. Stud. 58, 697–718 (2003)
Skitka, L.J., Mosier, K.L., Burdick, M.: Does automation bias decision-making? Int. J. Hum. Comput Stud. 51, 991–1006 (1999)
Mosier, K.L., Skitka, L.J., Heers, S., Burdick, M.: Automation bias: decision making and performance in high-tech Cockpits. Int. J. Aviat. Psychol. 8, 47–63 (1998)
de Visser, E.J., Krueger, F., McKnight, P.: The world is not enough: trust in cognitive agents. In: Proceedings of the Human Factors and Ergonomics Society Annual Meeting (2012)
Merritt, S.M., Heimbaugh, H., LaChapell, J., Lee, D.: I trust it, but I don’t know why. Hum. Factors J. Hum. Factors Ergon. Soc. 55, 520–534 (2013)
McKnight, D.H., Choudhury, V., Kacmar, C.: Developing and validating trust measures for e-commerce: an integrative typology. Inf. Syst. Res. 13, 334–359 (2002)
Komiak, S.Y.X., Wang, W., Benbasat, I.: Trust building in virtual salespersons versus in human salespersons: similarities and differences. e-Service J. 3, 49–64 (2004)
Rousseau, D.M., Sitkin, S.B., Burt, R.S.: Not so different after all: a cross-discipline view of trust. Acad. Manag. Rev. 23, 393–404 (1998)
Tomlinson, E.C., Mayer, R.C.: The role of causal attribution dimensions in trust repair. Acad. Manag. Rev. 34, 85–104 (2009)
Mayer, R.C., Davis, J.H., Schoorman, F.D.: An integrative model of organizational trust. Acad. Manag. Rev. 20, 709–734 (1995)
Barki, H., Robert, J., Dulipovici, A.: Reconceptualizing trust: a non-linear Boolean model. Inf. Manag. 52, 483–495 (2015)
Epley, N., Waytz, A., Cacioppo, J.T.: On seeing human: a three-factor theory of anthropomorphism. Psychol. Rev. 114, 864–886 (2007)
Waytz, A., Morewedge, C.K., Epley, N., Monteleone, G., Gao, J.-H., Cacioppo, J.T.: Making sense by making sentient: effectance motivation increases anthropomorphism. J. Pers. Soc. Psychol. 99, 410–435 (2010)
Riedl, R., Javor, A., Gefen, D., Felten, A., Reuter, M.: Oxytocin, trust, and trustworthiness: the moderating role of music (2017)
Qiu, L., Benbasat, I.: Evaluating anthropomorphic product recommendation agents: a social relationship perspective to designing information systems. J. Manag. Inf. Syst. 25, 145–182 (2009)
Al-Natour, S., Benbasat, I., Cenfetelli, R.T.: Creating rapport and intimate interactions with online virtual advisors. SIGHCI 2007 Proceedings (2007)
Al-Natour, S., Benbasat, I., Centefelli, R.: Trustworthy virtual advisors and enjoyable interactions: designing for expressiveness and transparency. In: Proceedings of the European Conference on Information Systems (ECIS). Regensburg, Germany (2010)
Benbasat, I., Dimoka, A., Pavlou, P.A., Qiu, L.: Incorporating social presence in the design of the anthropomorphic interface of recommendation agents: insights from an fMRI study. In: Proceedings of the International Conference on Information Systems (ICIS) (2010)
Cyr, D., Head, M., Larios, H., Pan, B.: Exploring human images in website design: a multi-method approach. MIS Q. 33, 539–566 (2009)
Pütten, von der, A.M., Krämer, N.C., Gratch, J., Kang, S.-H.: “It doesn’t matter what you are!” Explaining social effects of agents and avatars. Comput. Hum. Behav. 26, 1641–1650 (2010)
Appel, J., Pütten, von der, A., Krämer, N.C., Gratch, J.: Does humanity matter? Analyzing the importance of social cues and perceived agency of a computer system for the emergence of social reactions during human–computer interaction. Adv. Hum. Comput. Interact. pp. 1–10 (2012)
Fiske, S.T., Taylor, S.E.: Social Cognition: From Brains to Culture. Sage, London (2013)
Riedl, R., Mohr, P.N.C., Kenning, P.H., Davis, F.D., Heekeren, H.R.: Trusting humans and avatars: a brain imaging study based on evolution theory. J. Manag. Inf. Syst. 30, 83–114 (2014)
Krach, S., Hegel, F., Wrede, B., Sagerer, G., Binkofski, F., Kircher, T.: Can machines think? Interaction and perspective taking with robots investigated via fMRI. PLoS ONE 3, 1–11 (2008)
Poole, A., Ball, L.J.: Eye tracking in HCI and usability research. Encycl. Hum. Comput. Interact. 1, 211–219 (2006)
Just, M.A., Carpenter, P.A.: Eye fixations and cognitive processes. Cogn. Psychol. 8, 441–480 (1976)
Léger, P.M., Sénecal, S., Courtemanche, F.: Precision is in the eye of the beholder: application of eye fixation-related potentials to information systems research. J. Assoc. Inf. Syst. 15, 651–678 (2014)
Gilzenrat, M.S., Nieuwenhuis, S., Jepma, M., Cohen, J.D.: Pupil diameter tracks changes in control state predicted by the adaptive gain theory of locus coeruleus function. Cogn. Affect. Behav. Neurosci. 10, 252–269 (2010)
Reimer, J., Froudarakis, E., Cadwell, C.R., Yatsenko, D.: Pupil fluctuations track fast switching of cortical states during quiet wakefulness. Neuron 84, 355–362 (2014)
Satterthwaite, T.D., Green, L., Myerson, J., Parker, J.: Dissociable but inter-related systems of cognitive control and reward during decision making: evidence from pupillometry and event-related fMRI. Neuroimage. pp. 1017–1031 (2007)
Preuschoff, K., Hart, B.T.: Pupil dilation signals surprise: evidence for noradrenaline’s role in decision making. Front. Neurosci. 5, article 115 (2011)
Lavín, C., Martín, R.S., Jubai, E.R.: Pupil dilation signals uncertainty and surprise in a learning gambling task. Front. Behav. Neurosci 7, article 218 (2014)
Riedl, R., Davis, F.D., Hevner, A.R.: Towards a NeuroIS research methodology: intensifying the discussion on methods, tools, and measurement. J. Assoc. Inf. Syst. 15, i–xxxv (2014)
Xu, Q., Riedl, R.: Understanding online payment method choice: an eye-tracking study. In: Proceedings of the 32th International Conference on Information Systems, pp. 1–12, Shanghai (2011)
Riedl, R., Hubert, M., Kenning, P.: Are there neural gender differences in online trust? An fMRI study on the perceived trustworthiness of eBay offers. MIS Q. 34, 397–428 (2010)
Compeau, D.R., Higgins, C.A.: Computer self-efficacy: development of a measure and initial test. MIS Q. pp. 189–211 (1995)
Waytz, A., Cacioppo, J., Epley, N.: Who sees human? Perspect. Psychol. Sci. 5, 219–232 (2010)
MacDorman, K.F., Green, R.D., Ho, C.C., Koch, C.T.: Too real for comfort? Uncanny responses to computer generated faces. Comput. Hum. Behav. 25, 695–710 (2009)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer International Publishing AG
About this paper
Cite this paper
Seeger, AM., Heinzl, A. (2018). Human Versus Machine: Contingency Factors of Anthropomorphism as a Trust-Inducing Design Strategy for Conversational Agents. In: Davis, F., Riedl, R., vom Brocke, J., Léger, PM., Randolph, A. (eds) Information Systems and Neuroscience. Lecture Notes in Information Systems and Organisation, vol 25. Springer, Cham. https://doi.org/10.1007/978-3-319-67431-5_15
Download citation
DOI: https://doi.org/10.1007/978-3-319-67431-5_15
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-67430-8
Online ISBN: 978-3-319-67431-5
eBook Packages: Business and ManagementBusiness and Management (R0)