Abstract
Humans predict others’ behavior based on mental state inferences and expectations created on previous interactions. On the brink of the introduction of artificial agents in our social environment, the question of whether humans would use similar cognitive mechanisms to interact with these agents gains relevance. Recent research showed that people could indeed explain the behavior of a robot in mentalistic terms. However, there is scarce evidence regarding how expectations modulate the adoption of these mentalistic explanations. The present study aims at creating a questionnaire that measures expectations regarding the capabilities of the robot and testing whether these priors modulate the adoption of the intentional stance toward artificial agents. We found that individual expectations might influence the adoption of mentalistic explanations. After a show period of observation, participants with higher expectations tended to explain iCub’s behavior in mentalistic terms; meanwhile, participants with lower expectations maintained their mechanistic explanations of behavior. Our findings suggest that expectations about capabilities and purpose of the robot might modulate the adoption of intentional stance toward artificial agents.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Change history
21 November 2019
Unfortunately, the authors of this article had failed to add an acknowledgement to their contribution. This missing acknowledgement was added to the article and reads as follows:
Acknowledgement:
This work received support from the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation program (grant awarded to AW, titled “InStance: Intentional Stance for Social Attunement.” G.A. No: ERC-2016-StG-715058).
References
Marchesi, S., Ghiglino, D., Ciardo, F., Perez-Osorio, J., Baykara, E., Wykowska, A.: Do we adopt the intentional stance toward humanoid robots? Front. Psychol. 10, 450 (2019)
Dennett, D.C.: Intentional systems. J. Philos. 68(4), 87–106 (1971)
Searle, J.R.: Construction of Social Reality (1995)
Malle, B.F.: Attribution theories: how people make sense of behavior. Theor. Soc. Psychol. 23, 72–95 (2011)
Waytz, A., Epley, N., Cacioppo, J.T.: Social cognition unbound: insights into anthropomorphism and dehumanization. Curr. Dir. Psychol. Sci. 19(1), 58–62 (2010)
Krach, S., Hegel, F., Wrede, B., Sagerer, G., Binkofski, F., Kircher, T., et al.: Can machines think? Interaction and perspective taking with robots investigated via fMRI. PLoS One 3(7), e2597 (2008)
Chaminade, T., et al.: How do we think machines think? An fMRI study of alleged competition with an artificial intelligence. Front. Hum. Neurosci. 6, 103 (2012)
Gazzola, V., Rizzolatti, G., Wicker, B., Keysers, C.: The anthropomorphic brain: the mirror neuron system responds to human and robotic actions. Neuroimage 35(4), 1674–1684 (2007)
Oberman, L.M., McCleery, J.P., Ramachandran, V.S., Pineda, J.A.: EEG evidence for mirror neuron activity during the observation of human and robot actions: toward an analysis of the human qualities of interactive robots. Neurocomputing 70, 2194–2203 (2007)
Wykowska, A., Chellali, R., Al-Amin, M.M., Müller, H.J.: Implications of robot actions for human perception. How do we represent actions of the observed robots? Int. J. Soc. Robot. 6(3), 357–366 (2014)
Wykowska, A., Kajopoulos, J., Obando-Leitón, M., Chauhan, S.S., Cabibihan, J.J., Cheng, G.: Humans are well tuned to detecting agents among non-agents: examining the sensitivity of human perception to behavioral characteristics of intentional systems. Int. J. Soc. Robot. 7(5), 767–781 (2015)
Wiese, E., Metta, G., Wykowska, A.: Robots as intentional agents: using neuroscientific methods to make robots appear more social. Front. Psychol. 8, 1663 (2017)
Admoni, H., Srinivasa, S.: Predicting user intent through eye gaze for shared autonomy (2016)
Thellman, S., Silvervarg, A., Ziemke, T.: Folk-psychological interpretation of human vs. humanoid robot behavior: exploring the intentional stance toward robots. Front. Psychol. 8(Nov), 1–14 (2017)
Metta, G., et al.: The iCub humanoid robot: an open-systems platform for research in cognitive development. Neural Netw. 23(8–9), 1125–1134 (2010)
Natale, L., Bartolozzi, C., Pucci, D., Wykowska, A., Metta, G.: iCub: the not-yet-finished story of building a robot child. Sci. Robot. 2(13), eaaq1026 (2017)
Chaminade, T., et al.: Brain response to a humanoid robot in areas implicated in the perception of human emotional gestures. PLoS ONE 5(7), e11577 (2010)
Horstmann, A.C., Krämer, N.C.: Great expectations? Relation of previous experiences with social robots in real life or in the media and expectancies based on qualitative and quantitative assessment. Front. Psychol. 10, 939 (2019)
Dautenhahn, K., Woods, S., Kaouri, C., Walters, M.L., Koay, K.L., Werry, I.: What is a robot companion - friend, assistant or butler? In: 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS (2005)
Ray, C., Mondada, F., Siegwart, R.: What do people expect from robots? In: 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS (2008)
Goetz, J., Kiesler, S., Powers, A.: Matching robot appearance and behavior to tasks to improve human-robot cooperation. In: Proceedings - IEEE International Workshop on Robot and Human Interactive Communication (2003)
Arras, K.O., Cerqui, D.: Do we want to share our lives and bodies with robots? A 2000-people survey (2005)
Syrdal, D.S., Dautenhahn, K., Koay, K.L., Walters, M.L.: The negative attitudes towards robots scale and reactions to robot behaviour in a live human-robot interaction study. Adapt. Emergent Behav. Complex Syst. (2009)
Nomura, T.T., Syrdal, D.S., Dautenhahn, K.: Differences on social acceptance of humanoid robots between Japan and the UK. In: Salem, M., Weiss, A., Baxter, P., Dautenhahn, K. (eds.) Proceedings 4th International Symposium on New Frontiers in Human-Robot Interaction, pp. 115–120. The Society for the Study of Artificial Intelligence and the Simulation of Behaviour (AISB) (2015)
Carpinella, C.M., Wyman, A.B., Perez, M.A., Stroessner, S.J.: The robotic social attributes scale (RoSAS). In: Proceedings of the 2017 ACM/IEEE International Conference on Human-Robot Interaction - HRI 2017 (2017)
Bartneck, C., Kulić, D., Croft, E., Zoghbi, S.: Measurement instruments for the anthropomorphism, animacy, likeability, perceived intelligence, and perceived safety of robots. Int. J. Soc. Robot. 1(1), 71–81 (2009)
John, O.P., Srivastava, S.: Big five inventory (BFI). In: Handbook of Personality Second Edition: Theory and Research (1999)
Roncone, A., Pattacini, U., Metta, G., Natale, L.: A cartesian 6-DoF gaze controller for humanoid robots. In: Proceedings of Robotics: Science and Systems (2016)
Mathôt, S., Schreij, D., Theeuwes, J.: OpenSesame: an open-source, graphical experiment builder for the social sciences. Behav. Res. Methods 44(2), 314–324 (2012)
Metta, G., Fitzpatrick, P., Natale, L.: YARP: yet another robot platform. Int. J. Adv. Robot. Syst. 3(1), 8 (2006)
Frith, C.D., Frith, U.: How we predict what other people are going to do. Brain Res. 1079(1), 36–46 (2006)
Acknowledgement
This work received support from the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation program (grant awarded to AW, titled “InStance: Intentional Stance for Social Attunement.” G.A. No: ERC-2016-StG-715058).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Perez-Osorio, J., Marchesi, S., Ghiglino, D., Ince, M., Wykowska, A. (2019). More Than You Expect: Priors Influence on the Adoption of Intentional Stance Toward Humanoid Robots. In: Salichs, M., et al. Social Robotics. ICSR 2019. Lecture Notes in Computer Science(), vol 11876. Springer, Cham. https://doi.org/10.1007/978-3-030-35888-4_12
Download citation
DOI: https://doi.org/10.1007/978-3-030-35888-4_12
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-35887-7
Online ISBN: 978-3-030-35888-4
eBook Packages: Computer ScienceComputer Science (R0)