skip to main content
10.1145/3610978.3640702acmconferencesArticle/Chapter ViewAbstractPublication PageshriConference Proceedingsconference-collections
short-paper

Probing the Inductive Biases of a Gaze Model for Multi-party Interaction

Published:11 March 2024Publication History

ABSTRACT

The behavior management controls proposed for social robots are mostly designed for highly controlled scenarios. In the real world though, robots have to adapt to new situations, generalizing learned behaviors. To address this adaptation challenge, neural network models with embedding layers could be used. We present here an approach to better understand the inductive biases of our robotic gaze model. It was trained with multimodal features as inputs -- either endogenous or exogenous to the robot. Inductive biases were explored by observing feature representations in the embedding spaces. We found that the model was able to distinguish between the robot speech intentions that either request or provide information. Similarly, pairs of partners seem grouped according to their social behavior (speaking time, gaze). Finally, we checked that these groupings had a real impact on the model's performance. Driving these biases when facing new people should allow to generate adapted behavior.

Skip Supplemental Material Section

Supplemental Material

lbr1084.mp4

Supplemental video

mp4

9.1 MB

References

  1. Henny Admoni and Brian Scassellati. 2017. Social Eye Gaze in Human-Robot Interaction: A Review. J. Hum.-Robot Interact., Vol. 6, 1 (may 2017), 25--63.Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Remi Cambuzat, Frédéric Elisei, Gérard Bailly, Olivier Simonin, and Anne Spalanzani. 2018. Immersive Teleoperation of the Eye Gaze of Social Robots Assessing Gaze-Contingent Control of Vergence, Yaw and Pitch of Robotic Eyes. In ISR 2018 - 50th International Symposium on Robotics. VDE, Munich, Germany, 232--239.Google ScholarGoogle Scholar
  3. Mireille Fares, Catherine Pelachaud, and Nicolas Obin. 2023. Zero-shot style transfer for gesture animation driven by text and speech using adversarial disentanglement of multimodal style encoding. Frontiers in Artificial Intelligence, Vol. 6 (2023), 1142997.Google ScholarGoogle ScholarCross RefCross Ref
  4. Laurent Prévot Frédéric Elisei, Gérard Bailly. 2023. RoboTrio2. https://hdl.handle.net/11403/robotrio/v2 ORTOLANG (Open Resources and TOols for LANGuage) textendash www.ortolang.fr.Google ScholarGoogle Scholar
  5. Léa Haefflinger, Frédéric Elisei, Silvain Gerber, Béatrice Bouchot, Jean-Philippe Vigne, and Gérard Bailly. 2023. On the Benefit of Independent Control of Head and Eye Movements of a Social Robot for Multiparty Human-Robot Interaction. In Human-Computer Interaction, Masaaki Kurosu and Ayako Hashizume (Eds.). Springer Nature Switzerland, Cham, 450--466.Google ScholarGoogle Scholar
  6. Thilo Hagendorff and Sarah Fabi. 2023. Why we need biased ai: How including cognitive biases can enhance ai systems. Journal of Experimental & Theoretical Artificial Intelligence (2023), 1--14.Google ScholarGoogle Scholar
  7. Eyke Hüllermeier, Thomas Fober, and Marco Mernberger. 2013. Inductive Bias. Springer New York, New York, NY, 1018--1018.Google ScholarGoogle Scholar
  8. Carlos Toshinori Ishi and Taiken Shintani. 2021. Analysis of Eye Gaze Reasons and Gaze Aversions During Three-Party Conversations. In Proc. Interspeech 2021. 1972--1976.Google ScholarGoogle ScholarCross RefCross Ref
  9. Wolfgang Kabsch. 1976. A solution for the best rotation to relate two sets of vectors. Acta Crystallographica Section A: Crystal Physics, Diffraction, Theoretical and General Crystallography, Vol. 32, 5 (1976), 922--923.Google ScholarGoogle ScholarCross RefCross Ref
  10. Diederik Kingma and Jimmy Ba. 2014. Adam: A Method for Stochastic Optimization. International Conference on Learning Representations (2014).Google ScholarGoogle Scholar
  11. Chinmaya Mishra and Gabriel Skantze. 2022. Knowing Where to Look: A Planning-based Architecture to Automate the Gaze Behavior of Social Robots. In 2022 31st IEEE International Conference on Robot and Human Interactive Communication (RO-MAN). 1201--1208.Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Bilge Mutlu, Jodi Forlizzi, and Jessica Hodgins. 2006. A storytelling robot: Modeling and evaluation of human-like gaze behavior. In 2006 6th IEEE-RAS International Conference on Humanoid Robots. IEEE, 518--523.Google ScholarGoogle ScholarCross RefCross Ref
  13. Bilge Mutlu, Takayuki Kanda, Jodi Forlizzi, Jessica Hodgins, and Hiroshi Ishiguro. 2012. Conversational gaze mechanisms for humanlike robots. ACM Transactions on Interactive Intelligent Systems, Vol. 1 (01 2012), 12.Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Leann Myers and Maria J Sirois. 2004. Spearman correlation coefficients, differences between. Encyclopedia of statistical sciences, Vol. 12 (2004).Google ScholarGoogle Scholar
  15. Yukiko I. Nakano, Takashi Yoshino, Misato Yatsushiro, and Yutaka Takase. 2015. Generating Robot Gaze on the Basis of Participation Roles and Dominance Estimation in Multiparty Interaction. ACM Trans. Interact. Intell. Syst., Vol. 5, 4, Article 22 (dec 2015), 23 pages.Google ScholarGoogle Scholar
  16. Alberto Parmiggiani, Marco Randazzo, Marco Maggiali, Frederic Elisei, Gerard Bailly, and Giorgio Metta. 2014. An articulated talking face for the iCub. In 2014 IEEE-RAS International Conference on Humanoid Robots. 1--6. https://doi.org/10.1109/HUMANOIDS.2014.7041309Google ScholarGoogle ScholarCross RefCross Ref
  17. Taiken Shintani, Carlos T. Ishi, and Hiroshi Ishiguro. 2021. Analysis of Role-Based Gaze Behaviors and Gaze Aversions, and Implementation of Robot's Gaze Control for Multi-Party Dialogue. In Proceedings of the 9th International Conference on Human-Agent Interaction (Virtual Event, Japan) (HAI '21). Association for Computing Machinery, New York, NY, USA, 332--336.Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Gabriel Skantze. 2017. Predicting and Regulating Participation Equality in Human-Robot Conversations: Effects of Age and Gender. In Proceedings of the 2017 ACM/IEEE International Conference on Human-Robot Interaction (Vienna, Austria) (HRI '17). Association for Computing Machinery, New York, NY, USA, 196--204.Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Shirui Wang, Wenan Zhou, and Chao Jiang. 2020. A survey of word embeddings based on deep learning. Computing, Vol. 102 (2020), 717--740.Google ScholarGoogle ScholarCross RefCross Ref
  20. Svante Wold, Kim Esbensen, and Paul Geladi. 1987. Principal component analysis. Chemometrics and intelligent laboratory systems, Vol. 2, 1--3 (1987), 37--52.Google ScholarGoogle Scholar

Index Terms

  1. Probing the Inductive Biases of a Gaze Model for Multi-party Interaction

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      HRI '24: Companion of the 2024 ACM/IEEE International Conference on Human-Robot Interaction
      March 2024
      1408 pages
      ISBN:9798400703232
      DOI:10.1145/3610978

      Copyright © 2024 ACM

      Publication rights licensed to ACM. ACM acknowledges that this contribution was authored or co-authored by an employee, contractor or affiliate of a national government. As such, the Government retains a nonexclusive, royalty-free right to publish or reproduce this article, or to allow others to do so, for Government purposes only.

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 11 March 2024

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • short-paper

      Acceptance Rates

      Overall Acceptance Rate242of1,000submissions,24%
    • Article Metrics

      • Downloads (Last 12 months)17
      • Downloads (Last 6 weeks)12

      Other Metrics

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader