skip to main content
10.1145/3338286.3344396acmconferencesArticle/Chapter ViewAbstractPublication PagesmobilehciConference Proceedingsconference-collections
research-article

Semantic 3D gaze mapping for estimating focused objects

Published:01 October 2019Publication History

ABSTRACT

Eye-trackers are expected to be used in portable daily-use devices. However, it must register object information and define a unified coordinate system in advance for human--computer interaction and quantitative analysis. Therefore, we propose a semantic 3D gaze mapping to collect gaze information from multiple people on the unified map and detect focused objects automatically. The semantic 3D map can be reconstructed using keyframe-based semantic segmentation and structure-from-motion, and the 3D point-of-gaze can also be computed on the map. We confirmed that the fixation time of the focused object can be calculated through an experiment without prior information.

References

  1. Pablo F Alcantarilla and T Solutions. 2011. Fast explicit diffusion for accelerated features in nonlinear scale spaces. IEEE Trans. Patt. Anal. Mach. Intell 34, 7 (2011), 1281--1298.Google ScholarGoogle Scholar
  2. Kai Essig, Daniel Dornbusch, Daniel Prinzhorn, Helge Ritter, Jonathan Maycock, and Thomas Schack. 2012. Automatic analysis of 3D gaze coordinates on scene objects using data from eye-tracking and motion-capture systems. In Proceedings of the Symposium on Eye Tracking Research and Applications. ACM, 37--44.Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Kakeru Hagihara, Keiichiro Taniguchi, Irshad Abibouraguimane, Yuta Itoh, Keita Higuchi, Jiu Otsuka, Maki Sugimoto, and Yoichi Sato. 2018. Object-wise 3D Gaze Mapping in Physical Workspace. In Proceedings of the 9th Augmented Human International Conference. ACM, 25.Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Michael Maurus, Jan Hendrik Hammer, and Jürgen Beyerer. 2014. Realistic heatmap visualization for interactive analysis of 3D gaze data. In Proceedings of the Symposium on Eye Tracking Research and Applications. ACM, 295--298.Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Pierre Moulon, Pascal Monasse, Romuald Perrot, and Renaud Marlet. 2016. Openmvg: Open multiple view geometry. In International Workshop on Reproducible Research in Pattern Recognition. Springer, 60--74.Google ScholarGoogle Scholar
  6. Susan M Munn and Jeff B Pelz. 2008. 3D point-of-regard, position and head orientation from a portable monocular video-based eye tracker. In Proceedings of the 2008 symposium on Eye tracking research & applications. ACM, 181--188.Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Jeff B Pelz, Thomas B Kinsman, and Karen M Evans. 2011. Analyzing complex gaze behavior in the natural world. In Human Vision and Electronic Imaging XVI, Vol. 7865. International Society for Optics and Photonics, 78650Z.Google ScholarGoogle ScholarCross RefCross Ref
  8. Thies Pfeiffer. 2012. Measuring and visualizing attention in space with 3D attention volumes. In Proceedings of the Symposium on Eye Tracking Research and Applications. ACM, 29--36.Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Thies Pfeiffer, Patrick Renner, and Nadine Pfeiffer-Leßmann. 2016. EyeSee3D 2.0: model-based real-time analysis of mobile eye-tracking in static and dynamic three-dimensional scenes. In Proceedings of the Ninth Biennial ACM Symposium on Eye Tracking Research & Applications. ACM, 189--196.Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Kentaro Takemura, Yuji Kohashi, Tsuyoshi Suenaga, Jun Takamatsu, and Tsukasa Ogasawara. 2010. Estimating 3D point-of-regard and visualizing gaze trajectories under natural head movements. In Proceedings of the 2010 Symposium on Eye-Tracking Research & Applications. ACM, 157--160.Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Kentaro Takemura, Kenji Takahashi, Jun Takamatsu, and Tsukasa Ogasawara. 2014. Estimating 3-D point-of-regard in a real environment using a head-mounted eye-tracking system. IEEE Trans. Human-Machine Syst. 44, 4 (2014), 531--536.Google ScholarGoogle ScholarCross RefCross Ref
  12. Hengshuang Zhao, Jianping Shi, Xiaojuan Qi, Xiaogang Wang, and Jiaya Jia. 2017. Pyramid scene parsing network. In Proceedings of the IEEE conference on computer vision and pattern recognition. 2881--2890.Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Semantic 3D gaze mapping for estimating focused objects

      Recommendations

      Reviews

      Steve Carson

      Knowing where users focus their gaze in real, virtual, augmented, or mixed environments is essential for creating immersive applications in areas like gaming, product design, psychology, and advertising. In multiuser environments, knowing the 3D point-of-gaze (PoG) of each user is a key enabler for immersive interaction. Creating such systems is difficult if they require detailed modeling of a part of the real world and associating its elements with a virtual world, making systems that construct their own models from sensory data highly desirable. Previous research has used a wide variety of hardware (virtual reality headsets with eye trackers, head/body position trackers, video cameras) and software techniques (structure from motion, simultaneous localization and mapping, and key frames) to solve aspects of determining 3D PoG focused objects at the expense of complex and time-consuming setup. The authors cleverly leverage this previous research, combining results from multiple open-source software packages to derive semantic (that is, user attention to objects in the local environment) 3D gaze mapping for estimating user focus without prior information. The main contributions of this paper are its excellent descriptive figures, well-defined algorithms, logical expository flow, and presentation of the mathematical details of computing 3D PoG from multiple 2D PoGs projected into key frames by Delaunay triangulation. The authors also conduct an experiment that shows their method can compute approximate user fixation time on objects, making it suitable for marketing and advertising research.

      Access critical reviews of Computing literature here

      Become a reviewer for Computing Reviews.

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Conferences
        MobileHCI '19: Proceedings of the 21st International Conference on Human-Computer Interaction with Mobile Devices and Services
        October 2019
        646 pages
        ISBN:9781450368254
        DOI:10.1145/3338286

        Copyright © 2019 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 1 October 2019

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article
        • Research
        • Refereed limited

        Acceptance Rates

        Overall Acceptance Rate202of906submissions,22%

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader