skip to main content
10.1145/3615452.3617941acmconferencesArticle/Chapter ViewAbstractPublication PagesmobicomConference Proceedingsconference-collections
research-article

Toward Scalable and Controllable AR Experimentation

Published:06 October 2023Publication History

ABSTRACT

To understand how well a proposed augmented reality (AR) solution works, existing papers often conducted tailored and isolated evaluations for specific AR tasks, e.g., depth or lighting estimation, and compared them to easy-to-setup baselines, either using datasets or resorting to time-consuming data capturing. Conceptually simple, it can be extremely difficult to evaluate an AR system fairly and in scale to understand its real-world performance. The difficulties arise for three key reasons: lack of control of the physical environment, the time-consuming data capturing, and the difficulties to reproduce baseline results.

This paper presents our design of an AR experimentation platform, ExpAR, aiming to provide scalable and controllable AR experimentation. ExpAR is envisioned to operate as a standalone deployment or a federated platform; in the latter case, AR researchers can contribute physical resources, including scene setup and capturing devices, and allow others to time share these resources. Our design centers around the generic sensing-understanding-rendering pipeline and is driven by the evaluation limitations observed in recent AR systems papers. We demonstrate the feasibility of this vision with a preliminary prototype and our preliminary evaluations suggest the importance of further investigating different device capabilities to stream in 30 FPS.

References

  1. Kittipat Apicharttrisorn, Jiasi Chen, Vyas Sekar, Anthony Rowe, and Srikanth V. Krishnamurthy. 2023. Breaking Edge Shackles: Infrastructure-Free Collaborative Mobile Augmented Reality (SenSys '22). Association for Computing Machinery, New York, NY, USA, 1--15. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Ali J Ben Ali, Marziye Kouroshli, Sofiya Semenova, Zakieh Sadat Hashemifar, Steven Y Ko, and Karthik Dantu. 2022. Edge-SLAM: Edge-Assisted Visual Simultaneous Localization and Mapping. ACM Trans. Embed. Comput. Syst. 22, 1 (Oct. 2022), 1--31.Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Ying Chen, Hazer Inaltekin, and Maria Gorlatova. 2023. AdaptSLAM: Edge-Assisted Adaptive SLAM with Resource Constraints via Uncertainty Minimization. In Proc. IEEE INFOCOM.Google ScholarGoogle ScholarCross RefCross Ref
  4. Ruofei Du, Eric Turner, Maksym Dzitsiuk, Luca Prasso, Ivo Duarte, Jason Dourgarian, Joao Afonso, Jose Pascoal, Josh Gladstone, Nuno Cruces, Shahram Izadi, Adarsh Kowdle, Konstantine Tsotsos, and David Kim. 2020. DepthLab: Real-time 3D Interaction with Depth Maps for Mobile Augmented Reality. In Proceedings of the 33rd Annual ACM Symposium on User Interface Software and Technology (UIST '20).Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Dmitry Duplyakin, Robert Ricci, Aleksander Maricq, Gary Wong, Jonathon Duerig, Eric Eide, Leigh Stoller, Mike Hibler, David Johnson, Kirk Webb, Aditya Akella, Kuangching Wang, Glenn Ricart, Larry Landweber, Chip Elliott, Michael Zink, Emmanuel Cecchet, Snigdhaswin Kar, and Prabodh Mishra. 2019. The Design and Operation of CloudLab. In Proceedings of the USENIX Annual Technical Conference (ATC). 1--14.Google ScholarGoogle Scholar
  6. The Green Planet AR Experience. 2022. https://www.factory42.uk/greenplanetexperience.Google ScholarGoogle Scholar
  7. Hugging Face. 2023. https://huggingface.co/.Google ScholarGoogle Scholar
  8. AWS Device Farm. 2023. https://aws.amazon.com/device-farm/.Google ScholarGoogle Scholar
  9. Immersive Stream for XR (Preview). 2023. https://xr.withgoogle.com/.Google ScholarGoogle Scholar
  10. Yongjie Guan, Xueyu Hou, Nan Wu, Bo Han, and Tao Han. 2022. DeepMix: mobility-aware, lightweight, and hybrid 3D object detection for headsets. In Proceedings of the 20th Annual International Conference on Mobile Systems, Applications and Services (MobiSys '22). 28--41.Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Bo Han, Parth Pathak, Songqing Chen, and Lap-Fai Craig Yu. 2022. CoMIC: A Collaborative Mobile Immersive Computing Infrastructure for Conducting Multi-user XR Research. IEEE Netw. (2022), 1--9.Google ScholarGoogle Scholar
  12. Muhammad Huzaifa, Rishi Desai, Samuel Grayson, Xutao Jiang, Ying Jing, Jae Lee, Fang Lu, Yihan Pang, Joseph Ravichandran, Finn Sinclair, Boyuan Tian, Hengzhi Yuan, Jeffrey Zhang, and Sarita V Adve. 2021. ILLIXR: Enabling End-to-End Extended Reality Research. In 2021 IEEE International Symposium on Workload Characterization (IISWC). 24--38.Google ScholarGoogle Scholar
  13. Hyoukjun Kwon, Krishnakumar Nair, Jamin Seo, Jason Yik, Debabrata Mohapatra, Dongyuan Zhan, Jinook Song, Peter Capak, Peizhao Zhang, Peter Vajda, Colby Banbury, Mark Mazumder, Liangzhen Lai, Ashish Sirasao, Tushar Krishna, Harshit Khaitan, Vikas Chandra, and Vijay Janapa Reddi. 2023. XRBench: An extended reality (XR) machine learning benchmark suite for the metaverse. In Proceedings of Machine Learning and Systems, Vol. 5.Google ScholarGoogle Scholar
  14. Larry L. Peterson and Prof. David Culler. 2002. PlanetLab. https://planetlab.cs.princeton.edu/.Google ScholarGoogle Scholar
  15. Liangkai Liu, Ren Zhong, Aaron Willcock, Nathan Fisher, and Weisong Shi. 2023. An Open Approach to Energy-Efficient Autonomous Mobile Robots citation. In 2023 International Conference on Robotics and Automation (ICRA). IEEE Press, 7 pages.Google ScholarGoogle Scholar
  16. Z Liu, G Lan, J Stojkovic, Y Zhang, C Joe-Wong, and M Gorlatova. 2020. CollabAR: Edge-assisted Collaborative Image Recognition for Mobile Augmented Reality. In 2020 19th ACM/IEEE International Conference on Information Processing in Sensor Networks (IPSN). 301--312.Google ScholarGoogle Scholar
  17. Ajay Mandlekar, Yuke Zhu, Animesh Garg, Jonathan Booher, Max Spero, Albert Tung, Julian Gao, John Emmons, Anchit Gupta, Emre Orbay, Silvio Savarese, and Li Fei-Fei. 2018. ROBOTURK: A Crowd-sourcing Platform for Robotic Skill Learning through Imitation. In Conference on Robot Learning.Google ScholarGoogle Scholar
  18. Adithyavairavan Murali, Tao Chen, Kalyan Vasudev Alwala, Dhiraj Gandhi, Lerrel Pinto, Saurabh Gupta, and Abhinav Gupta. 2019. Py-Robot: An Open-source Robotics Framework for Research and Benchmarking. arXiv preprint arXiv:1906.08236 (2019).Google ScholarGoogle Scholar
  19. Oliver Kroemer. 2023. LoCoBot: An Open Source Low Cost Robot. http://www.locobot.org/.Google ScholarGoogle Scholar
  20. Siddhant Prakash, Alireza Bahremand, Linda D Nguyen, and Robert LiKamWa. 2019. GLEAM: An Illumination Estimation Framework for Real-time Photorealistic Augmented Reality on Mobile Devices. In Proceedings of the 17th Annual International Conference on Mobile Systems, Applications, and Services (Seoul, Republic of Korea) (MobiSys '19). Association for Computing Machinery, New York, NY, USA.Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. SunFounder. 2023. Raspberry Pi Ai Car Kit - PiCar-X. https://www.sunfounder.com/products/picar-x. Accessed: 2023-6-15.Google ScholarGoogle Scholar
  22. Albert Tung, Josiah Wong, Ajay Mandlekar, Roberto Martín-Martín, Yuke Zhu, Li Fei-Fei, and Silvio Savarese. 2021. Learning Multi-Arm Manipulation Through Collaborative Teleoperation. In 2021 IEEE International Conference on Robotics and Automation (ICRA). 9212--9219.Google ScholarGoogle Scholar
  23. Jamie Watson, Mohamed Sayed, Zawar Qureshi, Gabriel J. Brostow, Sara Vicente, Oisin Mac Aodha, and Michael Firman. 2023. Virtual Occlusions Through Implicit Depth. In CVPR.Google ScholarGoogle Scholar
  24. Jingao Xu, Guoxuan Chi, Zheng Yang, Danyang Li, Qian Zhang, Qiang Ma, and Xin Miao. 2021. FollowUpAR: enabling follow-up effects in mobile AR applications. In Proceedings of the 19th Annual International Conference on Mobile Systems, Applications, and Services (MobiSys '21). Association for Computing Machinery, 1--13.Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Juheon Yi and Youngki Lee. 2020. Heimdall: mobile GPU coordination platform for augmented reality applications. In Proceedings of the 26th Annual International Conference on Mobile Computing and Networking (London, United Kingdom) (MobiCom '20, Article 35). Association for Computing Machinery, New York, NY, USA, 1--14.Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. Jinrui Zhang, Huan Yang, Ju Ren, Deyu Zhang, Bangwen He, Ting Cao, Yuanchun Li, Yaoxue Zhang, and Yunxin Liu. 2022. MobiDepth: real-time depth estimation using on-device dual cameras. In Proceedings of the 28th Annual International Conference on Mobile Computing And Networking (MobiCom '22).Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. Wenxiao Zhang, Bo Han, and Pan Hui. 2022. SEAR: Scaling Experiences in Multi-user Augmented Reality. IEEE Trans. Vis. Comput. Graph. 28, 5 (May 2022), 1982--1992.Google ScholarGoogle ScholarCross RefCross Ref
  28. Yunfan Zhang, Tim Scargill, Ashutosh Vaishnav, Gopika Premsankar, Mario Di Francesco, and Maria Gorlatova. 2022. InDepth: Real-time Depth Inpainting for Mobile Augmented Reality. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 6, 1 (March 2022), 1--25.Google ScholarGoogle ScholarDigital LibraryDigital Library
  29. Yiqin Zhao, Sean Fanello, and Tian Guo. 2023. Multi-Camera Lighting Estimation for Photorealistic Front-Facing Mobile Augmented Reality. In Proceedings of the 24th International Workshop on Mobile Computing Systems and Applications (Newport Beach, California) (HotMobile '23). Association for Computing Machinery, New York, NY, USA, 68--73. Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. Yiqin Zhao and Tian Guo. 2020. PointAR: Efficient Lighting Estimation for Mobile Augmented Reality. In The European Conference on Computer Vision (ECCV '20). 678--693.Google ScholarGoogle Scholar
  31. Yiqin Zhao and Tian Guo. 2021. Xihe: a 3D vision-based lighting estimation framework for mobile augmented reality. In Proceedings of the 19th Annual International Conference on Mobile Systems, Applications, and Services (MobiSys '21). 28--40.Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. Yiqin Zhao, Chongyang Ma, Haibin Huang, and Tian Guo. 2022. LITAR: Visually Coherent Lighting for Mobile Augmented Reality. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 6, 3 (Sept. 2022), 1--29.Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Toward Scalable and Controllable AR Experimentation

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Conferences
        ImmerCom '23: Proceedings of the 1st ACM Workshop on Mobile Immersive Computing, Networking, and Systems
        October 2023
        83 pages
        ISBN:9798400703393
        DOI:10.1145/3615452

        Copyright © 2023 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 6 October 2023

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader