skip to main content
10.1145/3517428.3550387acmconferencesArticle/Chapter ViewAbstractPublication PagesassetsConference Proceedingsconference-collections
poster

Vibrotactile Navigation for Visually Impaired People

Published:22 October 2022Publication History

ABSTRACT

One of the largest impediments to autonomous grocery shopping for blind people is navigation. In a still ongoing multi-year research process we followed a user centred approach to explore how visually impaired people can be supported in indoor navigational issues, whether vibrotactiles are the right channel for this, where the actuators should sit on the body, whether the hardware solution would be acceptable for people to use on a daily basis in public and how they feel about the system after three weeks of usage.

References

  1. Apple Support. n.D.. Abrufen von Wegbeschreibungen auf der Apple Watch [retrieve directions on the Apple Watch]. https://support.apple.com/de-de/guide/watch/apdea7480950/watchosGoogle ScholarGoogle Scholar
  2. Sandra Birnstiel, Benedikt Steinmüller, Kerstin Bissinger, Simone Doll-Gerstendörfer, and Stephan Huber. 2019. Gartenfreund: Exploring the Botanical Garden with an Inclusive App. In Proceedings of Mensch und Computer 2019 (Hamburg, Germany). ACM, New York, NY, 499–502. https://doi.org/10.1145/3340764.3344446Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Roger Boldu, Denys J.C. Matthies, Haimo Zhang, and Suranga Nanayakkara. 2020. AiSee: An Assistive Wearable Device to Support Visually Impaired Grocery Shoppers. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 4, 4, Article 119 (dec 2020), 25 pages. https://doi.org/10.1145/3432196Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Vanessa Cobus, Bastian Ehrhardt, Susanne Boll, and Wilko Heuten. 2018. Vibrotactile Alarm Display for Critical Care. In Proceedings of the 7th ACM International Symposium on Pervasive Displays. ACM, New York, NY, 11. https://doi.org/10.1145/3205873.3205886Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Dominick Galli. 2016. Technologieakzeptanz von Tablet-PCs: Ein erweitertes Technologieakzeptanzmodell unter Einbeziehung des TAM 2 und der UTAUT: In Zusammenarbeit mit einem führenden Schweizer Textilmaschinenhersteller [Technology acceptance of tablet PCs: an extended technology acceptance model that includes TAM2 and UTAUT - in cooperation with leaders from Swiss textile machine industry. thesis. ZHAW Zürcher Hochschule für Angewandte Wissenschaften. https://doi.org/10.21256/zhaw-1246Google ScholarGoogle ScholarCross RefCross Ref
  6. João Guerreiro, Eshed Ohn-Bar, Dragan Ahmetovic, Kris Kitani, and Chieko Asakawa. 2018. How Context and User Behavior Affect Indoor Navigation Assistance for Blind People. In Proceedings of the 15th International Web for All Conference (Lyon, France). Association for Computing Machinery, New York, NY, Article 2. https://doi.org/10.1145/3192714.3192829Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Julia Hertel, Armin Schaare, Philipp Feuerbach, Oscar Ariza, and Frank Steinicke. 2019. STIC-Sensory and Tactile Improved Cane. In Proceedings of Mensch und Computer 2019 (Hamburg, Germany). ACM, New York, NY, 765–769. https://doi.org/10.1145/3340764.3344905Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Wilko Heuten, Niels Henze, Susanne Boll, and Martin Pielot. 2008. Tactile wayfinder: a non-visual support system for wayfinding. In Proceedings of the 5th Nordic conference on Human-computer interaction: building bridges (Lund, Sweden). Association for Computing Machinery, New York, NY, 172–181. https://doi.org/10.1145/1463160.1463179Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Karen Holtzblatt and Hugh Beyer. 2017. Contextual Design: Design for Life. Elsevier Science & Technology, San Francisco, CA.Google ScholarGoogle Scholar
  10. Jörn Hurtienne and Anja Naumann. 2010. QUESI—A questionnaire for measuring the subjective consequences of intuitive use. https://www.joernhurtienne.com/Publications_files/Abstract.pdfGoogle ScholarGoogle Scholar
  11. David Andrew Jeffrey. 2016. Testing the technology acceptance model 3 (tam 3) with the inclusion of change fatigue and overload, in the context of faculty from seventh-day adventist universities: A revised model. Ph. D. Dissertation. Andrews University.Google ScholarGoogle Scholar
  12. Vladimir Kulyukin and Aliasgar Kutiyanawala. 2010. From ShopTalk to ShopMobile: vision-based barcode scanning with mobile phones for independent blind grocery shopping., 5 pages. https://www.resna.org/sites/default/files/legacy/conference/proceedings/2010/TechnologyCognitiveSensory/KulyukinV1.htmlGoogle ScholarGoogle Scholar
  13. Jihyun Lee, Jinsol Kim, and Hyunggu Jung. 2020. Challenges and Design Opportunities for Easy, Economical, and Accessible Offline Shoppers with Visual Impairments. In Proceedings of the 2020 Symposium on Emerging Research from Asia and on Asian Contexts and Cultures (Honolulu, HI, USA) (AsianCHI ’20). Association for Computing Machinery, New York, NY, USA, 69–72. https://doi.org/10.1145/3391203.3391223Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Steve Mann, Jason Huang, Ryan Janzen, Raymond Lo, Valmiki Rampersad, Alexander Chen, and Taqveer Doha. 2011. Blind navigation with a wearable range camera and vibrotactile helmet. In Proceedings of the 19th ACM international conference on Multimedia (Scottsdale, AZ). Association for Computing Machinery, New York, NY, 1325–1328. https://doi.org/10.1145/2072298.2072005Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Anja Naumann and Jörn Hurtienne. 2010. Benchmarks for intuitive interaction with mobile devices. In Proceedings of the 12th international conference on Human computer interaction with mobile devices and services(Lisbon, Portugal). ACM, New York, NY, 401–402.Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Marko Periša, Dragan Peraković, Ivan Cvitić, and Marko Krstić. 2022. Innovative ecosystem for informing visual impaired person in smart shopping environment: InnIoTShop. Wireless Networks 28, 1 (2022), 469–479. https://doi.org/10.1007/s11276-021-02591-5Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Pablo-Alejandro Quinones, Tammy Greene, Rayoung Yang, and Mark Newman. 2011. Supporting visually impaired navigation: a needs-finding study. In CHI ’11 Extended Abstracts on Human Factors in Computing Systems (Vancouver, Canada). Association for Computing Machinery, New York, NY, 1645–1650. https://doi.org/10.1145/1979742.1979822Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Daisuke Sato, Uran Oh, Kakuya Naito, Hironobu Takagi, Kris Kitani, and Chieko Asakawa. 2017. NavCog3: An Evaluation of a Smartphone-Based Blind Indoor Navigation Assistant with Semantic Features in a Large-Scale Environment. In Proceedings of the 19th International ACM SIGACCESS Conference on Computers and Accessibility(Baltimore, MD). Association for Computing Machinery, New York, NY, 270–279. https://doi.org/10.1145/3132525.3132535Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Stefanie Schaack, George Chernyshov, Kirill Ragozin, Benjamin Tag, Roshan Peiris, and Kai Kunze. 2019. Haptic Collar: Vibrotactile Feedback around the Neck for Guidance Applications. In Proceedings of the 10th Augmented Human International Conference 2019 (Reims, France) (AH2019). Association for Computing Machinery, New York, NY, USA, Article 12, 4 pages. https://doi.org/10.1145/3311823.3311840Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Stefano Scheggi, A Talarico, and Domenico Prattichizzo. 2014. A remote guidance system for blind and visually impaired people via vibrotactile haptic feedback. In 22nd Mediterranean Conference on Control and Automation (Palermo, Italy). IEEE, Piscataway, NJ, 20–23. https://doi.org/10.1109/MED.2014.6961320Google ScholarGoogle ScholarCross RefCross Ref
  21. Wakana Sugiyama. 2019. Voice guidance in Maps, built for people with impaired vision. https://blog.google/products/maps/better-maps-for-people-with-vision-impairments/Google ScholarGoogle Scholar
  22. Takeshi Tanabe, Kiyohiko Nunokawa, Kouki Doi, and Shuichi Ino. 2022. Training System for White Cane Technique Using Illusory Pulling Cues Induced by Asymmetric Vibrations. IEEE Transactions on Neural Systems and Rehabilitation Engineering 30 (2022), 305–313. https://doi.org/10.1109/TNSRE.2022.3148770Google ScholarGoogle ScholarCross RefCross Ref
  23. Kaveri Thakoor, Nii Mante, Carey Zhang, Christian Siagian, James Weiland, Laurent Itti, and Gérard Medioni. 2015. A System for Assisting the Visually Impaired in Localization and Grasp of Desired Objects. In Computer Vision - ECCV 2014 Workshops, Lourdes Agapito, Michael M. Bronstein, and Carsten Rother (Eds.). Springer International Publishing, Heidelberg, Germany, 643–657. https://doi.org/10.1007/978-3-319-16199-0_45Google ScholarGoogle ScholarCross RefCross Ref
  24. Robert Tscharn, Tom Außenhofer, Dimitri Reisler, and Jörn Hurtienne. 2016. Turn left after the heater: Landmark navigation for visually impaired users. In Proceedings of the 18th International ACM SIGACCESS Conference on Computers and Accessibility(Reno, NV). ACM, New York, NY, 295–296. https://doi.org/10.1145/2982142.2982195Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Viswanath Venkatesh, Michael G Morris, Gordon B Davis, and Fred D Davis. 2003. User acceptance of information technology: Toward a unified view. MIS quarterly 27, 3 (2003), 425–478. https://doi.org/10.2307/30036540Google ScholarGoogle ScholarCross RefCross Ref
  26. Christopher D Wickens. 2008. Multiple resources and mental workload. Human factors 50, 3 (2008), 449–455.Google ScholarGoogle Scholar
  27. Chien Wen Yuan, Benjamin V. Hanrahan, Sooyeon Lee, Mary Beth Rosson, and John M. Carroll. 2017. I Didn’t Know That You Knew I Knew: Collaborative Shopping Practices between People with Visual Impairment and People with Vision. Proceedings of the ACM in Human-Computer Interaction 1, CSCW, Article 118 (dec 2017), 18 pages. https://doi.org/10.1145/3134753Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. Jialong Zhou, Chen Qiu, Bin Guo, and Zhiwen Yu. 2022. The challenges and future of navigation for people with disabilities. XRDS 28, 3 (2022), 32–36. https://doi.org/10.1145/3522688Google ScholarGoogle ScholarDigital LibraryDigital Library
  29. Peter A. Zientara, Sooyeon Lee, Gus H. Smith, Rorry Brenner, Laurent Itti, Mary B. Rosson, John M. Carroll, Kevin M. Irick, and Vijaykrishnan Narayanan. 2017. Third Eye: A Shopping Assistant for the Visually Impaired. Computer 50, 2 (2017), 16–24. https://doi.org/10.1109/MC.2017.36Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Vibrotactile Navigation for Visually Impaired People

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Conferences
        ASSETS '22: Proceedings of the 24th International ACM SIGACCESS Conference on Computers and Accessibility
        October 2022
        902 pages
        ISBN:9781450392587
        DOI:10.1145/3517428

        Copyright © 2022 Owner/Author

        Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 22 October 2022

        Check for updates

        Qualifiers

        • poster
        • Research
        • Refereed limited

        Acceptance Rates

        ASSETS '22 Paper Acceptance Rate35of132submissions,27%Overall Acceptance Rate436of1,556submissions,28%

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      HTML Format

      View this article in HTML Format .

      View HTML Format