Abstract
In multi-channel design, the combination of visual-audio dual channels is the most widely used way. This article elaborated on the characteristics of visual and audio channel, clarified the complementary relationship between visual and audio channels, and analysed the combination methods of visual and audio channels. EEG/ERP technology was used to carry out experiments and some alarm information designs of a real-time monitoring software system were used as an example to compare the cognitive differences caused by the combination of visual and audio channels’ combination. Through the quantitative analysis of indicators such as reaction time, brain wave peak value, and latency, the interface information design and evaluation methods for visual-audio dual channels were developed.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Li, T., Wang, D., Peng, C., Yu, C., Zhang, Y.: Speed-accuracy tradeoff of fingertip force control with visual/audio/haptic feedback. Int. J. Hum. Comput. Stud. 110, 33–44 (2018)
Grega, J., Christina, D., Jaka, S.: A user study of auditory, head-up and multi-modal displays in vehicles. Appl. Ergon., Part A 46, 184–192 (2015)
Andrea, B., Ian, O., Dong, S.K.: Counting clicks and beeps: exploring numerosity based haptic and audio PIN entry. Interact. Comput. 24(5), 409–422 (2012)
Hans-Jörg, S., Steffen, H.: Preset-based generation and exploration of visualization designs. J. Vis. Lang. Comput., Part A 31, 9–29 (2015)
Abbate, A.J., Bass, E.J.: A formal methods approach to semiotic engineering. Int. J. Hum. Comput. Stud. 115, 20–39 (2018)
Hooten, E.R., Hayes, S.T., Julie, A.A.: Communicative modalities for mobile device interaction. Int. J. Hum. Comput. Stud. 71(10), 988–1002 (2013)
Antonio, M.R.: A multimedia ontology model based on linguistic properties and audio-visual features. Inf. Sci. 277, 234–246 (2014)
Timothy, E.R., Ian, P., David, D.: Toward mobile entertainment: A paradigm for narrative-based audio only games. Sci. Comput. Program. 67(1), 76–90 (2007)
Gabrielle, L.H., Eric, L., Cara, E.S.: Effects of augmentative visual training on audio-motor mapping. Hum. Mov. Sci. 35, 145–155 (2014)
Jeffrey Jr., L.C., Robert, H.G., Brian, D.S.: The VERITAS facility: a virtual environment platform for human performance research. IFAC Proc. 46(15), 357–362 (2013)
Acknowledgments
This paper is supported by Science and Technology on Avionics Integration Laboratory and Aeronautical Science Fund (No. 20165569019).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer International Publishing AG, part of Springer Nature
About this paper
Cite this paper
Xi, W., Zhou, L., Chen, H., Ma, J., Chen, Y. (2019). Research on the Brain Mechanism of Visual-Audio Interface Channel Modes Affecting User Cognition. In: Ayaz, H., Mazur, L. (eds) Advances in Neuroergonomics and Cognitive Engineering. AHFE 2018. Advances in Intelligent Systems and Computing, vol 775. Springer, Cham. https://doi.org/10.1007/978-3-319-94866-9_20
Download citation
DOI: https://doi.org/10.1007/978-3-319-94866-9_20
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-94865-2
Online ISBN: 978-3-319-94866-9
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)