Next Article in Journal
Visual Feedback Control of a Rat Ankle Angle Using a Wirelessly Powered Two-Channel Neurostimulator
Previous Article in Journal
Latency Compensated Visual-Inertial Odometry for Agile Autonomous Flight
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Closed-Loop Attention Restoration Theory for Virtual Reality-Based Attentional Engagement Enhancement

Department of Micro/nano Electronics, Shanghai Jiao Tong University, Shanghai 200240, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(8), 2208; https://doi.org/10.3390/s20082208
Submission received: 1 March 2020 / Revised: 10 April 2020 / Accepted: 12 April 2020 / Published: 14 April 2020
(This article belongs to the Section Physical Sensors)

Abstract

:
Today, as media and technology multitasking becomes pervasive, the majority of young people face a challenge regarding their attentional engagement (that is, how well their attention can be maintained). While various approaches to improve attentional engagement exist, it is difficult to produce an effect in younger people, due to the inadequate attraction of these approaches themselves. Here, we show that a single 30-min engagement with an attention restoration theory (ART)-inspired closed-loop software program (Virtual ART) delivered on a consumer-friendly virtual reality head-mounted display (VR-HMD) could lead to improvements in both general attention level and the depth of engagement in young university students. These improvements were associated with positive changes in both behavioral (response time and response time variability) and key electroencephalography (EEG)-based neural metrics (frontal midline theta inter-trial coherence and parietal event-related potential P3b). All the results were based on the comparison of the standard Virtual ART tasks (control group, n = 15) and closed-loop Virtual ART tasks (treatment group, n = 15). This study provides the first case of EEG evidence of a VR-HMD-based closed-loop ART intervention generating enhanced attentional engagement.

1. Introduction

Today, as media and technology multitasking becomes pervasive, the majority of young people face a challenge regarding their attentional engagement (that is, how well their attention can be maintained) [1]. A study of min-by-min observations showed that the typical university student could not focus on their work for more than 3 to 5 min even during a short 15-min class session, because their minds were still thinking about what might be occurring in virtual worlds, coaxing them to get back to the smartphone, tablet, or laptop to “check in” [2]. Such constant attempts to multitask not only create challenges to the development of their cognitive control functions [3,4,5], but also have a negative impact on their real-world activities, including bad sleep [6], poorer school/workplace performance [7,8], and an increased level of stress and anxiety [9,10]. Therefore, there exists a need for new methods to enhance attentional engagement. A recent study showed that six weeks (20–30 min per week) of training with a closed-loop digital meditation software delivered on a smartphone/tablet improved attentional engagement in young adults [11]. Here, “closed-loop” refers to a novel meditation training approach where the training difficulty can be adjusted in a rapid manner (e.g., every dozen seconds), according to the participant’s performance. Actually, apart from meditation, attempts to boost attentional engagement in a drug-free manner also involved physical exercise, cognitive training, video games, brain stimulation, and exposure to nature [12]. All of these approaches are building on the foundation that our brain modifies itself—a phenomenon known as neuroplasticity in the context of cognitive neuroscience [12,13]. However, unlike those active interventions (e.g., physical exercise, cognitive training, video games, brain stimulation) that involve hard work of the participant, exposure to nature is an approach that is quite the opposite.
Exposure to nature is a kind of passive intervention. Walking and running in real nature, or passive viewing of natural scenes through images and videos are the widely seen forms in this approach [13]. Essentially, there are no demanding attentional tasks involved. Since the 1980s, accumulating evidence shows us that nature surroundings have restorative or stress-reducing effects [14,15,16,17]. Specifically, Kaplan et al. who developed attention restoration theory (ART) suggested that natural environments have properties that attract involuntary attention and, thus, allow a depleted directed attention capacity to recover so that cognitive fatigue can be reduced [18]. Here, the term “natural environments” must consist of the following four key components [19,20,21]: (1) “being away”, which refers to the sense of being psychologically detached from present worries and demands; (2) “soft fascination”, which involves the fascinating objects that capture one’s attention in a bottom-up fashion and generate minimal top-down responses; (3) “extent”, which refers to the degree that people feel immersive and engaging; (4) “compatibility”, which refers to our intrinsic motivation to stay in certain environment.
In a more recent review article about ART, Garside et al. showed some empirical evidence of attention-based benefits, such as improved working memory in both healthy people and those with psychological conditions, which were generated from “before and after” behavioral measurements [13]. However, in this review article, there were also some studies with mixed findings using different outcome measures of attention, as well as one that did not offer support for ART with outperformed control groups instead of the intervention groups [22]. It is likely that differences in the quality of the four aforementioned components themselves contribute to different outcomes [23]. For example, the usage of artificial settings, i.e., urban parks and gardens [24,25], apparently degraded the sense of “being away” and “soft fascination” if compared to real natural scene. Moreover, the awareness of “extent” makes it hard to believe that people would feel immersive and engaging by just viewing scenery images and videos. Furthermore, the recruited healthy population probably would not have a strong and “compatible” motivation if compared to the people who are suffering from cognitive fatigue. More importantly, none of them (all ART-related studies in Reference [13]) reported neural evidence. Thus, attentional engagement, a state that describes how well the attention-based benefits can be maintained, remains unclear.
Unlike previous ART studies that adopted ready-to-use natural resources, our approach involved designing, developing, and testing an ART-inspired virtual reality (VR) software program that integrates key restorative components of traditional ART with a real-time electroencephalography (EEG)-based closed-loop algorithm to monitor their vigilance levels—an indicator of hard mental work over prolonged periods of time [26]. When engaging with this program, users were firstly instructed to close their eyes for 3 min to get their EEG baseline parameter, and then they began “exposure to nature” while their vigilance levels were simultaneously monitored using EEG. They were not given explicit instructions on the best strategy via which to achieve a low vigilance level, but they did know that their EEG was being recorded and that it was being used to change the natural surroundings as a reward-based feedback for their vigilance levels (for example, fog disappears when low vigilance level detected). We define this closed-loop approach as CL-ART or “extent”-enhanced ART, which allows the interaction between human and natural environment in a personalized manner and offers regular feedback on the effectiveness of relaxing the mind from top-down demands through engagement in a strong bottom-up driven activity. In the context of cognitive neuroscience, attention is one of the three cognitive control abilities (the other two are working memory and goal management) [12]. Thus, the term “top-down” here refers to internal guidance of attention based on prior knowledge, willful plans, and current goals, while “bottom-up” refers to attentional guidance purely by externally driven factors to stimuli that are salient because of their inherent properties relative to the background [27]. Given that the core idea of ART is exposure to nature, ART-induced attention refers to bottom-up attention.
VR is a communication medium that leads an individual to perceive experiences as if they were physically present in that environment [28]. Such VR experiences not only provide easier access to difficult-to-arrange real-world situations, but also allow brain activity to be recorded in a controlled environment [29,30]. Therefore, VR environments are increasingly being used by researchers to simulate social interactions and natural events [31]. This trend is especially clear since the cost of VR was reduced from tens of thousands of United States (US) dollars a decade ago to the current hundreds of dollars with the development of consumer-friendly VR head-mounted display (HMD) technology in 2016 [32].
VR environments are designed using three technologies: non-immersive VR [33], semi-immersive VR [34], and immersive VR (IVR) [35]. There are two commonly used forms of IVR [35]: cave automatic virtual environments (CAVEs) and HMDs. A CAVE is a specially designed room in which the walls, ceiling, and/or floor are covered with a screen that can project virtual images or videos. An HMD is a VR headset that positions two small screens in front of both eyes, completely blocking out the physical world including the user’s body. All current state-of-the-art IVRs adopt consumer-friendly HMDs, including personal computer (PC)-powered HMDs (i.e., HTC Vive™), smartphone-based HMDs (i.e., Samsung Gear™), and all-in-one HMDs (i.e., Oculus Quest™), in order to achieve immersive effects in a manner that is both simple and inexpensive. More importantly, given that external factors may distract attention [36], VR-HMD has the inherent advantage of effectively limiting influences of external distraction on attention if compared to a conventional two-dimensional (2D) platform [37], highlighting the necessity of using VR-HMD in the current study.
The goal of this study was to assess both behavioral performance and EEG-based neural metrics on a perceptual discrimination task that was executed before and after two types of Virtual ART interventions: standard ART (ST-ART) and CL-ART. The two kinds of ART interventions were developed by our laboratory and delivered using a consumer-friendly VR-HMD platform, to quantitatively evaluate the impact of the “extent” component on attentional engagement with a high degree of ecological validity and experimental control. Given the higher degree of engagement that CL-ART is expected to offer (compared to ST-ART), we hypothesized that participants engaged in the CL-ART tasks would generate better post-intervention behavioral performance, as well as neural correlates associated with heightened attentional engagement.

2. Materials and Methods

2.1. Virtual ART Tasks

Virtual ART is a VR-HMD software developed at Shanghai Jiao Tong University (SJTU)’s bio-circuits and system laboratory to assess the effectiveness of ART with a high degree of ecological validity and experimental control. There were two types of virtual tasks in this study: ST-ART (control group) and CL-ART. Both were developed from the principles of Kaplan’s theory and delivered in HTC Vive™—a flagship consumer-friendly VR-HMD platform powered by an NVIDIA GeForce GTX 1070 graphic card. As can be seen in Table 1, to have a fair comparison of both kinds of tasks, all set-ups were in their best form. For example, head and limb movements were allowed in ST-ART tasks to achieve broader field of view and naturalistic interaction in the context of VR, while such movements were not allowed in CL-ART tasks to avoid the negative impact of serious motion artefacts on EEG. In the meanwhile, except for “extent”, we carefully kept other components identical. The duration of both tasks was the same 30 min. More technical details can be found in Figure A1. A few screenshots of ST-ART are presented in Figure A2. Additionally, a helpful video demo link for CL-ART can be found in Appendix A as well.
For CL-ART, EEG data were collected through an eight-channel wireless EEG recording device (StarStim 8™, Neuroelectrics Inc, Barcelona, Spain), which uses a high-resolution, high-speed analog-to-digital converter (24 bit at 500 sampling rate), and supports Bluetooth connection. The conventional wet electrodes were used and placed at O1, Oz, and O2 regions, which are proven locations that are highly correlated with the user’s vigilance level in the context of a real-time and real-world scenario [38,39]. The ground and reference electrodes were connected together and placed on the right earlobe using an ear clip. An external electrode was placed below the lower eyelid to record eye movements. The collected EEG data were firstly filtered using a 4–30-Hz bandpass filter and averaged across O1, Oz, and O2 channels. Then, two types of EEG-based feedback (as shown in Figure 1) were provided to participants: (1) real-time feedback, which was based on the values of relative α band power (RBP(α), see Equation (A1) in Appendix A) calculated every 2 s, indicating the dynamics of within-task vigilance level, and (2) punctuated feedback, which was based on the total counts of the values of 2-s RBP(α) being larger than 0.35 (35%) every 1 min, indicating the vigilance level between tasks. The specific thresholds between tasks were weighted values of the highest counts of RBP(α) > 0.35, which were determined using min-by-min observations during the initial 3-min EEG baseline testing session. The weights were 0.4, 0.6, and 0.8 for tasks 1, 2 and 3 respectively. Here, the values of weights and the value of 0.35 were based on feedback from our pilot testing, in which the principle of selecting these values was that the participants should not consider it too easy or too difficult. The lab streaming layer protocol [40] was used to synchronize the vigilance level and the Unreal Engine-based virtual natural scenes.

2.2. Outcome Measures

As shown in Table 2, behavioral performance and EEG-based neural metrics captured during a perceptual discrimination task (visual oddball) that was executed before and after the two ART interventions were utilized as our outcome measures to assess the potential benefits of attention. Specifically, the commonly used response time (RT) [41,42,43] and event-related potential (ERP: P3b latency) [44] were used to evaluate the general benefits of attention; then, the response time across trials (RTVar [11,45]) and inter-trial coherence at the frontal midline theta band (ITC(θ) [11,46]) were used to assess the attentional engagement. Finally, the long-range frontal-posterior inter-electrode coherence at the theta band (IEC(θ)) was used to investigate the brain functional connectivity [42].

2.2.1. Visual Oddball

For oddball testing, we used the same EEG cap (i.e., StarStim 8™). During the experiment, participants were instructed to keep their eyes fixated on a central fixation cross on the computer screen and keep their dominant-hand middle fingers on a home position on the keyboard until the appearance of a given stimulus type (target/distractor). Upon the appearance of these stimuli, participants were instructed to press the “enter” key using their middle fingers if it was a target (horizontal zebra stripes on the center of the screen; see Figure A3 in Appendix A) and to not release their middle fingers from the home position if it was a distractor (vertical zebra stripes on the center of the screen). In this kind of perceptual discrimination paradigm, the participant’s voluntary attention is directed to a rarely presented target stimulus while their EEG response to the target stimulus is simultaneously recorded; therefore, the neural evidence of the restored voluntary attention can be investigated. Each oddball task contained 300 trials and a ratio of 1/4 of target/distractor stimulus. The two trial types were presented randomly with no more than four consecutive trial types of either kind in a row. The inter-trial interval of 1500 ± 500 ms and a stimulus duration of 100 ms were used. Thus, the total time for each oddball task was approximately 10 min.

2.2.2. EEG Data Pre-Processing

A low-pass filter with a cutoff frequency of 30 Hz and a high-pass filter with a cutoff frequency of 0.5 Hz were applied to remove power line noise and the direct current drift, respectively. The filtered EEG signals were then corrected using the mean of each channel (including Ext, Fpz, Fz, Cz, Pz, O1, Oz, and O2), and decomposed into eight independent brain sources by independent component analysis (ICA). The prominent artefactual components, such as eye blinks, eye movements, and muscle activity were removed by using ADJUST version 1.1 (an automatic algorithm for ICA-based EEG artefact removal [47]). Next, the target epochs of −1000 ms to +1000 ms were created and further cleaned of excessive peak-to-peak deflections, amplifier clippings, and other artefacts, using a voltage threshold of 100 μV. All neural metrics were calculated using custom Matlab™ scripts and EEGLab v14.1.2. (an open-source Matlab plugin developed by Swartz Center for Computational Neuroscience, La Jolla, CA, USA; http://www.sccn.ucsd.edu/eeglab).

2.2.3. P3b

ERP time-locked to target trials were generated from the pre-processed EEG data recorded from participants while they performed the oddball task. All ERPs were baseline-corrected using a −200 to 0 ms time period, with the window of interest interrogated being 250–600 ms post-stimulus for P3b—an ERP component which is hypothesized to reflect the allocation of attention resources [44], and which was shown highly correlated with motor or cognitive action, such as pressing a button [48]. Given our focus on RT-based metrics for behavioral data, we focused on P3b latency in the Pz channel, which is the location where the P3b is commonly reported to reach its maximum amplitude [49].

2.2.4. ITC(θ) and Long-Range IEC(θ)

ITC is a measure that reflects the extent to which synchronization occurs from trial to trial in EEG at a particular frequency [46]. The frontal midline ITC in the theta frequency band, ITC(θ), is a widely used measure of electrophysiological response consistency [11,50]. Thus, we used it as our neural metric of the degree of attentional engagement. Here, the theta frequency band (4–7 Hz) at the Fz channel and the phase-locking value (PLVITC) were used as the input and output, respectively, in ITC(θ) analysis. The values of PLVITC ranged anywhere between 0 and 1. A larger value denotes perfect phase-locking synchronized across trials and, thus, a higher degree of attentional engagement.
IEC is a commonly used measure of phase consistency across electrodes [51]. Here, we calculated the PLV of the theta frequency for Fz–Pz coherence (PLVIEC) as our neural metric of brain frontal-posterior functional connectivity. Similarly, the values of PLVIEC ranged anywhere between 0 and 1. A larger value denotes perfect phase-locking synchronized across the Fz–Pz area and, thus, stronger brain frontal-posterior connectivity. All ITC(θ) and IEC(θ) values were baseline-corrected using a −200 to 0 ms time period, with the window of interest interrogated being 0–700 ms post-stimulus, as average RTs occur <700 ms.

2.3. Participants, Procedure, and Statistical Analysis

A total of 50 interested students of SJTU signed up this study through online advertisements (where we named the designed virtual environment as “the deep space of SJTU for mind relaxation”). All of them were screened online for moderate to high trait anxiety (a score greater than 40 on the trait-anxiety subscale of the State-Trait Anxiety Inventory questionnaire, STAI [52]—a standard clinical measure of trait and state anxiety). With two time points (“before and after” the ART interventions), two groups (ST-ART/CL-ART), five measurements (RT, RTVar, P3b latency, ITC(θ), and IEC(θ)), and assuming a moderate repeated measures correlation (r = 0.5) and 0.05 α level, we calculated that n = 30 would yield 93% power to detect a change with a medium effect size (0.5). This effect size is not uncommon, as shown in Reference [11].
Thus, a reasonable sample size (n = 30, 21.3 ± 1.58 years of age, 13 females) who met the trait anxiety criterion (STAI score: 48.1 ± 6.05) were further invited to schedule a lab session. All participants had normal or corrected-to-normal vision, had no history of stroke, traumatic brain injury, or psychiatric illness, and were not taking psychotropic medication. We only enrolled participants who did not have a history of VR experience to balance the impact of familiarity of user interface on task performance and neural metrics. All participants were paid $15/h for their participation and gave written informed consent before participation.
The recruited 30 students were randomized to either the ST-ART group (control group, n = 15) or the CL-ART group (treatment group, n = 15) using a random number generator (seed =123, https://stattrek.com/statistics/random-number-generator.aspx). Prior to virtual ART tasks, students completed the consent document and state-anxiety subscale of STAI. After that, they were equipped with an EEG cap to practice for 2 min before doing the real visual oddball task. Immediately after the oddball task, they put on the VR headset and were instructed to close their eyes for 3 min (i.e., sham/real baseline test for ST-/CL-ART tasks). Then, the software guided them in a 5-min practice session before experiencing either 30-min ST-ART or CL-ART tasks. After the ART tasks, they were instructed to take a 10-min break while the VR headset was taken off. Immediately after the break time, they did the oddball task again.
All behavioral and neural data were analyzed using standard one-way repeated ANOVA with ART task types (ST/CL) as the between-subject factor, where behavioral and neural data refer to the difference before and after the same ART type. We used the difference (diff) of “before–after” comparison if there was an expected decreased trend or “after–before” comparison if there was an expected increased trend (as shown in Equations (1) and (2)) to keep showing a positive difference. Paired t-tests were used to further compare the impact of the within-subject factor (before/after the intervention) on participant performance and neural response. All statistical analyses were done using SPSS 19.0 with an α level of 0.05.
Metrics _ Diff = M e t r i c s b e f o r e M e t r i c s a f t e r ,   w h e r e   M e t r i c s   { R T ,   R T V a r ,   P 3 b   l a t e n c y }
Metrics _ Diff = M e t r i c s a f t e r M e t r i c s b e f o r e ,   w h e r e   M e t r i c s { I T C ( θ ) ,   I E C ( θ ) }

2.4. Study Design

The study was designed in a double-blinded manner. Double-blinding began at the point of recruitment, when all participants were informed that they were being recruited for a study designed to test the efficacy of software interventions for mind relaxation. Thus, participants in both the CL-ART group and the ST-ART group thought that they were part of an active treatment group. One study coordinator (S.H.Z.) was in charge of the group assignments. M.Y.G. who collected data was blind to group assignment. All data analysis was done by Z.K. and G.L.

3. Results

First of all, the results of anxiety level were analyzed, in order to investigate the consistency of “compatibility” element between ST-ART and CL-ART group. Then, as shown in Table 2, the attention levels associated with behavioral metric, RT, and neural metric, P3b latency, are presented. Next, we analyzed the attentional engagement using the behavioral metric, RTVar, and neural metric, ITC(θ). It is important to note that the correlates of behavioral and neural metrics are presented. Finally, IEC(θ) was analyzed to explore the brain functional connectivity between the two groups before and after the 30-min VR experience.

3.1. Anxiety Level

To have a fair comparison, first of all, we analyzed the trait and state anxiety STAI scores between the ST-ART and CL-ART group. As shown in Figure 2a, there was no significant difference in both types of scores between the two groups (trait anxiety scores: F(1, 29) = 0.155, p = 0.697, mean (M) = 46.067 ± 1.714 for ST-ART and M = 47.067 ± 1.873 for CL-ART; state anxiety scores: F(1, 29) = 1.012, p = 0.323, M = 38 ± 2.434 for ST-ART and M = 41.2 ± 2.048 for CL-ART), indicating the same initial anxiety level or motivation (see “compatibility” element in Table 1) for each group. Moreover, we found that the on-site state anxiety scores highly correlated with trait anxiety scores (p = 0.016, r = 0.434, see Figure 2b), indicating the good consistency of the earlier sign-up session and later invited lab session.

3.2. Attention Level

For RT, we found a significant difference in P3b latency (F(1, 29) = 6.581, p = 0.016) between groups. The paired t-test of P3b latency before and after CL-ART intervention (t(14) = 3.481, p = 0.004) further shows that participants responded, on average, 62 ms faster in the post-CL-ART oddball task (M = 377.733 ± 17.116 ms) than in the pre-CL-ART condition (M = 439.867 ± 8.964 ms), indicating the neural evidence of enhanced attention level. This trend is clearly shown in Figure 3a and b by comparing the grand average of P3b between ST-ART and CL-ART groups. Furthermore, as shown in Figure 3c, we found a significant correlation between RT difference and P3b difference (p = 0.002, r = 0.541), indicating the highly correlated behavioral and neural evidence.

3.3. Attentional Engagement

We found a significant difference in RTVar data between groups (F(1, 29) = 4.256, p = 0.048), with M = 2.278 ± 10.532 ms for ST-ART and M = 25.932 ± 4.533 ms for CL-ART. The paired t-test (t(14) = 5.721, p < 0.001) further reveals that participants responded with less variability in RT in the post-CL-ART oddball task (M = 50.929 ± 3.940 ms) if compared to the pre-CL-ART condition (M = 76.862 ± 4.335 ms), as shown in Figure 4a. Moreover, we observed that there was a most prominent difference in averaged values at 300–400 ms post-stimulus ITC(θ) between groups (F(1, 29) = 14.215, p = 0.001). Such a group difference in ITC(θ), as shown in Figure 4b, was found by comparing the 100-ms bins one by one in our time window of interest (see Table 3). These behavioral and neural findings indicate that CL-ART may be indeed superior to ST-ART for improving attentional engagement. The highly correlated RTVar difference and ITC(θ) difference (p = 0.01, r = 0.463; see Figure 4c) further reveals the robustness of these findings.

3.4. Brain Functional Connectivity (IEC(θ))

Similar to ITC(θ), we investigated each 100-ms bin in our time window of interest (see Table 4); however, no significant difference in IEC(θ) was observed, indicating that there was no prominent improvement in frontal-posterior brain functional connectivity following the two kinds of ART experience.

4. Discussion

4.1. Comparison to Prior Work

Previously, the enhancement of general attention level was observed using traditional ART intervention and behavior-based metrics within a reported duration of as short as 15 min [53] or as long as weeks or years [54,55]. The upgrade from traditional ART intervention to this closed-loop approach aims to study EEG-based neural metrics and further verify the feasibility of improving attentional engagement by enhancing one of the restorative components, “extent”.
The present findings reveal that the attentional engagement is improved in young adults when a single 30-min ART-inspired closed-loop intervention is used as opposed to a standard ART intervention, evidenced via both behavioral and EEG-based neural metrics. This more engaging closed-loop approach was actually accompanied by improved sensitivity and greater fidelity in assessing cognitive abilities [56]. This was especially important when these abilities were often assessed with validated pencil-and-paper approaches or, now more commonly with these same paradigms deployed on either desktop or laptop computers, in a format that regularly reveals low test sensitivity in children [56,57]. In addition, the neural evidence-based attentional benefits occurring in such a short time period were observed in another study [58], in which the author reported that feedback training of α rhythms resulted in fMRI-proved changes in brain networks after a single 30-min session (i.e., the exactly same time period of intervention with current study). However, the difference between the current study and Reference [58] is that (1) we used ART-based natural scenes as a feedback paradigm rather than effort-demanding racing game, and (2) we used EEG to investigate the trial-to-trial neural metrics in attentional engagement instead of fMRI-based changes in brain structure or network. Of course, these approaches are not mutually exclusive; they are complementary, and participants will likely achieve the most beneficial outcomes if they use them concurrently. For example, the brain functional connectivity measured by IEC(θ) could perhaps be improved by using more effort-demanding tasks instead of ART-like natural scenes.

4.2. Significance for Future Study

The present findings demonstrate that the closed-loop personalized Virtual ART intervention achieved a higher degree of engagement compared to conventional Virtual ART intervention, highlighting the importance of real-time neurofeedback between human and virtual environment. This exploratory study is especially pertinent given that Jyoti et al. recently proposed the next generation of cognitive training approach that incorporates rapid, real-time, performance-driven, adaptive task challenges and performance feedback [59]. More importantly, with the development of miniatured bio-sensors and wireless VR-HMD, a unique and novel commercial VR platform [60], which combines EEG recording and eye-tracking technologies, emerged to facilitate Jyoti’s idea. Although this study was conducted in healthy adults only, related work demonstrated that populations with physical and intellectual disabilities can benefit from VR in capturing and maintaining attention [61,62]. Thus, the findings presented here have practical implications for assessing attention-related diseases, including sensory processing dysfunction [63], attention deficit hyperactivity disorder [64], and mild impaired cognition [65] in the context of clinical research. Given that previous work [63,64,65] showed the feasibility of using 2D-based attention assessments and interventions for these diseases, we surmise that the use of a more engaging and enjoyable VR-HMD may indeed follow a similar or even better results.

4.3. Limitations and Conclusion

While the improvements on attention level and attentional engagement were observed during a 10-min perceptual discrimination task (visual oddball) carried out 10 min after the CL-ART intervention, no conclusions can be made regarding the long-lasting effects of the observed attentional benefits. Moreover, without accompanying MRI neuroimaging to assess the neural mechanisms of the observed intervention effects, it is not possible to make strong conclusions as to the neural underpinnings of the CL-ART effects. Therefore, follow-up studies driven by the positive findings reported here are planned with (1) larger and more diverse populations, (2) advanced neuroimaging technologies, and (3) a greater variety of perceptual discrimination tasks.
To our knowledge, this is the first study to quantitatively evaluate the impact of the “extent” component on user attentional engagement after ST-ART and CL-ART interventions that were delivered on a consumer-friendly PC-powered VR-HMD platform with both behavioral and EEG measurements. The experimental results support the conclusion that the CL-ART tasks may indeed be superior to ST-ART tasks in terms of user engagement. This is especially pertinent given that a total of 14.35 million VR-HMD units were sold in 2017 to 2019 worldwide [66], highlighting the oncoming wave of such accessible technology for researchers and clinicians to utilize these tools in ways never before achieved. Furthermore, if one takes advantage of the burgeoning all-in-one VR-HMD platform (e.g., Oculus Quest), the benefits of this closed-loop ART approach can be extended to larger numbers and more diverse populations.

Author Contributions

Conceptualization, G.L. and S.Z.; methodology, G.L. and S.Z.; software, G.L.; validation, S.Z., Z.K., and M.G.; formal analysis, G.L. and S.Z.; investigation, G.L. and S.Z.; resources, G.L. and S.Z.; writing—original draft preparation, G.L., S.Z., Z.K., and M.G.; writing—review and editing, G.L.; supervision, G.L.; project administration, G.L.; funding acquisition, G.L. All authors read and agreed to the published version of the manuscript.

Funding

This research was sponsored by the Shanghai Sailing Program under Grant 17YF1426900.

Acknowledgments

The authors specially thank Hongwei Liu, from the Shanghai Zaichen Intelligent Technology, Co., Ltd., for his technical support in advanced VR programming for this study.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

This appendix contains some technical details about the Virtual ART, supplementary figures from Virtual ART and oddball tasks, and a link for the demo video of CL-ART.
1. Technical Details about Virtual ART
The contact quality of EEG electrodes and scalp was color-coded in a software, called NIC (NETM), which pairs with the EEG recording device (black: no signal; red: very poor signal; orange: fair signal; green: good signal). The collected EEG data were continuously displayed and stored on a PC by NIC. In the meanwhile, the C++ version of the lab streaming layer (LSL) protocol was used to transfer the EEG data from NIC to the Unreal game engine, and then the RBP(α) was calculated to control the natural scenes (see Figure A1 and Equation (A1)), where, Power is the fast Fourier transform (FFT) power, zi = {θ, α, β}.
Figure A1. The data flowchart of the combined VR–EEG system.
Figure A1. The data flowchart of the combined VR–EEG system.
Sensors 20 02208 g0a1
R B P ( α ) = P o w e r ( α ) i = 1 3 P o w e r ( Z i ) × 100 %
2. Supplementary Figures for Virtual ART
Figure A2. Screenshots from ST-ART showing (a) the task 1, (b) the reward following task 1 (clear sky), (c) the light snow during task 2, and (d) the reward following task 3 (aurora).
Figure A2. Screenshots from ST-ART showing (a) the task 1, (b) the reward following task 1 (clear sky), (c) the light snow during task 2, and (d) the reward following task 3 (aurora).
Sensors 20 02208 g0a2
3. Video Demo for CL-ART
4. Supplementary Figures for Visual Oddball Tasks
Figure A3. Visual oddball tasks.
Figure A3. Visual oddball tasks.
Sensors 20 02208 g0a3

References

  1. Cheever, N.A.; Peviani, K.; Rosen, L.D. Media Multitasking and Mental Health. In Technology and Adolescent Mental Health, 1st ed.; Moreno, M.A., Radovic, A., Eds.; Springer: Berlin/Heidelberg, Germany, 2018; pp. 101–112. [Google Scholar]
  2. Rosen, L.D.; Carrier, L.M.; Cheever, N.A. Facebook and Texting Made Me Do It: Media-Induced Task Switching While Studying. Comput. Hum. Behav. 2013, 29, 948–958. [Google Scholar] [CrossRef]
  3. Moisala, M.; Salmela, V.; Hietajärvi, L.; Salo, E.; Carlson, S.; Salonen, O.; Lonka, K.; Hakkarainen, K.; Salmela-Aro, K.; Alho, K. Media multitasking is associated with distractibility and increased prefrontal activity in adolescents and young adults. Neuroimage 2016, 134, 113–121. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Ophir, E.; Nass, C.; Wagner, A.D. Cognitive control in media multitaskers. Proc. Natl Acad. Sci. USA 2009, 106, 15583–15587. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Ralph, B.C.; Tomson, D.R.; Cheyne, J.A.; Smilek, D. Media multitasking and failures of attention in everyday life. Psychol. Res. 2014, 78, 661–669. [Google Scholar] [CrossRef]
  6. Rosen, L.; Carrier, L.M.; Miller, A.; Rokkum, J.; Ruzi. Sleeping with Technology: Cognitive, Affective, and Technology Usage Predictors of Sleep Problems among College Students. Sleep Health 2016, 2, 49–56. [Google Scholar] [CrossRef]
  7. Carrier, L.M.; Rosen, L.D.; Cheever, N.A.; Lim, A.F. Causes, Effects, and Practicalities of Everyday Multitasking. Dev. Rev. 2015, 35, 64–78. [Google Scholar] [CrossRef]
  8. Gonzalez, V.M.; Mark, G. Constant, Constant, Multitasking Craziness: Managing Multiple Working Spheres. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI’04), Vienna, Austria, 24–29 April 2004. [Google Scholar]
  9. Mark, G.; Wang, Y.; Niiya, M. Stress and Multitasking in Everyday College Life: An Empirical Study of Online Activity. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI’14), Toronto, ON, Canada, 26 April–1 May 2014. [Google Scholar]
  10. Cheever, N.A.; Rosen, L.D.; Carrier, L.M.; Chavez, A. Out of Sight Is not Out of Mind: The Impact of Restricting Wireless Mobile Device Use on Anxiety Levels among Low, Moderate, and High Users. Comput. Hum. Behav. 2014, 37, 290–297. [Google Scholar] [CrossRef]
  11. Ziegler, D.A.; Simon, A.J.; Gallen, C.L.; Skinner, S.; Janowich, J.R.; Volponi, J.J.; Rolle, C.E.; Mishra, J.; Kornfield, J.; Anguera, J.A.; et al. Closed-loop digital meditation improves sustained attention in young adults. Nat. Hum. Behav. 2019, 3, 746–757. [Google Scholar] [CrossRef]
  12. Gazzaley, A.; Rosen, L.D. Boosting Control. In The Distracted Mind, 1st ed.; MIT Press: Cambridge, MA, USA, 2017; pp. 181–211. [Google Scholar]
  13. Ohly, H.; White, M.P.; Wheeler, B.W.; Bethel, A.; Ukoumunne, O.C.; Nikolaou, V.; Garside, R. Attention Restoration Theory: A systematic review of the attention restoration potential of exposure to natural environments. J. Toxicol. Environ. Health A 2016, 19, 305–343. [Google Scholar] [CrossRef] [Green Version]
  14. Moore, E.O. A prison environment’s effect on health care service demands. J. Environ. Syst. 1981, 11, 17–34. [Google Scholar] [CrossRef]
  15. Verderber, S. Dimensions of person-window transactions in the hospital environment. Environ. Behav. 1986, 18, 450–466. [Google Scholar] [CrossRef]
  16. Ulrich, R.S. Biophilia, biophobia, and natural landscapes. In The Biophilia Hypothesis, 1st ed.; Kellert, S.R., Wilson, E.O., Eds.; Island Press: Washington, DC, USA, 1993; pp. 73–137. [Google Scholar]
  17. Tennessen, C.; Cimprich, G. Views to nature: Effects on attention. J. Environ. Psychol. 1995, 15, 77–85. [Google Scholar] [CrossRef]
  18. Kaplan, R.; Kaplan, S. Restorative Environment. In The Experience of Nature. A Psychological Perspective, 1st ed.; Cambridge University Press: Cambridge, UK, 1989; pp. 189–196. [Google Scholar]
  19. Kaplan, S. The restorative benefits of nature: Toward an integrative framework. J. Environ. Psychol. 1995, 15, 169–182. [Google Scholar] [CrossRef]
  20. Kaplan, R. The Nature of the View from Home: Psychological Benefits. Environ. Behav. 2001, 33, 507–542. [Google Scholar] [CrossRef]
  21. Daniel, R.M. The Effects of the Natural Environment of Attention Restoration. Master’s Thesis, Appalachian State University, Boone, NC, USA, May 2014. [Google Scholar]
  22. Hartig, T.; Book, A.; Garvill, J.; Olsson, T.; Garling, T. Environmental influences on psychological restoration. Scand. J. Psychol. 1996, 37, 378–393. [Google Scholar] [CrossRef]
  23. Herzog, T.R.; Colleen, P.M.; Nebel, M.B. Assessing the restorative components of environments. J. Environ. Psychol. 2003, 23, 159–170. [Google Scholar] [CrossRef]
  24. Bodin, M.; Hartig, T. Does the outdoor environment matter for psychological restoration gained through running? Psychol. Sport Exerc. 2003, 4, 141–153. [Google Scholar] [CrossRef]
  25. Ottosson, J.; Grahn, P. A comparison of leisure time spent in a garden with leisure time spent indoors: On measures of restoration in residents in geriatric care. Landsc. Res. 2007, 30, 23–55. [Google Scholar] [CrossRef]
  26. Warm, J.S.; Parasuraman, R.; Matthew, G. Vigilance requires hard mental work and is stressful. Hum. Factors 2008, 50, 433–441. [Google Scholar] [CrossRef]
  27. Katsuki, F.; Constantinidis, C. Bottom-Up and Top-Down Attention. Neuroscientist 2013, 20, 509–521. [Google Scholar] [CrossRef]
  28. Markowitz, D.M.; Laha, R.; Perone, B.P.; Pea, R.D.; Bailenson, J.N. Immersive virtual reality field trips facilitate learning about climate change. Front. Psychol. 2018, 9, 1–20. [Google Scholar] [CrossRef] [PubMed]
  29. Maltby, N.; Kirsch, I.; Mayers, M.; Allen, G.J. Virtual reality exposure therapy for the treatment of fear of flying: A controlled investigation. J. Consult. Clin. Psychol. 2002, 70, 1112–1118. [Google Scholar] [CrossRef] [PubMed]
  30. Cho, B.H.; Jang, D.P.; Lee, J.H.; Lee, J.M.; Kim, J.S.; Kim, S.I.; Ku, J.H.; Kim, I.Y. Attention Enhancement System using Virtual Reality and EEG Biofeedback. In Proceedings of the IEEE Conference on Virtual Reality, Orlando, FL, USA, 24–28 March 2002. [Google Scholar]
  31. Bohil, C.J.; Alicea, B.; Biocca, F.A. Virtual reality in neuroscience research and therapy. Nat. Rev. Neurosci. 2011, 12, 752–762. [Google Scholar] [CrossRef] [PubMed]
  32. Ewalt, D.M. Oculus Rift Review: The Beginning of the Age of VR. Available online: https://www.forbes.com/sites/davidewalt/2016/03/28/oculus-rift-review-the-beginning-of-the-age-of-vr/#62806e1d5aff (accessed on 29 February 2020).
  33. Robertson, G.; Card, S.K.; Mackinlay, J. Three views of virtual reality: Non-immersive virtual reality. Computer 1993, 26, 81. [Google Scholar] [CrossRef]
  34. Slobounov, S.M.; Ray, W.; Johnson, B.; Slobounov, E.; Newell, K.M. Modulation of cortical activity in 2D versus 3D virtual reality environments: An EEG study. Int. J. Psychophysiol. 2015, 95, 254–260. [Google Scholar] [CrossRef] [PubMed]
  35. Cummings, J.J.; Bailenson, J.N. How immersive is enough? A meta-analysis of the effect of immersive technology on user presence. Media Psychol. 2015, 19, 272–309. [Google Scholar] [CrossRef]
  36. Ziegler, D.A.; Janowich, J.R.; Gazzaley, A. Differential Impact of Interference on Internally- and Externally-Directed Attention. Sci. Rep. 2018, 8, 2498. [Google Scholar] [CrossRef]
  37. Li, G.; Khan, M.A. Deep learning on VR-induced attention. In Proceedings of the IEEE Conference on Artificial Intelligence and Virtual Reality, San Diego, CA, USA, 9–11 December 2019. [Google Scholar]
  38. Li, G.; Chung, W.Y. A Combined EEG-Gyroscope-tDCS Brain Machine Interface System for Early Management of Driver Drowsiness. IEEE Trans. Hum.-Mach. Syst. 2018, 48, 50–62. [Google Scholar] [CrossRef]
  39. Lin, C.T.; Chuang, C.H.; Wang, Y.K.; Tsai, S.F.; Chiu, T.C.; Ko, L.W. Neurocognitive characteristics of the driver: A review on drowsiness, distraction, navigation, and motion sickness. J. Neurosci. Neuroeng. 2012, 1, 61–81. [Google Scholar] [CrossRef]
  40. Kothe, C. Lab Streaming Layer (LSL)—A Software Framework for Synchronizing a Large Array of Data Collection and Stimulation Devices. The Swartz Center for Computational Neuroscience Software. Available online: https://code.google.com/archive/p/labstreaminglayer/ (accessed on 29 February 2020).
  41. Wenger, M.J.; Townsend, J.T. Basic Response Time Tools for Studying General Processing Capacity in Attention, Perception, and Cognition. J. Exp. Psychol. Gen. 1998, 127, 67–99. [Google Scholar] [CrossRef]
  42. Anguera, J.A.; Boccanfuso, J.; Rintoul, J.L.; Ai-Hashimi, O.; Faraji, F.; Kong, E.; Larraburo, Y.; Rolle, C.E.; Johnston, E.; Gazzaley, A. Video game training enhances cognitive control in older adults. Nature 2013, 501, 97–101. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  43. Rolle, C.E.; Anguera, J.A.; Skinner, S.N.; Voytek, B.; Gazzaley, A. Enhancing spatial attention and working memory in younger and older adults. J. Cogn. Neurosci. 2017, 29, 1483–1493. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Polich, J. Updating P300: An Integrative Theory of P3a and P3b. Clin. Neurophysiol. 2007, 118, 2128–2148. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Michael, J.K.; Mark, D.R.; Dustin, E.S.; Joseph, S.R.; Orban, S.A.; Lauren, M.F.; Ellen, G.K. Reaction time variability in ADHD: A meta-analytic review of studies. Clin. Psychol. Rev. 2011, 33, 795–811. [Google Scholar]
  46. Tallon-Baudry, C.; Bertrand, O.; Delpuech, C.; Pernier, J. Stimulus specifcity of phase-locked and non-phase-locked 40Hz visual responses in human. J. Neurosci. 1996, 16, 4240–4249. [Google Scholar] [CrossRef]
  47. Mognon, A.; Jovicich, J.; Bruzzone, L.; Buiatti, M. ADJUST: An automatic EEG artifact detector based on the joint use of spatial and temporal features. Psychophysiology 2011, 48, 229–240. [Google Scholar] [CrossRef]
  48. Kropotov, J.D. A P3b component as index of engagement operation. In Quantitative EEG, Event-Related Potentials and Neurotherapy, 1st ed.; Academic: San Diego, CA, USA, 2009; pp. 399–410. [Google Scholar]
  49. Johnson, R., Jr. On the neural generators of the P300 component of the event-related potential. Psychophysiology 1993, 30, 90–97. [Google Scholar] [CrossRef]
  50. Makeig, S.; Westerfield, M.; Jung, T.P.; Enghoff, S.; Townsend, J.; Courchesne, E.; Sejnowski, T.J. Dynamic brain sources of visual evoked responses. Science 2002, 295, 690–694. [Google Scholar] [CrossRef] [Green Version]
  51. Anguera, J.A.; Lyman, K.; Zanto, T.P.; Bollinger, J.; Gazzaley, A. Reconciling the influence of task-set switching and motor inhibition processes on stop signal after-effects. Front. Psychol. 2013, 4, 1–12. [Google Scholar] [CrossRef] [Green Version]
  52. Spielberger, C.D.; Gorsuch, R.L.; Lushene, R.E. Manual for the State-Trait Anxiety Inventory. American Psychology Association. Available online: https://www.apa.org/pi/about/publications/caregivers/practice-settings/assessment/tools/trait-state (accessed on 29 February 2020).
  53. Koelstra, S.; Muehl, C.; Soleymani, M.; Lee, J.-S.; Yazdani, A.; Ebrahimi, T.; Pun, T.; Nijholt, A.; Patras, I. DEAP: A Database for Emotion Analysis; Using Physiological Signals. IEEE Trans. Affect. Comput. 2012, 3, 18–31. [Google Scholar] [CrossRef] [Green Version]
  54. Mayer, F.S.; Frantz, C.M.; Bruehlman-Senecal, E.; Dolliver, K. Why Is Nature Beneficial?: The Role of Connectedness to Nature. Environ. Behav. 2008, 41, 607–643. [Google Scholar] [CrossRef]
  55. Kuo, F.E. Coping with poverty: Impacts of environment and attention in the inner city. Environ. Behav. 2001, 33, 5–34. [Google Scholar] [CrossRef]
  56. Anguera, J.A.; Brandes-Aitken, A.N.; Rolle, C.E.; Skinner, S.N.; Desai, S.S.; Bower, J.D.; Martucci, W.E.; Chung, W.K.; Sherr, E.H.; Marco, E.J. Characterizing cognitive control abilities in children with 16p11.2 deletion using adaptive ‘video game’ technology: A pilot study. Transl. Psychiatry 2016, 6, e893. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  57. Strauss, E.; Sherman, E.; Spreen, O. Compendium of Neuropsychological Tests: Administration Norms, and Commentary, 3rd ed.; Oxford Univeristy Press: New York, NY, USA, 2006. [Google Scholar]
  58. Ros, T.; Theberge, J.; Frewen, P.A.; Kluetsch, R.; Densmore, M.; Calhoun, V.D.; Lanius, R.A. Mind Over Chatter: Plastic Up-Regulation of the fMRI Salience Network Directly after EEG Neurofeedback. NeuroImage 2013, 65, 324–335. [Google Scholar] [CrossRef]
  59. Mishra, J.; Anguera, A.J.; Gazzaley, A. Video Games for Neuro-Cognitive Optimization. Neuro 2016, 90, 214–218. [Google Scholar] [CrossRef] [Green Version]
  60. VR Neurofeedback: A New Drug-free Treatment for Mental Disorders. Available online: https://arvrjourney.com/vr-neurofeedback-a-new-drug-free-treatment-for-mental-disorders-f721f652f159 (accessed on 24 March 2020).
  61. Weiss, P.L.; Bialik, P.; Kizony, R. Virtual reality provides leisure time opportunities for young adults with physical and intellectual disabilities. CyberPsychology Behav. 2003, 6, 335–342. [Google Scholar] [CrossRef] [Green Version]
  62. Mineo, B.A.; Ziegler, W.; Gill, S.; Salkin, D. Engagement with electronic screen media among students with autism spectrum disorders. J. Autism. Dev. Disord. 2009, 39, 172–187. [Google Scholar] [CrossRef]
  63. Anguera, J.A.; Brandes-Aitken, A.N.; Antovich, A.D.; Rolle, C.E.; Desai, S.S.; Marco, E.J. A pilot study to determine the feasibility of enhancing cognitive abilities in children with sensory processing dysfunction. PLoS ONE 2017, 12, e0172616. [Google Scholar] [CrossRef]
  64. Tucha, O.; Tucha, L.; Kaumann, G.; Konig, S.; Lange, K.M.; Stasik, D.; Streather, Z.; Engelschalk, T.; Lange, K.W. Training of attention functions in children with attention deficit hyperactivity disorder. Atten. Defic. Hyperact. Disord. 2011, 3, 271–283. [Google Scholar] [CrossRef] [Green Version]
  65. Li, H.; Li, J.; Li, N.; Li, B.; Wang, P.; Zhou, T. Cognitive intervention for persons with mild cognitive impairment: A meta-analysis. Ageing Res. Rev. 2011, 10, 285–296. [Google Scholar] [CrossRef]
  66. Units Shipments of Virtual Reality (VR) Devices Worldwide from 2017 to 2019. Available online: https://www.statista.com/statistics/671403/global-virtual-reality-device-shipments-by-vendor/ (accessed on 29 February 2020).
Figure 1. The thresholds for each task (left) during CL-ART experiment and the algorithm for CL-ART (right).
Figure 1. The thresholds for each task (left) during CL-ART experiment and the algorithm for CL-ART (right).
Sensors 20 02208 g001
Figure 2. (a) One-way ANOVA analysis of state and trait anxiety State-Trait Anxiety Inventory questionnaire (STAI) scores between ST-ART and CL-ART groups; (b) scatter plots for state and trait anxiety STAI scores.
Figure 2. (a) One-way ANOVA analysis of state and trait anxiety State-Trait Anxiety Inventory questionnaire (STAI) scores between ST-ART and CL-ART groups; (b) scatter plots for state and trait anxiety STAI scores.
Sensors 20 02208 g002
Figure 3. (a) The group grand average for P3b captured before and after the two kinds of ART interventions; (b) paired t-test and subject means for P3b; (c) scatter plots for RT difference and P3b difference.
Figure 3. (a) The group grand average for P3b captured before and after the two kinds of ART interventions; (b) paired t-test and subject means for P3b; (c) scatter plots for RT difference and P3b difference.
Sensors 20 02208 g003
Figure 4. (a) Paired t-test and subject means for RTVar; (b) the group grand average for ITC captured before and after the two kinds of ART interventions; as can be seen here, the highest phase-locking value (PLV; deep red) in the “after CL-ART” condition occurred in the bin of 200–300 ms, which was clearly faster than that in the bin of 300–400 ms in other conditions; (c) scatter plots for RTVar difference and ITC difference.
Figure 4. (a) Paired t-test and subject means for RTVar; (b) the group grand average for ITC captured before and after the two kinds of ART interventions; as can be seen here, the highest phase-locking value (PLV; deep red) in the “after CL-ART” condition occurred in the bin of 200–300 ms, which was clearly faster than that in the bin of 300–400 ms in other conditions; (c) scatter plots for RTVar difference and ITC difference.
Sensors 20 02208 g004aSensors 20 02208 g004b
Table 1. Comparison of standard attention restoration theory (ST-ART) and closed-loop attention restoration theory (CL-ART) tasks.
Table 1. Comparison of standard attention restoration theory (ST-ART) and closed-loop attention restoration theory (CL-ART) tasks.
Restorative ComponentsVirtual ART Tasks
ST-ARTCL-ART
Being awaySecluded home surrounded by snowy mountain, meadows, and pool.
Soft FascinationMorning (clouds and fog), afternoon (snowing), and night (cricket sound and aurora), as well as a shared scene from morning to night: motion of cherry blossom in the breeze.
CompatibilityAnxious university students who voluntarily participated in this study.
ExtentTask 1 (for morning scene)The thickness of fog is randomized. VR controller is used to capture stones on the table and throw them into the pool one by one.The thickness of fog is controlled by EEG at 2-s interval. Wireless Xbox 360 joystick is used to walk around in any accessible virtual space while relaxing mind to make fog vanish.
RewardBright sun makes fog vanish and task 2 starts.The degree of snow is controlled by EEG at 2-sec interval.
Task 2 (for afternoon scene)VR controller is used to turn on the music player on the table, listening to classic music while walking around (by teleport function) in any accessible virtual space until the end of music.Wireless Xbox 360 joystick is used to walk around in any accessible virtual space while relaxing mind to turn on the music player on the table, listening to classic music until the end of music.
RewardOnce the music is played, the default light snow converts into heavy snow. By the end of music, heavy snow stops and task 3 starts.The snow stops and task 3 starts.
Task 3 (for night scene)No specific task, just walking around (by teleport function) in any accessible virtual space.Wireless Xbox 360 joystick is used to walk around in any accessible virtual space while relaxing mind to call aurora.
RewardAurora appears. The range of the aurora is randomized, updating every 2 s.Aurora appears. The range of the aurora is controlled by EEG at 2-s interval.
Table 2. Summary of outcome measures. RT—response time; ITC—inter-trial coherence; IEC—inter-electrode coherence.
Table 2. Summary of outcome measures. RT—response time; ITC—inter-trial coherence; IEC—inter-electrode coherence.
TaskTypes of MeasureImplications
BehavioralNeural
Visual oddballRTP3b latencyGeneral attention level
RTVarITC(θ)Attentional engagement
/IEC(θ)Brain functional connectivity
Table 3. Summary of one-way ANOVA analysis results for ITC(θ) difference.
Table 3. Summary of one-way ANOVA analysis results for ITC(θ) difference.
Bins (ms)p-ValuePLV (Mean ± Standard Error)
CL-ARTST-ART
0–1000.0020.042 ± 0.1670.027 ± 0.011
100–2000.199−0.035 ± 0.0320.026 ± 0.034
200–3000.031−0.102 ± 0.0580.075 ± 0.052
300–4000.0010.041 ± 0.042−0.202 ± 0.493
400–5000.0080.032 ± 0.043−0.175 ± 0.058
500–6000.678−0.021 ± 0.052−0.051 ± 0.050
600–7000.898−0.050 ± 0.048−0.042 ± 0.038
Table 4. Summary of one-way ANOVA analysis results for IEC(θ) difference.
Table 4. Summary of one-way ANOVA analysis results for IEC(θ) difference.
Bins (ms)p-ValuePLV (Mean ± Standard Error)
CL-ARTST-ART
0–1000.1790.034 ± 0.026−0.027 ± 0.035
100–2000.1490.068 ± 0.032−0.021 ± 0.051
200–3000.4870.047 ± 0.031−0.006 ± 0.067
300–4000.467−0.004 ± 0.0430.054 ± 0.066
400–5000.7970.036 ± 0.0430.055 ± 0.058
500–6000.8410.053 ± 0.0530.070 ± 0.050
600–7000.7240.014 ± 0.0450.035 ± 0.038

Share and Cite

MDPI and ACS Style

Li, G.; Zhou, S.; Kong, Z.; Guo, M. Closed-Loop Attention Restoration Theory for Virtual Reality-Based Attentional Engagement Enhancement. Sensors 2020, 20, 2208. https://doi.org/10.3390/s20082208

AMA Style

Li G, Zhou S, Kong Z, Guo M. Closed-Loop Attention Restoration Theory for Virtual Reality-Based Attentional Engagement Enhancement. Sensors. 2020; 20(8):2208. https://doi.org/10.3390/s20082208

Chicago/Turabian Style

Li, Gang, Shihong Zhou, Zhen Kong, and Mengyuan Guo. 2020. "Closed-Loop Attention Restoration Theory for Virtual Reality-Based Attentional Engagement Enhancement" Sensors 20, no. 8: 2208. https://doi.org/10.3390/s20082208

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop