Skip to main content

ORIGINAL RESEARCH article

Front. Psychol., 17 March 2021
Sec. Cognition

Visual Search in 3D: Effects of Monoscopic and Stereoscopic Cues to Depth on the Validity of Feature Integration Theory and Perceptual Load Theory

  • School of Psychology, University College Dublin, Dublin, Ireland

Previous research has successfully used feature integration theory to operationalise the predictions of Perceptual Load Theory, while simultaneously testing the predictions of both models. Building on this work, we test the extent to which these models hold up in a 3D world. In two experiments, participants responded to a target stimulus within an array of shapes whose apparent depth was manipulated using a combination of monoscopic and stereoscopic cues. The search task was designed to test the predictions of (a) feature integration theory, as the target was identified by a single feature or a conjunction of features and embedded in search arrays of varying size, and (b) perceptual load theory, as the task included congruent and incongruent distractors presented alongside search tasks imposing high or low perceptual load. Findings from both experiments upheld the predictions of feature integration theory, regardless of 2D/3D condition. Longer search times in conditions with a combination of monoscopic and stereoscopic depth cues suggests that binding features into three-dimensional objects requires greater attentional effort. This additional effort should have implications for perceptual load theory, yet our findings did not uphold its predictions; the effect of incongruent distractors did not differ between conjunction search trials (conceptualised as high perceptual load) and feature search trials (low perceptual load). Individual differences in susceptibility to the effects of perceptual load were evident and likely explain the absence of load effects. Overall, our findings suggest that feature integration theory may be useful for predicting attentional performance in a 3D world.

Introduction

In order to be useful, cognitive theories must be simplified models of a complex world, from which we can derive specific predictions. In order to be meaningful, these models must also be scalable and applicable to real world situations. Models of visual search are principally based around perception of two-dimensional (2D) stimuli such as shapes on a computer screen, but are used to make inferences about our interaction with the three-dimensional (3D) world around us. When attempting to test the ecological validity of such models, researchers typically add complexity in a controlled manner, one variable at a time. In this paper we attempt to scale two models of visual search – Treisman’s feature integration theory and Lavie’s perceptual load theory– by incorporating monoscopic and stereoscopic cues to depth into an existing paradigm in which both models can be tested simultaneously. Both feature integration theory and perceptual load theory are simplified models of the type described above, from which useful predictions about the world have been generated. By the simple addition of a third dimension we literally add depth to the models, and test their validity in a three-dimensional world.

Feature Integration Theory

Feature integration theory (Treisman and Gelade, 1980; Treisman, 1988) is a model of visual attention in which perception occurs at the level of features rather than objects; once perceived, attentional resources are required to bind features into coherent objects. While feature-level perception can and does occur in parallel, the integration of features into objects occurs serially. The classic finding in feature integration theory is therefore that reaction times for identifying objects on the basis of a single feature (e.g., colour) are relatively unaffected by the number of distractor items in the search array. In contrast, when items must be identified on the basis of a conjunction of features (e.g., colour and shape) each item in the array must be processed in turn and reaction time to identify the target will therefore increase linearly with the number of items in the array (Treisman and Gelade, 1980). Feature integration theory has been supported by a plethora of research (see Quinlan, 2003; Humphreys, 2016 for reviews), and while it has been partially subsumed into more comprehensive theories (e.g., Duncan and Humphreys, 1989; Desimone and Duncan, 1995), it has undeniably made a significant contribution to our understanding of visual attention. While early work using this model was largely restricted to analysis of basic features such as colour and shape, subsequent work has included dynamic features such as direction of motion (Dick et al., 1987; Thornton and Zdravković, 2020) and axis of rotation (Schill et al., 2020). Recently, feature integration theory has been particularly useful in the development of a concrete operationalisation of perceptual load.

Perceptual Load Theory

The premise of perceptual load theory (Lavie, 1995; Lavie et al., 2004) is that perceptual capacity is limited, and that perceptual processing continues automatically until that capacity is filled. If a task or scene imposes high perceptual load then, according to the model, all available capacity will be used up in processing the central stimuli and no resources will be available to process additional distractor stimuli. In contrast, if the task imposes a low level of perceptual load, leftover attentional resources will “spill over” and process peripheral or distractor stimuli. Thus, perceptual load theory predicts that distractor perception will be reduced or eliminated under high perceptual load. This is typically assessed using flanker tasks in which a peripheral distractor stimulus requires a response that is either congruent or incongruent with the response to the target stimulus (Lavie, 2005). For example, a typical paradigm requires participants to search an array of letters and report the presence of either an X or an N. The search array is flanked by a distractor stimulus that may be response-congruent (e.g., an X when the target is an X) or response-incongruent (an X when the target is an N). Distractor processing is indicated by longer reaction times on incongruent trials because response competition between the target and distractor must be resolved before a response can be made. While a substantial congruency effect is expected under low perceptual load, this is predicted to be reduced or eliminated under high load. In this standard paradigm, perceptual load is manipulated by varying the features of the non-target items in the search array; in the low perceptual load condition, the target is presented in an array of O’s while in the high load condition the non-target items are angular letters such as M, H and Z.

Although perceptual load theory has been influential over the past twenty years, the concept of perceptual load has been criticised for its vagueness and lack of a clear operational definition (Murphy et al., 2016). Murphy and Greene (2017b) recently described a series of experiments demonstrating that the classic comparison in feature integration theory – between feature and conjunction searches – could act as an operationalisation of perceptual load theory. The five experiments reported in that paper tested the predictions of both feature integration theory and perceptual load theory within a single paradigm, and found that both were upheld. Following Lavie (1995), Murphy and Greene noted that feature and conjunction searches could be conceived of as low and high perceptual load tasks, respectively. In the most basic form of their paradigm, participants search a visual array for a specific target (a red triangle) and report its orientation while ignoring a distractor triangle whose orientation may be congruent (i.e., both pointing up) or incongruent with the target (one pointing up, one pointing down). The target is contained in a search array of between 2 and 10 items from which the target can be discriminated on the basis of one feature (e.g., a red triangle in a field of yellow triangles) or on the basis of a conjunction of features (e.g., a red triangle in a field of yellow triangles and red or yellow squares). They observed that response times in this paradigm followed the typical pattern expected of feature vs. conjunction searches, with fast responses to feature searches irrespective of the number of distractors, and slower responses to conjunction searches that required a serial inspection of each item in the array. The reaction time data also fit the predictions of perceptual load theory, with reduced distractor congruency effects under high load compared with low load. Subsequent experiments found similar results with this paradigm under a range of different conditions in which the stimulus presentation time, distractor location and stimulus complexity were varied, leading incrementally toward a more ecologically valid design. It has been suggested that the reduction in distractor interference under high load conditions may sometimes be caused by the presence in the array of neutral items which share visual features with the distractor, and thus dilute its effects (Tsal and Benoni, 2010; Benoni and Tsal, 2013). The paradigm described above controls for dilution as both low load and high load displays contain neutral stimuli which share a feature (shape) with the distractor triangle. Low load and high load displays are thus differentiated only by whether the target can be identified on the basis of a single feature or a conjunction of features [see Murphy and Greene (2017b) for further discussion of dilution as it pertains to this paradigm].

Ecological Validity in a 3D World

The lateral separation of human eyes means that each has a slightly different view of the world and the disparity between these views occurs along a single horizontal dimension. This binocular disparity is greater for objects that are closer to us than those that are farther away. Using this rule, the visual system can map objects along three dimensions of the visual field; a process referred to as stereoscopic vision or stereopsis (Qian, 1997). In an exploration of the function of stereoscopic depth perception, Read (2015) points to the understanding that predatory animals acquire superior ability to detect and distinguish camouflage in visually complex environments via binocular stereopsis. The importance of this ability is underscored by the finding that human participants with normal stereoscopic vision typically out-perform counterparts with reduced or no stereoscopic vision on visuomotor skills tasks (Mazyn et al., 2004; O’Connor et al., 2010). The influential two-stream hypothesis (Goodale and Milner, 1992) provided a neurological account of vision that separated such visuospatial information processing for visually guided behaviour (dorsal stream) from slower identification or recognition processes (ventral stream). Since then, reviews and critiques have rejected the extreme claim that these visual tasks are based on functionally independent neurological processing pathways (Milner and Goodale, 2008; McIntosh and Schenk, 2009). Despite mounting evidence for considerable crosstalk between streams (Milner, 2017), there is consensus that a less extreme version of the two-stream hypothesis can be a useful model for generating hypotheses about visual performance in the real world (Schenk and McIntosh, 2010) – the world is both something to be perceived and something to be acted upon.

Stereoscopic vision (requiring two eyes) is not the only way the visual system derives spatial and depth information from a visual scene. Monocular signals (such as shade colour, occlusion, size and elevation in a plane) are common and powerful cues to depth (Howard and Rogers, 2002), allowing us to infer the spatial relations of objects in three dimensions even from two dimensional representations, such as a photograph. A limited number of previous studies within the feature integration theory paradigm have previously shown how visual search can be improved by using monocular/pictorial cues to depth (e.g., an illustration of a cube in which the faces are coloured to create the illusion of light and shade; Enns and Rensink, 1990a,b; Zhang et al., 2015). These researchers compared response times to the same set of polygons (squares and diamonds) arranged to give the impression of a 3D cube or a 2D abstract pattern, and reported that search for the visually meaningful 3D cubes was faster and less effortful than search for the 2D shapes. However, others have found that visual search was negatively affected by the inclusion of monocular (pictorial) depth cues relative to simply presenting the stimuli in 2D (Kyritsis et al., 2013). In a closely related body of work using a change blindness paradigm, inclusion of monocular depth cues of relative size, saturation and brightness to a stereoscopic display have been shown to improve visual working memory for perceptually closer items (Qian et al., 2017, 2018).

Another approach exploring depth in feature integration theory has investigated the way in which apparent distance (or visual depth plane) acts as a visual feature (Plewan and Rinkenauer, 2016, 2017). For example, a target in a search array could be identified on the basis of a single feature (its apparent location relative to the viewer) or on the basis of a conjunction of features (e.g., a red shape at a particular distance from the viewer). These studies have indicated that objects that appear to be closer to the viewer are typically responded to more quickly. Yet, in a similar paradigm, Pomplun et al. (2013) report no substantial effects of depth of plane on visual search strategy derived from eye movement data. Other work has demonstrated that the apparent proximity of stimuli affects visual tracking (Viswanathan and Mingolla, 2002) and visual working memory (He and Nakayama, 1995; Xu and Nakayama, 2007; Qian et al., 2017, 2018; Chunharas et al., 2019). Finally a third body of work has explored visual search and foraging performance in dynamic virtual reality 3D environments (Prpic et al., 2019; Kristjánsson et al., 2020), sometimes compared to traditional 2D pictorial scenes (Li et al., 2016). The value of such studies includes the interaction of visual search with memory and proprioceptive feedback from self-initiated whole body motion.

Importantly, previous research has not yet compared the effects of viewing the same stimuli in monoscopic 2D and stereoscopic 3D on visual attention, or separated those effects from the role of monocular cues to depth such as colour and shading. If stereopsis and other binocular cues to depth (e.g., binocular convergence) improve our ability to act upon the world more precisely, they may benefit performance in visually demanding experimental trials. On the other hand, if additional visual features (depth plane, shading etc.) are required to produce a 3D percept, it follows that attention must be deployed to bind these features together; this would result in slowed visual search for stimuli containing these features.

The Present Study

In the present study we explore whether monocular and binocular cues to depth make unique demands of the visual system that limit the ecological validity of feature integration theory and perceptual load theory. We report two experiments, in which we replicated the paradigm used by Murphy and Greene (2017b) in two and three dimensions, and operationalised high and low perceptual load as conjunction and feature searches. In Experiment 1, dimensionality of the stimuli was manipulated using a combination of monocular and binocular cues to depth, while in Experiment 2 monocular and binocular cues were independently manipulated. We carried forward Murphy and Greene’s hypotheses, including that the predictions of both feature integration theory and perceptual load theory would be upheld. Specifically, feature integration theory predicts a significant interaction of search type and array size, such that reaction time would increase with set size for conjunction but not feature searches, and perceptual load theory predicts a significant interaction of perceptual load and congruency, such that distractor congruency effects would be reduced or eliminated under high perceptual load. It is unclear how the combination of monocular and binocular cues will affect search performance. While previous research suggests that depth cues offer advantages for visuospatial accuracy and visuomotor performance, results on the effects of depth on visual search identification are mixed. If the binding of features denoting depth requires attentional resources, this might constitute an additional source of perceptual load. Thus, we predict that the inclusion of depth cues will result in increased reaction times, especially in conditions where both monocular and binocular cues are provided.

Experiment 1

Method

Participants

Forty-two volunteers (17 male, 25 female) were recruited from the University College Dublin campus to participate in the experiment. Participants were aged between 18 and 52 years (M = 24.41 years, SD = 6.79 years). All had normal or corrected-to-normal vision and all were screened for colour-blindness and for stereoscopic depth perception.

Apparatus

The experiment was programmed and run using E-Prime 2.0 (Psychology Software Tools Inc.). A 47-inch LG Smart Cinema 3D television screen (resolution = 1920 × 1080) was used to present the stimuli in 2D or passive polarised stereoscopic 3D. For the stereoscopic 3D condition, the depth control was set to two points above the TV default, which creates a baseline crossed disparity of 12 pixels (∼0.62% of the screen width). All 3D stimuli appeared on a single depth plane with a sense of 3D space around them. To control for screen luminance effects, participants in both the 2D and 3D groups wore passive-polarised 3D glasses designed for use with this television.

Design

The experimental design followed Murphy and Greene (2017b) and comprised a four-way (2 × 2 × 5 × 2) factorial design with three within-subjects variables (search type: low load/feature search vs. high load/conjunction search), distractor congruency: congruent vs. incongruent, and array size: 2, 4, 6, 8, or 10 search items) and one between-subjects variable (2D vs. 3D). Participants were assigned to either the 2D or 3D condition through the use of a true randomisation technique, i.e., a coin flip (Vickers, 2006), with the exception of one participant who was assigned to the 2D group due to poor or no stereoacuity, and the final participant who was assigned to the 2D condition to ensure equal sample sizes across groups.

Task Stimuli

The experiment involved the search for a target stimulus within an array, while ignoring a congruent or incongruent distractor. In the 2D condition, participants were instructed to search for the target (red triangle) in an array of yellow and red shapes presented to the left of a central fixation point (a white cross), while ignoring a distractor to the right of the fixation point. Participants responded by indicating whether the target triangle was pointing up or down (using the arrows on a keyboard). In low load, feature search trials, the non-target stimuli in the search array were all yellow triangles, and the target was thus distinguished by a single feature (colour). In high load, conjunction search trials, the non-target stimuli were red squares, yellow squares and yellow triangles; a conjunction of colour (red) and shape (triangle) was thus required to distinguish the target. All shapes were created in Microsoft Office 365 ProPlus. The red shapes had an RGB value of (255, 0, 0) and the yellow shapes had an RGB value of (255, 255, 0).

To supplement the binocular stereoscopic effect in the 3D condition, the above described stimuli were adapted for the 3D condition to include apparent (monocular) depth cues; this was achieved by replacing the standard two-dimensional stimuli (squares and triangles) with two-dimensional representations of three dimensional objects (cubes and cones). Shading effects were used to create the illusion of depth, as shown in Figure 1. RGB values for the front surface of the images, on which light appeared to be shining, matched those of the 2D shapes (red: 255, 0, 0; yellow: 255, 255, 0). The ‘shadowed’ surfaces of the images had RGB values ranging from (255, 5, 5) to (184, 23, 23) for red shapes and from (255, 255, 5) to (124, 124, 0) for yellow shapes. At a viewing distance of 140 cm, each shape subtended 1.44 by 1.70 degrees of visual angle. The entire search array, presented to the left of fixation, subtended between 2.1 × 3.4 (array size 2) and 6.5 × 7 degrees (array size 10). The distractor triangle, presented to the right of fixation and subtending 1.08 by 1.34 degrees, was slightly larger than the other shapes to guarantee distinctiveness and reduce eccentricity effects.

FIGURE 1
www.frontiersin.org

Figure 1. Sample trials of low load/feature (top row) and high load/conjunction (bottom row) searches in the 2D and 3D conditions in Experiment 1. (A) 2D Feature search. The array size is 8 and the distractor orientation is incongruent with the target. (B) 3D Feature search. The array size is 2 and the distractor orientation is congruent with the target. (C) 2D Conjunction search. The array size is 4 and the distractor orientation is congruent with the target. (D) 3D Conjunction search. The array size is 6 and the distractor orientation is incongruent with the target.

The target red triangle/cone could appear in an array of 2, 4, 6, 8, or 10 shapes and the orientation of the distractor stimulus could be congruent (i.e., both pointing up or down) or incongruent with the target (i.e., one pointing up and other pointing down). See Figure 1 for sample trials.

Procedure

Participants were seated 140 cm from the television screen at a desk equipped with a keyboard. After providing written consent, participants were instructed to don the 3D glasses and the task was explained to them. All lighting in the laboratory was switched off during testing. Participants first completed 30 practice trials and were given the opportunity to clarify the requirements of the task. This also allowed the experimenter to confirm that the 3D effect was perceived by participants in the 3D condition. Participants then completed three blocks of 100 trials, comprising 15 repetitions of each combination of search type, congruency and array size. Trials were presented randomly within blocks, and participants were given the opportunity to take a break following completion of each block. Participants were instructed to maintain fixation on the central white cross, and to respond to the orientation of the target stimulus as quickly and accurately as possible. The stimuli in each trial remained onscreen until a response was recorded. No performance feedback was provided.

The study protocol was approved by the Human Research Ethics Committee (Humanities) of University College Dublin.

Results

All data are available in the Open Science Framework repository, at https://osf.io/ep5gv/. Data from one individual in the 3D condition were removed from analysis as their accuracy across all conditions was below chance (mean = 32%). Performance in the search task was generally very accurate, with mean accuracy among the remaining 41 participants of 96.5% (SD = 3.6%).

Median reaction time (RT) for each participant was extracted for correct search trials within each experimental condition. Median reaction time data are often preferred for such analyses as they are less susceptible than mean data to the effects of outliers (Ratcliff, 1993). A 2 (search type) × 2 (distractor congruency) × 5 (array size) × 2 (2D/3D condition) mixed within/between-subjects ANOVA was conducted. The full ANOVA results are shown in Table 1. The interested reader can find a similar analysis of accuracy data in Supplementary Tables 1, 21.

TABLE 1
www.frontiersin.org

Table 1. Results of the 4-way ANOVA examining effects of search type, distractor congruency, array size, and 2D/3D group on median search reaction time in Experiment 1.

A main effect of search type was observed such that responses were slower in high load/conjunction searches (M = 728.99 ms, SE = 13.66) than low load/feature searches (M = 677.98 ms, SE = 11.44). A main effect of array size was also observed; planned contrasts showed a significant linear trend, indicating that RT increased in line with the number of distractors in the array [F(1, 39) = 212.59, p < 0.001, ηp2 = 0.85]. These main effects were qualified by a significant interaction, depicted in Figure 2. To investigate the predictions of feature integration theory, a linear regression of response time against array size was conducted for each search condition. Array size was a significant predictor of RT in both the conjunction search task [intercept = 603.49ms; R2 = 0.26, F (1, 408) = 142.5, p < 0.001] and the feature search task [intercept = 636.35ms; R2 = 0.06, F (1, 408) = 24.62, p < 0.001], however, the conjunction search task had a slope of 20.85 ms/item, which was significantly steeper than the feature search slope of 6.75 ms/item, t(816) = 6.37, p < 0.001. Each additional distractor in the display therefore increased reaction time by a greater magnitude in the conjunction search than the feature search.

FIGURE 2
www.frontiersin.org

Figure 2. Median reaction time for feature and conjunction searches at different array sizes in Experiment 1. Error bars represent 95% confidence intervals.

There was no significant difference in mean reaction time between participants in the 2D (M = 703.4 ms, SE = 17.3) and 3D conditions (M = 703.58 ms, SE = 17.72). There was also no main effect of distractor congruency (congruent: M = 704.07 ms, SE = 12.68; incongruent: M = 702.92 ms, SE = 12.27) and no interaction between congruency and search type/perceptual load, as would be predicted by perceptual load theory. No other significant interaction effects were detected.

Experiment 1 Discussion

Although the effects of array size and search type predicted by feature integration theory were observed in this experiment, no effect of distractor congruency was observed. Decades of research have reported robust congruency effects in perceptual load theory paradigms (Murphy et al., 2016), suggesting that a visible distractor cannot be easily suppressed once perceived. We therefore concluded that the distractor stimuli were not reliably perceived in this experiment. This may have been due to the peripheral placement of the distractor; Beck and Lavie (2005) have observed that distractors at fixation produce larger interference effects than distractors in the periphery. Moreover, although participants were instructed to fixate on the central fixation cross throughout, it is possible that the placement of the distractor in the periphery of a large display may have allowed participants to shift their visual field (e.g., by turning their head) and exclude the peripheral distractor from view.

In addition, no main effect of 2D/3D group, and no interaction between group and any other variable were observed in this experiment. There are two potential explanations for this finding: first, it is possible that stereoscopic depth has no bearing on reaction time during visual search, and thus has no implications for the feature integration or perceptual load models. Alternatively, the peripheral positioning of the distractor may also have undermined the stereoscopic effect produced by the LG TV used in this experiment. Although the television manufacturers claim that the 3D effect should be visible from a wide range of horizontal angles, experimenters noted that the stereoscopic effect was best viewed from directly in front of the screen and was somewhat compromised when viewing at a different angle from behind the participant. To remedy these issues, we conducted a second experiment using a centrally presented distractor, as employed in Murphy and Greene (2017b) Experiment 3.

The design of Experiment 1 did not allow the effects of monocular and binocular cues to depth to be distinguished, as flat shapes were always presented in the 2D condition and shaded shapes were always presented in the 3D condition. In order to rule out any explanation of significant results in terms of one of these factors alone, Experiment 2 therefore includes independent manipulations of stereoscopic depth (2D vs. 3D) and stimulus shading (flat vs. shaded shapes).

Finally, Experiment 1 employed a relatively small sample size and was thus somewhat underpowered. This issue is also remedied in Experiment 2.

Experiment 2

Method

Participants

One hundred and thirty participants (51 male, 79 female; age range 18–54 years; M = 22.85, SD = 6.72) were recruited from the University College Dublin campus. All participants had normal or corrected-to-normal vision, none were colour-blind, and all were screened for stereoscopic depth perception to ensure that participants allocated to the 3D condition had stereoscopic vision. Power analysis was conducted in GPower 3.1 (Faul et al., 2007). As no previous studies have compared reaction times in feature/conjunction searches between subjects in 2D and 3D conditions, no a priori effect size estimates were available. The sample size was therefore selected on the basis that it provided 80% power to detect medium effects (f = 0.25; Cohen, 1988) in the between-subjects terms. This sample also provided 80% power to detect very small effects (f values ranging from.03 to.06) in the within/between-subjects interaction terms; this power analysis assumed an average correlation of 0.9 among repeated measures, based on data reported in Murphy and Greene (2017b).

Design

This experiment employed a five-way (2 × 2 × 5 × 2 × 2) factorial design with three within-subjects variables (search type, distractor congruency, and array size, as in Experiment 1) and two between-subjects variables (stereoscopic depth: 2D/3D and stimulus shading: flat/shaded). Participants were randomly allocated to view the task in monoscopic 2D or stereoscopic 3D, and to view either flat or shaded shapes.

Task Stimuli

In the task for Experiment 2, the distractor stimulus was presented in the centre of the screen surrounded by a white box, with the target appearing among an array of shapes presented in a ring around the distractor (see Figure 3). This adjusted the visual angle dimensions of the distractor which now subtended 3.83 by 3.01 degrees. The entire search array subtended 10.2 × 8.68 degrees of visual angle from a viewing distance of 140 cm.

FIGURE 3
www.frontiersin.org

Figure 3. Sample trials of low load/feature [top row, panels (A,B)] and high load/conjunction searches [bottom row, panels (C,D)] in Experiment 2, using either flat [panels (A,C)] or shaded shapes [panels (B,D)]. These stimuli were displayed under either monoscopic (2D) or stereoscopic (3D) conditions.

Procedure

All other features of the design, task and procedure were identical to those used in Experiment 1.

Results

All data are available in the Open Science Framework repository, at https://osf.io/ep5gv/. The data from one participant were removed from analysis due a technical error during data recording. Mean search accuracy for the remaining 129 participants was 97.3% (SD = 2.6%). Median RT for each participant was extracted for correct search trials within each experimental condition. Overall median RT was significantly longer in Experiment 2 (M = 731.94 ms, SD = 103.42) than Experiment 1 [M = 703.49, SD = 78.26; t(128) = 3.124, p = 0.002], suggesting that the central placement of the distractor increased the difficulty of the task.

A 2 (search type) × 2 (distractor congruency) × 5 (array size) × 2 (stereoscopic depth) × 2 (stimulus shading) ANOVA was performed. The full results of this analysis may be seen in Table 2. A similar analysis of accuracy data may be found in Supplementary Tables 3, 42. A main effect of search type was observed on median reaction time, such that search was slower in high load/conjunction search trials (M = 766.26ms, SE = 10.50) than in low load/feature search trials (M = 697.62ms, SE = 7.88). A main effect of array size was also observed, such that RT increased with each additional item in the array (2 items: M = 681.48ms, SE = 7.77; 4 items: M = 700.08ms, SE = 8.14; 6 items: M = 725.66ms, SE = 8.58; 8 items: M = 755.64ms, SE = 9.75; 10 items: M = 794.81ms, SE = 10.67). Critically, there was also a significant interaction between search type and array size. Linear regression analysis showed that RT increased with array size for conjunction trials [intercept = 633.03ms; R2 = 0.44, F(1,648) = 154.61, p < 0.001, slope = 26.13 ms/item] while the slope stayed relatively flat for feature search trials [intercept = 689.36ms; R2 = 0.10, F(1,648) = 6.87, p = 0.01, slope = 3.98 ms/item]; the slope of the line was significantly steeper in the conjunction search condition, t (1296) = 8.54, p < 0.001. Thus, the predictions of feature integration theory were upheld.

TABLE 2
www.frontiersin.org

Table 2. Results of the five-way ANOVA examining effects of search type, distractor congruency, array size, stereoscopic depth, and stimulus shading on median search reaction time in Experiment 2.

A significant effect of distractor congruency was observed; RTs were slower in incongruent trials (M = 744.47 ms, SE = 9.33) than congruent trials (M = 719.41 ms, SE = 9.02), indicating that the distractor stimuli were perceived and interfered with search performance. However, no significant interaction between load and congruency was observed, contrary to the predictions of perceptual load theory. Post hoc tests indicated a significant difference in RT between congruent and incongruent trials under both low load and high load conditions, whereas perceptual load theory would predict that this effect should be eliminated or greatly reduced under high load (see Figure 4). Additional analysis indicated considerable individual differences in this effect: the predictions of load theory were upheld for approximately half the sample, with many participants demonstrating reverse congruency effects (i.e., faster RTs during congruent trials relative to incongruent trials). See Supplementary Material for further details of these analyses.

FIGURE 4
www.frontiersin.org

Figure 4. Median RT in congruent and incongruent trials under high load (conjunction search) and low load (feature search) conditions in Experiment 2. **p < 0.001.

There was no main effect of stereoscopic depth (2D vs. 3D), however, a main effect of stimulus shading was observed, such that longer RTs were observed in response to shaded shapes (M = 753.63 ms, SE = 12.75) than flat shapes (M = 709.44 ms, SE = 12.06). These effects are qualified by an interaction between these two display variables. Post hoc Tukey tests revealed that RT in response to the 3D shaded condition (M = 787. 85 ms, SE = 17.54) was significantly higher than RT in the 3D flat (M = 702.56ms, SE = 18.75), 2D flat (M = 716.32, SE = 17.28) or 2D shaded conditions (M = 719.41, SE = 16.54, all p’s < 0.01); none of the other pairwise comparisons showed significant differences.

A number of three-way interactions were also observed, but were qualified by a four-way interaction of search type, array size, stereoscopic depth, and stimulus shading. Figure 5 depicts the overall pattern of effects; a clear interaction between search type and array size is evident across all display conditions, such that reaction time increases steeply with array size in conjunction search conditions, but not in feature search conditions. The RTs in each cell of this analysis were remarkably similar for the 2D flat, 2D shaded and 3D flat conditions, however, in the 3D shaded condition (depicted as solid lines in Panel B), a sharp increase in overall RT was observed. Bonferroni-corrected post hoc tests confirmed that average RT in the 3D shaded condition was significantly higher than that in the other three conditions which did not differ significantly from one another. Thus, the same pattern of interaction between search type and array size was seen under all conditions, but this effect was moderated by the combination of monoscopic and stereoscopic cues to depth, with slower overall reaction times in the 3D shaded condition.

FIGURE 5
www.frontiersin.org

Figure 5. Median reaction times during Experiment 2. Panel (A) depicts the monoscopic 2D conditions and panel (B) depicts the stereoscopic 3D conditions. Error bars represent 95% confidence intervals.

Discussion

We investigated the effects of monoscopic and stereoscopic cues to depth on visual search, with the aim of validating feature integration theory and perceptual load theory in a 3D world. Across two experiments, the predictions of feature integration theory were upheld under both 2D and 3D conditions; a significant interaction of search type and array size was observed, as the slope for feature search trials remained relatively flat while reaction times in the conjunction search trials increased linearly with number of distractors. This provides support for the classic feature integration idea that the target ‘popped out’ in the low load, feature search trials, but that a serial search was required to identify the target in the high load, conjunction search trials. These data provide support for the validity of feature integration theory under more ecologically valid conditions.

In Experiment 2, this interaction of array size and search type was moderated by a combination of monoscopic and stereoscopic cues to depth. The same pattern of a steep slope in conjunction search trials and flat slope in feature search trials was observed regardless of display condition. However, overall search time was significantly increased in the condition that combined stereoscopic 3D with shaded shapes. This slowed overall response was not seen with any other combination of depth cues; stereoscopic 3D presented with flat shapes resulted in similar RTs to monoscopic 2D with either flat or shaded shapes. Thus, the addition of stereoscopic 3D and monocular cues to depth – that is, the combination of depth cues typically found in the real world – slowed overall search reaction time, but did not interfere with the typical pattern of search observed in feature integration theory studies (Quinlan, 2003; Humphreys, 2016). Feature integration theory tells us that reaction times increase in conjunction searches because attention is required to bind the features of each item together, and the items are processed serially (Treisman, 1988). The present results indicate that binding the features of stereoscopic depth and stimulus shading into a single three-dimensional percept with colour and shape required extra attention, over and above that required by the addition of either monocular or binocular depth cues alone.

The finding that there was no difference in search RT between the 2D flat and 2D shaded conditions indicates that including monocular cues to depth alone does not affect search efficiency. These results are in contrast to previous findings that visual search was negatively affected by the inclusion of monocular (pictorial) depth cues relative to simply presenting the stimuli in 2D (Kyritsis et al., 2013); they also challenge research showing that search is facilitated – i.e., faster – when the constituent polygons of a target image are presented in the form of an apparently three-dimensional object, such as a cube, rather than rearranged into an abstract shape that appears two-dimensional (Enns and Rensink, 1990a,b; Zhang et al., 2015). We suggest that these latter findings may be attributed to the semantic meaning imposed by the organisation of the polygons into a recognisable shape, rather than the representation of depth. The flat stimuli in the present experiment were composed of two features: a shape (e.g., a square) and a colour (e.g., yellow). In contrast, the shaded stimuli were comprised of multiple facets and colours – for example, the red cube had three visible faces rendered in three different red hues to denote depth via light and shade. The parallel slopes during visual search for the flat and shaded shapes suggest that, when presented in 2D, the apparently three-dimensional cube was perceived as a single shape rather than a combination of polygons.

Similarly, the absence of a difference between the 2D flat and 3D flat conditions rules out an explanation in terms of stereopsis alone, or any methodological artefact arising from the use of the 3D TV screen. Computing depth from binocular disparity and creating a coherent percept from two distinct channels of visual input is an effortful and time-consuming process (Welchman, 2016). One marked characteristic of the 3D shaded stimuli was the illusion of volume and mass. These stimuli consequently afforded the sensation of being able to “reach out and grab” the shapes, which appeared to hang in mid-air. These action affordances during observation of apparently graspable objects engage early attentional processes (Proverbio et al., 2011) and require crosstalk between ventral and dorsal streams, which is increased when stereoscopic and pictorial depth cues are combined (Verhagen et al., 2008; Milner, 2017). Thus, we suggest that the provision of both monocular and binocular cues to depth in the current study may have required additional processing time during visual search, as participants computed motor-relevant information.

The predictions of perceptual load theory were not upheld in this study: in contrast with the experiments reported by Murphy and Greene (2017b) using the same paradigm, we did not observe a significant interaction of perceptual load and distractor congruency in either experiment. No main effect of distractor congruency was observed in Experiment 1, most likely due to the peripheral placement of the distractor stimulus, however, this was addressed in Experiment 2, in which the distractor stimulus was presented centrally and the search array items were presented in a ring around the distractor. A clear distractor congruency effect was observed here, suggesting that participants did process the centrally presented distractor. However, counter to the predictions of perceptual load theory (Lavie, 2005), distractor interference was not found to be reduced under high perceptual load. As the basic predictions of perceptual load theory were not upheld in this study, we were not able to validate the model under 3D conditions. No significant interactions were observed between the critical perceptual load theory variables (perceptual load and distractor congruency) and the display variables of stereoscopic depth and stimulus shading, however, in the absence of a significant interaction of load and congruency, these results cannot be reliably interpreted. The principles of perceptual load theory have previously been extended into complex multisensory tasks, including driving (Marciano and Yeshurun, 2012, 2015; Murphy and Greene, 2015, 2016b, 2017a), playing football (Furley et al., 2013) and eyewitness memory (Murphy and Greene, 2016a; Greene et al., 2017); nevertheless, the failure to replicate the basic distractor congruency effect presents a challenge to the model and precludes confirmation of its validity in more complex environments.

Marciano and Yeshurun (2017) recently reported large interindividual and intraindividual differences in distractor perception under high and low perceptual load, suggesting that the magnitude of the perceptual load effect may be dependent on specific individual characteristics that vary at both state and trait level. Here we report similar findings; just over half of the sample in Experiment 2 showed the expected reduction in distractor interference under high load, while the remainder experienced the opposite pattern, with larger distractor congruency effects under high load than low load. Combined with Marciano and Yeshurun’s results, the findings of the current study suggest that findings from the classic perceptual load theory experiments (e.g., Lavie, 1995; Lavie et al., 2004) may not be as robust as anticipated, or at least that they are dependent on some boundary conditions such as, for example, the display or type of stimuli. One difference between the present paradigm and much of the previous perceptual load theory literature is that the distractor stimulus was presented centrally rather than in the periphery. Beck and Lavie (2005) have noted that, while perceptual load effects are observed with both centrally and peripherally placed distractors, larger distractor effects are typically observed with foveal distractors. It is possible that the distractor placement may have affected the distribution of attention to target and distractor stimuli (though note that Murphy and Greene (2017b) observed perceptual load effects in a similar paradigm with centrally resented distractors). Future research may wish to investigate these boundary conditions and identify the circumstances under which perceptual load effects can and cannot be detected.

The data presented in this paper do, however, contribute to the validity of feature integration theory as a model for evaluating visual search in a three-dimensional world. The predictions of feature integration theory were upheld by findings from both experiments in the current study. Although this was the case regardless of 2D/3D condition, longer search times were observed in conditions with a combination of monoscopic and stereoscopic depth cues. This suggests that binding features into three-dimensional objects requires greater attentional effort. In the interests of ecological validity, we call for this line of work to be extended beyond the visual domain and to consider the integration of other sensory information as a part of a multimodal process.

Data Availability Statement

The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found below: Open Science Framework, https://osf.io/ep5gv/.

Ethics Statement

The studies involving human participants were reviewed and approved by the University College Dublin Human Research Ethics Committee (Humanities). The patients/participants provided their written informed consent to participate in this study.

Author Contributions

CG and BR designed the study, with the assistance of JB and AH. CG programmed the experimental paradigms and conducted the statistical analysis. JB, AH, SeK, SH, and StK collected the data and assisted with analysis. CG wrote the manuscript with BR. All authors approved the final manuscript.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

We would like to thank David Redmond, Cornelius De Beer, and Maria Fernanda Jara Rizzo for their assistance with data collection.

Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyg.2021.596511/full#supplementary-material

Footnotes

  1. ^ Some significant results were observed in this analysis, including a main effect of search type, main effect of array size and an interaction between search type and distractor congruency. As these results were based on an analysis of fewer than 4% of trials, they are not commented on further here.
  2. ^ Some significant results were observed in this analysis, including main effects of search type and distractor congruency, and interactions between search type, distractor congruency and array size, As these results were based on an analysis of fewer than 3% of trials, they are not commented on further here.

References

Beck, D. M., and Lavie, N. (2005). Look here but ignore what you see: effects of distractors at fixation. J. Exp. Psychol. Hum. Percept. Perform. 31, 592–607. doi: 10.1037/0096-1523.31.3.592

PubMed Abstract | CrossRef Full Text | Google Scholar

Benoni, H., and Tsal, Y. (2013). Conceptual and methodological concerns in the theory of perceptual load. Front. Psychol. 4:522. doi: 10.3389/fpsyg.2013.00522

PubMed Abstract | CrossRef Full Text | Google Scholar

Chunharas, C., Rademaker, R. L., Sprague, T. C., Brady, T. F., and Serences, J. T. (2019). Separating memoranda in depth increases visual working memory performance. J. Vis. 19:4. doi: 10.1167/19.1.4

CrossRef Full Text | Google Scholar

Cohen, J. (1988). Statistical Power Analysis for the Behavioral Sciences. Hove: Psychology Press.

Google Scholar

Desimone, R., and Duncan, J. (1995). Neural mechanisms of selective visual attention. Annu. Rev. Neurosci. 18, 193–222. doi: 10.1146/annurev.ne.18.030195.001205

PubMed Abstract | CrossRef Full Text | Google Scholar

Dick, M., Ullman, S., and Sagi, D. (1987). Parallel and serial processes in motion detection. Science 237, 400–402. doi: 10.1126/science.3603025

PubMed Abstract | CrossRef Full Text | Google Scholar

Duncan, J., and Humphreys, G. W. (1989). Visual search and stimulus similarity. Psychol. Rev. 96, 433–458. doi: 10.1037/0033-295x.96.3.433

PubMed Abstract | CrossRef Full Text | Google Scholar

Enns, J. T., and Rensink, R. A. (1990a). Influence of scene-based properties on visual search. Science 247, 721–723. doi: 10.1126/science.2300824

PubMed Abstract | CrossRef Full Text | Google Scholar

Enns, J. T., and Rensink, R. A. (1990b). Sensitivity to three-dimensional orientation in visual search. Psychol. Sci. 1, 323–326. doi: 10.1111/j.1467-9280.1990.tb00227.x

CrossRef Full Text | Google Scholar

Faul, F., Erdfelder, E., Lang, A.-G., and Buchner, A. (2007). GPower 3: a flexible statistical power analysis program for the social, behavioral, and biomedical sciences. Behav. Res. Methods 39, 175–191. doi: 10.3758/bf03193146

PubMed Abstract | CrossRef Full Text | Google Scholar

Furley, P., Memmert, D., and Schmid, S. (2013). Perceptual load in sport and the heuristic value of the perceptual load paradigm in examining expertise-related perceptual-cognitive adaptations. Cogn. Process. 14, 31–42. doi: 10.1007/s10339-012-0529-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Goodale, M. A., and Milner, A. D. (1992). Separate visual pathways for perception and action. Trends Neurosci. 15, 20–25. doi: 10.1016/0166-2236(92)90344-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Greene, C. M., Murphy, G., and Januszewski, J. (2017). Under high perceptual load, observers look but do not see. Appl. Cogn. Psychol. 31, 431–437. doi: 10.1002/acp.3335

CrossRef Full Text | Google Scholar

He, Z. J., and Nakayama, K. (1995). Visual attention to surfaces in three-dimensional space. Proc. Natl. Acad. Sci. U.S.A. 92, 11155–11159. doi: 10.1073/pnas.92.24.11155

PubMed Abstract | CrossRef Full Text | Google Scholar

Howard, I. P., and Rogers, B. J. (2002). Seeing in Depth, Vol. 2: Depth Perception. Toronto: University of Toronto Press.

Google Scholar

Humphreys, G. W. (2016). Feature confirmation in object perception: feature integration theory 26 years on from the treisman bartlett lecture. Q. J. Exp. Psychol. 69, 1910–1940. doi: 10.1080/17470218.2014.988736

PubMed Abstract | CrossRef Full Text | Google Scholar

Kristjánsson, T., Draschkow, D., Pálsson, Á, Haraldsson, D., Jónsson, P. Ö, and Kristjánsson, Á (2020). Moving foraging into three dimensions: feature- versus conjunction-based foraging in virtual reality. Q. J. Exp. Psychol. doi: 10.1177/1747021820937020 [Epub ahead of print].

CrossRef Full Text | Google Scholar

Kyritsis, M., Gulliver, S. R., Morar, S., and Stevens, R. (2013). “Issues and benefits of using 3D interfaces: visual and verbal tasks,” in Proceedings of the Fifth International Conference on Management of Emergent Digital EcoSystems, (New York, NY: Association for Computing Machinery).

Google Scholar

Lavie, N. (1995). Perceptual load as a necessary condition for selective attention. J. Exp. Psychol. Hum. Percept. Perform. 21, 451–468. doi: 10.1037/0096-1523.21.3.451

PubMed Abstract | CrossRef Full Text | Google Scholar

Lavie, N. (2005). Distracted and confused: selective attention under load. Trends Cogn. Sci. 9, 75–82. doi: 10.1016/j.tics.2004.12.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Lavie, N., Hirst, A., de Fockert, J. W., and Viding, E. (2004). Load theory of selective attention and cognitive control. J. Exp. Psychol. Gen. 133, 339–354. doi: 10.1037/0096-3445.133.3.339

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, C.-L., Aivar, M. P., Kit, D. M., Tong, M. H., and Hayhoe, M. M. (2016). Memory and visual search in naturalistic 2D and 3D environments. J. Vis. 16:9. doi: 10.1167/16.8.9

CrossRef Full Text | Google Scholar

Marciano, H., and Yeshurun, Y. (2012). Perceptual load in central and peripheral regions and its effects on driving performance: advertizing billboards. Work 41(Suppl. 1), 3181–3188. doi: 10.3233/wor-2012-0580-3181

PubMed Abstract | CrossRef Full Text | Google Scholar

Marciano, H., and Yeshurun, Y. (2015). Perceptual load in different regions of the visual scene and its relevance for driving. Hum. Factors 57, 701–716. doi: 10.1177/0018720814556309

PubMed Abstract | CrossRef Full Text | Google Scholar

Marciano, H., and Yeshurun, Y. (2017). Large inter-individual and intra-individual variability in the effect of perceptual load. PLoS One 12:e0175060. doi: 10.1371/journal.pone.0175060

PubMed Abstract | CrossRef Full Text | Google Scholar

Mazyn, L. I., Lenoir, M., Montagne, G., and Savelsbergh, G. J. (2004). The contribution of stereo vision to one-handed catching. Exp. Brain Res. 157, 383–390.

Google Scholar

McIntosh, R. D., and Schenk, T. (2009). Two visual streams for perception and action: current trends. Neuropsychologia 47, 1391–1396. doi: 10.1016/j.neuropsychologia.2009.02.009

PubMed Abstract | CrossRef Full Text | Google Scholar

Milner, A. D. (2017). How do the two visual streams interact with each other? Exp. Brain Res. 235, 1297–1308. doi: 10.1007/s00221-017-4917-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Milner, A. D., and Goodale, M. A. (2008). Two visual systems re-viewed. Neuropsychologia 46, 774–785. doi: 10.1016/j.neuropsychologia.2007.10.005

PubMed Abstract | CrossRef Full Text | Google Scholar

Murphy, G., and Greene, C. M. (2015). High perceptual load causes inattentional blindness and deafness in drivers. Vis. Cogn. 23, 810–814. doi: 10.1080/13506285.2015.1093245

CrossRef Full Text | Google Scholar

Murphy, G., and Greene, C. M. (2016a). Perceptual load affects eyewitness accuracy and susceptibility to leading questions. Front. Psychol. 7:1322. doi: 10.3389/fpsyg.2016.01322

PubMed Abstract | CrossRef Full Text | Google Scholar

Murphy, G., and Greene, C. M. (2016b). Perceptual load induces inattentional blindness in drivers. Appl. Cogn. Psychol. 30, 479–483. doi: 10.1002/acp.3216

CrossRef Full Text | Google Scholar

Murphy, G., and Greene, C. M. (2017a). Load theory behind the wheel; perceptual and cognitive load effects. Can. J. Exp. Psychol. 71, 191–202. doi: 10.1037/cep0000107

PubMed Abstract | CrossRef Full Text | Google Scholar

Murphy, G., and Greene, C. M. (2017b). Visual feature vs. conjunction searches as a robust manipulation of perceptual load. J. Cogn. Psychol. 29, 305–326. doi: 10.1080/20445911.2016.1256888

CrossRef Full Text | Google Scholar

Murphy, G., Groeger, J. A., and Greene, C. M. (2016). Twenty years of load theory - where are we now, and where should we go next? Psychon. Bull. Rev. 23, 1316–1340. doi: 10.3758/s13423-015-0982-5

PubMed Abstract | CrossRef Full Text | Google Scholar

O’Connor, A. R., Birch, E. E., Anderson, S., and Draper, H. (2010). The functional significance of stereopsis. Invest. Ophthalmol. Vis. Sci. 51, 2019–2023. doi: 10.1167/iovs.09-4434

PubMed Abstract | CrossRef Full Text | Google Scholar

Plewan, T., and Rinkenauer, G. (2016). Fast and forceful: modulation of response activation induced by shifts of perceived depth in virtual 3D space. Front. Psychol. 7:1939. doi: 10.3389/fpsyg.2016.01939

PubMed Abstract | CrossRef Full Text | Google Scholar

Plewan, T., and Rinkenauer, G. (2017). Simple reaction time and size–distance integration in virtual 3D space. Psychol. Res. 81, 653–663. doi: 10.1007/s00426-016-0769-y

PubMed Abstract | CrossRef Full Text | Google Scholar

Pomplun, M., Garaas, T. W., and Carrasco, M. (2013). The effects of task difficulty on visual search strategy in virtual 3D displays. J. Vis. 13, 24–24. doi: 10.1167/13.3.24

PubMed Abstract | CrossRef Full Text | Google Scholar

Proverbio, A. M., Adorni, R., and D’Aniello, G. E. (2011). 250 ms to code for action affordance during observation of manipulable objects. Neuropsychologia 49, 2711–2717. doi: 10.1016/j.neuropsychologia.2011.05.019

PubMed Abstract | CrossRef Full Text | Google Scholar

Prpic, V., Kniestedt, I., Camilleri, E., Maureira, M. G., Kristjánsson, Á, and Thornton, I. M. (2019). A serious game to explore human foraging in a 3D environment. PLoS One 14:e0219827. doi: 10.1371/journal.pone.0219827

PubMed Abstract | CrossRef Full Text | Google Scholar

Qian, J., Li, J., Wang, K., Liu, S., and Lei, Q. (2017). Evidence for the effect of depth on visual working memory. Sci. Rep. 7, 1–10.

Google Scholar

Qian, J., Zhang, K., Wang, K., Li, J., and Lei, Q. (2018). Saturation and brightness modulate the effect of depth on visual working memory. J. Vis. 18, 16–16. doi: 10.1167/18.9.16

CrossRef Full Text | Google Scholar

Qian, N. (1997). Binocular disparity and the perception of depth. Neuron 18, 359–368. doi: 10.1016/s0896-6273(00)81238-6

CrossRef Full Text | Google Scholar

Quinlan, P. T. (2003). Visual feature integration theory: past, present, and future. Psychol. Bull. 129, 643–673. doi: 10.1037/0033-2909.129.5.643

PubMed Abstract | CrossRef Full Text | Google Scholar

Ratcliff, R. (1993). Methods for dealing with reaction time outliers. Psychol. Bull. 114, 510–532. doi: 10.1037/0033-2909.114.3.510

PubMed Abstract | CrossRef Full Text | Google Scholar

Read, J. C. A. (2015). What is stereoscopic vision good for? Proc. SPIE IS T Electron. Imaging 9391:93910N. doi: 10.1117/12.2184988

CrossRef Full Text | Google Scholar

Schenk, T., and McIntosh, R. D. (2010). Do we have independent visual streams for perception and action? Cogn. Neurosci. 1, 52–62. doi: 10.1080/17588920903388950

PubMed Abstract | CrossRef Full Text | Google Scholar

Schill, H. M., Cain, M. S., Josephs, E. L., and Wolfe, J. M. (2020). Axis of rotation as a basic feature in visual search. Attent. Percept. Psychophys. 82, 31–43. doi: 10.3758/s13414-019-01834-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Thornton, I. M., and Zdravković, S. (2020). Searching for illusory motion. Attent. Percept. Psychophys. 82, 44–62. doi: 10.3758/s13414-019-01750-3

PubMed Abstract | CrossRef Full Text | Google Scholar

Treisman, A. (1988). Features and objects: the fourteenth bartlett memorial lecture. Q. J. Exp. Psychol. Section A 40, 201–237. doi: 10.1080/02724988843000104

PubMed Abstract | CrossRef Full Text | Google Scholar

Treisman, A. M., and Gelade, G. (1980). A feature-integration theory of attention. Cogn. Psychol. 12, 97–136. doi: 10.1016/0010-0285(80)90005-5

PubMed Abstract | CrossRef Full Text | Google Scholar

Tsal, Y., and Benoni, H. (2010). Diluting the burden of load: perceptual load effects are simply dilution effects. J. Exp. Psychol. Hum. Percept. Perform. 36, 1645–1656. doi: 10.1037/a0018172

PubMed Abstract | CrossRef Full Text | Google Scholar

Verhagen, L., Dijkerman, H. C., Grol, M. J., and Toni, I. (2008). Perceptuo-motor interactions during prehension movements. J. Neurosci. 28, 4726–4735. doi: 10.1523/jneurosci.0057-08.2008

PubMed Abstract | CrossRef Full Text | Google Scholar

Vickers, A. J. (2006). How to randomize. J. Soc. Integr. Oncol. 4, 194–198. doi: 10.2310/7200.2006.023

PubMed Abstract | CrossRef Full Text | Google Scholar

Viswanathan, L., and Mingolla, E. (2002). Dynamics of attention in depth: evidence from multi-element tracking. Perception 31, 1415–1437. doi: 10.1068/p3432

PubMed Abstract | CrossRef Full Text | Google Scholar

Welchman, A. E. (2016). The human brain in depth: how we see in 3D. Annu. Rev. Vis. Sci. 2, 345–376. doi: 10.1146/annurev-vision-111815-114605

PubMed Abstract | CrossRef Full Text | Google Scholar

Xu, Y., and Nakayama, K. (2007). Visual short-term memory benefit for objects on different 3-D surfaces. J. Exp. Psychol. Gen. 136:653. doi: 10.1037/0096-3445.136.4.653

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, X., Huang, J., Yigit-Elliott, S., and Rosenholtz, R. (2015). Cube search, revisited. J. Vis. 15:9. doi: 10.1167/15.3.9

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: visual search, 3D, stereoscopic depth, monoscopic depth, feature integration, distractor congruency, perceptual load theory

Citation: Greene CM, Broughan J, Hanlon A, Keane S, Hanrahan S, Kerr S and Rooney B (2021) Visual Search in 3D: Effects of Monoscopic and Stereoscopic Cues to Depth on the Validity of Feature Integration Theory and Perceptual Load Theory. Front. Psychol. 12:596511. doi: 10.3389/fpsyg.2021.596511

Received: 19 August 2020; Accepted: 22 February 2021;
Published: 17 March 2021.

Edited by:

Paula Goolkasian, University of North Carolina at Charlotte, United States

Reviewed by:

Raoul Bell, Heinrich Heine University of Düsseldorf, Germany
Jiehui Qian, Sun Yat-sen University, China

Copyright © 2021 Greene, Broughan, Hanlon, Keane, Hanrahan, Kerr and Rooney. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Ciara M. Greene, ciara.greene@ucd.ie

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.