Statistical Learning of Incidental Perceptual Regularities Induces Sensory Conditioned Cortical Responses

Simple Summary Our study demonstrated neural encoding of incidental sensory regularities leading to modulation of cortical responses to both predictive and predicted sensory stimuli. As in the case of goal-directed behavior, such task-irrelevant predictive mechanisms might result from the brain’s intrinsic drive of reducing uncertainty about state transition dynamics of the environment. Abstract Statistical learning of sensory patterns can lead to predictive neural processes enhancing stimulus perception and enabling fast deviancy detection. Predictive processes have been extensively demonstrated when environmental statistical regularities are relevant to task execution. Preliminary evidence indicates that statistical learning can even occur independently of task relevance and top-down attention, although the temporal profile and neural mechanisms underlying sensory predictions and error signals induced by statistical learning of incidental sensory regularities remain unclear. In our study, we adopted an implicit sensory conditioning paradigm that elicited the generation of specific perceptual priors in relation to task-irrelevant audio–visual associations, while recording Electroencephalography (EEG). Our results showed that learning task-irrelevant associations between audio–visual stimuli resulted in anticipatory neural responses to predictive auditory stimuli conveying anticipatory signals of expected visual stimulus presence or absence. Moreover, we observed specific modulation of cortical responses to probabilistic visual stimulus presentation or omission. Pattern similarity analysis indicated that predictive auditory stimuli tended to resemble the response to expected visual stimulus presence or absence. Remarkably, Hierarchical Gaussian filter modeling estimating dynamic changes of prediction error signals in relation to differential probabilistic occurrences of audio–visual stimuli further demonstrated instantiation of predictive neural signals by showing distinct neural processing of prediction error in relation to violation of expected visual stimulus presence or absence. Overall, our findings indicated that statistical learning of non-salient and task-irrelevant perceptual regularities could induce the generation of neural priors at the time of predictive stimulus presentation, possibly conveying sensory-specific information about the predicted consecutive stimulus.


Introduction
Human sensory perception is posited to depend upon the rapid encoding of probabilistic occurrences of perceptual information and the consequent timely generation of specific sensory priors [1][2][3][4][5][6][7].A number of studies indicated that statistical learning of sensory Biology 2024, 13, 576 2 of 19 patterns leads to predictive neural processes, enhancing visual perception and enabling fast deviancy detection [8][9][10][11][12][13][14][15].Predictive models of perception postulate that high-level generative neural signals provide anticipatory neural representations of sensory signals, reducing perceptual surprisal, and minimizing computational effort [3,[16][17][18][19].Accordingly, rapid learning of relevant perceptual regularities has been shown to induce attenuation of neural population reaction to predictable sensory stimuli, as well as enhanced responses to unpredictable information [9,[20][21][22][23][24][25].The neural expectation suppression effect [22] is proposed to arise either from the dampening of the stimulus response as a result of a decrease in global surprise signals [26,27] or from a sharpening of cortical response to specific sensory input and an inherent reduction of prediction errors [10,28].In line with this last assumption, prior expectations can facilitate sensory processing by increasing fine-tuning of the primary sensory cortex [10,21,28,29].Moreover, the encoding of predictable sensory patterns was also associated with anticipatory perceptual processes such as pre-stimulus-specific baseline shifts [30,31], and the pre-activation of the primary sensory cortex immediately before stimulus presentation, resulting in timely cortical instantiation of the expected representational content for the stimulus [32][33][34], as well as attenuation of alpha-band oscillations and the contingent negative variation [35].On the other hand, expectation-related effects are still under debate, and in some statistical learning studies, these effects do not emerge when checking for potential confounding factors such as repetition suppression, adaptation, and novelty effects [36][37][38][39][40][41].
In addition, learned perceptual statistical regularities and resulting predictive neural processes are significantly influenced by task relevance and attention [32,[42][43][44][45][46].However, statistical learning in the absence of explicit top-down attention can also occur, and this leads to an attentional suppression effect [47].In addition, attenuation of cortical fMRI responses to predicted stimuli was also observed in cases of task-irrelevant sensory input [9,21,48].Stimulus relevance and probability appear to have dissociable effects on visual processing [14,45].For instance, stimulus relevance can enhance the precision of stimulus processing by suppressing internal noise, whereas sensory signal probability would bias stimulus detection by increasing the baseline activity of signal-selective units during early visual processing [14].Notably, the learned probability of sensory stimulus occurrence can significantly and differentially impact both late and early response stages [14,32].
Overall, whether the brain can learn statistical associations of task-irrelevant stimuli remains contentious.Some studies provided evidence supporting this associative learning mechanism [20,21,48], whereas other research suggested that an association of task-irrelevant sensory stimuli does not occur [49][50][51].In addition, while the processes associated with responses to predictable information, such as cortical response attenuation, are well documented, the neural mechanisms underlying the postulated anticipatory processing of predictive and predicted task-irrelevant sensory information have yet to be described.
Here, we assumed that the learning of non-salient sensory regularities can be shaped through associative mechanisms that might lead to sensory-conditioned stimuli capable of inducing anticipatory responses similar to unconditioned responses [52][53][54][55].In previous sensory pre-conditioning studies, an initial sensory conditioning phase typically preceded Pavlovian conditioning; this aspect raises the question of whether it was in fact the consecutive conditioning phase involving potent biological stimuli that significantly influenced learned associations between neutral sensory stimuli.On the other hand, functional neuroimaging studies suggested that learned associations of incidental audio-visual regularities can occur independently of stimulus salience and relevance and induce modifications of neural responses to paired sensory inputs [21,29].Associative effects of such neural responses were explained in terms of predictions and prediction errors of expected sensory regularities [21].,However, as of yet, the temporal and representational aspects of the neural responses underlying predictions and prediction error signals related to learned associations of non-salient and task-irrelevant, yet statistically regular, sensory stimuli have not been elucidated.We based our hypothesis in our investigation on predictive processing principles.We hypothesized that associative learning of non-relevant, but probabilistically interrelated, neutral sensory stimuli resulted in early neural responses to predictive sensory stimuli associated with anticipatory signals which are informative of the subsequent sensory input, as well as a specific response modulation to the predicted stimulus.To this aim, we collected electroencephalographic (EEG) data during a novel implicit sensory conditioning paradigm that, throughout its unfolding, was expected to induce increasingly specific perceptual priors relative to probabilistic but task-irrelevant audio-visual associations.The experimental protocol entailed participants being exposed to incidental probabilistic associations of non-salient audio-visual patterns while engaged in a main stimulus detection task (Figure 1).We specifically assumed that the implicit learning of task-irrelevant audio-visual associations would result in anticipatory brain activity associated with auditory stimuli (audio period).This anticipatory activity was expected to convey information about the consecutive visual stimulus presentation (post-audio period), possibly being evidenced by the increased similarity between the pattern of neural response to the predictive auditory stimulus and that to the predicted visual stimulus.We also conjectured that such an anticipatory forward mechanism should be differentially implemented in cases of probabilistic visual stimulus presence and absence and would also result in distinct prediction error assessment.In our experimental protocol-learned perceptual associations were assumed to occur exclusively due to repeated exposure to statistical regularity of sensory stimuli.To test our hypotheses, we examined neural activity during sensory conditioning by combining ERP analysis and multivariate classification.We then performed Pattern Similarity Analysis [56,57] to assess the relationship of activation pattern evoked by a predictive auditory stimulus with that evoked by expected visual stimulus presentation or omission, by computing the cross-validated Mahalanobis distance (cvMD) [58].Finally, we modeled the evolution of prediction error signals during implicit learning of incidental perceptual regularities [11,59,60] using the Hierarchical Gaussian filter (HGF) modeling, a computational approach that was shown to successfully explain associative learning in terms of probabilistic perceptual priors [13].The HGF, a Bayesian ideal observer model for characterizing inferences of uncertain perceptual inputs [61], allowed us to estimate dynamic changes of prediction error signals in relation to differential probabilistic occurrences of visual input presentation and omission during sensory conditioning.Overall, our study indicates that the human brain rapidly encodes task-irrelevant pairings of sensory stimuli, and it shows that neural responses to learned associations of paired audio-visual stimuli can be explained within the predictive coding framework.

Figure 1. (A).
The experimental procedure entailed passive exposition to a stream of task-irrelevant auditory stimuli, A1 and A2.These consist of low and high-frequency pure tones, 250 Hz and 500 Hz, respectively (counterbalanced across participants).These were paired with task-irrelevant visual stimuli, V1 and V2, consisting of two white Gabor patches with a 45° and a 135° orientation, presented against a grey background.The main task entailed detecting and pressing a button in response to two specific targets only, consisting of an auditory stimulus combining both A1 and A2 and a visual stimulus combining both V1 and V2.(B).Contingency table showing the probabilistic occurrences of visual stimuli presentation given each auditory stimulus, and the resulting conditional probabilities of the four different trials.V0 refers to the absence of a visual stimulus.(C).Trial structure implied fixation cross presentation for 100 ms, followed 500 ms later by the presentation of one of two equally probable auditory stimuli for 600 ms, this, in turn, being followed 50 ms after its cessation by the presentation of one of two Gabor patches for 500 ms.Four target stimuli were also interleaved in a block after the ITI period.The representation of the two trials' time course shows that the target stimulus (task-relevant, lower line) was not always preceded by the audio-visual association (task-irrelevant, upper line).(D).Graphical description of the Hierarchical Gaussian Filter (HGF) model adopted to depict individual trajectories of precisionweighted prediction error on the basis of ongoing variability in weighting between sensory evidence and perceptual beliefs.(E).Exemplary single subject precision-weighted prediction error trajectory in relation to V1, V0|A1 condition.(F).The EEG montage we adopted shows the position of the 27 channels.

Participants
Twenty-one volunteers (13 females, range 19-32, mean age 24.3 ± 3.4 (SD)) participated in the study.All were right-handed with normal or corrected-to-normal vision and normal hearing, they had no history of neurological disorders, and they were not taking any neurological medications.All participants gave informed written consent.The study was conducted in accordance with the Declaration of Helsinki and approved by the Ethics Committee of the University of Trento.The experimental procedure entailed passive exposition to a stream of task-irrelevant auditory stimuli, A1 and A2.These consist of low and high-frequency pure tones, 250 Hz and 500 Hz, respectively (counterbalanced across participants).These were paired with task-irrelevant visual stimuli, V1 and V2, consisting of two white Gabor patches with a 45 • and a 135 • orientation, presented against a grey background.The main task entailed detecting and pressing a button in response to two specific targets only, consisting of an auditory stimulus combining both A1 and A2 and a visual stimulus combining both V1 and V2.(B).Contingency table showing the probabilistic occurrences of visual stimuli presentation given each auditory stimulus, and the resulting conditional probabilities of the four different trials.V0 refers to the absence of a visual stimulus.(C).Trial structure implied fixation cross presentation for 100 ms, followed 500 ms later by the presentation of one of two equally probable auditory stimuli for 600 ms, this, in turn, being followed 50 ms after its cessation by the presentation of one of two Gabor patches for 500 ms.Four target stimuli were also interleaved in a block after the ITI period.The representation of the two trials' time course shows that the target stimulus (task-relevant, lower line) was not always preceded by the audio-visual association (task-irrelevant, upper line).(D).Graphical description of the Hierarchical Gaussian Filter (HGF) model adopted to depict individual trajectories of precision-weighted prediction error on the basis of ongoing variability in weighting between sensory evidence and perceptual beliefs.(E).Exemplary single subject precision-weighted prediction error trajectory in relation to V1, V0|A1 condition.(F).The EEG montage we adopted shows the position of the 27 channels.

Participants
Twenty-one volunteers (13 females, range 19-32, mean age 24.3 ± 3.4 (SD)) participated in the study.All were right-handed with normal or corrected-to-normal vision and normal hearing, they had no history of neurological disorders, and they were not taking any neurological medications.All participants gave informed written consent.The study was conducted in accordance with the Declaration of Helsinki and approved by the Ethics Committee of the University of Trento.

Procedure
During the experimental procedure participants were exposed to a stream of auditory and visual stimuli while sitting in a dimly lit booth at a distance of 1 m from the monitor (22.5"VIEWPixx; resolution: 1024 × 768 pixels; refresh rate: 100 Hz; screen width: 50 cm).Participants were informed that they were involved in an audio-visual detection task consisting of a button response to the target stimuli presentation only.Auditory stimuli A1 and A2 consisted of low and high-frequency pure tones, 250 Hz and 500 Hz, respectively, whereas visual stimuli V1 and V2 consisted of two white-colored Gabor patches, sinusoidal gratings with a Gaussian envelope, with 45 • and 135 • orientation (4.4 • × 3.4 • visual angle, generated with Gaussian envelope, standard deviation = 18.0, spatial frequency = 0.08 cycles/pixel), presented against a grey background (Figure 1A).In each trial, auditory stimuli were followed by the presentation of visual stimuli according to an equivalent temporal sequence with two opposite probability distributions resulting in high frequent and low frequent visual stimulus occurrence (Figure 1B).In the condition of highly frequent visual stimulus occurrence, A1 stimulus was followed by V1 Gabor patch (V1|A1) 90% of the time and 10% by visual stimulus absence, V0|A1.In the condition of low frequent visual stimulus occurrence, the A2 stimulus was followed by a visual stimulus (V2|A2) 10% of the time and by visual stimulus absence 90% of the time, V0|A2.Pairings of the auditory and visual stimuli were counterbalanced across participants.Each trial started with a fixation cross presented for 100 ms, followed 500 ms later by the presentation of one of two equally probable auditory stimuli for 600 ms, and that, in turn, was followed 50 ms after its cessation by the presentation of one of two Gabor patches for 500 ms (Figure 1C).Trials of V0|A1 and V0|A2 conditions, which were of equal length, entailed no Gabor patch presentation.Trials were interspersed with an inter-trial interval (ITI) of 2500 ms ± 500 ms.The main audio-visual target detection task consisted of a button press response only when presented with specific target stimuli represented by an auditory target combining both the A1 and A2 stimuli, and a visual target combining both the V1 and V2 stimuli.Both of these lasted 500 ms and were followed by equal ITI.The experimental session consisted of 400 trials presented in 10 blocks, each including the random presentation of 4 perceptual targets, with a total duration of about 40 min.The experimental protocol was implemented using OpenSesame (v.2.8) and PsychoPy (v.2.1) as backend [62].The experimental procedure aimed at inducing more attentional resource allocation to task-relevant perceptual targets.Importantly, probabilistic contingencies of audio-visual pairing were completely irrelevant to the audio-visual target detection task.

EEG Data Acquisition and Preprocessing
EEG data were recorded with a standard 10-5 system and 27 Ag/AgCl electrodes cap (EasyCap, Brain Products, Gilching, Germany) at a sampling rate of 1 kHz.Impedance was kept below 10 kΩ for all channels.AFz was used as the ground and the right mastoid was used as a reference.Electrodes were approximately evenly spaced and positioned at the following scalp sites: Fpz, Fz, F3, F4, F7, F8, F9, F10, FC5, FC6, T7, C3, Cz, C4, T8, CP5, CP6, P7, P3, Pz, P4, P8, PO7, PO8, O1, Oz, and O2 (Figure 1F).All preprocessing steps were conducted using EEGLAB [63] in accordance with guidelines and recommendations for EEG data preprocessing such as HAPPE [64], which are also applicable in the case of low-density recordings, HAPPILEE [65].Spherical interpolation was carried out on a limited number of bad channels on the basis of channel correlation lower than 0.85 on average with respect to its neighbors and guided by visual inspection (average number of interpolated channels: 0.74, range: 0-3).Data were down-sampled at 250 Hz, high-pass filtered at 0.1 Hz and low-pass filtered at 80 Hz, using a Butterworth IIR filter with model order 2. CleanLine (https://github.com/sccn/cleanlineaccessed on 23 July 2023) with default parameters was used to remove power line 50 Hz noise and its harmonics up to 200 Hz.The data were then re-referenced to a common average reference [65] and epoched between −300 ms and 1300 ms relative to the onset of the auditory stimulus with a baseline correction between −300 ms and 0 ms.Artifact rejection was performed through visual inspection and by an automatic procedure excluding epochs with very large signal amplitudes (detection threshold = ±500).The average number of trials rejected per participant was 1.1% (SD = 2.1%, range 0-7.3%).Stereotypical artefacts, including eyeblinks, eye movements, and muscle artefacts, were detected via independent component analysis using the extended Infomax algorithm [66].A rejection strategy based on ICLabel [67] and visual inspection resulted in the removal of an average number of independent components equal to 9.33 (±3.48 SD).Finally, the data were converted to Fieldtrip format [68] for subsequent analyses.

EEG Data Analysis
Data analysis aimed at assessing implicit associative learning of paired audio-visual stimuli by investigating neural responses evoked by both predictive auditory stimuli and predicted visual stimulus presentation or omission.The analysis focused on two main epochs by performing grand averaging, considering auditory and visual stimuli separately: 0-650 ms relative to auditory stimuli presentation (audio period), and 650-1300 ms relative to Gabor patch presence or absence (post-audio period).Trials were divided into three equivalent groups to analyze the initial (first third of trials) and final phases (last third of trials) of sensory conditioning, and each phase consisted of 54 trials (except possible rejection of trials due to artifacts).Conventional event-related potential (ERP) analysis was first performed on both audio and post-audio periods considering initial and final trials.All trials of A1, A2, V1|A1, and V0|A2 conditions were averaged separately in relation to initial and final phases, whereas trials of V0|A1 and V2|A2 conditions were not analyzed due to their limited number required by the specific contingencies schedule of our conditioning paradigm.As in previous related studies [51,69], for ERP analysis we adopted an approach that considers the average neural response over comparable predefined regions of interest (ROI) using frontal (Fpz, Fz, F3, F4, F7, F8, F9, F10), temporo-parietal (FC5, FC6, T7, C3, Cz, C4, T8, CP5, CP6, P7, P3, Pz), and parieto-occipital channels (P4, P8, PO7, PO8, O1, Oz, O2), and thus enables more robust statistics through cluster-based correction.In addition, multivariate pattern classification based on Linear Discriminant Analysis (LDA) examining EEG signal differences between A1 and A2 at the subject level was performed considering all conditioning trials and channels as samples and features, respectively, with the MVPA-Light toolbox [70] and custom MATLAB (Mathworks, Natick, MA, USA) scripts.Z-scoring was applied across samples for each time point separately to normalize channel variances and remove baseline shifts.A 5-fold cross validation scheme was adopted and the Area Under the Curve (AUC) was used as a performance measure of LDA.An empirical chance level was obtained by running twice the same classification analysis with the same hyperparameter but with permuted labels.Statistical significance of model performance with respect to empirical chance level was assessed at the group level (paired permutation t-test two tailed, α = 0.05) using mass univariate cluster-based permutation tests (10,000 iterations) and maxsum as cluster statistic, a valid and powerful way of dealing with the problem of multiple comparisons [71,72].Effect size was estimated using Cohen's d (d) and Scaled Jeffrey-Zellner-Siow Bayes Factor t-test (BF 10 , Cauchy prior with a scale parameter equal to √ 2/2), reporting the peak value inside the significant cluster.Analysis of relevant channels for classification was performed by converting the estimated weights of the LDA model at each fold into interpretable activation patterns [73].Furthermore, we performed pattern similarity analysis to estimate the similarity of brain responses evoked by predictive A1 and A2 stimuli with those evoked by expected V1 and V0, by calculating Pattern Dissimilarity Matrices based on the crossvalidated Mahalanobis distance (cvMD) [58] for each participant.This measure consists of splitting the trials in train and test sets, subtracting the average multivariate pattern across channels between conditions in both train and test sets, and finally performing a matrix multiplication between, the difference vector from the train set, the inverse of the covariance matrix estimated on the train set and the difference vector from the test set [58].We computed the cvMD dissimilarity measure across trials between the auditory period and the visual period at each time-point of the selected time window and at each ROI.A 5-fold cross validation scheme was applied to assess the similarity of EEG signal between audio (200-650 ms) and post-audio (0-550 ms) periods, considering time points (EEG samples) of initial and final trials.These time windows were selected as a result of the ERP analysis and MVPA.Statistical significance of Dissimilarity Matrices (DM) differences between initial and final trials was tested at group level with cluster-based permutation tests with the same hyperparameters as described above.

Hierarchical Gaussian Filter Modelling
The HGF is a Bayesian generative model [61,74] of perceptual inference on a changing environment based on sequential input [13,75].The HGF consists of perceptual and response models, representing a Bayesian ideal observer who receives a sequence of inputs and generates behavioral responses.Since our experimental design deliberately precluded behavioral responses, we used only the perceptual model [76].In this framework, a perceptual model comprised 3 hierarchal hidden states (x), which accounted for a multilevel belief updating process of the hierarchically related environmental states giving rise to sensory inputs, and the observed input (y) representing the actual occurrence of a stimulus in a given trial (Figure 1D).
Our HGF model assumed that environmental hidden states evolved conditionally on the states at the immediately higher level.The hidden states processed at the first level of the perceptual model represented a sequence of beliefs (x t 1 ) about stimulus occurrence, that is, whether a visual stimulus was presented (y t = 1) or absent (y t = 0) at trial t, and was modelled as follows: x where s x t 2 := 1 + exp x t 2 −1 is the logistic sigmoid function.Here, the hidden states at the second level (x t 2 ) is an unbounded real parameter of the probability that x t 1 = 1, thus representing the current belief of the probability that a given stimulus occurred.Such a hidden state process evolves according to a Gaussian random walk: which depends on both its value at a previous trial t, and the hidden state at the third level of the hierarchy.In particular, the higher-level hidden state process (x 3 ) determines the log-volatility of the hidden state process at the second level, thus codifying the volatility of the environment during the time course of the experiment.This process evolves according to a Gaussian random walk: The parameter set (κ, ω, ϑ) determined the dispersion of the random walks at different levels of the hierarchy and allowed us to shape individual difference in learning.By inverting the generative model, given a sequence of observations (y), it was possible to obtain the updating process of the trial-by-trial estimates of the hidden state variables.
The update rules shared a common structure across the model's hierarchy: at any level i the update of the posterior mean µ t i of the state x i , that represented the belief on trial k, was proportional to the precision-weighted prediction error (pwPE) ε t i as follows: Biology 2024, 13, 576 As shown in Equations ( 4)-( 6), in each trial, a belief update µ t−1 i − µ t i is proportional to the prediction error at the level below δ t i−1 .The pwPE is the product of the prediction error δ t i−1 and a precision ratio ψ t i that depends on the precision (inverse variance, Equation ( 5)) of the prediction at the level below πt i−1 and the current level π t i .In this application, we were interested in the updated equations of the hidden states at the second level, which have a general form similar to those of traditional reinforcement learning models, such as the Rescorla-Wagner model [77].The pwPE on the second level, was thus assumed to be responsible for the learned perceptual associations.The nature of the pwPE could be described through the following update equation of the mean of the second level: where the last term represents the prediction error at the first level weighted by the precision term σ t 2 (see [61] for a general derivation and more mathematical details).Individual trajectories of pwPEs with separate models for V1, V0|A1 and V2, V0|A2 were calculated by estimating the parameters that minimized Bayesian Surprise using the Broyden-Fletcher-Goldfar-Shannon (BFGS) quasi-Newton optimization algorithm.We determined these Bayes optimal perceptual parameters by inverting the perceptual model based on the stimulus sequence alone and a predefined prior for each parameter (HGF toolbox, version 5.2 implemented via the Translational Algorithms for Psychiatry Advancing Science toolbox).These model-derived trajectories of pwPEs (Figure 1E) from the second level were used as regressors in a general linear model (GLM) applied to each channel-time point pair for each participant.We used the R 2 measure for evaluating the goodness of fit and averaged these values over the selected ROIs.Statistical significance was tested at the group level with cluster-based permutation tests and using the same hyperparameters as previously described.

Results
We collected EEG and behavioral data from 21 human volunteers exposed to a stream of non-target auditory and visual stimuli, whilst being involved in a main audio-visual detection task consisting of button press responses to target stimuli presentation (Figure 1A).Crucially, we manipulated the transition probabilities between the non-target stimuli so that non-target audio-visual co-occurrences had no predictable effects on target presentation, thus making the learning of these associations task-irrelevant (Figure 1B,C).Participants debriefed at the end of the experiment reported not being consciously aware of stimuli pairings.When specifically questioned about possible stimuli associations they reported not to have noticed either any regularities of stimuli presentation or specific audio-visual pairings.

Event-Related Potentials
We performed conventional event-related potential (ERP) analysis on both audio and post-audio periods considering the initial and final trials so as to assess the evoked activity throughout the learning of statistical associations of non-target auditory and visual stimuli.

Multivariate Classification
Multivariate decoding, performed using Linear Discriminant Analysis-based classification, aimed to assess EEG signal differences between A1 and A2 during the whole conditioning session.Such expected differences should reflect differential processing of equivalent auditory stimuli that distinctively anticipated visual stimulus presence or absence.Our results showed significant discrimination performance between A1 and A2

Multivariate Classification
Multivariate decoding, performed using Linear Discriminant Analysis-based classification, aimed to assess EEG signal differences between A1 and A2 during the whole conditioning session.Such expected differences should reflect differential processing of equivalent auditory stimuli that distinctively anticipated visual stimulus presence or absence.Our results showed significant discrimination performance between A1 and A2 with respect to chance level in three different time windows (Figure 3A): 192-340 ms (AUC peak = 0.56, d = 1.17,BF 10 = 771.98,p < 0.0001, cluster corrected), 344-444 ms (AUC peak = 0.53, d = 0.85, BF 10 = 39.42,p = 0.0017, cluster corrected), and 500-540 ms (AUC peak = 0.52, d = 0.73, BF 10 = 13.04,p = 0.0484, cluster corrected), that is immediately preceding probabilistic visual stimulus presence or absence.These results thus indicated distinct processing of equivalent A1 and A2 stimuli, differentially predicting V1 and V0, respectively.Relevant channels for classification were located in the frontal regions for the first significant time window and mostly in the temporo-occipital regions for the other two time windows.
with respect to chance level in three different time windows (Figure 3A): 192-340 ms (  = 0.56,  = 1.17,  = 771.98, < 0.0001 , cluster corrected), 344-444 ms ( = 0.53,  = 0.85,  = 39.42, = 0.0017 , cluster corrected), and 500-540 ms (  = 0.52,  = 0.73,  = 13.04, = 0.0484 , cluster corrected), that is immediately preceding probabilistic visual stimulus presence or absence.These results thus indicated distinct processing of equivalent A1 and A2 stimuli, differentially predicting V1 and V0, respectively.Relevant channels for classification were located in the frontal regions for the first significant time window and mostly in the temporo-occipital regions for the other two time windows.For the visual stimulus, the temporal interval ranges between visual stimulus onset and its offset, whereas for the auditory stimulus, it ranges between 200 ms after the auditory stimulus onset (this time was set on the basis of ERP and MVPA classification results) and its offset.

Pattern Similarity Analysis
Pattern similarity analysis aimed to test whether the pattern elicited by the predictive auditory stimulus increasingly resembled the pattern elicited by the expected visual stimulus during sensory conditioning.Pattern Similarity Analysis (PSA) was performed by calculating a dissimilarity matrix (DM) between audio and post-audio periods to specifically assess the relationship of brain responses evoked by predictive A1 and A2 stimuli with those evoked by predicted V1 and V0, respectively (Figure 3B).DM, estimated on the basis of the cross-validated Mahalanobis distance (cvMD), showed that in the V1|A1 condition the cvMD was significantly lower in the final trials with respect to initial trials (left) and V0|A2 (right) condition, for each selected ROI.For each condition, the first column refers to cvMD during initial trials, the second column refers to cvMD during final trials, and the third column depicts statistical differences between initial and final trials.Outlined regions indicate statistical significance.X and Y axes depict the time course of visual and auditory stimulus, respectively.For the visual stimulus, the temporal interval ranges between visual stimulus onset and its offset, whereas for the auditory stimulus, it ranges between 200 ms after the auditory stimulus onset (this time was set on the basis of ERP and MVPA classification results) and its offset.

Pattern Similarity Analysis
Pattern similarity analysis aimed to test whether the pattern elicited by the predictive auditory stimulus increasingly resembled the pattern elicited by the expected visual stimulus during sensory conditioning.Pattern Similarity Analysis (PSA) was performed by calculating a dissimilarity matrix (DM) between audio and post-audio periods to specifically assess the relationship of brain responses evoked by predictive A1 and A2 stimuli with those evoked by predicted V1 and V0, respectively (Figure 3B).DM, estimated on the basis of the cross-validated Mahalanobis distance (cvMD), showed that in the V1|A1 condition the cvMD was significantly lower in the final trials with respect to initial trials in several clusters in frontal, temporo-parietal, and occipital ROIs.In the frontal ROI two main significant clusters were observed, one corresponding to ~300-500 ms in the audio period and ~0-200 ms in the post-audio period (d = 1.36,BF 10 = 4608.53,p < 0.0001, cluster corrected), and another one corresponding to ~300-400 ms in the audio period and ~250-400 ms in the post-audio period (d = 0.84, BF 10 = 36.18,p = 0.0009, cluster corrected).In the temporo-parietal (d = 1.02,BF 10 = 202.98,p = 0.005, cluster corrected) and parieto-occipital (d = 0.99, BF 10 = 161.58,p = 0.015, cluster corrected), ROIs the significant cluster corresponded to ~300-400 ms in the audio period and ~0-150 ms in the post-audio period.In V0|A2 condition, the cvMD was also significantly lower in the final trials with respect to initial trials in several clusters in frontal, temporo-parietal, and occipital ROIs.In the frontal ROI two main significant clusters were observed, one corresponding to ~300-400 ms in the audio period and ~50-400 ms in the post-audio period (d = 1.07,BF 10 = 313.13,p = 0.0048, cluster corrected), and another one corresponding to ~400-500 ms in the audio period and ~50-350 ms in the post-audio period (d = 1.21,BF 10 = 1132.09,p = 0.005, cluster corrected).In the temporo-parietal ROI two significant clusters were also observed, one corresponding to ~200-250 ms in the audio period and ~0-400 ms in the post-audio period (d = 1.10,BF 10 = 434.36,p = 0.002, cluster corrected), and another one corresponding to ~280-380 ms in the audio period and ~0-500 ms in the post-audio period (d = 1.03,BF 10 = 219.11,p = 0.0019, cluster corrected).In addition, a significant cluster was observed in the occipital ROI corresponding to ~320-400 ms in the audio period and ~0-300 ms in the post-audio period (d = 1.21,BF 10 = 1124.35,p = 0.007, cluster corrected).Altogether these results showed decreased dissimilarity for both V1|A1, an V0|A2 conditions during sensory conditioning.

HGF Modelling
We modeled individual trajectories of precision weighted prediction error (pwPE) on the basis of ongoing variability in weighting between sensory evidence and perceptual beliefs using the Hierarchical Gaussian filter (HGF), a Bayesian ideal observer model that attempts to predict future stimuli occurrence on the basis of the history and uncertainty of contextual events.This analysis allowed us to assess differential prediction error processing associated with violation of expected presentation or omission of visual stimulus, and to further demonstrate that associative learning of perceptual stimuli can occur even when perceptual stimuli are tasirrelevant.We thus fed into the HGF model the same sequence of stimuli that every subject was exposed to, and then we obtained the prediction error trajectory.We used these ideal error trajectories as regressors in a general linear model (GLM) applied to each channel-time point pair, for each participant.HGF analysis showed that pwPE, significant for both V1, V0|A1 and V2, V0|A2, was mediated by all selected brain regions, although the largest effect size was in the occipital ROI (Figure 4).Regression analysis resulted in a significant effect for both V1, V0|A1 and V2, V0|A2 (Figure 4) in the interval ~250-550 ms of post-audio period in the frontal (V1, V0|A1: d = 0.66, BF 10 = 7.29, p = 0.0007, cluster corrected; V2, V0|A2: d = 0.72, BF 10 = 11.84,p = 0.0007, cluster corrected) and temporo-parietal ROIs (V1, V0|A1: d = 0.81, BF 10 = 26.01,p = 0.0003, cluster corrected; V2, V0|A2: d = 0.69, BF 10 = 8.89, p = 0.0005, cluster corrected), and ~150-550 ms in the occipital ROI (V1, V0|A1: d = 0.72, BF 10 = 11.96,p = 0.0386, cluster corrected; V2, V0|A2: d = 0.77, BF 10 = 18.89, p < 0.0001, cluster corrected).Finally, R 2 was significantly larger for V1, V0|A1 than for V2, V0|A2 (Figure 4) in the interval ~370-400 ms in the post-audio period in the occipital ROI (d = 0.64, BF 10 = 6.03, p = 0.0398, cluster corrected), indicating differential neural processing of PE in relation to violation of predicted visual stimulus presentation and omission.

Discussion
In our study, we investigated whether incidental but recurrent exposure to non-salient and task-irrelevant probabilistic audio-visual associations induced implicit learning of perceptual patterns resulting in conditioned cortical responses, possibly representing early instantiations of predicted stimulus-related sensory information.To this aim, we assessed both temporal and representational aspects of neural signals associated with the learned conditional probability of paired audio-visual stimuli.Our analyses showed that incidental perceptual regularities were rapidly encoded, leading to differential neural responses to anticipatory auditory stimuli predictive of either the presence or absence of a visual stimulus.ERPs results revealed that in the final trials of sensorial conditioning brain responses to equivalent A1 and A2 stimuli, anticipating probabilistic visual stimulus presence and absence, respectively, were significantly different around 200 ms after stimulus onset, in both the frontal and temporo-parietal regions (Figure 2A,B).Specifically, a response attenuation in the temporo-parietal region was observed in response to A2 but not to A1.This suggests that A1, evoking anticipation of V1 being presented in contrast to A2 preceding an absence of visual stimulus, required differential processing.
In line with the known audiovisual cross-modal effect resulting in visual cortex activity during auditory perception [78,79], increased signal amplitude in the parieto-occipital region was observed in response to both A1 and A2 across all trials.However, a significant attenuation over time in the parieto-occipital region was measured in response to A1, but not to A2.The point of repetition suppression effect only being observed for A1

Discussion
In our study, we investigated whether incidental but recurrent exposure to nonsalient and task-irrelevant probabilistic audio-visual associations induced implicit learning of perceptual patterns resulting in conditioned cortical responses, possibly representing early instantiations of predicted stimulus-related sensory information.To this aim, we assessed both temporal and representational aspects of neural signals associated with the learned conditional probability of paired audio-visual stimuli.Our analyses showed that incidental perceptual regularities were rapidly encoded, leading to differential neural responses to anticipatory auditory stimuli predictive of either the presence or absence of a visual stimulus.ERPs results revealed that in the final trials of sensorial conditioning brain responses to equivalent A1 and A2 stimuli, anticipating probabilistic visual stimulus presence and absence, respectively, were significantly different around 200 ms after stimulus onset, in both the frontal and temporo-parietal regions (Figure 2A,B).Specifically, a response attenuation in the temporo-parietal region was observed in response to A2 but not to A1.This suggests that A1, evoking anticipation of V1 being presented in contrast to A2 preceding an absence of visual stimulus, required differential processing.
In line with the known audiovisual cross-modal effect resulting in visual cortex activity during auditory perception [78,79], increased signal amplitude in the parieto-occipital region was observed in response to both A1 and A2 across all trials.However, a significant attenuation over time in the parieto-occipital region was measured in response to A1, but not to A2.The point of repetition suppression effect only being observed for A1 indicates differential predictive processes in relation to expected visual stimulus presentation and omission.
According to the assumption that the expectation suppression effect reflects decreased surprise signals, we would have expected to observe a similar attenuation for both A1 and A2.Vice versa, in line with the interpretation of sensory attenuation resultant of the sharpening of stimulus-specific representation [2,9,28], we interpreted the parieto-occipital differential response to A1 with respect to A2 as increased response tuning.This is because A1, in contrast to A2, carries information about consecutive V1 presentations.Our results are also in line with previous evidence of enhanced stimulus-specific baseline activity during early sensory processing in relation to probabilistic signal occurrence [14].Multivariate pattern analysis corroborated ERP results showing differential neural responses to A1 and A2, in particular in a temporal interval immediately preceding probabilistic visual stimulus presence or absence and disclosed that such divergent responses manifested early in the frontal region and later in the temporo-occipital areas (Figure 3A).These results might also be potentially ascribable to the inherent difference of auditory stimuli.However, the analysis of feature relevance, showing that occipital channels were particularly involved in later stages of the classification (350-600 ms), suggests that the two auditory stimuli might indeed convey differential predictive information of consequent stimulus presentation.Pattern similarity analysis revealed that for the V1|A1 condition the cvMD between audio and post-audio period, although initially large, decreased over time in the frontal, parieto-occipital, and temporo-parietal regions, resulting in decreased dissimilarity of neural activity between ~0-150 ms after V1 onset and ~250-400 ms after A1 onset (Figure 3B).Decreased cvMD was also observed in the frontal region between neural activity ~400-500 ms after A1 onset and ~0-150 ms after V1 onset (Figure 3B).
Altogether, our findings indicated that over time the pattern of neural activity elicited by A1 increasingly resembled that elicited by V1, and thus suggest that some kind of stimulus-specific perceptual priors relative to probabilistic but task-irrelevant audio-visual association were instantiated at the cortical level during the unfolding of sensory conditioning.Previous studies showed that prior expectations can elicit anticipatory sensory-like neural representations of predictable sensory information, in particular when this is somehow related to the task [28,32].In particular, Kok and colleagues reported that a predictive auditory cue evoked early sensory representation in the primary visual cortex immediately before the expected visual stimulus [32], or at the time of the expected but omitted visual stimulus [28].In our study, in line with previous findings showing an expectation-induced decrease of primary visual cortex activity [9,10,28,80], we also measured an attenuated response over time during V1|A1, and immediately preceding V1, in the parieto-occipital and frontal regions.Similarity analysis did not show evidence of pre-stimulus neural instantiation of expected sensory input immediately before visual stimulus presentation (−50 to 0 ms) as in Kok et al. [32] but a similar effect occurred earlier at the presentation of the predictive auditory stimulus.
Notably, the cvMD also decreased over time in relation to V0|A2 condition in the frontal, parieto-occipital, and temporo-parietal regions.In particular, decreased A2-related cvMD in the temporo-parietal region was observed between neural activity around 200-250 ms and 280-350 ms in the audio period, and in the V0-V2 post-audio period (0-400 ms).These results are possibly ascribable to the attenuated response to A2, resulting in increased similarity with the response to frequent V2 omission.On the other hand, as for the A1 stimulus, these effects might also reflect the tuning of A2 response progressively incorporating perceptual priors of V2 occurrence's low probability.Accordingly, decreased dissimilarity was also observed in the parieto-occipital region between neural activity occurring 320-400 ms after A2 and during the V0-V2 time period.Moreover, neural attenuation over time was also observed for the V0|A2 condition, in both frontal and parieto-occipital regions, in a time window anticipating and also corresponding to V2 presentation.This further suggests an increased encoding of low V2 occurrence probability.
Overall, our results indicated that forward neural processes anticipating some specific aspects of consecutive visual stimulus presentation might occur earlier during the initial processing of predictive stimulus; however, due to limited spatial resolution, the exact representational content of such anticipatory activity has yet to be fully clarified.
As reported in previous studies [10,21], we observed anticipatory neural mechanisms independently of stimulus salience, as our stimuli consisted of non-salient abstracts and auditory and visual stimuli.In addition, the observed effects were task-independent, since statistical regularity of paired audiovisual stimuli was not necessary and irrelevant to detecting the target stimuli.This result is also in line with studies showing that task-irrelevant visual perceptual learning can occur as a result of mere exposure to perceptual features.
It has been proposed that several aspects of perceptual and statistical learning might be unified in the framework of Hierarchical Bayesian [81].In both perceptual and statistical learning, attention plays an important role [45,82].In visual perceptual learning, attention enhances bottom-up signals from task-relevant features, whereas it decreases signals from task-irrelevant features; however, visual perceptual learning of task-irrelevant features can also occur as far as they can be optimally expected (suprathreshold presentation) [82].Similarly, in statistical learning, attenuation of neural responses to predictable stimuli vanishes when they are not expected [45].In our study, as non-target stimuli shared both auditory and visual features with target stimuli, we supposed that they were likely expected, however overt encoding of the statistical regularity of audio-visual associations was not necessarily completely irrelevant to the task.
Finally, the observed conditioned responses in the sensory cortices appeared to be mediated by frontal and prefrontal areas.The differential responses to A1 and A2 in the frontal ROI, preceding changes in temporo-occipital ROIs suggest that prefrontal regions might support mutual information exchange between auditory and visual cortices [83], likely in relation to temporal aspects of perceptual regularities [84][85][86] for timely instantiation of specific perceptual priors.Sensory priors inducing auditory-cued shaping of visual cortex responses in both V1|A1 and V0|A2 conditions might then be mediated by direct interactions between auditory and visual cortices [80].
Remarkably, computational modeling of pwPE trajectories further demonstrated the instantiation of predictive neural signals by showing distinct neural processing of prediction error in relation to violations of expected visual stimulus presence or absence.Moreover, differential neural processing of PE correlated with activity in frontal, temporal, and occipital areas (Figure 4) at latencies corresponding to those of typical event-related potentials elicited by deviant stimuli [87,88].The precision of predictions for audiovisual patterns, likely mediated by temporal regions such as the superior temporal gyrus [89,90], might trigger a gradual update of prefrontal regions-mediated cortical representations of expected V1 and V0 [49,91].Indeed, analysis of pwPE trajectories revealed a significant difference of V1|A1 with respect to V0|A2 occurring about 300-400 ms after V1 and V0 onset in the occipital region, indicating stimulus-specific differential processing of prediction violation [92,93].

Conclusions
Our study demonstrated rapid encoding of incidental but probabilistic audio-visual regularities leading to stimulus-specific modulation of cortical responses to both predictive and predicted sensory stimuli.These findings corroborate previous evidence of statistical learning of task-irrelevant sensory regularities [21,29] and extend it by showing that the acquired sensory associations lead to specific predictive processes resembling those observed in relation to learned task-relevant sensory patterns.In particular, we showed that sensory conditioning of task-irrelevant audio-visual associations appeared to induce increased similarity between the neural response to predictive auditory stimuli and the response to predicted visual stimulus presence or absence.Remarkably, Hierarchical Gaussian filter modeling, estimating dynamic changes of prediction error signals in relation to differential probabilistic occurrences of audio-visual stimuli, further demonstrated instantiation of predictive neural signals by showing distinct neural processing of prediction error in relation to violations of expected visual stimulus presence or absence.Overall, our findings suggest that statistical learning of non-salient and task-irrelevant perceptual regularities might induce early generation of neural priors at the time of predictive stimulus presentation conveying sensory-specific information of the predicted consecutive stimulus.However, the exact nature of sensory-specific priors is yet to be elucidated, in particular to what extent predictive stimuli anticipate the expected consecutive visual stimulus.Further studies using more conventional conditioning paradigms and using multiple predictable sensory stimuli with differential probabilistic schemes, might help to clarify this aspect.Moreover, advanced techniques such as high-resolution EEG and fMRI retinotopic mapping might definitely permit delineating a clearer representational pattern elicited by the predictive neural signals.Finally, it remains unclear whether encoding of statistical regularities of sensory stimuli occurred implicitly or explicitly.Explicit attention to task-irrelevant audiovisual stimuli was unnecessary to perform our target detection task, on the other hand, our target stimuli shared some features with either auditory or visual stimuli that possibly biased attention.
Nevertheless, here as in the case of goal-directed behavior where predictive mechanisms were related to information-seeking random exploration [91,94], the intrinsic neural need to reduce uncertainty about state transition dynamics of the environment might also explain learned incidental probabilistic sensory patterns.Learning sensory representations might then depend on automatic perceptual processes exploiting either reward statistics of past experience or beliefs about future representations [95] that optimize neural computations for adaptive behavior.
Ultimately, in accordance with predictive brain principles, our results suggest that associative processes might even occur exclusively at the perceptual level, possibly as a consequence of Hebbian neural plasticity [96], and that stimulus salience that is typically considered a critical element for learning in classical conditioning might not, in fact, be strictly required [97,98].

Figure 1 .
Figure 1.(A).The experimental procedure entailed passive exposition to a stream of task-irrelevant auditory stimuli, A1 and A2.These consist of low and high-frequency pure tones, 250 Hz and 500 Hz, respectively (counterbalanced across participants).These were paired with task-irrelevant visual stimuli, V1 and V2, consisting of two white Gabor patches with a 45 • and a 135 • orientation, presented against a grey background.The main task entailed detecting and pressing a button in response to two specific targets only, consisting of an auditory stimulus combining both A1 and A2 and a visual stimulus combining both V1 and V2.(B).Contingency table showing the probabilistic occurrences of visual stimuli presentation given each auditory stimulus, and the resulting conditional probabilities of the four different trials.V0 refers to the absence of a visual stimulus.(C).Trial structure implied fixation cross presentation for 100 ms, followed 500 ms later by the presentation of one of two equally probable auditory stimuli for 600 ms, this, in turn, being followed 50 ms after its cessation by the presentation of one of two Gabor patches for 500 ms.Four target stimuli were also interleaved in a block after the ITI period.The representation of the two trials' time course shows that the target stimulus (task-relevant, lower line) was not always preceded by the audio-visual association (task-irrelevant, upper line).(D).Graphical description of the Hierarchical Gaussian Filter (HGF) model adopted to depict individual trajectories of precision-weighted prediction error on the basis of ongoing variability in weighting between sensory evidence and perceptual beliefs.(E).Exemplary single subject precision-weighted prediction error trajectory in relation to V1, V0|A1 condition.(F).The EEG montage we adopted shows the position of the 27 channels.

Figure 2 .
Figure 2. (A).Top, ERPs of initial (A1 light blue, A2 light red) and final (A1 dark blue, A2 dark red) trials in relation to auditory stimuli onset for each selected ROI.Shading refers to SEM across participants, horizontal bars refer to statistical significance.Bottom, topographic maps depicting wholebrain spatial distribution of EEG signal across 100 ms intervals after auditory stimuli onset.In the legend, filled squares indicate the initial or final blocks of trials from which the ERPs are computed (B).Top, Full time-course ERPs of initial (V1|A1 light blue, V0|A2 light red) and final (V1|A1 dark blue, V0|A2 dark red) trials including both auditory and visual stimuli presentation for each selected ROI.Bottom, Topographic maps depicting whole-brain spatial distribution of EEG signal across 100 ms intervals after auditory stimuli onset.

Figure 2 .
Figure 2. (A).Top, ERPs of initial (A1 light blue, A2 light red) and final (A1 dark blue, A2 dark red) trials in relation to auditory stimuli onset for each selected ROI.Shading refers to SEM across participants, horizontal bars refer to statistical significance.Bottom, topographic maps depicting whole-brain spatial distribution of EEG signal across 100 ms intervals after auditory stimuli onset.In the legend, filled squares indicate the initial or final blocks of trials from which the ERPs are computed (B).Top, Full time-course ERPs of initial (V1|A1 light blue, V0|A2 light red) and final (V1|A1 dark blue, V0|A2 dark red) trials including both auditory and visual stimuli presentation for each selected ROI.Bottom, Topographic maps depicting whole-brain spatial distribution of EEG signal across 100 ms intervals after auditory stimuli onset.

Figure 3 .
Figure 3. (A).Left, EEG signals classification performance (Area Under the Curve) of A1 versus A2 during audio period (green) against estimated chance level (orange).Shading indicates SEM across folds.Green horizontal lines indicate statistical significance.Right, topographic maps of activation patterns computed from the model's parameters, representing feature importance for classification performance in relation to EEG channels and time windows.(B).Pattern Similarity Analysis.PSA aimed to test whether the pattern elicited by the predictive auditory stimulus increasingly resembled the pattern elicited by the expected visual stimulus during sensory conditioning.The images show Pattern Dissimilarity Matrices of EEG signal considering audio and post-audio periods for V1|A1 (left) and V0|A2 (right) condition, for each selected ROI.For each condition, the first column refers to cvMD during initial trials, the second column refers to cvMD during final trials, and the third column depicts statistical differences between initial and final trials.Outlined regions indicate statistical significance.X and Y axes depict the time course of visual and auditory stimulus, respectively.For the visual stimulus, the temporal interval ranges between visual stimulus onset and its offset, whereas for the auditory stimulus, it ranges between 200 ms after the auditory stimulus onset (this time was set on the basis of ERP and MVPA classification results) and its offset.

Figure 3 .
Figure 3. (A).Left, EEG signals classification performance (Area Under the Curve) of A1 versus A2 during audio period (green) against estimated chance level (orange).Shading indicates SEM across folds.Green horizontal lines indicate statistical significance.Right, topographic maps of activation patterns computed from the model's parameters, representing feature importance for classification performance in relation to EEG channels and time windows.(B).Pattern Similarity Analysis.PSA aimed to test whether the pattern elicited by the predictive auditory stimulus increasingly resembled the pattern elicited by the expected visual stimulus during sensory conditioning.The images show Pattern Dissimilarity Matrices of EEG signal considering audio and post-audio periods for V1|A1(left) and V0|A2 (right) condition, for each selected ROI.For each condition, the first column refers to cvMD during initial trials, the second column refers to cvMD during final trials, and the third column depicts statistical differences between initial and final trials.Outlined regions indicate statistical significance.X and Y axes depict the time course of visual and auditory stimulus, respectively.For the visual stimulus, the temporal interval ranges between visual stimulus onset and its offset, whereas for the auditory stimulus, it ranges between 200 ms after the auditory stimulus onset (this time was set on the basis of ERP and MVPA classification results) and its offset.

Figure 4 .
Figure 4. (Top).Line plots of GLM performance ( ) fitting HGF model-derived precision-weighted prediction error trajectories to each EEG channel, averaged across ROIs, for V1, V0|A1 (solid blue) and V2, V0|A2 (solid red) conditions.Dotted lines represent average performances during the baseline period (preceding the onset of visual stimuli).Shading indicates SEM across participants.Horizontal lines indicate the statistical significance of differences between each model and baseline (blue and red for V1, V0|A1 and V2, V0|A2, respectively), and between V1, V0|A1 and V2, V0|A2 (grey).(Bottom).Topographic maps depicting whole-brain spatial distribution of GLM performance for each model across 50 ms intervals in the post-audio period.

Figure 4 .
Figure 4. (Top).Line plots of GLM performance (R 2 ) fitting HGF model-derived precision-weighted prediction error trajectories to each EEG channel, averaged across ROIs, for V1, V0|A1 (solid blue) and V2, V0|A2 (solid red) conditions.Dotted lines represent average performances during the baseline period (preceding the onset of visual stimuli).Shading indicates SEM across participants.Horizontal lines indicate the statistical significance of differences between each model and baseline (blue and red for V1, V0|A1 and V2, V0|A2, respectively), and between V1, V0|A1 and V2, V0|A2 (grey).(Bottom).Topographic maps depicting whole-brain spatial distribution of GLM performance for each model across 50 ms intervals in the post-audio period.