# Multivoxel Pattern of Blood Oxygen Level Dependent Activity can be sensitive to stimulus specific fine scale responses

## Abstract

At ultra-high field, fMRI voxels can span the sub-millimeter range, allowing the recording of blood oxygenation level dependent (BOLD) responses at the level of fundamental units of neural computation, such as cortical columns and layers. This sub-millimeter resolution, however, is only nominal in nature as a number of factors limit the spatial acuity of functional voxels. Multivoxel Pattern Analysis (MVPA) may provide a means to detect information at finer spatial scales that may otherwise not be visible at the single voxel level due to limitations in sensitivity and specificity. Here, we evaluate the spatial scale of stimuli specific BOLD responses in multivoxel patterns exploited by linear Support Vector Machine, Linear Discriminant Analysis and Naïve Bayesian classifiers across cortical depths in V1. To this end, we artificially misaligned the testing relative to the training portion of the data in increasing spatial steps, then investigated the breakdown of the classifiers’ performances. A one voxel shift led to a significant decrease in decoding accuracy (p < 0.05) across all cortical depths, indicating that stimulus specific responses in a multivoxel pattern of BOLD activity exploited by multivariate decoders can be as precise as the nominal resolution of single voxels (here 0.8 mm isotropic). Our results further indicate that large draining vessels, prominently residing in proximity of the pial surface, do not, in this case, hinder the ability of MVPA to exploit fine scale patterns of BOLD signals. We argue that tailored analytical approaches can help overcoming limitations in high-resolution fMRI and permit studying the mesoscale organization of the human brain with higher sensitivities.

## Introduction

Largely due to the ability to achieve relatively high spatial and temporal resolution functional images simultaneously across the whole brain, functional magnetic resonance imaging (fMRI) has become one of the most powerful tools to study the human brain non-invasively over the last 25 years. At ultra-high field (UHF, 7 Tesla and above), functional voxels span the sub-millimeter range, measuring 0.8 mm isotropic (e.g.1 for a review see2), 0.65 mm isotropic over small regions3, or even 0.45 mm using super resolution techniques (e.g.4). These high-resolution images allow the recording of blood oxygenation level dependent (BOLD5) responses at the level of cortical layers and columns (e.g.6,7,8,9,10,11,12,13). UHF fMRI therefore provides the unique opportunity to investigate the organizing principles of the human cortex at the mesoscale level, narrowing the gap between invasive animal electrophysiology and human neuroimaging14.

However, this sub-millimeter resolution is only nominal in nature, because a number of factors limit the point spread function of gradient echo (GE) BOLD responses and, ultimately, sensitivity to fine-grained functional structures. These factors include voxel blurring along the phase encoding direction and proximity to large draining blood vessels. Studies investigating the point spread function of GE BOLD responses at 7 T have shown that it spreads beyond the millimeter range, with an upper limit of approximately 2 mm15,16. More recently though, it has been argued that these estimates fail to account for the spread of the neuronal response as it relates to the size of receptive fields and their scatteredness in V117. Chaimow et al.17 suggest that when minimizing the contribution of macroscopic vessels, the point spread at 7 T for GE BOLD acquisitions is closer to 1 mm than 2 mm, approaching the nominal resolution of single voxels (in this case 0.8 mm isotropic).

Moreover, large veins also significantly modulate BOLD amplitude, leading to an increase in signal towards the pial surface, especially for GE recordings (e.g.18,19,20,21,22,23).

The implementation of appropriate analytical strategies may help to circumvent the impact of large draining vessels on biasing BOLD signal responses to outer cortical layers. For example, differential mapping techniques, along with the presence of pseudo-periodic functional structures24,25, can permit the mapping of orientation preference columns with high field spin echo (SE) fMRI, despite the limited spatial resolution and/or functional precision.

For the highly desirable GE BOLD signal, however, it remains to be determined whether and how neuroscientists can fully exploit the high spatial resolution data achievable at UHF, in order to investigate functional profiles of human cortical columns and layers. To this end, analyzing the information contained in voxel populations using multivoxel pattern analysis (MVPA) as opposed to average response amplitudes may represent an appealing analytical strategy to maximize fMRI sensitivity to fine-grained cortical features26. Kamitani and Tong27 employed MVPA to successfully decode orientation tuning in human V1 at 3 Tesla, with a voxel resolution that spanned well beyond the millimeter range (see also28). With the promise of retrieving information that would otherwise remain inaccessible, MVPA, such as linear support vector machines (SVM), have become widely used.

This account has been challenged recently, and the nature and spatial scale of the information exploited by the MVPA called into question. A number of studies have argued that orientation decoding may rely on coarser global maps that co-vary with micro-scale features (e.g.29,30,31,32). An example of such a coarse-scale organization that could account for orientation decoding in V1 is radial-preference retinotopic maps32. The somewhat unresolved debate sparked by these opposing views has motivated several neuroimaging studies to assess whether MVPA is sensitive to fine or coarse spatial patterns of multivoxel BOLD activity (e.g.29,30,31,33,34,35), with the possibility that orientation decoding could be underpinned by both coarse as well as fine scale structures.

With the growing availability of UHF scanners, the question of whether MVPA decoding relies on fine-grained spatial information becomes topical for the neuroimaging community. As mentioned above, GE BOLD is limited in spatial specificity, casting doubts on the spatial integrity of sub-millimeter fMRI maps. However, a demonstration that, unlike univariate amplitudes, MVPA effectively exploits finer grained information from GE-BOLD data, taking full advantage of the sub-millimeter resolutions, stands to increase the utility of high field high resolution GE BOLD. Within the context of this paper, we define univariate BOLD as the average BOLD response (or the contrast of the average BOLD responses elicited by 2 conditions) across all voxels within a given region of interest (ROI).

In this work, we re-analyzed feedforward and feedback cortical depth dependent previously acquired 7 T GE EPI data with 0.8 mm isotropic functional voxels1 (Fig. 1) to determine the spatial scale of stimulus specific responses to which MVPA is sensitive. To this end, we artificially misaligned the pattern structure in increasing spatial steps and investigated the breakdown of the classifier performance. We reasoned that if the spatial scale of stimulus specific responses to which decoding is sensitive is that of a single voxel, even a 1 voxel misalignment should lead to a significant decrease in decoding accuracy. To test this hypothesis, we first ran a simulation on synthetic data generated with realistic signal-to-noise (SNR) properties.

We then tested the performance of 3 classifiers: linear SVM; Linear Discriminant Analysis (LDA); and Naïve Bayes Classifier (NBC) on real data. It should be noted that comparable, albeit slightly different approaches have previously been implemented to assess the scale of information exploited by MVPA during orientation decoding in V1 at 3 T. Alink et al.36, for example, implemented an analysis strategy comparable to the one proposed here. They spatially offset the testing relative to the training set by 1, 2, 4, and 6 mm and measured the impact on decoding. Freeman et al.37 also implemented a comparable approach, however, their spatial shift occurred during the acquisition phase, rather than in the analyses stages. Moreover, to gain insights into whether venous-related amplitude increases also affect the magnitude of univariate differences and decoding accuracy, we directly assessed the relationship of these measures and univariate BOLD across cortical depths by means of correlational analyses.

## Methods

### Subjects

We reused data acquired from 4 subjects at the Center for Magnetic Resonance Research (CMRR, Minneapolis, MN, USA) and described in1. All subjects were healthy volunteers with normal or corrected visual acuity. Subjects gave written informed consent and received financial compensation for their participation. All methods were performed in accordance with the relevant guidelines and regulations. The institutional review board for human subject research at the University of Minnesota approved the study.

### Stimuli

Stimuli are described in the original study (see1). In summary, subjects viewed three visual scenes (‘car on street’, ‘people at market’, ‘ship in harbor’, see1,38). We controlled the images for global luminance, contrast and energy using matlab shine-toolbox39. Scenes were presented in full (‘feedforward’ condition) or masked with an occluder over the lower right visual field (‘feedback’ condition). We presented a set of contrast-reversing checkerboard mapping stimuli for ‘target’ and ‘surround’ regions in each run, and in a separate localizer run. The surround checkerboards mapped the outer 2 degrees of the white occluder and the target mapped the remaining inner section of the occluder (Fig. 1). The design of the experiment was comparable to our previous study38, but with the visual stimuli reduced in size by 20% to fit the smaller MRI bore due to the use of a head gradient insert (see below). We kept the width of the ‘surround’ mapping stimulus at 1 degree of visual angle and added an additional 1-degree border between the ‘surround’ stimulus and the edge of the ‘target’ stimulus region. We conducted a separate phase-encoded retinotopic mapping experiment40,41,42. Stimuli consisted of a wedge-shaped (22.5 degrees) checkerboard rotating slowly (64 s for full 360-degree rotation) around the fixation point in the middle of the screen. A white ‘spider web’ configuration was presented in the background to stabilize fixation together with a center fixation color change task43.

As described in1, the experiment comprised four functional runs of 350 volumes each. An experimental condition was presented for 6 volumes (12 s), and each of 6 experimental conditions was presented in a randomized order within a block followed by 12 volumes (24 s) of baseline (6 × 12 s + 24 = 96 s per block). Mapping blocks consisting of 2 conditions (‘target’, ‘surround’) were presented for 6 volumes (12 s) interleaved with 6 volumes (12 s) of baseline in between conditions and 12 volumes (24 s) at the end of the block (2 × 12 s +12 s + 24 s = 60 s). A functional run consisted of 6 experimental blocks and 2 mapping blocks and an additional baseline of 2 volumes (4 s) at the start of the run (6 × 96 s + 2 × 48 s + 4 s = 700 s = 350 volumes). Therefore, each experimental condition was presented 24 times across four runs.

The retinotopic mapping run comprised 12 repetitions of a full rotation lasting 32 volumes (64 s), with an extended baseline of 10 volumes (20 s) at the beginning and 12 volumes (24 s) at the end of the run (resulting in 406 volumes: 12 × 64 s + 20 s + 24 s = 812 s). An additional localizer run comprised 12 repetitions of ‘target’ and 12 repetitions of ‘surround’ mapping, with 25 baseline periods in between, all of which lasted for 6 volumes (12 s), resulting in 294 volumes ((12 + 12 + 25) × 12s = 588 s).

Subjects viewed the visual stimuli on a projection screen mounted to the rear end of the head coil using a head-coil mounted mirror. A video projector combined with a mirror projected the stimuli onto the screen. Stimuli were presented using Presentation software (Neurobehavioral Systems, CA, USA) for the experiments, and for retinotopic mapping with StimulGL (custom-built stimulation software, Maastricht University, Maastricht, NL). We instructed the subjects to keep fixation to the center of the screen and to perform a color-change detection task at the center of the screen, during both the experimental runs and retinotopic mapping.

### MRI Acquisition

MRI data for the first experiment was conducted on an ultra-high magnetic field (7 Tesla, 90 cm bore, Magnex Scientific, Abingdon, UK) at the CMRR in Minneapolis (MN, USA). The scanner was driven with a Siemens console (Erlangen, Germany) and used a head gradient insert with a 6-channel receive (1 Tx) array RF coil that covered only the visual areas.

Functional scans were recorded using GE-EPI at high resolution (nominal resolution, isotropic 0.8 mm3, TE = 17 ms, maximum flip angle (determined by a flip angle map) = 85°, slices = 38, TR = 2000ms, FOV = 128 ×128 mm2, matrix = 160 ×160, IPAT = 2, partial Fourier = 6/8, pixel bandwidth = 1375 Hz/pixel). Anatomical scans were acquired at 1mm3 using a MPRAGE sequence optimized for T1-weighted (3D MPRAGE) and proton density (PD)-weighted contrast (176 slices, FOV = 136×256 mm2, matrix = 136×256, voxel size = 1×1×1 mm3).

### Anatomical data analysis - cortical depth sampling

All data were analyzed with BrainVoyager QX 2.8. Proton density scans with identical slice positioning were used to remove spatial intensity inhomogeneities from T1 scans by dividing the T1 by the PD images44. We manually adjusted inner and outer grey matter boundaries along the local intensity values to eliminate pial blood vessels and to correct for GE-EPI distortions. We used relative cortical depth values to create Laplace-based equipotential grid-lines (i.e. solving the Laplace equation to obtain a smooth vector field and then create smooth meshes directly within the grey matter boundaries – e.g.45) at six depths (from inner to outer 90%, 74%, 58%, 42%, 26%, and 10% depths, Fig. 1). The gridlines were calculated smoothly at a highly up-sampled spatial coordinate system46. In a subsequent step, we used smooth gridlines to assign voxels to a respective cortical depth. Individual voxels were allowed to belong to adjacent depths. The depth gridlines covered the cortical representation of the occluded image section in the lower right visual field quadrant of retinotopic area V1d (Fig. 1). We saved the layered regions of interest as BrainVoyager QX VOI files (volume of interest).

### Functional data analysis

We pre-processed the fMRI data using slice scan timing corrections (sinc interpolation), 3D rigid body motion correction (sinc interpolation), intra-session alignment to the functional data of the last run, and temporal high pass filtering of 4 cycles. We aligned functional data to anatomical data with manual adjustments and iterative optimizations. We used activation maps of retinotopic mapping to optimize segmentation and alignment. Specifically, after T1 based segmentation, the grey matter borders were projected into EPI space and locally optimized according to the mean EPI image, which provides additional information regarding the spatial location of grey and white matter boundaries and outer pial surface. To further assess the quality of alignment and segmentation, we projected the activation maps from the retinotopic mapping runs to the segmented cortical ribbon. We visually inspected the quality of alignment and segmentation and optimized either or both accordingly. We implemented this procedure on the assumption that activity originates in the grey matter.

Analysis of functional data included general linear model (GLM) estimation of averaged conditions and single trials. We generated design matrices by the convolution of a double gamma function with a “boxcar” function (representing onset and offset of the image stimuli).

Independently per voxel and functional run, we implemented a classic general linear model (GLM) analysis (least squared minimization stress) to estimate the activation triggered by each single image block (i.e. single trial estimation modeling). We computed design matrices by convolving a double gamma function with a “boxcar” function (representing onset and offset of the image stimuli). Each design matrix thus consisted of 350 rows, representing the runs’ temporal dimension (i.e. volumes), and 41 columns, one per each visual stimulation trial plus the intercept term. Of 40 visual trials, 36 represented our experimental conditions: 6 trials x 2 experimental conditions (i.e. full and occluded images) x 3 visual scenes; and 4 represented the mapping stimulus: 2 trials x 2 mapping conditions (i.e. ‘target’ and ‘surround’ checkerboard stimuli). Methods up to this point are described in1.

### General decoding procedure

Linear support vector machine (SVM) decoding analysis was performed using SVM algorithms as implemented by the LIBSVM toolbox47, with default parameters (notably C = 1). Linear discriminant analysis (LDA) was implemented using the “fitcdiscr” and “predict” functions and the “cvshrink” function to implement cross-validated regularization inbuilt in Matlab’s statistic toolbox (Matlab, The Mathworks Inc, 2014). Naïve Bayes classifier was implemented using the function Classify from Matlab’s statistic toolbox with the ‘diaglinear’ option. Note that, before being input to the classifiers, the activity of each voxel was scaled using the same scaling factors for training and testing sets47. Firstly, the training portion of the data was normalized within a range of −1 to 1. This normalization was achieved by subtracting from the training data set its minimum value (to set the minimum to 0), dividing the resulting data set by its maximum (to scale the data between 0 and 1), multiplying by 2 and subtracting 1 (to scale the data between −1 and 1). The testing portion of the data was then scaled with the same procedure, but using scaling factors obtained from the training portion of the data47. All decoding analyses were performed only on voxels responding to target more than to surround (defined by the contrast ttarget > tsurruond ∩ ttarget > 0). We have described the decoding analyses in more detail previously38.

In brief, we trained all classifiers (linear pattern) to map between activation patterns from three scenes (full feedforward images in experiment 1) or between occluded scenes. We tested the trained classifiers on independent data (leave one run out cross validation). We measured the classifier performance of each cortical depth independently and we tested the single trial classification for significance using permutation testing (10,000 iterations of randomly assigned labels). To determine the empirical chance level, we implemented the following procedure. Independently per subject cortical depth, signal and misalignment extent, we randomly shuffled the labels of the classifiers’ input prior to the training phase. We then performed training and testing with shuffled labels. We repeated this procedure 10,000 times to produce a null distribution of decoding accuracy. We then sorted the label shuffled accuracy scores and selected the 95% largest score as the empirical chance level. Statistical significance was inferred when the low confidence interval of the unshuffled decoding accuracy (computed across cross-validated folds) was larger than empirical chance (i.e. p < 0.05).

### Artificial misalignment

To directly measure whether MVPA is capable of relying on stimulus specific fine scale responses, we developed a simple data driven approach that builds upon the impact of misalignment between training and test ROIs on decoding accuracy. Independently per cortical depth, we trained a classifier on the original ROI and tested its accuracy/performance on a number of misaligned sites.

We parametrically shifted the test site 0 (i.e. no misalignment) to 5 voxels relative to the training site.

To determine whether the drop in decoding accuracy following misalignment can be directly related to the spatial precision of stimulus specific responses, we generated synthetic sets of data while parametrically varying the precision of 3D patterns of simulated BOLD activity via 3D convolution (Fig. 2).

We began by computing a baseline volume for each participant by extracting all fixation volumes across the 4 runs and concatenating them to achieve the same number of time points as a “real” functional run. We regressed out the mean over time and computed the mean over time for the “demeaned” fixation time series. The resulting volume was used as the static component for our synthetic data. For each subject and cortical depth, we then generated synthetic data mimicking the BOLD activation triggered by 2 visual conditions (see Fig. 2). We began by estimating mean betas and variance across runs and trials. We then generated 2 (i.e. one per visual condition) 3D pseudo-random high-resolution patterns of normally distributed white noise with mean of 0 and variance of 1 (using the “randn” function in matlab). These 2 3D textures represent the true (i.e. no noise) multivoxel pattern of our synthetic conditions. We then added the previously estimated betas’ mean to each condition to attain comparable mean activation. Next, we generated patterns of noise. Using a comparable procedure, we proceeded to produce as many normally distributed high-resolution white noise patterns (with mean 0 and variance of 1) as the total number of experimental trials (in this case 48 – i.e. 6 trials × 4 runs × 2 conditions). We then used the previously estimated variance to scale the noise patterns so that the variance across the 24 3D noise textures of each image was comparable to that estimated for our real data set. We added each scaled noise texture to the signal patterns, producing 24 trials per condition with the same underlying spatial structure and different amounts of noise. We then smoothed the 3D multivoxel patterns of each simulated trial by convolving it with a 3D Gaussian kernel with a full-width half-max (FWHM) of 0 (i.e. no smoothing), 1, 2, 3 and 4 voxels, to simulate voxel resolutions of 0.8, 1.6, 2.4, 3.2 and 4 mm isotropic respectively. We finally added the noise and signal texture patterns to our baseline volume.

This process led to the generation of 2 synthetic multivoxel patterns of betas per simulated voxel resolution with comparable means and distinct spatial patterns of activation, faithfully reproducing the activation profile of our real data. We then carried out our misalignment approach on the synthetic data sets.

We then proceeded to implement this analysis to our data. We used two different, yet complementary, approaches: 1) volumetric driven and 2) surface grid driven misalignment. While misalignment was performed in volume space in both approaches, unlike the volumetric misalignment, the surface grid driven approach ensured that all spatial offsets of the training ROI were confined within a given cortical depth (Fig. 3, see below for more details).

All misalignments were performed in Matlab. We imported the data in Matlab using the BVQX toolbox and used a number of in-house tools to misalign the test site relative to the training site.

#### Volumetric misalignment

This approach consists of shifting the test ROI 0 to 5 voxels along the 3 axes (i.e. x, y and z) and 2 directions (i.e. positive and negative). Volumetric misalignment can be conceptualized by thinking of the training site as denoting the “origin” of a three-dimensional discrete Cartesian space, while the test site represents a point within said space whose coordinates differ from the origin along one of the 3 dimensions. Starting on the x axis, for example, we moved the test ROI 1 voxel in one direction (e.g. positive relative to the origin). We then tested the accuracy of the classifiers, trained on the original ROI, with the misaligned site and stored that value. We then moved the test site again (by one additional voxel) along the same dimension and direction, and tested the accuracy of the classifiers’ models on this newly shifted ROI. We repeated this procedure for all other dimensions (i.e. y and z) and directions (i.e. positive and negative) to reach a total of 5 voxel shifts for all axes and directions. This approach led to a set of 6 (shifts i.e. 5 misaligned plus the original site) by 3 (dimensions) by 2 (directions) accuracy scores for each subject. We then computed the mean across dimensions and directions independently per voxel shift, leading to 6 accuracy scores (one per voxel shift). Importantly, misaligning the test site in volumetric space allows the inclusion of voxels belonging to neighboring layers in the shifted ROI.

#### Surface grid driven misalignment

As the title suggests, we used the spatial coordinates of Laplace-based equipotential grids to spatially guide the misalignment of the test site in volume space. First, we identified the voxels belonging to a single cortical depth as indicated by the 3D coordinates of the Laplacian grids. Prior to misaligning the test ROI, we removed up to 5 voxels per grid row at the medial most edge of the representation of the occluded quadrant. This procedure was implemented to allow shifting of the test site up to 5 voxels while still remaining within the boundaries of the representation of the occluded quadrant within each cortical depth, and ensuring that the number of voxels remained constant across misaligned ROIs. We trained the classifiers and tested their performance on the original ROI and on its shifted versions. As in the volume-based misalignment, the test ROI was shifted 1 to 5 voxels. Importantly, to ensure that the misaligned test ROI only included voxels within a given cortical depth, misalignment only occurred in one direction, specifically away from the medial portion of the occluded quadrant. This procedure ensured that displacement of the test ROI remained confined within the retinotopic representation of the occluded quadrant, to avoid potential confounds in our decoding results related to the inclusion of BOLD activity elicited by the stimulated portion of V1.

### Univariate analyses

To test whether the 3 different images elicited different univariate BOLD amplitudes (defined as the mean activity across all voxels within a given cortical depth), we carried out the following statistical tests. We performed a 2 (signals) by 3 (images) by 6 (cortical depths) Linear Mixed Model (Matlab, The Mathworks Inc, 2014) with the mean BOLD response as a dependent variable. We combined the data from 4 runs and 4 participants (i.e. using 16 data points). Random variation across runs within each subject was accounted for by considering the subjects as random effects. This was implemented in Matlab using the following equation:

$${\rm{Data}} \sim {{\rm{Signals}}}^{\ast }{{\rm{Shifts}}}^{\ast }\,{\rm{Layers}}+(1|{\rm{Sbj}})$$

We estimated our fixed effects coefficients by means of maximum likelihood estimation. We computed hierarchical 95% bootstrap confidence intervals post hoc on significant main effects and interactions. We constructed a bootstrap distribution as follows: first, for a given subject, we computed the difference between the accuracies estimated on the unshifted ROI and those estimated for e.g. 1 voxel shift for all runs. For each bootstrap iteration, we then sampled with replacement the runs, computed the mean across the sampled runs and stored the value. We repeated this procedure for the remaining subjects, leading to 4 mean values (one per subject). We then sampled with replacement subjects and computed the mean, this time across subjects. We repeated this operation 2000 times. This procedure allowed us to construct a bootstrap distribution that is not limited by the factorial of N (number of subjects), as the same random sample of subjects would produce different bootstrap mean values due to the different runs sampled. We therefore computed the 95% confidence interval (Bonferroni corrected by adjusting the alpha threshold by the number of comparisons, in this case 5 – i.e. the number of voxels shifts). Statistical significance was inferred when 95% bootstrap confidence interval did not overlap with 0.

### Univariate vs. multivariate

To assess the relationship between univariate differences, univariate amplitudes and decoding accuracies across cortical depths, we performed a Spearman correlation analysis amongst these 5 measures (i.e. 3 decoding accuracies, univariate differences and univariate amplitudes) on the feed-forward signal for the original, un-shifted ROI. To quantify the differences in activation across conditions, we computed univariate differences as follows: for each subject, run, image and trial, we computed the mean across the beta weights of each voxel (representing BOLD percent signal change, or PSC, amplitude). We then L2 normalized these values as follows: for each subject we put runs, images and trials in a vector and computed the L2 norm according to the following equation:

$$\Vert L2\Vert =\sqrt{\mathop{\sum }\limits_{i=1}^{n}{({a}_{i})}^{2}}$$

where n represents the length of vector (i.e. 3 images × 4 runs × 6 trials). We divided the values of each subject by the L2 norm and computed the mean across trials. This procedure was implemented because we were interested in the differentials pattern across cortical depths, regardless of the raw differences in PSC BOLD amplitude across subjects. We then calculated the root square differences (RSD) of these normalized measures between each pair of images according to the following equation:

$$RSD=\sqrt{{(im{g}_{i}-im{g}_{j})}^{2}}$$

where imgi and imgj represent the L2 normalized activity elicited by 2 given images. We calculated the mean across the differences for all runs and image pairings. We choose RSD to quantify univariate differences in amplitude elicited by the 3 image stimuli because we wanted a measure that was insensitive to the sign of the difference, thus allowing averaging to quantify the mean difference across all image pairs.

We therefore computed Spearman correlation coefficients between: a) the 6 accuracy scores (i.e. one per cortical depth) for each of the tested classifiers and the 6 average univariate differences; b) the 6 accuracy scores for each of the tested classifiers and the 6 average univariate amplitudes; and c) the 6 average univariate amplitudes and the 6 average univariate differences.

Moreover, to assess the relationship between univariate amplitude and decoding accuracies and determine whether the former can explain the pattern of accuracies across cortical depths and misalignment extents, we carried out further correlational analyses. Independently per signal and layer, we calculated the mean across voxels, conditions, runs and trials for the original ROI and all its misaligned versions. We thus compared the resulting 6 (layers) by 6 (shifts) univariate amplitude matrices with the 6 by 6 matrices of the SVM accuracy scores, with the 6 by 6 matrices of the LDA accuracy scores, and with the 6 by 6 matrices of the NBC accuracy scores. To quantify the similarity between the decoding accuracies of all classifiers and univariate BOLD activation, we computed Spearman correlation coefficient between the accuracy scores and the univariate BOLD matrices.

### Inferential statistic on decoding accuracies

For each of the 3 classifiers, to test the main effects and interactions between signals (feedforward and feedback), misalignment extents (1 to 6) and cortical depths (1 to 6), we performed a 2 (signals) by 6 (cortical depths) by 6 (voxel shifts) Linear Mixed Model (Matlab, The Mathworks Inc, 2014) with the accuracy scores as a dependent variable, as explained above. Moreover, we computed hierarchical 95% bootstrap confidence intervals post hoc on significant main effects and interactions as also described above. Specifically, we computed hierarchical 95% bootstrap confidence intervals post hoc on significant main effects and interactions on the difference between the accuracies of the original ROI and its misaligned counterparts. We constructed a bootstrap distribution as follows: first, for a given subject, we computed the difference between the accuracies estimated on the unshifted ROI and those estimated for e.g. 1 voxel shift for all runs. For each bootstrap iteration, we then sampled with replacement these differences over the runs, computed the mean across the sampled runs and stored the value. We repeated this procedure for the remaining subjects, leading to 4 mean values (one per subject). We then sampled with replacement subjects and computed the mean, this time across subjects. We repeated this operation 2000 times. This procedure allowed us to construct a bootstrap distribution that is not limited by the factorial of N (number of subjects), as the same random sample of subjects would produce different bootstrap mean values due to the different runs sampled. We therefore computed the 95% confidence interval (Bonferroni corrected by adjusting the alpha threshold by the number of comparisons, in this case 5 – i.e. the number of voxels shifts) for these bootstrapped differences. Statistical significance was inferred when 95% bootstrap confidence interval did not overlap with 0. Only differences between the accuracies on the unshifted ROI and those on its displaced version were computed.

## Results

### Univariate analysis

The 2 (signals) by 6 (cortical depths) by 3 (images) linear mixed model performed on BOLD amplitude averaged across all voxels within a given ROI cortical depth showed a significant main effect (p < 0.01) of signal (F(1,552) = 14.554) driven by larger amplitudes elicited by feedforward stimulus conditions compared to feedback; and a significant (p < 0.05) interaction between signal and cortical depths (F(5,552) = 2.342), driven by the fact that different cortical depths elicit significantly different amplitudes (increasing as we approach the pial surface) for the feedforward but not the feedback condition (see Fig. 4A). Importantly, this analysis did not reveal significant differences between univariate BOLD elicited by the 3 different images. No additional significant main effects or interactions were observed.

We further report a strong positive Spearman correlation (rho(5) = 0.94; p= 0.0167) across cortical depths between feed-forward univariate averages (i.e. mean across images, trials and voxels), amplitudes and differences (Figs. 4B and 4C); no significant correlations (p > 0.05) were observed between feed-forward decoding accuracies (Figs. 4D, 4E and 4F) and either univariate amplitudes or average differences.

### Decoding analysis

As previously reported (see1), for the decoding analyses on the original un-shifted ROI, single-block classification was significant at each depth for each classifier and each subject (permutation tested at 5%; no corrections) during feedforward stimulation of V1. For the feedback condition (i.e. the occluded images), only the superficial outermost depth (10%) was significant in all four subjects and classifiers. The second-most outer depth (26%) was significant in three of four subjects for all classifiers, and the mid-depth (42%) was significant in two of four for SVM and LDA, and for 1 out of 4 for NBC. No subjects showed significant above chance decoding at the 58% mid-depth across all classifiers.

### Misalignment on synthetic data

The results of the simulations are shown in Fig. 5.

We found that the resolution of the mutlivoxel pattern significantly modulates the impact of misalignment on SVM decoding accuracy: bonferroni corrected 95% bootstrap confidence intervals show that with no smoothing, a one voxel misalignment completely destroys the highly specific spatial structure of the patterns, rendering the model computed on the original ROI (i.e. training phase) redundant and causing decoding accuracy to immediately drop to baseline. Misaligning a mutlivoxel pattern smoothed with a 3D Gaussian kernel with a FWHM of one voxel required a 2 voxel shift to impair decoding accuracy; patterns smoothed with a FWHM of 2 voxels required a 3 voxel shift; FWHM of 3 voxels, a 4 voxel shift; and a FWHM of 4 voxels required a 5 voxel misalignment to significantly impair decoding accuracy (Fig. 5).

### Misalignment on real data

Artificial misalignment either led to a decrease or to no significant change in the accuracies of all classifiers for all layers, signals, and misalignment scenarios (see sections below).

While we carried out a fully parameterized linear mixed model, we were specifically interested in the 3-way interaction between cortical depths, signals, and voxel shifts. In the following sections, F values for all significant main effects and interactions captured by the model are reported; however, a more in-depth discussion of these numbers is outside the scope of this work. We focus on the interactions amongst cortical depths, signal and voxel shifts, and the related post-hoc bootstrap tests.

#### Volumetric misalignment

The 2 (signals) by 6 (cortical depths) by 6 (voxel shifts) linear mixed model carried out on SVM accuracy showed significant main effects (p < 0.01) of signals (F(1,1080) = 289.04) and depths (F(5,1080) = 24.806) as well as significant (p < 0.01) interactions between signals and cortical depths (F(5,1080) = 12.995), signals and voxel shifts (F(5,1080) = 37.749), cortical depths and voxel shifts (F(25,1080) = 3.2061) and signals, cortical depths and shifts (F(25,1080) = 1.7484, p < 0.05).

The same analysis carried out on LDA accuracies showed significant main effects (p < 0.01) of signals (F(1,1080) = 241.92) and depths (F(5,1080) = 24.69) as well as significant (p < 0.01) interactions between signals and cortical depths (F(5,1080)=16.067), signals and voxel shifts (F(5,1080) = 30.983), cortical depths and voxel shifts (F(25,1080) = 3.252) and signals, cortical depths and shifts (F(25,1080) = 2.1793, p < 0.01).

The linear mixed model for NBC also showed significant main effects (p < 0.01) of signal (F(1,1080) = 175.92) and depths (F(5,1080) = 15.715) as well as significant (p < 0.01) interactions between signal and cortical depths (F(5,1080) = 8.756), signal and voxel shifts (F(5,1080) = 19.865), and signals, cortical depths and voxel shifts (F(25,1080) = 1.8607).

For the feed-forward signal, post-hoc 95% bootstrap confidence interval (btCI) revealed that a one voxel shift produces a significant decrease in decoding accuracies of all classifiers for all cortical depths (Fig. 6). Conversely, for the feedback signal, 95% btCI showed that for the outermost cortical depth (i.e. 10%) a 2 voxel shift is required before observing a significant decrease in decoding accuracies for all classifiers; for LDA, a 2 voxel shift also led to a significant drop in decoding accuracy for the second and third outermost depths (i.e. 26% and 42%) (Fig. 6).

Spearman coefficients were computed to quantify the similarity between decoding accuracies and univariate BOLD activation (Fig. 7), by correlating decoding accuracies for all classifiers and univariate BOLD responses at all misalignment extents independently per cortical depth and signal. We observed no significant correlations (q > 0.05 FDR corrected).

In Table 1, we report the total number of voxels independently per subject and cortical depth, after performing the contrast ttarget > tsurround ∩ ttarget > 0 ∩ tfull-stimulus > toccluded-stimulus.

#### Surface grid driven misalignment

The 2 (signals) by 6 (cortical depths) by 6 (voxel shifts) linear mixed model carried out on SVM accuracy showed significant main effects (p < 0.01) of signal (F(1,1080) = 131.1) and depths (F(5,1080) = 10.291), as well as significant (p < 0.01) interactions between signal and cortical depths (F(5,1080) = 7.635), signal and voxel shifts (F(5,1080) = 18.302), cortical depth and voxel shift (F(25,1080) = 1.989), and signal, cortical depths and shift (F(25,1080) = 1.886, p < 0.01).

The same analysis carried out on LDA accuracies showed significant main effects (p < 0.01) of signal (F(1,1080) = 98.074) and depths (F(5,1080) = 8.606) as well as significant (p < 0.01) interactions between signal and cortical depths (F(5,1080) = 9.585), signal and voxel shifts (F(5,1080) = 12.651), cortical depth and voxel shift (F(25,1080) = 2.539) and signal, cortical depths and shift (F(25,1080) = 2.513, p < 0.01).

The linear mixed model for NBC also showed significant main effects (p < 0.01) of signal (F(1,1080) = 112.17) and depths (F(5,1080) = 6.854) as well as significant (p < 0.01) interactions between signal and cortical depths (F(5,1080) = 6.665), signal and voxel shifts (F(5,1080) = 14.653) and signal, cortical depths and shift (F(25,1080) = 1.663, p < 0.05).

For the feed-forward signal, post-hoc 95% bootstrap confidence interval (btCI) revealed that one voxel shift produced a significant decrease in decoding accuracies of all classifiers for all cortical depths (Fig. 8). For the feedback signal, 95% btCI showed that for the 2 outermost cortical depths only (i.e. 10% and 26%), a 1 voxel shift produced a significant decrease in SVM accuracies (Fig. 8); for LDA accuracy, a 1 voxel shift led to a significant decrease for the 2nd outermost depth only (i.e. 26%), while for the outermost depth, a 2 voxel shift was necessary to significantly impair decoding accuracy; for NBC accuracy, a 2 voxel shift led to a significant decrease for the outermost depth only (i.e. 10%).

We further carried out Spearman correlations to quantify the similarity between the MVPA decoding accuracy and univariate BOLD activation (Fig. 9). Spearman coefficients, computed independently per cortical depth and signal by correlating decoding accuracies and univariate BOLD response at all misalignment extents, revealed no significant correlations (q > 0.05 FDR corrected).

The total number of voxels after performing the contrast ttarget > tsurround ∩ ttarget > 0 ∩ tfull-stimulus > toccluded-stimulus independently per subject and cortical depth are reported in Table 1.

## Discussion

The ability to exploit the sub-millimeter resolution achievable with UHF fMRI is critical for advancing cortical depth dependent functional investigations in humans. This is particularly true for the widely used GE BOLD contrast, which has high signal-to-noise ratio, but limited spatial acuity. We measured whether MVPA is capable of relying on stimulus specific fine scale responses.

With previously collected data from1, we parcellated the cortical sheet into 6 equally spaced depths, ranging from 10% to 90% distance from the pial surface. We analyzed feed-back and feed-forward signals triggered by images of natural scenes in V11 independently per cortical depth.

To assess whether MVPA relies on fine scale stimulus specific responses, we systematically misaligned voxels between the training and test ROI. We trained decoding algorithms (linear Support Vector Machine (SVM), Linear Discriminant Analysis (LDA) and Naïve Bayes Classifier (NBC)) on a given cortical depth and tested their performances on a ROI that was misaligned anywhere from 0 to 5 voxels relative to the training site. This approach allowed us to assess whether information decoded with MVPA is at least as precise as the nominal resolution of single voxels. We hypothesized that a negligible decrease in decoding accuracy following the spatial offset of the test ROI relative to the training ROI would suggest that the exact correspondence of spatial structures is not necessary to achieve the highest decoding accuracy, indicating that the multivoxel pattern is blurred and/or that the information decoded exists at a scale coarser than the tested offset. Conversely, a significant decrease in decoding accuracy following a 1 voxel misalignment would indicate that the exact correspondence of spatial structures is necessary to achieve the highest decoding accuracy, suggesting that the scale of responses exploited by multivoxel decoders and associated with the tested stimuli is at least as precise as the nominal resolution of single voxels (here 0.8 mm isotropic). The results of our simulation on synthetic data support our hypothesis, suggesting that the spatial scale of the multivoxel pattern of BOLD activation modulates the drop in decoding accuracy and, therefore, that our method does measure stimulus specific scale of BOLD responses. While noise characteristic of our synthetic data is not realistic and therefore limits the interpretation of our simulation results, the purpose of the simulation was only to test the validity of our method. Within this context, the use of white uncorrelated noise may represent an advantage as it confers more control over the data.

When we applied the misalignment approach to real data we found that as little as a one voxel misalignment led to a significant decrease in decoding accuracy across all cortical depths. We argue that multivoxel activity patterns carry a substantial amount of spatially precise information at the nominal resolution of single voxels. Our result suggests that multivoxel decoding can enhance the relevance of voxels that are not corrupted by reduced specificity, thereby increasing the sensitivity of the multivoxel approach to fine-grained spatial responses.

### Distinguishing different sources of spatial resolution

We first briefly distinguish the related, yet different sources of spatial resolution of interest. We identify two sources of spatial resolution: 1) those related to acquisition, including the nominal resolution of BOLD images and the functional resolution of single voxels (for example, as measured by PSF); and 2) those related to post-processing or analytical operations, such as the resolution of the multivoxel pattern exploited by MVPA.

Our focus is to measure the resolution exploited by multivoxel decoders, and to understand whether MVPA profits from the nominal resolution afforded by UHF fMRI (here 0.8 mm isotropic). The question was partly motivated by the observation that point spread measurements of GE BOLD recordings suggest that the point spread function of GE BOLD responses is above the millimeter range (e.g.15; see “Relation to BOLD PSF” below). These reports not only challenge the feasibility of imaging human cortical layers and columns, but also question the usefulness of recording BOLD images with nominal sub-millimeter resolution. But is the resolution of the multivoxel pattern of BOLD activity exploited by multivariate decoding also impacted by the limited precision of BOLD measurements? Here we show that this is not the case, and argue that multivariate decoders exploit fine-grained information contained in multivoxel patterns of activity. Our finding that a 1 voxel shift leads to a significant impairment in decoding accuracy indicates that: 1) the resolution of the multivoxel pattern of BOLD responses exploited by MVPA is at least as precise as the nominal resolution of single voxels (here 0.8 mm iso); 2) while BOLD PSF results render the investigation of the mesoscale organization of the human cortex in the sub-millimeter range challenging at best, alternative analytical strategies, such as MVPA or differential maps (e.g.24,25), seem to permit this submillimeter scale; 3) using MVPA we can fully exploit the nominal resolution of functional voxels (here 0.8 mm isotropic); and 4) decoding of complex image stimuli, containing a mixture of low and high spatial frequencies, relies both on fine and coarse patterns of multivoxel activity. These arguments are now discussed in more detail.

### Relation to BOLD PSF

As suggested by studies measuring the GE BOLD point spread function (PSF) of single voxels at 7 T (e.g.15,17,48,49), the spatial specificity of individual functional voxels extends beyond the millimeter range, despite their nominal resolution. The GE BOLD PSF for a single condition in humans has been reported to have an upper limit of 2 mm when avoiding large vessels15 (but see17,48). If we take this measurement at face value, shifting the multivoxel pattern of activity by 0.8 mm (i.e. one voxel) should have a negligible impact on decoding accuracy. This is because, despite the 0.8 mm nominal resolution, neighboring voxels should yield highly correlated signals, effectively decreasing the functional resolution of the voxel population. Our results demonstrate, however, that a one-voxel misalignment significantly impacts decoding accuracy, indicating that MVPA decoding operates at a resolution that is at least as precise as the nominal voxel size (i.e. 0.8 mm isotropic), thus fully exploiting the increase in spatial resolution affordable with GE sequences at 7 T, and further motivating the need for sub-millimeter GE recording. These results are somewhat in line with those reported by Chaimow et al.17,48, who estimate a PSF of 1.02 mm for GE BOLD. This however is not to say that the BOLD information content is at the same spatial precision, as a number of factors (for example, vascular heterogeneity) may be contributing to the pattern of results observed here.

In the above paragraph we have discussed the implications of our findings in relation to point spread function (PSF) measurements. To this end, it is worth noting that, while interesting, a direct comparison between the spatial precision of a pattern of responses as computed here (and defined as the amount of misalignment tolerable by the classifiers) and that of single voxels measured by PSF measurement is challenging. This is because while the latter can provide an estimate for the precision of a functional response for a single voxel, the former refers instead to the precision of the information contained in a population of voxels, and, as shown here, it can include coarse and fine spatial scales.

### Pulse sequence limitations in high-resolution fMRI

It is well known that spatial specificity of the BOLD response is strongly modulated by large draining vessels15,25,50,51,52,53,54,55,56,57,58,59,60,61,62,63,64,65,66,67,68. This modulation is, to an extent, dependent on the type of contrast used. Spin echo (SE) based acquisitions are less susceptible to large veins and thus less affected by venous artifacts compared to gradient echo (GE) sequences25,66,68 and therefore seemingly represent the ideal choice to maximize BOLD spatial specificity. However, SE based acquisitions are limited in terms of coverage, SNR, and CNR. In this respect, GE acquisitions represent an appealing choice for high resolution fMRI at high fields. However, the high sensitivity of GE BOLD data are related to the impact of large draining veins15,25,55,56,57,58,59,61,62,63,64,65,66,67,68. Venous BOLD signal, demonstrated in both humans59,69 and animals70,71, produces a larger response compared to that of tissue67,72. Moreover, the vascular architecture of the human cortex, characterized by a higher concentration of large draining veins in proximity of the pial surface, represents an additional challenge in imaging layers using GE sequences. The typical GE laminar BOLD response is characterized by a ramping-like profile with larger amplitude, SNR, and CNR in outer layers (e.g.18,19,20,21,22,23; however,20 and21 also report a significantly larger activation in the middle of the cortical ribbon). Such a bias makes investigating depth dependent functional responses challenging, especially when only evaluating mean univariate amplitudes. At the multivoxel level, however, we found that decoding accuracy in human primary visual cortex does not co-vary with univariate BOLD amplitude and amplitude differences. We observed that decoding accuracy, at least for the feed-forward signal, peaks over the mid layers, while the univariate BOLD response and its differences peak in the outer depths. To further evaluate this uncoupling, we directly compared the relationship between univariate amplitude and MVPA decoding accuracy by correlating these 2 metrics for all cortical depths and misalignment extents (Figs. 6 and 8). We observed no significant correlation across layers and signals. Differential mapping can help avoid the effects of larger vessels and overcome the lack of high spatial specificity typically observed in GE data (e.g.7,24,25,73,74,75,76). We therefore wanted to assess whether the impact of large draining vessels on mean univariate differences is comparable to that observed on mean univariate BOLD amplitudes. Large vessels, prominently distributed on the pial surface, are known to increase BOLD response, giving rise to the widely observed ramping pattern of BOLD amplitudes across cortical depths (i.e. larger at the outer compared to the inner depths). As indicated by a strong positive correlation between univaritate differences and average amplitudes, we report a comparable, albeit shallower, profile of the average univaritate differences, magnitudes and BOLD amplitudes across cortical depths, both peaking in the outer depths. Conversely, no correlation was observed between mean univariate amplitudes or their differences and MVPA decoding accuracy, with the latter peaking over the mid cortical depths. Moreover, as mentioned, large vessels compromise the spatial specificity of GE BOLD. Yet, for the feedforward signal, a 1 voxel misalignment on decoding accuracy always leads to a significant decrease in decoding accuracy across all depths, in spite of the diverse distribution of vein across the cortical ribbon. These results indicate a comparable spatial precision of multivoxel patterns of activity across depths.

Taken alone, the layer profile observed for decoding accuracy (peaking in the mid, rather than outer depths) may reflect hypersensitivity of MVPA to physiological vascular noise, prominent in outer depths due to the high concentration of large vessels. Such noise may limit the performance of MVPA in outer depths. While this still represents a plausible scenario, the decoding layer profile, together with the comparable impact of misalignment across depths, suggest that, unlike univariate analyses, MVPA is relatively less contaminated by the effects of large vessels. It is worth noting that univariate differences as computed here, although comparable, are not the same as differential mapping. The latter does not imply spatial averaging, preserving the spatial structure of the BOLD response. Differential mapping, like MVPA, thus exploits the pattern of activation, as opposed to its average, and could therefore be less susceptible to the effect of large vessels25. While the debate regarding the optimal sequence choice to study layers and columns is not yet settled, we argue that clever post-processing and analytical tools represent a viable path to overcome sequence-based resolution and specificity issues.

### Univariate amplitude vs. multivoxel decoding accuracy

The feed-forward pattern of decoding accuracy observed here, peaking over the mid cortical depths, together with the observation that a 1 voxel shift equally impacts MVPA decoding accuracy across depths, demonstrates that univariate BOLD amplitude does not (in this case) modulate multivoxel decoding accuracy. This result is in line with a previous report showing that, while macroscopic vessels can carry neuronal-specific information, their contribution to multivoxel decoding accuracy may be redundant77,78.

Moreover, unlike multivoxel decoding, we report that standard univariate contrast, defined as the average BOLD amplitude across all voxels within a given cortical depth, does not show significant differences between the amplitudes elicited by different images for either the feed-forward or the feed-back conditions (Fig. 4). This observation could stem from a number of points: 1) different images elicit differently retinotopically distributed maps of activation, and these differences are obscured following voxels averaging (for a review of the differences between multivariate and univariate analysis see79); 2) MVPA carries information at a finer scale, one that is not visible by means of average univariate analyses; and 3) MVPA decoding can be more sensitive (i.e. higher statistical power) than average univariate analyses, probably due to the lesser impact of noisy voxels that can be “ignored” by the classifier.

### Finer vs. Coarser scale information in MVPA

Importantly, while misaligning the test ROI by 1 voxel negatively impacts decoding accuracy, as demonstrated by the volumetric approach, significant above chance decoding can still be achieved with as many as three voxel shifts. This result indicates that the multivoxel pattern of activity carries both coarser- and finer- grained information.

Several approaches have estimated whether MVPA relies on stimulus specific fine scale responses, ranging from smoothing31 or spatially filtering80 the activation maps in order to degrade the resolution of the information, to shifting the slice positioning by 1 mm (i.e. half a voxel) during the acquisition37. While spatial filtering is a generally useful strategy, we argue that, within the context of this work, it does not represent a straightforward advantage over misalignment for a number of reasons. Low pass filtering, for example, increases SNR and decreases run to run variation (e.g.81), potentially boosting cross validated decoding accuracy (see Figure S4). Additionally, smoothing introduces artificial, spurious correlations across voxels (e.g.82) and this effect and its impact on cross-validated decoding accuracy is difficult to quantify. Moreover, in light of the observed reliance of decoding on both finer and coarser spatial information, together with the afore mentioned low-pass filtering induced increase in SNR, down-sampling the input pattern will not necessarily lead to a drop in accuracy, even if multivoxel decoding does rely on spatially precise patterns of activation. The complex and poorly understood interplay of these forces would therefore render down-sampling related modulations on cross-validated multivoxel decoding accuracy difficult to interpret.

It is worth making one more consideration regarding the performance of the different classifiers implemented here. All classifiers led to very similar results, with only minor differences. The most obvious difference is that NBC led to overall lower decoding accuracy compared to SVM and LDA. This is probably due to the fact that NBC assumes orthogonality amongst the variables within a class (i.e. zero off-diagonal covariance) and it is thus less flexible than SVM and LDA. However, the impact of misalignment on decoding accuracies led to consistent results for all 3 classifiers, demonstrating that the results are likely to be a property of the resolution of the multivoxel pattern exploited by MVPA rather than specific to a classification algorithm.

It should be noted that classifiers such as those implemented here will depend on both signal and noise. If such noise is spatially or temporally structured, multivoxel decoders can exploit it to maximize performance (e.g.85). The observed drop in decoding accuracy following a one voxel misalignment could therefore be reflecting fine spatial neuronal structure, or fine spatial noise structure. We therefore argue that, regardless of its origins, our results indicate that multivoxel patterns of BOLD responses can carry spatial structure that is as fine as the nominal resolution of functional voxels (i.e. 0.8 mm iso).

### Differences between volume and surface based misalignments

We implemented artificial misalignment using 2 approaches: 1) volume-based misalignment, where/when the test ROI was misaligned along all dimensions and directions, thus allowing trespassing into neighboring cortical depths; and 2) surface grid-driven misalignment, where/when misalignment occurred only within a given cortical depth (Fig. 4). While both approaches show that a single voxel misalignment leads to a significant decrease in decoding accuracy, we observe differences between the 2 techniques. Volumetric misalignment yields a smoother function of accuracy across space or voxel shifts (Figs. 7 and 9), taking at least 3 shifts before decoding accuracy drops to chance level. Conversely, grid-driven misalignment yields a sharper function, where, across layers and signals, one voxel shift leads to a drastic decrease in decoding accuracy. We hypothesize that the difference between the approaches is related to: 1) a sparser voxel sampling for the grid compared to the volumetric-based misalignment (as indicated by Figure S2); and 2) by constraining the spatial offset of the training ROI within a given depth we effectively avoid large penetrating draining vessels perpendicular to the cortical surface. The sparser sampling is a direct outcome of constraining the misalignment within a layer. As depicted in Fig. 4 and S2, for the volumetric misalignment the distance between neighboring voxels is invariant and equates to the width of a voxel (in this case 0.8 mm), whereas for the surface grid driven approach, the distance between 2 voxels can be greater in at least 2 scenarios: 1) when 2 voxels that are adjacent to one another in volume space belong to different depths, therefore introducing a one voxel “gap” between 2 neighboring voxels within a cortical depth; and 2) when 2 adjacent voxels belonging to the same depth do not lie on the same plane, rendering the distance between them equal to the square root of the sum of their squared edge length (i.e. the voxel’s diagonal). Sparser sampling leads to a greater impact of one voxel shift on decoding accuracies because when only considering activity within a given layer, voxels are spatially farther apart and thus less correlated. Spatially offsetting the test relative to the training ROI would therefore lead to greater misalignment for the grid-driven compared to the volumetric approach. The distance between voxels before and after a one voxel misalignment for the surface-based approach is summarized in figure S2. In surface space, the flattening of the cortical curvature led to a sparser voxel sampling and a consequent varying distance between voxels in the original and misaligned ROI. As indicated in figure S2, in the surface driven regime, a one voxel misalignment leads to an average distance of roughly 1 mm. This alone may not be sufficient to explain the sharp drop to baseline in decoding accuracy. However, as shown by the histograms in figure S2, there are a number of voxels that move up to over 3 mm away (comparable to a 3 or even 4 voxel shift in the volumetric approach). These may account for the observed drop to baseline after a 1 voxel shift in the surface driven misalignment.

Moreover, given that during surface grid-driven misalignment the spatial offset of the training relative to the test ROI only occurs within a depth (see Fig. 4) and therefore tangentially to the cortical surface, this approach is less affected by the effect of large penetrating vessels that contribute to blurring single voxels’ functional specificity. This claim is supported by the cross-layer decoding analysis performed in1. In1 we trained linear SVM on a given depth and tested its performance on all other depths. This approach showed that, unlike the surface driven regime, when misalignment occurs orthogonally to the surface, above chance decoding accuracy can be achieved even when the training and testing ROIs are highly misaligned (e.g. training on the outermost depth and testing on the innermost - see Fig. 1D in1). We further observe that volumetric misalignment has a lesser impact on decoding the feedback compared to the feedforward signal, requiring a two-voxel shift before significantly decreasing feedback accuracy. As previously argued1, this observation could suggest that the feed-back signals from higher-level regions with larger receptive fields carry information that is more abstract and therefore spatially coarser than its feed-forward counterpart. While this claim is supported by a recent study directly investigating the contribution of spatial frequencies to feedback signal86, as indicated by the size of the error bars, this finding could simply be related to noisier signal rather than an apparent coarser resolution. Moreover, as previously mentioned, we observed that decoding accuracy is differentially modulated by misalignment for feedback and feedforward signals. Unlike feedforward, for the feedback signal of the outermost depth only (i.e. where decoding accuracy was significantly above chance for all subjects and classifiers) decoding accuracy shows no significant decrease following a one voxel shift. The differential modulatory impact of misalignment on feedback and feedforward signals represents a further indication that our results are not a mere property of decoding analyses (or else we would expect a comparable pattern of accuracy decrease across signals, see also figure S1 in the supplementary section).

## Conclusion

We showed that the multivoxel pattern of activity exploited by MVPA decoding carries information about the visual experimental condition on both coarse and, importantly, fine scale, one that here is at least as precise as the nominal resolution of functional voxels (i.e. 0.8 mm iso). Unlike average univariate analyses, MVPA can successfully distinguish the patterns of BOLD responses elicited by our 3 images. These findings are promising for future fMRI studies of cortical layers and columns, as it indicates that, when the spatial precision of mean univariate amplitude is corrupted by macroscopic biases such as, for example, large draining vessels or blurring in the phase encoding direction, MVPA can potentially circumvent sensitivity and specificity limits of the GE BOLD signal. While there are several pulse sequence variants that could reduce the large vessel biases present in high field GE BOLD data, such as SE or VASO, they are costly in terms of efficiency and sensitivity. As an alternative, intelligent analysis strategies provide benefits in enhancing the spatial precision of the information in fMRI signals.

## References

1. 1.

Muckli, L. et al. Contextual Feedback to Superficial Layers of V1. Curr Biol 25, 2690–2695, https://doi.org/10.1016/j.cub.2015.08.057 (2015).

2. 2.

Lawrence, S. J. D., Formisano, E., Muckli, L. & de Lange, F. P. Laminar fMRI: Applications for cognitive neuroscience. Neuroimage 197, 785–791, https://doi.org/10.1016/j.neuroimage.2017.07.004 (2019).

3. 3.

Heidemann, R. M. et al. Isotropic submillimeter fMRI in the human brain at 7 T: combining reduced field-of-view imaging and partially parallel acquisitions. Magn Reson Med 68, 1506–1516, https://doi.org/10.1002/mrm.24156 (2012).

4. 4.

Vu, A. T., Beckett, A., Setsompop, K. & Feinberg, D. A. Evaluation of SLIce Dithered Enhanced Resolution Simultaneous MultiSlice (SLIDER-SMS) for human fMRI. Neuroimage 164, 164–171, https://doi.org/10.1016/j.neuroimage.2017.02.001 (2018).

5. 5.

Ogawa, S. et al. Intrinsic signal changes accompanying sensory stimulation: functional brain mapping with magnetic resonance imaging. Proc Natl Acad Sci USA 89, 5951–5955, https://doi.org/10.1073/pnas.89.13.5951 (1992).

6. 6.

Kok, P., Bains, L. J., van Mourik, T., Norris, D. G. & de Lange, F. P. Selective Activation of the Deep Layers of the Human Primary Visual Cortex by Top-Down Feedback. Curr Biol 26, 371–376, https://doi.org/10.1016/j.cub.2015.12.038 (2016).

7. 7.

Nasr, S., Polimeni, J. R. & Tootell, R. B. Interdigitated Color- and Disparity-Selective Columns within Human Visual Cortical Areas V2 and V3. J Neurosci 36, 1841–1857, https://doi.org/10.1523/JNEUROSCI.3518-15.2016 (2016).

8. 8.

Kashyap, S. et al. Resolving laminar activation in human V1 using ultra-high spatial resolution fMRI at 7T. Sci Rep 8, 17063, https://doi.org/10.1038/s41598-018-35333-3 (2018).

9. 9.

Kashyap, S., Ivanov, D., Havlicek, M., Poser, B. A. & Uludag, K. Impact of acquisition and analysis strategies on cortical depth-dependent fMRI. Neuroimage 168, 332–344, https://doi.org/10.1016/j.neuroimage.2017.05.022 (2018).

10. 10.

Koopmans, P. J. & Yacoub, E. Strategies and prospects for cortical depth dependent T2 and T2* weighted BOLD fMRI studies. Neuroimage 197, 668–676, https://doi.org/10.1016/j.neuroimage.2019.03.024 (2019).

11. 11.

Norris, D. G. & Polimeni, J. R. Laminar (f)MRI: A short history and future prospects. Neuroimage 197, 643–649, https://doi.org/10.1016/j.neuroimage.2019.04.082 (2019).

12. 12.

Poser, B. A. & Setsompop, K. Pulse sequences and parallel imaging for high spatiotemporal resolution MRI at ultra-high field. Neuroimage 168, 101–118, https://doi.org/10.1016/j.neuroimage.2017.04.006 (2018).

13. 13.

van der Zwaag, W., Schafer, A., Marques, J. P., Turner, R. & Trampel, R. Recent applications of UHF-MRI in the study of human brain function and structure: a review. NMR Biomed 29, 1274–1288, https://doi.org/10.1002/nbm.3275 (2016).

14. 14.

De Martino, F. et al. The impact of ultra-high field MRI on cognitive and computational neuroimaging. Neuroimage 168, 366–382, https://doi.org/10.1016/j.neuroimage.2017.03.060 (2018).

15. 15.

Shmuel, A., Yacoub, E., Chaimow, D., Logothetis, N. K. & Ugurbil, K. Spatio-temporal point-spread function of fMRI signal in human gray matter at 7 Tesla. Neuroimage 35, 539–552, https://doi.org/10.1016/j.neuroimage.2006.12.030 (2007).

16. 16.

Uludag, K. & Blinder, P. Linking brain vascular physiology to hemodynamic response in ultra-high field MRI. Neuroimage 168, 279–295, https://doi.org/10.1016/j.neuroimage.2017.02.063 (2018).

17. 17.

Chaimow, D., Yacoub, E., Ugurbil, K. & Shmuel, A. Spatial specificity of the functional MRI blood oxygenation response relative to neuronal activity. Neuroimage 164, 32–47, https://doi.org/10.1016/j.neuroimage.2017.08.077 (2018).

18. 18.

Goense, J. B. & Logothetis, N. K. Laminar specificity in monkey V1 using high-resolution SE-fMRI. Magn Reson Imaging 24, 381–392, https://doi.org/10.1016/j.mri.2005.12.032 (2006).

19. 19.

Goense, J. B., Zappe, A. C. & Logothetis, N. K. High-resolution fMRI of macaque V1. Magn Reson Imaging 25, 740–747, https://doi.org/10.1016/j.mri.2007.02.013 (2007).

20. 20.

Koopmans, P. J., Barth, M. & Norris, D. G. Layer-specific BOLD activation in human V1. Hum Brain Mapp 31, 1297–1304, https://doi.org/10.1002/hbm.20936 (2010).

21. 21.

Koopmans, P. J., Barth, M., Orzada, S. & Norris, D. G. Multi-echo fMRI of the cortical laminae in humans at 7 T. Neuroimage 56, 1276–1285, https://doi.org/10.1016/j.neuroimage.2011.02.042 (2011).

22. 22.

Polimeni, J. R., Fischl, B., Greve, D. N. & Wald, L. L. Laminar analysis of 7T BOLD using an imposed spatial activation pattern in human V1. Neuroimage 52, 1334–1346, https://doi.org/10.1016/j.neuroimage.2010.05.005 (2010).

23. 23.

Ress, D., Glover, G. H., Liu, J. & Wandell, B. Laminar profiles of functional activity in the human brain. Neuroimage 34, 74–84, https://doi.org/10.1016/j.neuroimage.2006.08.020 (2007).

24. 24.

Yacoub, E., Harel, N. & Ugurbil, K. High-field fMRI unveils orientation columns in humans. Proc Natl Acad Sci USA 105, 10607–10612, https://doi.org/10.1073/pnas.0804110105 (2008).

25. 25.

Yacoub, E., Shmuel, A., Logothetis, N. & Ugurbil, K. Robust detection of ocular dominance columns in humans using Hahn Spin Echo BOLD functional MRI at 7 Tesla. Neuroimage 37, 1161–1177, https://doi.org/10.1016/j.neuroimage.2007.05.020 (2007).

26. 26.

Kriegeskorte, N. & Bandettini, P. Analyzing for information, not activation, to exploit high-resolution fMRI. Neuroimage 38, 649–662, https://doi.org/10.1016/j.neuroimage.2007.02.022 (2007).

27. 27.

Kamitani, Y. & Tong, F. Decoding the visual and subjective contents of the human brain. Nat Neurosci 8, 679–685, https://doi.org/10.1038/nn1444 (2005).

28. 28.

Haynes, J. D. & Rees, G. Predicting the orientation of invisible stimuli from activity in human primary visual cortex. Nat Neurosci 8, 686–691, https://doi.org/10.1038/nn1445 (2005).

29. 29.

Freeman, J., Brouwer, G. J., Heeger, D. J. & Merriam, E. P. Orientation decoding depends on maps, not columns. J Neurosci 31, 4792–4804, https://doi.org/10.1523/JNEUROSCI.5160-10.2011 (2011).

30. 30.

Mannion, D. J. & McDonald, J. S. & Clifford, C. W. The influence of global form on local orientation anisotropies in human visual cortex. Neuroimage 52, 600–605, https://doi.org/10.1016/j.neuroimage.2010.04.248 (2010).

31. 31.

Op de Beeck, H. P. Against hyperacuity in brain reading: spatial smoothing does not hurt multivariate fMRI analyses? Neuroimage 49, 1943–1948, https://doi.org/10.1016/j.neuroimage.2009.02.047 (2010).

32. 32.

Sasaki, Y. et al. The radial bias: a different slant on visual orientation sensitivity in human and nonhuman primates. Neuron 51, 661–670, https://doi.org/10.1016/j.neuron.2006.07.021 (2006).

33. 33.

Alink, A., Krugliak, A., Walther, A. & Kriegeskorte, N. fMRI orientation decoding in V1 does not require global maps or globally coherent orientation stimuli. Front Psychol 4, 493, https://doi.org/10.3389/fpsyg.2013.00493 (2013).

34. 34.

Chaimow, D., Yacoub, E., Ugurbil, K. & Shmuel, A. Modeling and analysis of mechanisms underlying fMRI-based decoding of information conveyed in cortical columns. Neuroimage 56, 627–642, https://doi.org/10.1016/j.neuroimage.2010.09.037 (2011).

35. 35.

Seymour, K., Clifford, C. W., Logothetis, N. K. & Bartels, A. Coding and binding of color and form in visual cortex. Cereb Cortex 20, 1946–1954, https://doi.org/10.1093/cercor/bhp265 (2010).

36. 36.

Alink, A., Walther, A., Krugliak, A. & Kriegeskorte, N. Local opposite orientation preferences in V1: fMRI sensitivity to fine-grained pattern information. Sci Rep 7, 7128, https://doi.org/10.1038/s41598-017-07036-8 (2017).

37. 37.

Freeman, J., Heeger, D. J. & Merriam, E. P. Coarse-scale biases for spirals and orientation in human visual cortex. J Neurosci 33, 19695–19703, https://doi.org/10.1523/JNEUROSCI.0889-13.2013 (2013).

38. 38.

Smith, F. W. & Muckli, L. Nonstimulated early visual areas carry information about surrounding context. Proc Natl Acad Sci USA 107, 20099–20103, https://doi.org/10.1073/pnas.1000233107 (2010).

39. 39.

Willenbockel, V. et al. Controlling low-level image properties: the SHINE toolbox. Behav Res Methods 42, 671–684, https://doi.org/10.3758/BRM.42.3.671 (2010).

40. 40.

Muckli, L. & Petro, L. S. Network interactions: non-geniculate input to V1. Curr Opin Neurobiol 23, 195–201, https://doi.org/10.1016/j.conb.2013.01.020 (2013).

41. 41.

Petro, L. S., Smith, F. W., Schyns, P. G. & Muckli, L. Decoding face categories in diagnostic subregions of primary visual cortex. Eur J Neurosci 37, 1130–1139, https://doi.org/10.1111/ejn.12129 (2013).

42. 42.

Sereno, M. I. et al. Borders of multiple visual areas in humans revealed by functional magnetic resonance imaging. Science 268, 889–893, https://doi.org/10.1126/science.7754376 (1995).

43. 43.

Schira, M. M., Tyler, C. W., Breakspear, M. & Spehar, B. The foveal confluence in human visual cortex. J Neurosci 29, 9050–9058, https://doi.org/10.1523/JNEUROSCI.1760-09.2009 (2009).

44. 44.

Van de Moortele, P. F. et al. T1 weighted brain images at 7 Tesla unbiased for Proton Density, T2* contrast and RF coil receive B1 sensitivity with simultaneous vessel visualization. Neuroimage 46, 432–446, https://doi.org/10.1016/j.neuroimage.2009.02.009 (2009).

45. 45.

Kemper, V. G., De Martino, F., Emmerling, T. C., Yacoub, E. & Goebel, R. High resolution data analysis strategies for mesoscale human functional MRI at 7 and 9.4T. Neuroimage 164, 48–58, https://doi.org/10.1016/j.neuroimage.2017.03.058 (2018).

46. 46.

De Martino, F. et al. High-Resolution Mapping of Myeloarchitecture In Vivo: Localization of Auditory Areas in the Human Brain. Cereb Cortex 25, 3394–3405, https://doi.org/10.1093/cercor/bhu150 (2015).

47. 47.

Chang, C. Ca. L. C.J. LIBSVM: a library for support vector machines. ACM Transactions on Intelligent Systems and Technology 2(27), 1–27 (2011).

48. 48.

Chaimow, D., Ugurbil, K. & Shmuel, A. Optimization of functional MRI for detection, decoding and high-resolution imaging of the response patterns of cortical columns. Neuroimage 164, 67–99, https://doi.org/10.1016/j.neuroimage.2017.04.011 (2018).

49. 49.

Parkes, L. M. et al. Quantifying the spatial resolution of the gradient echo and spin echo BOLD response at 3 Tesla. Magn Reson Med 54, 1465–1472, https://doi.org/10.1002/mrm.20712 (2005).

50. 50.

Bandettini, P. A., Wong, E. C., Jesmanowicz, A., Hinks, R. S. & Hyde, J. S. Spin-echo and gradient-echo EPI of human brain activation using BOLD contrast: a comparative study at 1.5 T. NMR Biomed 7, 12–20 (1994).

51. 51.

Boxerman, J. L. et al. The intravascular contribution to fMRI signal change: Monte Carlo modeling and diffusion-weighted studies in vivo. Magn Reson Med 34, 4–10, https://doi.org/10.1002/mrm.1910340103 (1995).

52. 52.

Constable, R. T., McCarthy, G., Allison, T., Anderson, A. W. & Gore, J. C. Functional brain imaging at 1.5 T using conventional gradient echo MR imaging techniques. Magn Reson Imaging 11, 451–459, https://doi.org/10.1016/0730-725x(93)90463-n (1993).

53. 53.

Duong, T. Q. et al. Microvascular BOLD contribution at 4 and 7 T in the human brain: gradient-echo and spin-echo fMRI with suppression of blood effects. Magn Reson Med 49, 1019–1027, https://doi.org/10.1002/mrm.10472 (2003).

54. 54.

Duong, T. Q. et al. High-resolution, spin-echo BOLD, and CBF fMRI at 4 and 7 T. Magn Reson Med 48, 589–593, https://doi.org/10.1002/mrm.10252 (2002).

55. 55.

Duyn, J. H., Moonen, C. T., van Yperen, G. H., de Boer, R. W. & Luyten, P. R. Inflow versus deoxyhemoglobin effects in BOLD functional MRI using gradient echoes at 1.5 T. NMR Biomed 7, 83–88 (1994).

56. 56.

Frahm, J., Merboldt, K. D., Hanicke, W., Kleinschmidt, A. & Boecker, H. Brain or vein–oxygenation or flow? On signal physiology in functional MRI of human brain activation. NMR Biomed 7, 45–53 (1994).

57. 57.

Kim, S. G., Hendrich, K., Hu, X., Merkle, H. & Ugurbil, K. Potential pitfalls of functional MRI using conventional gradient-recalled echo techniques. NMR Biomed 7, 69–74 (1994).

58. 58.

Lai, S. et al. Identification of vascular structures as a major source of signal contrast in high resolution 2D and 3D functional activation imaging of the motor cortex at 1.5T: preliminary results. Magn Reson Med 30, 387–392, https://doi.org/10.1002/mrm.1910300318 (1993).

59. 59.

Lee, A. T., Glover, G. H. & Meyer, C. H. Discrimination of large venous vessels in time-course spiral blood-oxygen-level-dependent magnetic-resonance functional neuroimaging. Magn Reson Med 33, 745–754, https://doi.org/10.1002/mrm.1910330602 (1995).

60. 60.

Menon, R. S. et al. BOLD based functional MRI at 4 Tesla includes a capillary bed contribution: echo-planar imaging correlates with previous optical imaging using intrinsic signals. Magn Reson Med 33, 453–459, https://doi.org/10.1002/mrm.1910330323 (1995).

61. 61.

Menon, R. S., Ogawa, S., Tank, D. W. & Ugurbil, K. Tesla gradient recalled echo characteristics of photic stimulation-induced signal changes in the human primary visual cortex. Magn Reson Med 30, 380–386, https://doi.org/10.1002/mrm.1910300317 (1993).

62. 62.

Segebarth, C. et al. Functional MRI of the human brain: predominance of signals from extracerebral veins. Neuroreport 5, 813–816, https://doi.org/10.1097/00001756-199403000-00019 (1994).

63. 63.

Song, A. W., Wong, E. C., Tan, S. G. & Hyde, J. S. Diffusion weighted fMRI at 1.5 T. Magn Reson Med 35, 155–158, https://doi.org/10.1002/mrm.1910350204 (1996).

64. 64.

Ugurbil, K. et al. Functional mapping in the human brain using high magnetic fields. Philos Trans R Soc Lond B Biol Sci 354, 1195–1213, https://doi.org/10.1098/rstb.1999.0474 (1999).

65. 65.

Uludag, K., Muller-Bierl, B. & Ugurbil, K. An integrative model for neuronal activity-induced signal changes for gradient and spin echo functional imaging. Neuroimage 48, 150–165, https://doi.org/10.1016/j.neuroimage.2009.05.051 (2009).

66. 66.

Yacoub, E. et al. Spin-echo fMRI in humans using high spatial resolutions and high magnetic fields. Magn Reson Med 49, 655–664, https://doi.org/10.1002/mrm.10433 (2003).

67. 67.

Yacoub, E. et al. Imaging brain function in humans at 7 Tesla. Magn Reson Med 45, 588–594, https://doi.org/10.1002/mrm.1080 (2001).

68. 68.

Yacoub, E., V D Moortele, P. F., Shmuel, A. & Ugurbil, K. Signal and noise characteristics of Hahn SE and GE BOLD fMRI at 7 T in humans. Neuroimage 24, 738–750, https://doi.org/10.1016/j.neuroimage.2004.09.002 (2005).

69. 69.

Krings, T., Erberich, S. G., Roessler, F., Reul, J. & Thron, A. MR blood oxygenation level-dependent signal differences in parenchymal and large draining vessels: implications for functional MR imaging. AJNR Am J Neuroradiol 20, 1907–1914 (1999).

70. 70.

Keilholz, S. D., Silva, A. C., Raman, M., Merkle, H. & Koretsky, A. P. BOLD and CBV-weighted functional magnetic resonance imaging of the rat somatosensory system. Magn Reson Med 55, 316–324, https://doi.org/10.1002/mrm.20744 (2006).

71. 71.

Silva, A. C., Koretsky, A. P. & Duyn, J. H. Functional MRI impulse response for BOLD and CBV contrast in rat somatosensory cortex. Magn Reson Med 57, 1110–1118, https://doi.org/10.1002/mrm.21246 (2007).

72. 72.

Gati, J. S., Menon, R. S., Ugurbil, K. & Rutt, B. K. Experimental determination of the BOLD field strength dependence in vessels and tissue. Magn Reson Med 38, 296–302, https://doi.org/10.1002/mrm.1910380220 (1997).

73. 73.

Cheng, K., Waggoner, R. A. & Tanaka, K. Human ocular dominance columns as revealed by high-field functional magnetic resonance imaging. Neuron 32, 359–374, https://doi.org/10.1016/s0896-6273(01)00477-9 (2001).

74. 74.

Dechent, P. & Frahm, J. Direct mapping of ocular dominance columns in human primary visual cortex. Neuroreport 11, 3247–3249, https://doi.org/10.1097/00001756-200009280-00039 (2000).

75. 75.

Goodyear, B. G. & Menon, R. S. Brief visual stimulation allows mapping of ocular dominance in visual cortex using fMRI. Hum Brain Mapp 14, 210–217 (2001).

76. 76.

Menon, R. S., Ogawa, S., Strupp, J. P. & Ugurbil, K. Ocular dominance in human V1 demonstrated by functional magnetic resonance imaging. J Neurophysiol 77, 2780–2787, https://doi.org/10.1152/jn.1997.77.5.2780 (1997).

77. 77.

Shmuel, A., Chaimow, D., Raddatz, G., Ugurbil, K. & Yacoub, E. Mechanisms underlying decoding at 7 T: ocular dominance columns, broad structures, and macroscopic blood vessels in V1 convey information on the stimulated eye. Neuroimage 49, 1957–1964, https://doi.org/10.1016/j.neuroimage.2009.08.040 (2010).

78. 78.

Yao, Z., et al Mechanisms of Decoding Oriented Grating Stimuli Investigated by Optical Imaging of Cat Area 18. OHBM (2017).

79. 79.

Davis, T. et al. What do differences between multi-voxel and univariate analysis mean? How subject-, voxel-, and trial-level variance impact fMRI analysis. Neuroimage 97, 271–283, https://doi.org/10.1016/j.neuroimage.2014.04.037 (2014).

80. 80.

Swisher, J. D. et al. Multiscale pattern analysis of orientation-selective activity in the primary visual cortex. J Neurosci 30, 325–330, https://doi.org/10.1523/JNEUROSCI.4811-09.2010 (2010).

81. 81.

Alakorkko, T., Saarimaki, H., Glerean, E., Saramaki, J. & Korhonen, O. Effects of spatial smoothing on functional brain networks. Eur J Neurosci 46, 2471–2480, https://doi.org/10.1111/ejn.13717 (2017).

82. 82.

Korhonen, O., Saarimaki, H., Glerean, E., Sams, M. & Saramaki, J. Consistency of Regions of Interest as nodes of fMRI functional brain networks. Netw Neurosci 1, 254–274, https://doi.org/10.1162/NETN_a_00013 (2017).

83. 83.

Ugurbil, K. What is feasible with imaging human brain function and connectivity using functional magnetic resonance imaging. Philos Trans R Soc Lond B Biol Sci 371, https://doi.org/10.1098/rstb.2015.0361 (2016).

84. 84.

Pratte, M. S., Sy, J. L., Swisher, J. D. & Tong, F. Radial bias is not necessary for orientation decoding. Neuroimage 127, 23–33, https://doi.org/10.1016/j.neuroimage.2015.11.066 (2016).

85. 85.

Bejjanki, V. R., da Silveira, R. A., Cohen, J. D. & Turk-Browne, N. B. Noise correlations in the human brain and their impact on pattern classification. PLoS Comput Biol 13, e1005674, https://doi.org/10.1371/journal.pcbi.1005674 (2017).

86. 86.

Revina, Y., Petro, L. S. & Muckli, L. Cortical feedback signals generalise across different spatial frequencies of feedforward inputs. Neuroimage 180, 280–290, https://doi.org/10.1016/j.neuroimage.2017.09.047 (2018).

## Acknowledgements

We thank Dr. Junpeng Lao for the useful discussions on analytical procedures, and Yulia Revina for help with reference data. This project has received funding from the European Union’s Horizon 2020 Framework Program for Research and Innovation under the Specific Grant Agreement No. 720270 and 785907 (Human Brain Project SGA1 and SGA2) and European Research Council (ERC StG 2012_311751-“Brain reading of contextual feedback and predictions” to LM).

## Author information

Authors

### Contributions

All co-authors carefully read the final text of this manuscript, made edits and provided critical feedback on results and interpretation. L.V. performed all analyses, formulated figures and wrote the manuscript. L.M. provided the initial concept for the ms. F.d.M. collected the data. L.M and E.Y. provided guidance and supervision throughout all stages of this study.

### Corresponding author

Correspondence to Luca Vizioli.

## Ethics declarations

### Competing interests

The authors declare no competing interests.

Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

## Rights and permissions

Reprints and Permissions

Vizioli, L., De Martino, F., Petro, L.S. et al. Multivoxel Pattern of Blood Oxygen Level Dependent Activity can be sensitive to stimulus specific fine scale responses. Sci Rep 10, 7565 (2020). https://doi.org/10.1038/s41598-020-64044-x

• Accepted:

• Published:

• ### Recurrent Processing Drives Perceptual Plasticity

• Ke Jia
• , Elisa Zamboni
• , Valentin Kemper
• , Catarina Rua
• , Nuno Reis Goncalves
• , Adrian Ka Tsun Ng
• , Christopher T. Rodgers
• , Guy Williams
• , Rainer Goebel
•  & Zoe Kourtzi

Current Biology (2020)