Representations of imaginary scenes and their properties in cortical alpha activity

Imagining natural scenes enables us to engage with a myriad of simulated environments. How do our brains generate such complex mental images? Recent research suggests that cortical alpha activity carries information about individual objects during visual imagery. However, it remains unclear if more complex imagined contents such as natural scenes are similarly represented in alpha activity. Here, we answer this question by decoding the contents of imagined scenes from rhythmic cortical activity patterns. In an EEG experiment, participants imagined natural scenes based on detailed written descriptions, which conveyed four complementary scene properties: openness, naturalness, clutter level and brightness. By conducting classification analyses on EEG power patterns across neural frequencies, we were able to decode both individual imagined scenes as well as their properties from the alpha band, showing that also the contents of complex visual images are represented in alpha rhythms. A cross-classification analysis between alpha power patterns during the imagery task and during a perception task, in which participants were presented images of the described scenes, showed that scene representations in the alpha band are partly shared between imagery and late stages of perception. This suggests that alpha activity mediates the top-down re-activation of scene-related visual contents during imagery.

In the present study, we thus aimed to answer the question if natural scenes and their properties are represented in cortical alpha activity during visual imagery and if these representations are shared with perceptual processing.To that end, we conducted an EEG experiment in which participants imagined natural scenes based on a written description and viewed images of the same scenes in a separate task.The scenes varied in four properties which have previously been investigated in the scene literature: openness, naturalness, clutter level and brightness 26,27 .We employed frequency-resolved multivariate pattern classification to track the representations of scenes across neural rhythms, and imagery-perception cross-classification to investigate whether scene representations are similarly coded in rhythmic cortical activity during imagery and perception.
We show that imagined natural scenes and their properties are represented in cortical alpha activity and that scene representations in the alpha frequency band are partly shared between scene imagery and late stages of scene perception.Our results indicate that cortical alpha activity mediates the top-down re-instantiation of complex natural scenes during visual imagery.

Results
In order to investigate the neural representations of imagined and perceived scenes and their properties in neural rhythms, we conducted two experimental tasks while participants' neural activity was recorded via EEG (see Fig. 1c for a schematic of the tasks).In the imagery task, participants imagined natural scenes for 4000 ms based on a detailed, three-sentence description of each scene (see Fig. 1a for an example).The 16 described scenes varied independently in four properties: openness, naturalness, clutter level and brightness.After completing the imagery task, participants were asked to rate the imagined scenes in regard to the four properties on a scale from 1 to 7.These ratings confirmed that, at least on average, the properties of the imagined scenes aligned with those conveyed by the descriptions (see Fig. 1b).In the subsequent perception task, participants viewed images that matched the scene descriptions (3 images per scene; see Fig. 1a) with each image being presented for 1000 ms.

Mean pairwise scene decoding
To identify at which neural frequencies information related to individual imagined scenes can be found, we conducted a mean pairwise frequency searchlight decoding analysis.We transformed the EEG signals across the entire imagery period into the frequency domain and trained classifiers to distinguish between each possible pair of imagined scenes based on power patterns across channels at each frequency from 4 to 30 Hz (see Fig. 2a for a schematic).Averaging the pairwise decoding accuracies across pairs yielded a measure of information content regarding the individual imagined scenes at each frequency.We found that the individual imagined scenes could be discriminated the best in the alpha frequency range (see Fig. 3a), with significant mean pairwise scene decoding from 8 to 13 Hz, peaking at 11 Hz (p < 0.001).There was also some weaker, but sporadically significant mean pairwise scene decoding in the beta band (significant at 18 and 21 Hz), indicating that some information might also be contained there.These results suggest that individual imagined scenes are represented most prominently in the alpha frequency band.

Scene property decoding
We further examined how the four investigated properties (openness, naturalness, clutter level and brightness) of the imagined scenes are represented across the frequency domain.Using the same frequency searchlight approach as above, we had classifiers predict for each property which property category the imagined scene belonged to (e.g., for naturalness, if the scene was natural or man-made).This analysis revealed that all four properties could exclusively be decoded from the alpha band (see Fig. 3b).The decoding accuracy for each property peaked at around 10 Hz (openness: 9 Hz, p = 0.009; naturalness: 11 Hz; p < 0.001, clutter level: 10 Hz, p = 0.038; brightness: 10 Hz, p = 0.031).Aligning with our decoding analysis of the individual scenes, these findings suggest that the properties of imagined scenes are also represented in cortical alpha activity.

Imagery-perception cross-decoding in the alpha frequency band
Next, we investigated if any of the of representations of the individual scenes or scene properties we found in the alpha band are shared between imagery and different stages of the perceptual processing hierarchy.To that end, we conducted the same mean pairwise and property decoding analyses, but trained the classifiers on alpha power patterns in the frequency-resolved imagery data and tested them on alpha power patterns at each time point in the time-frequency-resolved perception data and vice versa (see Fig. 2b for a schematic).We chose to assess these shared alpha representations across the entire imagery period while maintaining temporal resolution for the perceptual data in order to increase the power of the analysis, since imagery representations were found to be relatively invariable across time 18,28,29 , whereas perceptual representations are thought to be more temporally variable [29][30][31] .This yielded a time-resolved measure of shared representations in the alpha band between scene imagery and each stage in the processing hierarchy of scene perception, in which temporal representational variations index the procession of perceptual processing 32 .In the mean pairwise scene cross-decoding analysis, we identified an increase in mean pairwise scene cross-decoding accuracy starting at around 600 ms during perceptual processing that was marginally significant (p = 0.068 at peak) at 750-800 ms (see Fig. 3c).This trend suggests that there are representations of individual scenes in the alpha band that are shared between scene imagery and late stages of scene perception.In the property cross-decoding analysis, we found relatively low but significant cross-classification performance for openness at 750 ms (p = 0.012) and for clutter level at 800-850 ms (p = 0.002 at peak) as well as 950-1000 ms (p = 0.008 at peak) in the perceptual processing hierarchy (see Fig. 3d).There also was marginally significant (p = 0.057) cross-decoding performance for brightness at 400 ms, but no significant cross-decoding performance for naturalness.The peak in cross-decoding accuracy for openness was very similar to the first peak in cross-decoding accuracy for clutter level and both aligned temporally almost perfectly.They both also temporally overlapped with the marginally significant peak in the mean pairwise cross-decoding for the individual scenes.The results of our property cross-decoding analysis indicate that scene imagery shares representations with late scene perception in the alpha band, at least for some properties.We exploratively conducted all imagery-perception cross-decoding analyses in the theta and beta bands as well, but there was no solid evidence of shared scene representations in those frequency bands (see supplementary Fig. S3).Average ratings of all four properties for each imagined scene.After finishing the imagery task, participants rated each imagined scene on a scale from 1 to 7 regarding the four investigated properties openness, naturalness, clutter level and brightness.On average, property ratings of imagined scenes aligned with the properties conveyed by the scene descriptions (e.g. the mental images of open scenes were also rated as high on the openness dimension and those of closed scenes were rated as low).Large circles: property category mean rating.Small circles: mean rating of individual scenes in the respective property category.(c) Experimental tasks.In the imagery task, participants were presented with a scene description surrounded by a black frame until they proceeded with a button press.A black fixation dot appeared and after a jittered interval of 1000-2000 ms, the fixation dot turned red which was their cue to imagine the scene within the surrounding frame.They were instructed to maintain the mental image of the scene while fixating the red dot until it turned black again after 4000 ms.In the perception task, participants were presented images for 1000 ms that matched the scene descriptions in the imagery task and were only tasked with attentively viewing them.

Shuffled property decoding
In a control analysis, we assessed to what extent the representations the classifiers utilized during scene property decoding encode differences in individual scene features or property category information.We performed a shuffled property decoding analysis for both the imagery property decoding and imagery-perception property cross-decoding in which all scenes were randomly assigned to two mock property categories for all possible permutations and classifiers were trained to distinguish between these categories.Since property category information is randomized in this decoding scheme, the classifiers should be limited to differentiate between these mock categories based on the features of the individual scenes they happen to encompass.If the property decoding did not only use differences in individual scenes, but also more abstract information on property categories, shuffled property decoding performance should be reduced in comparison to the original property decoding performance, since the classifiers had no access to this additional source of information in the shuffled analysis.
To test this, we compared the peak (cross-)decoding accuracy for each property that was discriminable in the original analyses to the shuffled property (cross-)decoding accuracy at the respective frequency or time point.
For the imagery property decoding analysis, we found higher peak decoding accuracies compared to the shuffled property decoding accuracies for all properties (see Fig. 4a).This difference was significant for openness (p = 0.015) and naturalness (p = 0.031), marginally significant for clutter level (p = 0.072) and not significant for brightness (p = 0.102).This implies that for openness, naturalness and potentially clutter level, there are property category representations in the alpha band during scene imagery.We also conducted the shuffled property decoding at each frequency during imagery (see supplementary Fig. S4).While slightly lower in overall accuracy, the decoding performance profile looks strikingly similar to that in our mean pairwise scene decoding, further corroborating that the shuffled property decoding mainly reflects neural discriminability based on individual scene information.
For the imagery-perception property cross-decoding analysis, we conducted the comparison between property cross-decoding and shuffled property cross-decoding for all properties except naturalness, since we did not find any interpretable cross-decoding accuracy peak for this property.We found higher cross-decoding accuracies Figure 2. Decoding approaches.We conducted two main decoding analyses.In the imagery frequency decoding analysis (a), we investigated rhythmic scene representations during imagery by training classifiers on power patterns across all channels at each frequency from 4 to 30 Hz in the frequency-resolved imagery EEG data.In the imagery-perception cross-decoding analysis (b), we investigated shared scene representations between imagery and perception in the alpha band, by training the classifiers on alpha power patterns across all channels in the frequency-resolved imagery data and testing them on alpha power patterns across all channels at each time point in the time-frequency-resolved perception data and vice versa.We employed two decoding schemes in both analyses: a mean pairwise scene decoding and a scene property decoding.In the mean pairwise scene decoding, classifiers were trained to distinguish between each possible pair of scenes and decoding accuracies were averaged across each scene pair, as a measure of neural discriminability among individual scenes.In the scene property decoding, classifiers were trained to predict in which of two property categories a scene belonged for each property dimension the scenes varied in.In an additional control analysis, we re-conducted the scene property (cross-) decoding analyses by training classifiers to discriminate between mock property categories with randomly assigned scenes as a measure of neural discriminability between property categories based on individual scene features.www.nature.com/scientificreports/compared to the shuffled property cross-decoding accuracies for all three investigated properties (see Fig. 4b).This was significant for all properties (openness: p = 0.008, clutter level: p = 0.004, brightness: p = 0.006), suggesting that there are category representations of these properties in the alpha band that are shared between imagery and late stages of perceptual processing.When conducting the shuffled property cross-decoding at each time point during perception (see supplementary Fig. S4), we found significant above-chance cross-decoding performance starting at about 750 ms, providing further evidence of shared scene representations between imagery and late perception in the alpha band.

Relationship between imagery vividness and neural representations in the alpha band
In an exploratory analysis, we investigated if there is a relationship between the participants' ability to evoke vivid mental images and the neural representations in the alpha band of such mental images.We correlated their scores in the Vividness of Visual Imagery Questionnaire (VVIQ) 33 , which they provided online during recruitment, with their peak in mean pairwise scene decoding and cross-decoding accuracy in the alpha band.While we did not find a meaningful correlation between the VVIQ and the peak in the mean pairwise decoding (Pearson's r = 0.074, p = 0.307), we did find a marginally significant positive correlation with the peak in the mean pairwise cross-decoding (Pearson's r = 0.2, p = 0.084).This marginal trend suggests that increased representational overlap between imagery and perception in the alpha band leads to a more vivid imagery experience, which aligns with previous research showing the same finding in the fMRI 34 .To test whether this relationship is solid, future studies could use trial-wise ratings of imagery vividness, which tend to be a more reliable measure 34,35 .

Discussion
In the present study, we investigated the representations of imagined and perceived natural scenes in rhythmic cortical activity.We found, as hypothesized, that both individual scenes as well as scene properties are represented in cortical alpha activity during visual imagery.We also found evidence that scene representations in the alpha frequency band are partly shared between imagery and late stages of perceptual processing.These results indicate that the top-down reactivation of scene representations during visual imagery is enabled by cortical alpha activity.This aligns well with studies showing that alpha rhythms play a role in visual imagery (e.g. 11,12), and specifically with the notion that imagery-related alpha oscillations are a top-down signal that represents the imagined visual contents 18 .In more general terms, our results also support theories that postulate that top-down information flows are mediated by alpha dynamics in visual cortex (e.g. 9 ).
Our decoding analyses revealed that all four investigated scene properties were discriminable from cortical alpha activity during visual imagery and that for all scene properties except naturalness (and brightness being only marginally significant) these alpha representations were shared with late stages of scene perception.Comparing the initial imagery property decoding to a decoding scheme in which property categories were randomized showed that there were genuine and abstract representations of scene properties in the alpha band, which were observed for all properties except brightness during imagery (clutter level being only marginally significant).Extending this comparison to the imagery-perception cross-decoding, we found evidence of shared alpha representations of abstract property information for openness, clutter level and brightness.These results suggest that, during imagery, alpha activity enables the top-down reactivation of scene property representations, some of which are shared with late stages of scene perception.This further implies that the representational division into (global) scene properties found during perception 36 also holds during imagery.One exception was, however, that we did not find any shared alpha representations for naturalness.A feasible explanation might be that, while participants rated the naturalness of the imagined scenes as expected, the natural and man-made contents they imagined might have differed from the natural and man-made contents in the images they viewed (e.g.different types of natural or man-made objects), resulting in different neural representations being recruited during imagery and perception.The properties for which we did find evidence of shared representations (openness, clutter level and brightness) are much less dependent on the specific types of imagined objects.www.nature.com/scientificreports/ We found evidence of shared scene representations in the alpha band between imagery and perception from around 400 ms (for brightness) until 1000 ms (for clutter level) after stimulus onset.This is in alignment with previous studies employing cross-decoding techniques which have also reported late shared representations during perception.Xie et al. 18 , who originally found shared alpha representations between object imagery and perception, also reported late timings during perceptual processing with the strongest correspondence with imagery for perception emerging after 400 ms.Dijkstra et al. 29 reported shared representations between imagery and perception up until 1000 ms during perception.Why would imagery reactivate representations that occur so late during perceptual processing?One potential explanation is that imagery and perception share fewer representations in low-level and more in high-level visual areas 5 , making it more likely that shared representations occur during later perceptual processing.This can be explained by the prominent conceptualization of imagery as a reverse reactivation of the perceptual hierarchy starting from high-level visual cortex 37,38 .Following this notion, the representational format of cortical brain areas in late stages of perceptual processing (i.e.high-level visual cortex) is thought to be more similar to those in imagery since they are closer to the trigger source of the imagery signal 1 .In alignment with this, Xie et al. found that the late shared alpha representations were best explained by complex visual features analyzed in high-level visual cortex.Thus, the shared alpha representations in our results might also reflect late processing in high-level visual areas.This is supported by our shuffled property control analysis that yielded that some of the shared property representations in the alpha band encode category information, which is typically represented in high-level visual cortex 39 .
However, even if late shared representations between imagery and perception are not unexpected, the timings of our results are still quite late, given that processing of scenes and (global) scene properties (specifically openness, naturalness and clutter level) has been shown to be rapid and already occurs within the first 250 ms after stimulus onset 26,27,[40][41][42] .Since most research on the temporal dynamics of scene processing has focused on comparatively early neural signatures 43 , what happens during such very late stages of scene processing is still largely unknown.Given that the perceived scene images in our study were presented throughout the entire analysis time window, a possible explanation is that the alpha representations scene perception shares with scene imagery in our data reflect recurrent processing of the scenes and their properties after the first feed-forward sweep 37 .During recurrent processing, the perceptual representational format might be altered in a way that makes it more similar to imagery representations.Future studies could clarify to what extent recurrent processes shape the late representations in perception that generalize to imagery.
A final caveat in our results are the low decoding accuracies.Imagery-related brain signals tend to have a low signal-to-noise-ratio (e.g. 44) which results in lower decoding accuracies in imagery studies that employ MVPA 18,29 .Furthermore, our imagery task was designed to ensure that the imagined scenes sufficiently differed from the perceived scenes in terms of their low-level features.We had participants imagine the scenes based on descriptions that allow for variability in the generated mental images and only presented them the actual images after, so that if we did find shared representations between imagery and perception, they would not be based on similarities in low-level features.However, a side effect of this might have been reduced cross-decoding performance since the classifiers could not exploit such low-level features to a great extent.The cross-decoding performance might also have been impacted by the limited range of images in the perception task which might not fully cover the variability in the imagined visual contents.This could be remedied in future studies by generating a large image set based on the imagined prompts using text-to-image algorithms 45 .In addition, due to the relatively long trial duration, we only had 192 imagery trials of training data per participant, which further limited classifier performance.As a result, in particular the very low cross-decoding accuracies (being less than 1% above chance) need to be interpreted with caution.Nevertheless, there are multiple factors that point towards the cross-decoding results being a true effect.First, due to the low temporal resolution typically employed in time-frequency decomposition, there were only 20 post-stimulus time points in our perception data, which is considerably less than the hundreds of time points that typically require multiple comparison correction in temporally resolved decoding analyses 16 and we did appropriately correct for multiple comparisons.Second, we found the same latency of roughly 750-800 ms across four different analyses that investigated shared scene representations using three different decoding schemes: the mean-pairwise cross-decoding (Fig. 3c), the openness and clutter level property cross-decoding (Fig. 3d) and the shuffled property cross-decoding (Fig. S4).Third, the late timings in the cross-decoding analyses roughly align with late timings reported in previous imagery-perception cross-decoding studies as discussed above 18,29 .Fourth, the results of the shuffled property cross-decoding control analysis suggested that there are genuine property representations at the peak cross-decoding time points for openness, clutter level and brightness.Finally, decoding accuracies are considered a poor measure of effect size and low decoding accuracies can still constitute meaningful effects, indicating that information is represented consistently in neural response patterns across participants 46,47 .
Overall, our results suggest that the top-down reactivation of scene representations during visual imagery is mediated by cortical alpha activity and that the re-instantiated alpha representations are partly shared with late stages of scene perception.They show that alpha dynamics are not only critical for generating mental images of individual objects, but also mediate the creation of complex natural environments in our mind's eye.

Methods
Participants 50 participants (25 male; mean age = 25.74 years, SD = 6.31) with normal or corrected-to-normal eyesight took part in the experiment.One participant was excluded from all analyses because they did not complete the imagery task due to a technical error during the EEG recording.During recruitment, participants filled in a German translation of the Vividness of Visual Imagery Questionnaire (VVIQ) 33 , a common measure of a person's aptitude at evoking mental images, on Limesurvey (https:// www.limes urvey.org/ en/).The scale of the VVIQ was reversed www.nature.com/scientificreports/

Frequency decomposition
All of our frequency decompositions were conducted separately for each trial and each channel.We transformed the EEG signals within the entire 4000 ms imagery period into the frequency domain.For the perception data, the period from − 200 to 1000 ms was transformed into the time-frequency domain using a fixed-size 500 ms sliding window with 50 ms steps.For both decompositions, we utilized multitapers (15 DPSS tapers for imagery and 3 for perception) with constant 2 Hz frequency smoothing as implemented in FieldTrip.We chose multitapers in particular to increase power in our frequency-based analyses.Imagery data tends to be noisy (e.g. 18,44) and the multitaper approach typically increases the signal-to-noise ratio in frequency-resolved data at the expense of increased temporal and frequency smoothing 51 .In addition, we decided to omit temporal resolution of the frequency decomposition of the imagery data while maintaining it for the perception data in order to further boost power, since imagery representations have been shown to be relatively invariable across time 18,28,29 while perceptual representations have been shown to be temporally variable as a function of the different processing stages in the visual hierarchy [29][30][31][32] .Extracted frequencies ranged from 4 to 30 Hz, thus covering the theta (4-7 Hz), alpha (8-13 Hz) and beta (14-30 Hz) frequency bands.

Decoding analyses
All decoding analyses were conducted using CoSMoMVPA 52 .We employed linear discriminant analysis (LDA) classifiers which were trained within-subject on power patterns across all channels (see Fig. 2).
In order to assess how the individual imagined scenes and their properties are represented in neural activity patterns, we employed two decoding approaches and solely altered the features on which we conducted these analyses to answer different questions.The first approach was a mean pairwise scene decoding analysis in which classifiers were trained to discriminate between each possible pair of scenes and the resulting pairwise decoding accuracies were averaged across pairs, yielding a measure of discriminability among individual scenes from neural responses.The second approach was a scene property decoding analysis in which classifiers had to distinguish for each property in which of two property categories a scene belonged (e.g. for naturalness, if the imagined scene was a natural or a man-made scene).
First, we assessed at which neural frequencies scene information is represented by running the mean pairwise scene decoding and the scene property decoding on power patterns at each individual frequency from 4 to 30 Hz in the frequency-resolved imagery EEG data (see Fig. 2a).Classifiers were trained using a leave-one-trial-out cross-validation scheme in which one trial per stimulus was left out to avoid imbalance between conditions.Decoding accuracies were calculated as the mean of all cross-validation fold accuracies.For the mean pairwise scene decoding, this resulted in one mean pairwise scene decoding accuracy at each frequency for each participant.For the scene property decoding, this yielded one decoding accuracy at each frequency for each property and each participant.
Second, we examined if there are scene representations in the alpha frequency band that are shared between imagery and different stages of perceptual processing across time.We again conducted the mean pairwise and property decoding analyses, but trained the classifiers on power patterns across the entire alpha frequency range (8-13 Hz) in the frequency-resolved imagery data and tested them on the alpha power patterns at each time point in the time-frequency resolved perception data and vice versa (see Fig. 2b).The decoding accuracies of both train-test directions were averaged, which resulted in one mean pairwise cross-decoding accuracy time course as well as four property category cross-decoding accuracy time courses for each participant.We also exploratively conducted all imagery-perception cross-decoding analyses in the theta and beta bands (see supplementary Fig. S3).
Finally, it is possible that during the property decoding analyses the classifiers did not utilize property category information, but just exploited differences in features of individual imagined and perceived scenes.To investigate this, we conducted a shuffled property decoding analysis in which we estimated how well the classifiers perform if they are constrained to individual scene feature information, without access to property category information and compared this performance to the original property decoding.Within each participant, the 16 scenes were randomly assigned to two mock property categories for all possible 16 8 = 12,780 permutations and at each permutation, classifiers were trained to distinguish between the property categories.Decoding accuracies were then averaged across permutations.Since in this decoding scheme the property category information was randomized, classifiers were limited to discriminate based on differences in the individual scene features in each category.If the property decoding only exploited individual scene features, the shuffled property decoding performance should be identical or highly similar to it.If, however, property category information was also used for property discrimination in the original analysis, the shuffled property decoding performance should be reduced in comparison since the classifiers in the shuffled analysis had no access to this additional source of information.We applied the shuffled property decoding scheme to both the frequency-resolved imagery property decoding and our imagery-perception property cross-decoding.We tested the difference between property decoding and shuffled property decoding by assessing if the peak decoding accuracy of each property in the imagery property decoding and imagery-perception property cross-decoding is greater than the decoding accuracy at the respective frequency or time point in the shuffled property decoding.For the imagery-perception cross-decoding, this comparison was omitted for naturalness since we found no interpretable above-chance cross-decoding performance for this property in the original analysis.We also conducted the shuffled property decoding across all frequencies in the imagery property decoding and all time points in the imagery-perception property cross-decoding (see supplementary Fig. S4) in order to investigate if the shuffled property (cross-) decoding was equal to or exceeded the peak property (cross-)decoding at the frequencies or time points at which we did not conduct the comparison.These analyses showed that this was not the case.
We investigated the temporal dynamics of imaginary scene representations as well.However, when conducting the aforementioned mean pairwise and property decoding analyses on broadband EEG responses at each time point, consistent with a previous imagery study 18,44 , we did not find robust above-chance decoding performance (see supplementary Fig. S5).

Statistical testing
Decoding accuracies in all frequency-resolved and time-resolved analyses were tested against chance level (50%) using threshold-free cluster enhancement (TFCE) 53 as implemented in CoSMoMVPA.Multiple comparison correction was conducted by comparing actual TFCE statistics to a null distribution of maximum TFCE statistics, estimated using a permutation test with 10,000 sign permutations.The resulting z-scores were converted to p-values and thresholded at p < 0.05 (one-tailed).In our time-resolved analyses, only the post-stimulus time points were tested for significance.We compared the peak property (cross-)decoding accuracies to the shuffled property (cross-)decoding accuracies at the respective frequency or time point using paired, one-tailed Wilcoxon signed rank tests.All statistical tests were conducted on the full sample of n = 49.

Figure 1 .
Figure 1.Stimuli and Paradigm.(a) Stimuli used in the EEG experiment.Participants imagined 16 natural scenes according to detailed three-sentence descriptions (right) and in a separate task viewed images that matched these descriptions (three per scene; left).The scenes varied in their openness, naturalness, clutter level and brightness.(b)Average ratings of all four properties for each imagined scene.After finishing the imagery task, participants rated each imagined scene on a scale from 1 to 7 regarding the four investigated properties openness, naturalness, clutter level and brightness.On average, property ratings of imagined scenes aligned with the properties conveyed by the scene descriptions (e.g. the mental images of open scenes were also rated as high on the openness dimension and those of closed scenes were rated as low).Large circles: property category mean rating.Small circles: mean rating of individual scenes in the respective property category.(c) Experimental tasks.In the imagery task, participants were presented with a scene description surrounded by a black frame until they proceeded with a button press.A black fixation dot appeared and after a jittered interval of 1000-2000 ms, the fixation dot turned red which was their cue to imagine the scene within the surrounding frame.They were instructed to maintain the mental image of the scene while fixating the red dot until it turned black again after 4000 ms.In the perception task, participants were presented images for 1000 ms that matched the scene descriptions in the imagery task and were only tasked with attentively viewing them.

Figure 3 .
Figure 3. Decoding of imagined scenes and their properties based on spatially distributed power patterns in rhythmic neural activity.(a) Mean pairwise decoding of imagined scenes at each frequency from 4 to 30 Hz. Individual imagined scenes could be discriminated the best from alpha band activity (8-13 Hz).(b) Scene property decoding.All properties could exclusively be decoded from the alpha frequency band.(c) Mean pairwise imagery-perception scene cross-decoding in the alpha frequency band at each time point during scene perception.A marginally significant trend suggests that there are shared representations of individual scenes in the alpha frequency band between imagery and late stages of perception.d) Imagery-perception scene property cross-decoding in the alpha frequency band at each time point during scene perception.For openness, clutter level and brightness (brightness being only marginally significant) we found evidence of shared representations in the alpha frequency band between imagery and late stages of perception.Error margins reflect the standard error of the mean.Square markers indicate significance at p < 0.05, cross markers indicate marginal significance at p < 0.1 (both corrected for multiple comparisons).

Figure 4 .
Figure 4.Comparison between peak scene property decoding accuracies and decoding accuracies with randomized property assignment (shuffled property decoding).(a) Comparison at the peak decoding frequency for each property in the imagery property decoding.Property decoding accuracies exceeded shuffled decoding accuracies for all properties.This was significant for openness, naturalness and marginally for clutter level, suggesting that there are representations of property category information of these properties in the alpha band.(b) Comparison at the peak imagery-perception property cross-decoding time points for all properties that were discriminable in the original analysis.For each property, property cross-decoding accuracies significantly exceeded the shuffled accuracies, suggesting that there are property category representations in the alpha band that are shared between scene imagery and late stages of scene perception.Error bars reflect the standard error of the mean.Asterisks indicate significance at p < 0.05, cross markers indicate marginal significance at p < 0.1. https://doi.org/10.1038/s41598-024-63320-4 )